Next Article in Journal
SOAM Block: A Scale–Orientation-Aware Module for Efficient Object Detection in Remote Sensing Imagery
Previous Article in Journal
LISA-YOLO: A Symmetry-Guided Lightweight Small Object Detection Framework for Thyroid Ultrasound Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Super Resolution for Mangrove UAV Remote Sensing Images

School of Electronic Information, Guilin University of Electronic Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(8), 1250; https://doi.org/10.3390/sym17081250
Submission received: 3 June 2025 / Revised: 10 July 2025 / Accepted: 21 July 2025 / Published: 6 August 2025
(This article belongs to the Section Computer)

Abstract

Mangroves play a crucial role in ecosystems, and the accurate classification and real-time monitoring of mangrove species are essential for their protection and restoration. To improve the segmentation performance of mangrove UAV remote sensing images, this study performs species segmentation after the super-resolution (SR) reconstruction of images. Therefore, we propose SwinNET, an SR reconstruction network. We design a convolutional enhanced channel attention (CEA) module within a network to enhance feature reconstruction through channel attention. Additionally, the Neighborhood Attention Transformer (NAT) is introduced to help the model better focus on domain features, aiming to improve the reconstruction of leaf details. These two attention mechanisms are symmetrically integrated within the network to jointly capture complementary information from spatial and channel dimensions. The experimental results demonstrate that SwinNET not only achieves superior performance in SR tasks but also significantly enhances the segmentation accuracy of mangrove species.

1. Introduction

Mangroves are communities of salt-tolerant evergreen trees and other plant species found in tropical and subtropical intertidal zones. They provide important ecosystem services such as nutrient cycling, carbon sequestration, and mitigation of coastal disasters [1]. The ecological functions of mangroves have been degrading over the past few decades due to the impacts of climate change, natural disasters, and human disturbances [2,3]. Therefore, accurate classification and real-time monitoring of mangrove species are crucial for their protection and restoration. Traditionally, obtaining information on mangrove species requires expensive, laborious, and time-consuming field surveys, and surveyors often find it difficult to access mangrove areas [4]. In recent years, remote sensing technologies have been widely used for mangrove monitoring and assessment due to their excellent spatial and textural features and high-resolution multispectral imagery [5]. These technologies include spectrometer measurements, high-resolution aerial imagery, medium- to high-spatial-resolution satellite remote sensing imagery, and hyperspectral imagery [6]. Most remote sensing technologies are used for large-scale forest resource surveys and cannot capture detailed distributions of mangrove species [7]. In recent years, the rapid development of UAV technology has provided a new data source for the classification of mangrove communities [8,9]. Its flexibility, ability to operate below clouds, low cost, and centimeter-level high spatial resolution make it a complement to satellite remote sensing with broad development prospects. However, accurate classification of mangrove species remains challenging due to the diversity of species and canopy structures. Even with high-spectral-resolution and high-spatial-resolution remote sensing data, it is difficult to distinguish between different tree species [10].
With the development of deep learning [11,12,13], SR reconstruction technology can enhance image details in the network, improving the effectiveness of image segmentation [14].
Based on this technology, this study conducted research on mangrove species recognition using SR reconstruction of UAV remote sensing images. This study proposed SwinNET, an SR reconstruction network of mangrove UAV remote sensing images based on the improved SwinIR [15]. In recent years, in the field of image SR, the effectiveness of Single-Image SR (SISR) based on Transformer [16] models has gradually surpassed traditional Convolutional Neural Network (CNN) methods [17]. The Swin Transformer [18] is a visual model based on the Transformer structure, which has good characteristics such as global perception and dynamic weighting, making it perform well in various visual tasks. SwinIR, as an SR model based on the Swin Transformer, has also demonstrated excellent performance.
Based on this, this study proposed SwinNET, an improved mangrove UAV remote sensing image reconstruction network based on SwinIR. Specifically, similarly to SwinIR, this network consists of shallow feature extraction, deep feature extraction, and high-quality image reconstruction modules. The shallow feature extraction module uses convolutional layers to extract shallow features, which are directly transmitted to the reconstruction module to preserve low-frequency information. The deep feature extraction module is mainly composed of residual attention modules. We design the CEA in the residual attention module to enhance the network’s feature reconstruction on channel attention. Additionally, we introduce the NAT [19] into the network to help the model better focus on leaf detail features. CEA and NAT are symmetrically integrated within the attention mechanism to capture complementary information from both the channel and spatial dimensions, reflecting a balanced and structured design. Finally, a convolution layer is added to the end of the block for feature enhancement. In the image reconstruction layer, the Pixel-Shuffle method is used to merge shallow and deep features. The experimental results show that our network performs well in the SR reconstruction network, and the images reconstructed after SR significantly improve the mangrove image segmentation performance.The contributions of this study are as follows:
(1)
Designed a CEA channel attention module, which combines ECA attention to enhance the network’s feature extraction on channel attention.
(2)
Introduced the NAT module into the SR network to enhance the network’s ability to extract leaf detailed features.
(3)
The images reconstructed using an SR reconstruction network enhance the segmentation performance of mangrove species in UAV remote sensing images.

2. Methods

2.1. Network Architecture

To improve the species recognition performance of mangrove UAV remote sensing images, we propose a novel super-resolution reconstruction network based on an improved SwinIR architecture. As illustrated in Figure 1, the overall framework consists of three main components: shallow feature extraction, deep feature extraction, and image reconstruction. In the shallow feature extraction stage, a convolutional layer is applied to extract low-level features from the input image. The deep feature extraction stage leverages a series of Residual Mixed Attention Modules (RMAMs) to enhance texture and structural information. Each RMAM contains six Mixed Attention Modules (MAMs), which symmetrically integrate both CEA and NAT to capture channel dependencies and the detailed textural features of mangrove leaf structures from complementary perspectives. Finally, the image reconstruction stage fuses the extracted shallow and deep features and upsamples them using pixel shuffle to generate the high-resolution output.
Specifically, in the shallow feature extraction stage, similarly to SwinIR, a convolutional layer is used to extract shallow features to address edge and texture feature extraction in remote sensing images. This provides more effective input for subsequent feature extraction processing.
In the deep feature extraction stage, this study designed RMAMs to extract deep features F D R H × W × C from shallow features. The formula is as follows:
F D = H D ( F S )
where H D ( · ) denotes the deep feature extraction module.
Specifically, the final deep feature F D is extracted by each residual attention module and the final convolutional layer. The formula is as follows:
F i = H R M A M ( F i ) , i = 1 , 2 , 3 , · · · , 6 .
where H R M A M ( · ) represents the i-th RMAM module.
The RMAM is mainly composed of the MAM, and the MAM is mainly composed of the CEA module and the NAT module. Given an input feature F k , the formula is as follows:
F n = L N ( F k )
F m = C E A ( F n ) + N A T ( F n ) + F k
F i = M L P ( L N ( F m ) ) + F m
where F n and F m are intermediate features.
In the image reconstruction layer, the Pixel-Shuffle method is used to merge shallow and deep features. The upsampling method helps effectively retain information for image detail recovery, thereby improving the quality of image reconstruction.

2.2. Residual Mixed Attention Module

The RMAM mainly consists of MAM, which integrate the CEA and NAT modules in a symmetric structure to extract complementary features from the channel and spatial dimensions.
The CEA module, as shown in Figure 1, is designed to enhance the model’s capability in capturing rich channel-wise features, thereby improving the extraction of textural features of mangrove leaves during super-resolution reconstruction. The CEA module strengthens the representation of the important textural and structural features of mangrove canopies, which contributes to generating clearer and more informative high-resolution images.
In particular, the initial convolutional layers extract richer and more abstract features from the input feature maps while simultaneously reducing their channel dimensions. Nonlinear activation functions are applied after each convolution to introduce nonlinearity, enabling the model to capture more complex patterns and relationships. The subsequent convolutional layers further refine the features and apply Batch Normalization to stabilize training and accelerate convergence. Finally, the ECA module introduces a lightweight yet effective channel attention mechanism that adaptively emphasizes informative channels, enhancing the accuracy and efficiency of feature representation. Through the integration of these architectural components, the module significantly improves the extraction of fine-grained venation patterns and structural details, thereby enhancing the super-resolution reconstruction quality of UAV-based mangrove remote sensing imagery.
In addition, NAT is integrated into the network to enhance feature extraction and detail capture capabilities, thereby improving the modeling of fine spatial details. Given the complex and diverse texture patterns present in mangrove images, a detail-sensitive mechanism like NAT is particularly effective in enhancing the reconstruction quality of mangrove leaf textures.
Specifically, NAT employs an innovative neighborhood attention mechanism that enables the more flexible and localized processing of each pixel, while maintaining high computational efficiency. This allows the model to more accurately capture and reconstruct fine image details. Moreover, NAT preserves translational invariance, which is essential for maintaining spatial consistency in mangrove images. At the core of NAT is its ability to restrict self-attention to the local neighborhood of each pixel, rather than computing attention globally across the entire image. This approach significantly reduces computational complexity while retaining critical spatial information.Additionally, NAT inherits the local perceptual capabilities of traditional convolutional networks, enabling the model to be more sensitive and accurate in capturing subtle structural variations. This design makes NAT particularly well-suited for UAV-based mangrove super-resolution reconstruction tasks, where a balance between efficiency and detail fidelity is essential. By incorporating NAT, the network gains a more powerful and flexible means of enhancing complex textures and fine-grained features in high-resolution mangrove imagery.

3. Dataset

The study area selected for this research is the Shankou Mangrove Nature Reserve located in Hepu County, Beihai City, Guangxi, China. It is primarily distributed along the eastern and western coastal zones of the Shatian Peninsula, covering an area of approximately 8000 hectares. The reserve is situated in a humid subtropical climate zone of southern Asia and consists of marine areas, land areas, and extensive intertidal zones on both the eastern and western sides of the peninsula. The location of the study area is shown in Figure 2.
Throughout the aerial imaging process, this study carefully considered natural environmental factors that could interfere with image acquisition to ensure image quality and data consistency. In particular, attention was paid to weather conditions, wind speed, and tidal variations. Aerial missions were preferentially scheduled during periods of overcast weather with soft lighting and low wind conditions to reduce issues such as overexposure, shadow interference, and abnormal image contrast caused by direct sunlight. Additionally, low wind speeds contributed to the stable flight of the UAV platform and minimized the shaking of mangrove leaves, thereby improving image clarity and spatial coverage accuracy and avoiding blurring or displacement caused by motion.
To obtain the data, this study used DJI Mavic series drones equipped with high-resolution RGB camera sensors to capture images at a flying height of 10 m. A flight altitude of 10 m was chosen to reduce the impact of UAV airflow on mangrove leaves. During the data processing stage, with the assistance of mangrove species experts, the research team carefully selected 500 high-quality mangrove images from the original UAV imagery. These images have a resolution of 5280 × 3956 pixels, resulting in a Ground Sampling Distance (GSD) of approximately 2.78 mm per pixel, as shown in Figure 3. To construct a high-quality dataset suitable for super-resolution image training and to ensure the preservation of edge regions during cropping, a sliding window strategy was employed. Specifically, each image was divided into multiple sub-images, each with a size of 480 × 480 pixels. These images cover a variety of typical mangrove species, including Bruguiera gymnorrhiza, Rhizophora stylosa, and Aegiceras corniculatum, featuring rich species diversity and structural characteristics. They provide a comprehensive representation of the ecological composition of the mangrove forests within the study area.

4. Experiments

4.1. Experimental Setup

In our experiments, this study employed the AdamW optimization algorithm with β 1 set to 0.9, β 2 set to 0.999, and a batch size of 32. In this study, the size N of the network is set to 6, and the initial learning rate is set to 2.5 × 10 4 ; we introduced a weight decay rate of 0.01 to effectively prevent overfitting. During the training phase, this study utilized the MultiStepLR strategy to efficiently adjust the learning rate at different training stages. For example, in the case of a scaling factor of 2, the model underwent a total of 1,500,000 iterations to ensure sufficient learning and optimization. Specifically, we multiplied the learning rate by 0.5 at training steps 300,000, 500,000, 700,000, 1,000,000, 1,300,000, and 1,450,000 to achieve timely decay, helping the model maintain optimal performance at different training stages.

4.2. Results on Image SR

Table 1 shows the comparison of SwinNET with FSRCNN [20], VDSR [21], EDSR [22], RDN [23], RCAN [24], DRLN [25], SAN [26], IGNN [27], ELAN [28], and SwinIR on the dataset of UAV mangrove remote sensing images. The experimental results demonstrate that at scaling factors of ×2, ×3, and ×4, our method achieves 34.04/94.43, 30.59/88.55, and 28.52/83.35, respectively. Compared to SwinIR, our method achieves performance improvements of +1.39/3.19, +0.18/0.10, and +0.04/1.66. To better illustrate the comparison results, the visual comparisons corresponding to Table 1 are presented in Figure 4.
Table 2 presents a performance comparison between the proposed method, SwinNET, and a series of other methods on the DF2K (DIV2K and Flicker2K [29]) dataset, including traditional interpolation methods, early classic SR algorithms, and current mainstream methods such as SwinIR. The experimental results show that SwinNET achieves excellent PSNR (dB) and SSIM (%) across all test datasets. At a scaling factor of ×4, SwinNET’s PSNR/SSIM values on the Set5, Set14, BSD100, Urban100, Manga109, and mangrove datasets are 32.79/90.32, 29.23/79.54, 28.13/75.43, 27.84/83.22, 32.16/92.61, and 28.52/83.35, respectively. Compared to SwinIR, our method achieves performance improvements of +0.14/0.08, +0.18/0.10, +0.24/0.55, +0.38/0.70, and +0.11/0.13 on these datasets. Our method also performs well at other scaling factors.

4.3. Visual Comparison

A visual comparison of the super-resolution results of six mangrove UAV images is presented in Figure 5. The SwinNET-based approach demonstrates superior performance in restoring fine textures and structural details by effectively leveraging both channel and neighborhood features. In particular, the reconstructed results show more accurate texture recovery and more complete structural information, resulting in clearer and more natural visual quality compared to other models. As shown in Figure 5a, the results from other methods appear noticeably blurred and fail to reconstruct the leaf edges, whereas the proposed method successfully preserves fine boundary details and the intricate structure of the mangrove canopy.

4.4. Results of Mangrove Species Classification

The mangrove images captured by the drone were processed using the labeling tool LabelMe. Under the guidance of mangrove species experts, this study meticulously annotated different tree species in the images, such as Rhizophora stylosa, Avicennia marina, and Bruguiera gymnorrhiza. To meet the input requirements of the deep learning model, this study cropped the selected regions of the images to 512 × 512 pixels and used a 128-pixel sliding window method to extract 1320 images. Among these, 924 images were used for training, 132 for testing, and 264 for validation.
To evaluate the enhancement effect of super-resolution reconstruction on the original images, this study employed the classical and relatively simple segmentation network FPN for comparative experiments. The purpose was not to pursue state-of-the-art segmentation accuracy but rather to isolate and demonstrate the improvement brought by the SR reconstruction itself.
The experimental results are shown in Table 3, the performance of the FPN network in the segmentation task of different mangrove tree species is significantly different. Under the conditions of original HR images and SR images that are enlarged four times, the segmentation performance of Rhizophora stylosa, Avicennia marina, and Bruguiera gymnorrhiza all improved. For Rhizophora stylosa, the Intersection over Union(IoU) was 91.85% and accuracy (ACC) was 97.79% in the original HR image, which increased to 93.42% IoU and 98.34% accuracy in the ×4 image. The segmentation performance of Avicennia marina improved significantly, with an IoU of 66.35% and ACC of 69.09% in the original HR image, which increased to 74.13% IoU and 80.26% ACC in the ×4 image. Bruguiera gymnorrhiza had an IoU of 74.36% and ACC of 74.64% in the original HR image, which increased to 95.56% IoU and 96.86% ACC in the ×4 image.
The experimental results indicate that using ×4 enlarged images significantly improves the performance of the FPN network in mangrove species segmentation tasks, especially in the segmentation of Avicennia marina and Bruguiera gymnorrhiza, which show particularly outstanding performance. Especially for Bruguiera gymnorrhiza, SR processing significantly improved the segmentation effect, indicating that reasonable SR processing can significantly improve the segmentation performance of mangrove images.

5. Conclusions

This study proposes a method for mangrove species recognition based on drone remote sensing images. By combining SR reconstruction technology with deep learning, the resolution of mangrove images is improved, enabling the accurate identification of mangrove species. Our model, based on the SwinIR-improved SwinNET for mangrove drone remote sensing image reconstruction, thoroughly exploits the channel and neighborhood features of images, accurately restoring the details and textures of the original high-resolution images. The experimental results of the drone-based mangrove remote sensing image dataset show that SwinNET achieves high performance at scale factors of 2×, 3×, and 4×, with improvements over SwinIR, demonstrating better SR reconstruction effects. Additionally, this study conducted mangrove species segmentation experiments using the classic segmentation network FPN, which shows significantly improved performance in the segmentation tasks of Rhizophora stylosa, Avicennia marina, and Bruguiera gymnorrhiza when using 4× enlarged images. The proposed method achieved certain results in the processing and species recognition of mangrove drone remote sensing images, providing reference for the protection and restoration of mangrove ecosystems.
Although this study has achieved promising results, it is limited by the scope of the study area and the inability to distinguish all mangrove species, particularly under complex environmental conditions such as adverse weather. Additionally, the generalizability of the model to other mangrove sites with different ecological characteristics remains to be validated. Moreover, this study only employed the classical FPN segmentation network for species classification, which may limit the comprehensiveness of the evaluation. In future work, we plan to integrate multispectral imagery and other complementary data sources to enhance species classification and improve model robustness. We also acknowledge the need to evaluate our approach using additional segmentation models, such as U-Net and DeepLabV3+, to further validate the effectiveness of super-resolution reconstruction across a broader range of architectures. Furthermore, the potential for real-time, onboard UAV processing will be explored to enable more efficient and responsive monitoring of mangrove ecosystems.

Author Contributions

Conceptualization, Q.Q. and W.D.; methodology, W.D.; software, Q.Q. and W.D.; validation, Q.Q. and W.D.; formal analysis, Q.Q.; investigation, Q.Q.; resources, X.W.; data curation, X.W.; writing—original draft preparation, W.D.; writing—review and editing, W.D.; visualization, Q.Q.; supervision, X.W.; project administration, Q.Q.; funding acquisition, X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The mangrove datasets, implementation code, and trained models for this study are available from the corresponding author upon reasonable request.

Acknowledgments

We sincerely thank the anonymous reviewers for their valuable insights and constructive comments, which have greatly enhanced the quality and clarity of this paper. We are also very grateful to all individuals and institutions that have provided support and assistance throughout the research process. We especially thank the Institute of Marine Electronics and Information Technology, South Campus of Guilin University of Electronic Technology, for providing an excellent experimental environment and important resource support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jia, M.; Liu, M.; Wang, Z.; Mao, D.; Ren, C.; Cui, H. Evaluating the effectiveness of conservation on mangroves: A remote sensing-based comparison for two adjacent protected areas in Shenzhen and Hong Kong, China. Remote Sens. 2016, 8, 627. [Google Scholar] [CrossRef]
  2. Shiwen, W.; Hongchang, H.; Bolin, F. Analysis of physiological structure parameters of Shankou mangrove based on Sentinel-2 data and space-time characteristics. Sci. Technol. Eng. 2021, 21, 3698–3707. [Google Scholar]
  3. Kabiri, K.; Abedi, E. Rapid mangrove dieback in the northern Persian Gulf driven by anthropogenic activities and environmental stressors. Discov. Environ. 2025, 3, 22. [Google Scholar] [CrossRef]
  4. Cao, J.; Leng, W.; Liu, K.; Liu, L.; He, Z.; Zhu, Y. Object-based mangrove species classification using unmanned aerial vehicle hyperspectral images and digital surface models. Remote Sens. 2018, 10, 89. [Google Scholar] [CrossRef]
  5. Maurya, K.; Mahajan, S.; Chaube, N. Remote sensing techniques: Mapping and monitoring of mangrove ecosystem—A review. Complex Intell. Syst. 2021, 7, 2797–2818. [Google Scholar] [CrossRef]
  6. Kai, L.; Hui, G.; Jingjing, C.; Yuanhui, Z. Comparison of mangrove remote sensing classification based on multi-type UAV data. Trop. Geogr. 2019, 39, 492–501. [Google Scholar]
  7. You, H.; Liu, Y.; Lei, P.; Qin, Z.; You, Q. Segmentation of individual mangrove trees using UAV-based LiDAR data. Ecol. Inform. 2023, 77, 102200. [Google Scholar] [CrossRef]
  8. Wang, X.; Zhang, Y.; Ca, J.; Qin, Q.; Feng, Y.; Yan, J. Semantic segmentation network for mangrove tree species based on UAV remote sensing images. Sci. Rep. 2024, 14, 29860. [Google Scholar] [CrossRef] [PubMed]
  9. Kabiri, K. Mapping coastal ecosystems and features using a low-cost standard drone: Case study, Nayband Bay, Persian gulf, Iran. J. Coast. Conserv. 2020, 24, 62. [Google Scholar] [CrossRef]
  10. Wen, X.; Jia, M.; Li, X.; Wang, Z.; Zhong, C.; Feng, E. Identification of mangrove canopy species based on visible unmanned aerial vehicle images. J. For. Environ. 2020, 40, 486–496. [Google Scholar]
  11. Jin, B.; Gonçalves, N.; Cruz, L.; Medvedev, I.; Yu, Y.; Wang, J. Simulated multimodal deep facial diagnosis. Expert Syst. Appl. 2024, 252, 123881. [Google Scholar] [CrossRef]
  12. Cheng, Y.; Yan, J.; Zhang, F.; Li, M.; Zhou, N.; Shi, C.; Jin, B.; Zhang, W. Surrogate modeling of pantograph-catenary system interactions. Mech. Syst. Signal Process. 2025, 224, 112134. [Google Scholar] [CrossRef]
  13. Yan, J.; Cheng, Y.; Wang, Q.; Liu, L.; Zhang, W.; Jin, B. Transformer and graph convolution-based unsupervised detection of machine anomalous sound under domain shifts. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 2827–2842. [Google Scholar] [CrossRef]
  14. Zhang, Q.; Yang, G.; Zhang, G. Collaborative network for super-resolution and semantic segmentation of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4404512. [Google Scholar] [CrossRef]
  15. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  16. Vaswani, A. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  17. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  18. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  19. Hassani, A.; Walton, S.; Li, J.; Li, S.; Shi, H. Neighborhood attention transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 6185–6194. [Google Scholar]
  20. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  21. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  22. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  23. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2472–2481. [Google Scholar]
  24. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  25. Anwar, S.; Barnes, N. Densely residual laplacian super-resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1192–1204. [Google Scholar] [CrossRef] [PubMed]
  26. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 11065–11074. [Google Scholar]
  27. Zhou, S.; Zhang, J.; Zuo, W.; Loy, C.C. Cross-scale internal graph neural network for image super-resolution. Adv. Neural Inf. Process. Syst. 2020, 33, 3499–3509. [Google Scholar]
  28. Zhang, X.; Zeng, H.; Guo, S.; Zhang, L. Efficient long-range attention network for image super-resolution. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 649–667. [Google Scholar]
  29. Timofte, R.; Agustsson, E.; Van Gool, L.; Yang, M.H.; Zhang, L. Ntire 2017 challenge on single image super-resolution: Methods and results. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 114–125. [Google Scholar]
Figure 1. Overall architecture of the proposed SwinNET. The network consists of shallow feature extraction, deep feature extraction, and image reconstruction layers.
Figure 1. Overall architecture of the proposed SwinNET. The network consists of shallow feature extraction, deep feature extraction, and image reconstruction layers.
Symmetry 17 01250 g001
Figure 2. Location of Shankou Mangrove Nature Reserve. (a) The location of Guangxi in China. (b) The location of Shankou Mangrove Nature Reserve in Guangxi. (c) A detailed location of Shankou Mangrove Nature Reserve.
Figure 2. Location of Shankou Mangrove Nature Reserve. (a) The location of Guangxi in China. (b) The location of Shankou Mangrove Nature Reserve in Guangxi. (c) A detailed location of Shankou Mangrove Nature Reserve.
Symmetry 17 01250 g002
Figure 3. Several representative mangrove canopy images acquired from Shankou Mangrove Nature Reserve.
Figure 3. Several representative mangrove canopy images acquired from Shankou Mangrove Nature Reserve.
Symmetry 17 01250 g003
Figure 4. Quantitative comparison of SSIM (a) and PSNR (b) values for different SR reconstruction methods at scaling factors of ×2, ×3, and ×4 on the UAV mangrove remote sensing dataset.
Figure 4. Quantitative comparison of SSIM (a) and PSNR (b) values for different SR reconstruction methods at scaling factors of ×2, ×3, and ×4 on the UAV mangrove remote sensing dataset.
Symmetry 17 01250 g004
Figure 5. Visual comparison on ×4 SR.A total of six mangrove image patches, labeled from (af), were selected for visual comparison. The regions used for comparison in the original image are marked with red boxes.
Figure 5. Visual comparison on ×4 SR.A total of six mangrove image patches, labeled from (af), were selected for visual comparison. The regions used for comparison in the original image are marked with red boxes.
Symmetry 17 01250 g005
Table 1. Comparison between SwinNET and other SR reconstruction algorithms in the UAV remote sensing mangrove SR reconstruction dataset.
Table 1. Comparison between SwinNET and other SR reconstruction algorithms in the UAV remote sensing mangrove SR reconstruction dataset.
MethodTraining
Dataset
×2×3×4
PSNR (dB) SSIM (%) PSNR (dB) SSIM (%) PSNR (dB) SSIM (%)
FSRCNNmangrove31.3190.1928.1685.9226.0779.30
VDSRmangrove31.8790.6428.5786.0226.5679.97
EDSRmangrove32.2591.1129.0386.3826.9180.11
RDNmangrove32.4490.7028.6286.8126.6679.61
RCANmangrove32.5791.0829.1287.3026.9080.02
DRLNmangrove32.3891.2329.7588.3926.8180.16
SANmangrove32.0690.1130.1389.4526.8680.21
IGNNmangrove32.1390.6530.0489.5526.7381.33
ELANmangrove32.1790.8030.3290.0726.6482.09
SwinIRmangrove32.6591.2430.5590.3227.0282.09
SwinNET (ours)mangrove34.0494.4330.5990.2128.5283.35
The best results have been marked in bold.
Table 2. Comparison of SwinNET and other SR reconstruction algorithms in DF2K dataset.
Table 2. Comparison of SwinNET and other SR reconstruction algorithms in DF2K dataset.
MethodScaleSet5Set14BSDS100Urban100Manga109
PSNR SSIM PSNR SSIM PSNR SSIM PSNR SSIM PSNR SSIM
SRCNN×236.5995.3132.4490.6131.3188.7229.4989.4135.5196.60
FSRCNN36.9595.5532.5990.9031.4489.1029.7890.1236.6197.02
VDSR37.4795.7933.0291.2131.8889.5330.6891.3237.1697.40
EDSR37.0395.9633.8991.9532.2790.1232.8293.4039.0097.68
RDN37.6296.0333.5192.0632.3390.1432.8293.4839.0697.62
RCAN37.2696.1433.0892.0732.3790.2533.2893.7639.3997.58
DRLN37.2496.1033.2292.0532.3490.2533.3093.8239.1297.79
SAN37.8896.1233.6792.0232.4090.2033.0893.6339.2297.85
IGNN38.0396.1033.0791.8732.3990.1533.2393.7339.2697.78
ELAN38.0996.1133.1792.0332.3890.3033.2693.8639.4197.84
SwinIR38.1596.1233.3392.0532.3690.2733.3193.2339.1297.89
SwinNET38.1096.1533.9192.0632.9791.3333.3594.1939.4498.02
SRCNN×332.7290.8429.2582.0628.3678.6126.2279.8730.4491.14
FSRCNN33.0991.3329.3182.0828.5079.0526.4380.7631.0992.04
VDSR33.5992.1029.6883.1528.8179.8427.1482.9032.0193.34
EDSR34.1492.1330.1384.1229.2280.8827.7384.4733.1693.69
RDN34.1692.1830.2384.1329.2280.9127.7584.4933.0793.74
RCAN34.2392.3930.1884.1329.3181.0827.0784.9533.4493.99
DRLN34.1893.0130.1484.1029.3381.1727.1584.1233.6593.09
SAN34.1392.1930.2184.1329.2881.0227.8484.6133.2393.90
IGNN34.1592.3230.2484.1829.2480.9527.9484.8633.3894.92
ELAN34.1593.0730.2584.1929.3581.2328.3084.3833.6993.09
SwinIR34.2092.3530.2484.1329.2080.3828.2684.1433.3893.89
SwinNET34.3592.6230.3084.2029.3782.3128.5185.8833.7694.58
SRCNN×430.3886.2727.4375.0926.8070.9524.4372.1527.5485.49
FSRCNN30.7286.5127.5575.4526.9371.4724.5872.7727.8486.08
VDSR31.3188.2527.9376.7527.2372.2625.0975.3028.7288.60
EDSR32.4383.5928.7078.7327.6474.1026.5680.3030.9591.42
RDN32.3989.8428.8078.6427.6774.1626.6080.1930.9991.43
RCAN32.5689.9528.8578.8927.7774.3426.7680.8131.2091.65
DRLN32.5590.0128.8678.9127.8074.4026.9481.1431.4491.89
SAN32.5790.0028.8978.8427.7774.2826.7080.6131.1291.64
IGNN32.4789.9828.7578.8527.7574.3426.8380.8231.2791.81
ELAN32.6490.1828.8879.0627.7774.5227.1381.5831.6192.22
SwinIR32.6590.2429.0579.4427.8974.8827.4682.5232.0592.48
SwinNET32.7990.3229.2379.5428.1375.4327.8483.2232.1692.61
The best results are marked in bold.
Table 3. Segmentation results of drone-based HR original images and SR reconstructed mangrove remote sensing images.
Table 3. Segmentation results of drone-based HR original images and SR reconstructed mangrove remote sensing images.
HR ImagesSwinNET
ClassIOUACCIOUACC
Rhizophora stylosa91.8597.7993.4298.34
Bruguiera gymnorhiza66.3569.0974.1380.26
Aegiceras corniculata74.3674.6495.5696.86
The three classified mangrove species are Rhizophora stylosa, Avicennia marina, and Bruguiera gymnorrhiza.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, Q.; Dai, W.; Wang, X. Super Resolution for Mangrove UAV Remote Sensing Images. Symmetry 2025, 17, 1250. https://doi.org/10.3390/sym17081250

AMA Style

Qin Q, Dai W, Wang X. Super Resolution for Mangrove UAV Remote Sensing Images. Symmetry. 2025; 17(8):1250. https://doi.org/10.3390/sym17081250

Chicago/Turabian Style

Qin, Qin, Wenlong Dai, and Xin Wang. 2025. "Super Resolution for Mangrove UAV Remote Sensing Images" Symmetry 17, no. 8: 1250. https://doi.org/10.3390/sym17081250

APA Style

Qin, Q., Dai, W., & Wang, X. (2025). Super Resolution for Mangrove UAV Remote Sensing Images. Symmetry, 17(8), 1250. https://doi.org/10.3390/sym17081250

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop