Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (37)

Search Parameters:
Keywords = ship target perception

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 3734 KB  
Article
CLEAR: A Cognitive LLM-Empowered Adaptive Restoration Framework for Robust Ship Detection in Complex Maritime Scenarios
by Min Li, Xinyu Zhao and Yunfeng Wan
Remote Sens. 2026, 18(8), 1142; https://doi.org/10.3390/rs18081142 - 12 Apr 2026
Viewed by 431
Abstract
Ship detection in remote sensing imagery serves as a cornerstone of modern maritime surveillance. Existing visible light detectors suffer from severe performance degradation in adverse environmental conditions (e.g., fog, low light) due to domain gaps. Traditional global enhancement methods often lack adaptability, leading [...] Read more.
Ship detection in remote sensing imagery serves as a cornerstone of modern maritime surveillance. Existing visible light detectors suffer from severe performance degradation in adverse environmental conditions (e.g., fog, low light) due to domain gaps. Traditional global enhancement methods often lack adaptability, leading to “negative transfer”—where artifacts are introduced into clean images or mismatched with degradation types. To address these challenges, we propose CLEAR (Cognitive Large Language Model (LLM)-Empowered Adaptive Restoration) framework. Inspired by the dual-process theory of cognition, we introduce a dynamic switching mechanism between fast perception and deep reasoning. Rather than processing all images indiscriminately, it utilizes a hybrid gating mechanism to efficiently filter nominal samples, triggering Vision–Language Model (VLM) only when necessary to diagnose degradation and dispatch targeted restoration operators. Extensive experiments on the constructed HRSC-Robust dataset demonstrate that CLEAR achieves an overall mean Average Precision (mAP) at 0.5 Intersection-over-Union (IoU) of 86.92%, outperforming the baseline by 7.74%. Notably, it establishes a “fail-safe” mechanism for optical degradations. By adaptively resolving fog and low-light, it effectively mitigates detector blindness—exemplified by a doubled Recall rate (52.52%) in dark scenarios. Furthermore, a confidence-based sparse triggering strategy ensures operational efficiency, maintaining a throughput of ~11.8 FPS in nominal conditions. This work validates the potential of VLMs for interpretable and robust remote sensing tasks. Full article
Show Figures

Figure 1

25 pages, 6369 KB  
Article
A Lightweight Attention-Guided and Geometry-Aware Framework for Robust Maritime Ship Detection in Complex Electro-Optical Environments
by Zhe Zhang, Chang Lin and Bing Fang
Automation 2026, 7(2), 48; https://doi.org/10.3390/automation7020048 - 12 Mar 2026
Viewed by 434
Abstract
Reliable ship detection in complex maritime optical imagery is a fundamental requirement for intelligent maritime monitoring and maritime automation systems. However, severe image degradation, large-scale variations, and background clutter often lead to feature ambiguity and unstable detection performance in real-world maritime environments. To [...] Read more.
Reliable ship detection in complex maritime optical imagery is a fundamental requirement for intelligent maritime monitoring and maritime automation systems. However, severe image degradation, large-scale variations, and background clutter often lead to feature ambiguity and unstable detection performance in real-world maritime environments. To address these challenges, this paper proposes a lightweight one-stage ship detection framework designed for robust real-time perception under degraded maritime sensing conditions. The proposed method incorporates an Adaptive Expert Selection Attention (AESA) mechanism to perform adaptive feature selection and background suppression under visually degraded conditions, together with a Geometry-Aware MultiScale Fusion (GAMF) module that enables orientation-aware aggregation of contextual information for elongated ship targets near complex sea–sky boundaries. In addition, a geometry-aware bounding box regression refinement is introduced to improve localization consistency in image space. Extensive experiments conducted on a unified real-world maritime benchmark demonstrate that the proposed framework consistently outperforms the baseline YOLO11n model by approximately 2–5 percentage points in terms of mAP@0.5 and mAP@0.5:0.95, while maintaining moderate computational complexity and real-time inference capability. These results indicate that the proposed method provides a practical and deployment-oriented perception solution for maritime automation applications, including onboard electro-optical sensing and coastal surveillance. Full article
Show Figures

Figure 1

27 pages, 12041 KB  
Article
FPGA-Based CNN Acceleration on Zynq-7020 for Embedded Ship Recognition in Unmanned Surface Vehicles
by Abdelilah Haijoub, Aissam Bekkari, Anas Hatim, Mounir Arioua, Mohamed Nabil Srifi and Antonio Guerrero-Gonzalez
Sensors 2026, 26(5), 1626; https://doi.org/10.3390/s26051626 - 5 Mar 2026
Viewed by 629
Abstract
Unmanned surface vehicles (USVs) increasingly rely on vision-based perception for safe navigation and maritime surveillance, while onboard computing is constrained by strict size, weight, and power (SWaP) budgets. Although deep convolutional neural networks (CNNs) offer strong recognition performance, their computational and memory requirements [...] Read more.
Unmanned surface vehicles (USVs) increasingly rely on vision-based perception for safe navigation and maritime surveillance, while onboard computing is constrained by strict size, weight, and power (SWaP) budgets. Although deep convolutional neural networks (CNNs) offer strong recognition performance, their computational and memory requirements pose significant challenges for deployment on low-cost embedded platforms. This paper presents a hardware–software co-design architecture and deployment study for CNN acceleration on a heterogeneous ARM–FPGA system, targeting energy-efficient near-sensor processing for embedded maritime applications. The proposed approach exploits a fully streaming hardware architecture in the FPGA fabric, based on line-buffered convolutions and AXI-Stream dataflow, while the ARM processing system is responsible for lightweight configuration, scheduling, and data movement. The architecture was evaluated using representative CNN models trained on a maritime ship dataset. Our experimental results on a Zynq-7020 system-on-chip demonstrate that the proposed co-design strategy achieves a balanced trade-off between throughput, resource utilisation, and power consumption under tight embedded constraints, highlighting its suitability as a practical building block for onboard perception in USVs. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

23 pages, 54902 KB  
Article
RSAND: A Fine-Grained Dataset and Benchmark for AtoN Detection in River–Sea Intermodal and Complex Estuarine Environments
by Qi Chen, Mingyang Pan, Zongying Liu, Ruolan Zhang, Fei Yan, Xiaofeng Pan, Yang Zhang and Chao Li
J. Mar. Sci. Eng. 2026, 14(5), 422; https://doi.org/10.3390/jmse14050422 - 25 Feb 2026
Viewed by 383
Abstract
Robust visual perception of Aids to Navigation (AtoN) is essential for Maritime Autonomous Surface Ships (MASS) operating in restricted navigational waters, where estuarine clutter, fog, glare, and dense traffic can severely degrade detection reliability. Existing maritime vision datasets largely emphasize open-sea targets or [...] Read more.
Robust visual perception of Aids to Navigation (AtoN) is essential for Maritime Autonomous Surface Ships (MASS) operating in restricted navigational waters, where estuarine clutter, fog, glare, and dense traffic can severely degrade detection reliability. Existing maritime vision datasets largely emphasize open-sea targets or coarse AtoN categories, leaving a granularity gap for IALA-compliant fine-grained understanding in river–sea transition and port-approach channels. The River–Sea AtoN Navigation Dataset (RSAND) is introduced, a large-scale benchmark collected along the Yangtze River Deepwater Channel from inland corridors to open estuarine waters. RSAND contains 39,926 images with expert-verified bounding-box annotations and a hierarchical taxonomy that jointly captures AtoN location, shape, and functional semantics across 29 categories. To support both realistic long-tailed evaluation and standardized model comparison, two protocols are provided: RSAND-Full (29 categories) and RSAND-Balanced (10 critical categories). All quantitative benchmarking results in this paper are reported on RSAND-Balanced, while RSAND-Full is released for future large-scale long-tailed robustness studies. Benchmarking experiments on 14 state-of-the-art detectors demonstrate that YOLOv12x achieves superior performance with an mAP50-95 of 80.7%, significantly outperforming previous baselines. However, the analysis reveals persistent challenges in detecting small, distant targets and distinguishing visually similar functional markers. RSAND and the accompanying evaluation toolkit are released to facilitate reproducible research toward safer and smarter marine and coastal navigation. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

23 pages, 3059 KB  
Article
Research on Ship Target Detection in Complex Sea Surface Scenarios Based on Improved YOLOv7
by Zhuang Cai and Weina Zhou
Appl. Sci. 2026, 16(4), 1769; https://doi.org/10.3390/app16041769 - 11 Feb 2026
Viewed by 444
Abstract
Ships target detection plays a crucial role in safeguarding maritime transportation. However, affected by factors such as ocean waves, extreme weather, and target diversity (e.g., large size differences, arbitrary rotation, and occlusion), existing deep learning-based detection methods struggle to achieve a satisfactory balance [...] Read more.
Ships target detection plays a crucial role in safeguarding maritime transportation. However, affected by factors such as ocean waves, extreme weather, and target diversity (e.g., large size differences, arbitrary rotation, and occlusion), existing deep learning-based detection methods struggle to achieve a satisfactory balance among accuracy, speed, and model size in complex marine environments. To address this challenge, this paper proposes a real-time ship detection algorithm (C-YOLO) integrating global perception and multi-scale feature enhancement. First, a Transformer encoder is added before the detection head, which suppresses interference from sea clutter and cloud mist occlusion through long-range dependency modeling, improving the detection of small and occluded ships. Second, a Dual-Effect Focused Residual Fusion Module is designed to replace the backbone’s multi-scale pooling structure, combining the advantages of CBAM (background noise suppression) and SK-Net (dynamic scale adaptation) to simultaneously capture features of ships of different sizes. Finally, a CZIoU loss function is proposed, which integrates constraints on angle, center point, vertex, and area to address rotation, deformation, and multi-scale issues in ship detection. Experimental results on the SeaShips 7000 dataset show that the proposed C-YOLO achieves a Recall of 0.842, mAP@50 of 0.797, and mAP@50:95 of 0.552, outperforming mainstream algorithms such as YOLOv7 (Recall = 0.785, mAP@50 = 0.781), YOLOv9s (Recall = 0.819, mAP@50 = 0.755), and SSD (Recall = 0.802, mAP@50 = 0.833). With 76.75 M parameters and an inference speed of 119 FPS, the model maintains efficient real-time performance while ensuring detection accuracy. This method effectively reduces false detection and missed detection rates in complex scenarios such as port monitoring and maritime traffic control, providing a reliable technical solution for intelligent maritime surveillance and safe navigation—with significant practical value for improving maritime transportation efficiency and reducing safety risks. Full article
(This article belongs to the Special Issue Advances in Computer Vision and Digital Image Processing)
Show Figures

Figure 1

21 pages, 5664 KB  
Article
M2S-YOLOv8: Multi-Scale and Asymmetry-Aware Ship Detection for Marine Environments
by Peizheng Li, Dayong Qiao, Jianyi Mu and Linlin Qi
Sensors 2026, 26(2), 502; https://doi.org/10.3390/s26020502 - 12 Jan 2026
Viewed by 541
Abstract
Ship detection serves as a core foundational task for marine environmental perception. However, in real marine scenarios, dense vessel traffic often causes severe target occlusion while multi-scale targets, asymmetric vessel geometries, and harsh conditions (e.g., haze, low illumination) further degrade image quality. These [...] Read more.
Ship detection serves as a core foundational task for marine environmental perception. However, in real marine scenarios, dense vessel traffic often causes severe target occlusion while multi-scale targets, asymmetric vessel geometries, and harsh conditions (e.g., haze, low illumination) further degrade image quality. These factors pose significant challenges to vision-based ship detection methods. To address these issues, we propose M2S-YOLOv8, an improved framework based on YOLOv8, which integrates three key enhancements: First, a Multi-Scale Asymmetry-aware Parallelized Patch-wise Attention (MSA-PPA) module is designed in the backbone to strengthen the perception of multi-scale and geometrically asymmetric vessel targets. Second, a Deformable Convolutional Upsampling (DCNUpsample) operator is introduced in the Neck network to enable adaptive feature fusion with high computational efficiency. Third, a Wasserstein-Distance-Based Weighted Normalized CIoU (WA-CIoU) loss function is developed to alleviate gradient imbalance in small-target regression, thereby improving localization stability. Experimental results on the Unmanned Vessel Zhoushan Perception Dataset (UZPD) and the open-source Singapore Maritime Dataset (SMD) demonstrate that M2S-YOLOv8 achieves a balanced performance between lightweight design and real-time inference, showcasing strong potential for reliable deployment on edge devices of unmanned marine platforms. Full article
(This article belongs to the Section Environmental Sensing)
Show Figures

Figure 1

27 pages, 5048 KB  
Article
MCB-RT-DETR: A Real-Time Vessel Detection Method for UAV Maritime Operations
by Fang Liu, Yongpeng Wei, Aruhan Yan, Tiezhu Cao and Xinghai Xie
Drones 2026, 10(1), 13; https://doi.org/10.3390/drones10010013 - 27 Dec 2025
Viewed by 1098
Abstract
Maritime UAV operations face challenges in real-time ship detection. Complex ocean backgrounds, drastic scale variations, and prevalent distant small targets create difficulties. We propose MCB-RT-DETR, a real-time detection transformer enhanced by multi-component boosting. This method builds upon the RT-DETR architecture. It significantly improves [...] Read more.
Maritime UAV operations face challenges in real-time ship detection. Complex ocean backgrounds, drastic scale variations, and prevalent distant small targets create difficulties. We propose MCB-RT-DETR, a real-time detection transformer enhanced by multi-component boosting. This method builds upon the RT-DETR architecture. It significantly improves detection under wave interference, lighting changes, and scale differences. Key innovations address these challenges. An Orthogonal Channel Attention (Ortho) mechanism preserves high-frequency edge details in the backbone network. Receptive Field Attention Convolution (RFAConv) enhances robustness against background clutter. A Small Object Detail Enhancement Pyramid (SOD-EPN) strengthens small-target representation. SOD-EPN combines SPDConv with multi-scale CSP-OmniKernel transformations. The neck network integrates ultra-lightweight DySample upsampling. This enables content-aware sampling for precise multi-scale localization. The method maintains high computational efficiency. Experiments on the SeaDronesSee dataset show significant improvements. MCB-RT-DETR achieves 82.9% mAP@0.5 and 49.7% mAP@0.5:0.95. These correspond to improvements of 4.5% and 3.4% relative to the baseline model. Inference speed maintains 50 FPS for real-time processing. The outstanding performance in cross-dataset tests further validates the algorithm’s strong generalization capability on DIOR remote sensing images and VisDrone2019 aerial scenes. The method provides a reliable visual perception solution for autonomous maritime UAV operations. Full article
Show Figures

Figure 1

22 pages, 12312 KB  
Article
ES-YOLO: Multi-Scale Port Ship Detection Combined with Attention Mechanism in Complex Scenes
by Lixiang Cao, Jia Xi, Zixuan Xie, Teng Feng and Xiaomin Tian
Sensors 2025, 25(24), 7630; https://doi.org/10.3390/s25247630 - 16 Dec 2025
Viewed by 731
Abstract
With the rapid development of remote sensing technology and deep learning, the port ship detection based on a single-stage algorithm has achieved remarkable results in optical imagery. However, most of the existing methods are designed and verified in specific scenes, such as fixed [...] Read more.
With the rapid development of remote sensing technology and deep learning, the port ship detection based on a single-stage algorithm has achieved remarkable results in optical imagery. However, most of the existing methods are designed and verified in specific scenes, such as fixed viewing angle, uniform background, or open sea, which makes it difficult to deal with the problem of ship detection in complex environments, such as cloud occlusion, wave fluctuation, complex buildings in the harbor, and multi-ship aggregation. To this end, ES-YOLO framework is proposed to solve the limitations of ship detection. A novel edge perception channel, Spatial Attention Mechanism (EACSA), is proposed to enhance the extraction of edge information and improve the ability to capture feature details. A lightweight spatial–channel decoupled down-sampling module (LSCD) is designed to replace the down-sampling structure of the original network and reduce the complexity of the down-sampling stage. A new hierarchical scale structure is designed to balance the detection effect of different scale differences. In this paper, a remote sensing ship dataset, TJShip, is constructed based on Gaofen-2 images, which covers multi-scale targets from small fishing boats to large cargo ships. The TJShip dataset was adopted as the data source, and the ES-YOLO model was employed to conduct ablation and comparison experiments. The results show that the introduction of EACSA attention mechanism, LSCD, and multi-scale structure improves the mAP of ship detection by 0.83%, 0.54%, and 1.06%, respectively, compared with the baseline model, also performing well in precision, recall and F1. Compared with Faster R-CNN, RetinaNet, YOLOv5, YOLOv7, and YOLOv8 methods, the results show that the ES-YOLO model improves the mAP by 46.87%, 8.14%, 1.85%, 1.75%, and 0.86%, respectively, under the same experimental conditions, which provides research ideas for ship detection. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

23 pages, 41707 KB  
Article
Ship Detection in SAR Images Using Sparse R-CNN with Wavelet Deformable Convolution and Attention Mechanism
by Zhiqiang Zeng, Zongsi Chen, Junjun Yin and Huiping Lin
Remote Sens. 2025, 17(23), 3794; https://doi.org/10.3390/rs17233794 - 22 Nov 2025
Viewed by 1202
Abstract
This paper proposes a synthetic aperture radar (SAR) ship detection method based on wavelet-domain deformable convolution (WDC) and multi-head attention, built upon the Sparse R-CNN framework. First, a wavelet-domain convolution module is introduced to enhance the modeling of ship targets with diverse scales [...] Read more.
This paper proposes a synthetic aperture radar (SAR) ship detection method based on wavelet-domain deformable convolution (WDC) and multi-head attention, built upon the Sparse R-CNN framework. First, a wavelet-domain convolution module is introduced to enhance the modeling of ship targets with diverse scales and shapes while incorporating frequency-domain information. Deformable convolution adaptively adjusts sampling locations, overcoming the limitations of traditional convolution in capturing target edges and blurred boundaries. Next, a position encoding module is employed to normalize candidate bounding box coordinates and integrate them into region-of-interest features. By providing spatial context, position encoding strengthens spatial perception and enables the subsequent multi-head attention mechanism to more effectively capture associations between targets and candidate regions, thereby improving localization accuracy under arbitrary spatial distributions. Furthermore, the original dynamic head is replaced with a multi-head attention mechanism. Through position-encoded multi-head attention, the model more accurately emphasizes regions with spatial and semantic correlations to the target, enhancing both focus and discrimination for sparse targets. Extensive experiments conducted on two benchmark datasets (SSDD and HRSID) demonstrate the effectiveness and superiority of the proposed method. Overall, the method significantly improves the detection of sparse, multi-scale, and randomly distributed ship targets in SAR images. Full article
(This article belongs to the Special Issue Microwave Remote Sensing on Ocean Observation)
Show Figures

Figure 1

26 pages, 4332 KB  
Article
CDSANet: A CNN-ViT-Attention Network for Ship Instance Segmentation
by Weidong Zhu, Piao Wang and Kuifeng Luan
J. Imaging 2025, 11(11), 383; https://doi.org/10.3390/jimaging11110383 - 31 Oct 2025
Viewed by 872
Abstract
Ship instance segmentation in remote sensing images is essential for maritime applications such as intelligent surveillance and port management. However, this task remains challenging due to dense target distributions, large variations in ship scales and shapes, and limited high-quality datasets. The existing YOLOv8 [...] Read more.
Ship instance segmentation in remote sensing images is essential for maritime applications such as intelligent surveillance and port management. However, this task remains challenging due to dense target distributions, large variations in ship scales and shapes, and limited high-quality datasets. The existing YOLOv8 framework mainly relies on convolutional neural networks and CIoU loss, which are less effective in modeling global–local interactions and producing accurate mask boundaries. To address these issues, we propose CDSANet, a novel one-stage ship instance segmentation network. CDSANet integrates convolutional operations, Vision Transformers, and attention mechanisms within a unified architecture. The backbone adopts a Convolutional Vision Transformer Attention (CVTA) module to enhance both local feature extraction and global context perception. The neck employs dynamic-weighted DOWConv to adaptively handle multi-scale ship instances, while SIoU loss improves localization accuracy and orientation robustness. Additionally, CBAM enhances the network’s focus on salient regions, and a MixUp-based augmentation strategy is used to improve model generalization. Extensive experiments on the proposed VLRSSD dataset demonstrate that CDSANet achieves state-of-the-art performance with a mask AP (50–95) of 75.9%, surpassing the YOLOv8 baseline by 1.8%. Full article
Show Figures

Figure 1

24 pages, 50939 KB  
Article
A Progressive Saliency-Guided Small Ship Detection Method for Large-Scene SAR Images
by Hanying Zhu, Dong Li, Haoran Wang, Ruquan Yang, Jishen Liang, Shuang Liu and Jun Wan
Remote Sens. 2025, 17(17), 3085; https://doi.org/10.3390/rs17173085 - 4 Sep 2025
Cited by 2 | Viewed by 1499
Abstract
Large-scene space-borne SAR images with a high resolution are particularly effective for monitoring vast oceanic areas globally. However, ships are easily overlooked in such large scenes due to their small size and cluttered backgrounds, making SAR ship detection challenging for the existing methods. [...] Read more.
Large-scene space-borne SAR images with a high resolution are particularly effective for monitoring vast oceanic areas globally. However, ships are easily overlooked in such large scenes due to their small size and cluttered backgrounds, making SAR ship detection challenging for the existing methods. To address this challenge, we propose a progressive saliency-guided (PSG) method, which uses saliency-derived positional priors to guide the model in focusing on small targets and extracting their features. Specifically, a dual-guided perception enhancement (DGPE) module is developed, which introduces additional target saliency maps as prior information to cross-guide and highlight key regions in SAR images at the feature level, enhancing small object feature representation. Additionally, a saliency confidence aware assessment (SCAA) mechanism is designed to strengthen small object proposal learning at the proposal level, guided by classification and localization scores at key locations. The DGPE and SCAA modules jointly enhance small object learning across different network levels. Extensive experiments demonstrate that the PSG method significantly improves the detection performance (+4.38% AP on LS-SSDD and +4.35% on HRSID) for small ships in large-scene SAR images compared to that of the baseline, providing an effective solution for small ship detection in large scenes. Full article
Show Figures

Figure 1

27 pages, 9566 KB  
Article
CSBBNet: A Specialized Detection Method for Corner Reflector Targets via a Cross-Shaped Bounding Box Network
by Wangshuo Tang, Yuexin Gao, Mengdao Xing, Min Xue, Huitao Liu and Guangcai Sun
Remote Sens. 2025, 17(16), 2760; https://doi.org/10.3390/rs17162760 - 8 Aug 2025
Viewed by 1236
Abstract
In synthetic aperture radar (SAR) maritime target detection tasks, corner reflector targets (CRTs) and their arrays can easily interfere with the accurate detection of ship targets, significantly increasing the misdetection rate and false alarm rate of detectors. Current deep learning-based research on SAR [...] Read more.
In synthetic aperture radar (SAR) maritime target detection tasks, corner reflector targets (CRTs) and their arrays can easily interfere with the accurate detection of ship targets, significantly increasing the misdetection rate and false alarm rate of detectors. Current deep learning-based research on SAR maritime target detection primarily focuses on ship targets, while dedicated detection methods addressing corner reflector interference have not yet established a comprehensive research framework. There remains a lack of theoretical innovation in detection principles for such targets. To address these issues, utilizing the prior knowledge of cross-shaped structures exhibited by marine CRTs in SAR images, we propose an innovative cross-shaped bounding box (CSBB) annotation strategy and design a novel dedicated detection network CSBBNet. The proposed method is constructed through three innovative component modules, namely the cross-shaped spatial feature perception (CSSFP) module, the wavelet cross-shaped attention downsampling (WCSAD) module, and the cross-shaped attention detection head (CSAD-Head). Additionally, to ensure effective training, we propose a cross-shaped intersection over union (CS-IoU) loss function. Comparative experiments with state-of-the-art methods demonstrate that our approach exhibits efficient detection capabilities for CRTs. Ablation experiment results validate the effectiveness of the proposed component architectures. Full article
Show Figures

Figure 1

20 pages, 1865 KB  
Article
A Robust Cross-Band Network for Blind Source Separation of Underwater Acoustic Mixed Signals
by Xingmei Wang, Peiran Wu, Haisu Wei, Yuezhu Xu and Siyu Wang
J. Mar. Sci. Eng. 2025, 13(7), 1334; https://doi.org/10.3390/jmse13071334 - 11 Jul 2025
Viewed by 1108
Abstract
Blind source separation (BSS) of underwater acoustic mixed signals aims to improve signal clarity by separating noise components from aliased underwater signal sources. This enhancement directly increases target detection accuracy in underwater acoustic perception systems, particularly in scenarios involving multi-vessel interference or biological [...] Read more.
Blind source separation (BSS) of underwater acoustic mixed signals aims to improve signal clarity by separating noise components from aliased underwater signal sources. This enhancement directly increases target detection accuracy in underwater acoustic perception systems, particularly in scenarios involving multi-vessel interference or biological sound coexistence. Deep learning-based BSS methods have gained wide attention for their superior nonlinear modeling capabilities. However, existing approaches in underwater acoustic scenarios still face two key challenges: limited feature discrimination and inadequate robustness against non-stationary noise. To overcome these limitations, we propose a novel Robust Cross-Band Network (RCBNet) for the BSS of underwater acoustic mixed signals. To address insufficient feature discrimination, we decompose mixed signals into sub-bands aligned with ship noise harmonics. For intra-band modeling, we apply a parallel gating mechanism that strengthens long-range dependency learning so as to enhance robustness against non-stationary noise. For inter-band modeling, we design a bidirectional-frequency RNN to capture the global dependency relationships of the same signal across sub-bands. Our experiment demonstrates that RCBNet achieves a 0.779 dB improvement in the SDR compared to the advanced model. Additionally, the anti-noise experiment demonstrates that RCBNet exhibits satisfactory robustness across varying noise environments. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

26 pages, 7701 KB  
Article
YOLO-StarLS: A Ship Detection Algorithm Based on Wavelet Transform and Multi-Scale Feature Extraction for Complex Environments
by Yihan Wang, Shuang Zhang, Jianhao Xu, Zhenwen Cheng and Gang Du
Symmetry 2025, 17(7), 1116; https://doi.org/10.3390/sym17071116 - 11 Jul 2025
Cited by 3 | Viewed by 1213
Abstract
Ship detection in complex environments presents challenges such as sea surface reflections, wave interference, variations in illumination, and a range of target scales. The interaction between symmetric ship structures and wave patterns challenges conventional algorithms, particularly in maritime wireless networks. This study presents [...] Read more.
Ship detection in complex environments presents challenges such as sea surface reflections, wave interference, variations in illumination, and a range of target scales. The interaction between symmetric ship structures and wave patterns challenges conventional algorithms, particularly in maritime wireless networks. This study presents YOLO-StarLS (You Only Look Once with Star-topology Lightweight Ship detection), a detection framework leveraging wavelet transforms and multi-scale feature extraction through three core modules. We developed a Wavelet Multi-scale Feature Extraction Network (WMFEN) utilizing adaptive Haar wavelet decomposition with star-topology extraction to preserve multi-frequency information while minimizing detail loss. We introduced a Cross-axis Spatial Attention Refinement module (CSAR), which integrates star structures with cross-axis attention mechanisms to enhance spatial perception. We constructed an Efficient Detail-Preserving Detection head (EDPD) combining differential and shared convolutions to enhance edge detection while reducing computational complexity. Evaluation on the SeaShips dataset demonstrated YOLO-StarLS achieved superior performance for both mAP50 and mAP50–95 metrics, improving by 2.21% and 2.42% over the baseline YOLO11. The approach achieved significant efficiency, with a 36% reduction in the number of parameters to 1.67 M, a 34% decrease in complexity to 4.3 GFLOPs, and an inference speed of 162.0 FPS. Comparative analysis against eight algorithms confirmed the superiority in symmetric target detection. This work enhances real-time ship detection and provides foundations for maritime wireless surveillance networks. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

24 pages, 40762 KB  
Article
Multiscale Task-Decoupled Oriented SAR Ship Detection Network Based on Size-Aware Balanced Strategy
by Shun He, Ruirui Yuan, Zhiwei Yang and Jiaxue Liu
Remote Sens. 2025, 17(13), 2257; https://doi.org/10.3390/rs17132257 - 30 Jun 2025
Cited by 2 | Viewed by 1105
Abstract
Current synthetic aperture radar (SAR) ship datasets exhibit a notable disparity in the distribution of large, medium, and small ship targets. This imbalance makes it difficult for a relatively small number of large and medium-sized ships to be effectively trained, resulting in many [...] Read more.
Current synthetic aperture radar (SAR) ship datasets exhibit a notable disparity in the distribution of large, medium, and small ship targets. This imbalance makes it difficult for a relatively small number of large and medium-sized ships to be effectively trained, resulting in many false alarms. Therefore, to address the issues of scale diversity, intra-class imbalance in ship data, and the feature conflict problem associated with traditional coupled detection heads, we propose an SAR image multiscale task-decoupled oriented ship target detector based on a size-aware balanced strategy. First, the multiscale target features are extracted using the multikernel heterogeneous perception module (MKHP). Meanwhile, the triple-attention module is introduced to establish the remote channel dependence to alleviate the issue of small target feature annihilation, which can effectively enhance the feature characterization ability of the model. Second, given the differences in the demand for feature information between the detection and classification tasks, a channel attention-based task decoupling dual-head (CAT2D) detector head structure is introduced to address the inherent conflict between classification and localization tasks. Finally, a new size-aware balanced (SAB) loss strategy is proposed to guide the network in focusing on the scarce targets in training to alleviate the intra-class imbalance problem during the training process. The ablation experiments on SSDD+ reflect the contribution of each component, and the results of the comparison experiments on the RSDD-SAR and HRSID datasets show that the proposed method achieves state-of-the-art performance compared to other state-of-the-art detection models. Furthermore, our approach exhibits superior detection coverage for both offshore and inshore scenarios for ship detection tasks. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Graphical abstract

Back to TopTop