Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (47)

Search Parameters:
Keywords = maritime target recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 6806 KiB  
Article
Fine Recognition of MEO SAR Ship Targets Based on a Multi-Level Focusing-Classification Strategy
by Zhaohong Li, Wei Yang, Can Su, Hongcheng Zeng, Yamin Wang, Jiayi Guo and Huaping Xu
Remote Sens. 2025, 17(15), 2599; https://doi.org/10.3390/rs17152599 - 26 Jul 2025
Viewed by 258
Abstract
The Medium Earth Orbit (MEO) spaceborne Synthetic Aperture Radar (SAR) has great coverage ability, which can improve maritime ship target surveillance performance significantly. However, due to the huge computational load required for imaging processing and the severe defocusing caused by ship motions, traditional [...] Read more.
The Medium Earth Orbit (MEO) spaceborne Synthetic Aperture Radar (SAR) has great coverage ability, which can improve maritime ship target surveillance performance significantly. However, due to the huge computational load required for imaging processing and the severe defocusing caused by ship motions, traditional ship recognition conducted in focused image domains cannot process MEO SAR data efficiently. To address this issue, a multi-level focusing-classification strategy for MEO SAR ship recognition is proposed, which is applied to the range-compressed ship data domain. Firstly, global fast coarse-focusing is conducted to compensate for sailing motion errors. Then, a coarse-classification network is designed to realize major target category classification, based on which local region image slices are extracted. Next, fine-focusing is performed to correct high-order motion errors, followed by applying fine-classification applied to the image slices to realize final ship classification. Equivalent MEO SAR ship images generated by real LEO SAR data are utilized to construct training and testing datasets. Simulated MEO SAR ship data are also used to evaluate the generalization of the whole method. The experimental results demonstrate that the proposed method can achieve high classification precision. Since only local region slices are used during the second-level processing step, the complex computations induced by fine-focusing for the full image can be avoided, thereby significantly improving overall efficiency. Full article
(This article belongs to the Special Issue Advances in Remote Sensing Image Target Detection and Recognition)
Show Figures

Figure 1

19 pages, 3520 KiB  
Article
Vision-Guided Maritime UAV Rescue System with Optimized GPS Path Planning and Dual-Target Tracking
by Suli Wang, Yang Zhao, Chang Zhou, Xiaodong Ma, Zijun Jiao, Zesheng Zhou, Xiaolu Liu, Tianhai Peng and Changxing Shao
Drones 2025, 9(7), 502; https://doi.org/10.3390/drones9070502 - 16 Jul 2025
Viewed by 457
Abstract
With the global increase in maritime activities, the frequency of maritime accidents has risen, underscoring the urgent need for faster and more efficient search and rescue (SAR) solutions. This study presents an intelligent unmanned aerial vehicle (UAV)-based maritime rescue system that combines GPS-driven [...] Read more.
With the global increase in maritime activities, the frequency of maritime accidents has risen, underscoring the urgent need for faster and more efficient search and rescue (SAR) solutions. This study presents an intelligent unmanned aerial vehicle (UAV)-based maritime rescue system that combines GPS-driven dynamic path planning with vision-based dual-target detection and tracking. Developed within the Gazebo simulation environment and based on modular ROS architecture, the system supports stable takeoff and smooth transitions between multi-rotor and fixed-wing flight modes. An external command module enables real-time waypoint updates. This study proposes three path-planning schemes based on the characteristics of drones. Comparative experiments have demonstrated that the triangular path is the optimal route. Compared with the other schemes, this path reduces the flight distance by 30–40%. Robust target recognition is achieved using a darknet-ROS implementation of the YOLOv4 model, enhanced with data augmentation to improve performance in complex maritime conditions. A monocular vision-based ranging algorithm ensures accurate distance estimation and continuous tracking of rescue vessels. Furthermore, a dual-target-tracking algorithm—integrating motion prediction with color-based landing zone recognition—achieves a 96% success rate in precision landings under dynamic conditions. Experimental results show a 4% increase in the overall mission success rate compared to traditional SAR methods, along with significant gains in responsiveness and reliability. This research delivers a technically innovative and cost-effective UAV solution, offering strong potential for real-world maritime emergency response applications. Full article
Show Figures

Figure 1

25 pages, 9677 KiB  
Article
YOLO-SEA: An Enhanced Detection Framework for Multi-Scale Maritime Targets in Complex Sea States and Adverse Weather
by Hongmei Deng, Shuaiqun Wang, Xinyao Wang, Wen Zheng and Yanli Xu
Entropy 2025, 27(7), 667; https://doi.org/10.3390/e27070667 - 22 Jun 2025
Viewed by 573
Abstract
Maritime object detection is essential for resource monitoring, maritime defense, and public safety, yet detecting diverse targets beyond ships remains challenging. This paper presents YOLO-SEA, an efficient detection framework based on the enhanced YOLOv8 architecture. The model incorporates the SESA (SimAM-Enhanced SENetV2 Attention) [...] Read more.
Maritime object detection is essential for resource monitoring, maritime defense, and public safety, yet detecting diverse targets beyond ships remains challenging. This paper presents YOLO-SEA, an efficient detection framework based on the enhanced YOLOv8 architecture. The model incorporates the SESA (SimAM-Enhanced SENetV2 Attention) module, which integrates the channel-adaptive weight adjustment of SENetV2 with the parameter-free spatial-channel modeling of SimAM to enhance feature representation. An improved BiFPN (Bidirectional Feature Pyramid Network) structure enhances multi-scale fusion, particularly for small object detection. In the post-processing stage, Soft-NMS (Soft Non-Maximum Suppression) replaces traditional NMS to reduce false suppression in dense scenes. YOLO-SEA detects eight maritime object types. Experiments show it achieves a 5.8% improvement in mAP@0.5 and 7.2% improvement in mAP@0.5:0.95 over the baseline, demonstrating enhanced accuracy and robustness in complex marine environments. Full article
(This article belongs to the Special Issue Space-Air-Ground-Sea Integrated Communication Networks)
Show Figures

Figure 1

19 pages, 8033 KiB  
Article
SR-DETR: Target Detection in Maritime Rescue from UAV Imagery
by Yuling Liu and Yan Wei
Remote Sens. 2025, 17(12), 2026; https://doi.org/10.3390/rs17122026 - 12 Jun 2025
Viewed by 982
Abstract
The growth of maritime transportation has been accompanied by a gradual increase in accident rates, drawing greater attention to the critical issue of man-overboard incidents and drowning. Traditional maritime search-and-rescue (SAR) methods are often constrained by limited efficiency and high operational costs. Over [...] Read more.
The growth of maritime transportation has been accompanied by a gradual increase in accident rates, drawing greater attention to the critical issue of man-overboard incidents and drowning. Traditional maritime search-and-rescue (SAR) methods are often constrained by limited efficiency and high operational costs. Over the past few years, drones have demonstrated significant promise in improving the effectiveness of search-and-rescue operations. This is largely due to their exceptional ability to move freely and their capacity for wide-area monitoring. This study proposes an enhanced SR-DETR algorithm aimed at improving the detection of individuals who have fallen overboard. Specifically, the conventional multi-head self-attention (MHSA) mechanism is replaced with Efficient Additive Attention (EAA), which facilitates more efficient feature interaction while substantially reducing computational complexity. Moreover, we introduce a new feature aggregation module called the Cross-Stage Partial Parallel Atrous Feature Pyramid Network (CPAFPN). By refining spatial attention mechanisms, the module significantly boosts cross-scale target recognition capabilities in the model, especially offering advantages for detecting smaller objects. To improve localization precision, we develop a novel loss function for bounding box regression, named Focaler-GIoU, which performs particularly well when handling densely packed and small-scale objects. The proposed approach is validated through experiments and achieves an mAP of 86.5%, which surpasses the baseline RT-DETR model’s performance of 83.2%. These outcomes highlight the practicality and reliability of our method in detecting individuals overboard, contributing to more precise and resource-efficient solutions for real-time maritime rescue efforts. Full article
Show Figures

Figure 1

28 pages, 4962 KiB  
Article
YOLO-Ssboat: Super-Small Ship Detection Network for Large-Scale Aerial and Remote Sensing Scenes
by Yiliang Zeng, Xiuhong Wang, Jinlin Zou and Hongtao Wu
Remote Sens. 2025, 17(11), 1948; https://doi.org/10.3390/rs17111948 - 4 Jun 2025
Viewed by 748
Abstract
Enhancing the detection capabilities of marine vessels is crucial for maritime security and intelligence acquisition. However, accurately identifying small ships in complex oceanic environments remains a significant challenge, as these targets are frequently obscured by ocean waves and other disturbances, compromising recognition accuracy [...] Read more.
Enhancing the detection capabilities of marine vessels is crucial for maritime security and intelligence acquisition. However, accurately identifying small ships in complex oceanic environments remains a significant challenge, as these targets are frequently obscured by ocean waves and other disturbances, compromising recognition accuracy and stability. To address this issue, we propose YOLO-ssboat, a novel small-target ship recognition algorithm based on the YOLOv8 framework. YOLO-ssboat integrates the C2f_DCNv3 module to extract fine-grained features of small vessels while mitigating background interference and preserving critical target details. Additionally, it employs a high-resolution feature layer and incorporates a Multi-Scale Weighted Pyramid Network (MSWPN) to enhance feature diversity. The algorithm further leverages an improved multi-attention detection head, Dyhead_v3, to refine the representation of small-target features. To tackle the challenge of wake waves from moving ships obscuring small targets, we introduce a gradient flow mechanism that improves detection efficiency under dynamic conditions. The Tail Wave Detection Method synergistically integrates gradient computation with target detection techniques. Furthermore, adversarial training enhances the network’s robustness and ensures greater stability. Experimental evaluations on the Ship_detection and Vessel datasets demonstrate that YOLO-ssboat outperforms state-of-the-art detection algorithms in both accuracy and stability. Notably, the gradient flow mechanism enriches target feature extraction for moving vessels, thereby improving detection accuracy in wake-disturbed scenarios, while adversarial training further fortifies model resilience. These advancements offer significant implications for the long-range monitoring and detection of maritime vessels, contributing to enhanced situational awareness in expansive oceanic environments. Full article
Show Figures

Figure 1

21 pages, 52785 KiB  
Article
MC-ASFF-ShipYOLO: Improved Algorithm for Small-Target and Multi-Scale Ship Detection for Synthetic Aperture Radar (SAR) Images
by Yubin Xu, Haiyan Pan, Lingqun Wang and Ran Zou
Sensors 2025, 25(9), 2940; https://doi.org/10.3390/s25092940 - 7 May 2025
Viewed by 771
Abstract
Synthetic aperture radar (SAR) ship detection holds significant application value in maritime monitoring, marine traffic management, and safety maintenance. Despite remarkable advances in deep-learning-based detection methods, performance remains constrained by the vast size differences between ships, limited feature information of small targets, and [...] Read more.
Synthetic aperture radar (SAR) ship detection holds significant application value in maritime monitoring, marine traffic management, and safety maintenance. Despite remarkable advances in deep-learning-based detection methods, performance remains constrained by the vast size differences between ships, limited feature information of small targets, and complex environmental interference in SAR imagery. Although many studies have separately tackled small target identification and multi-scale detection in SAR imagery, integrated approaches that jointly address both challenges within a unified framework for SAR ship detection are still relatively scarce. This study presents MC-ASFF-ShipYOLO (Monte Carlo Attention—Adaptively Spatial Feature Fusion—ShipYOLO), a novel framework addressing both small target recognition and multi-scale ship detection challenges. Two key innovations distinguish our approach: (1) We introduce a Monte Carlo Attention (MCAttn) module into the backbone network that employs random sampling pooling operations to generate attention maps for feature map weighting, enhancing focus on small targets and improving their detection performance. (2) We add Adaptively Spatial Feature Fusion (ASFF) modules to the detection head that adaptively learn spatial fusion weights across feature layers and perform dynamic feature fusion, ensuring consistent ship representations across scales and mitigating feature conflicts, thereby enhancing multi-scale detection capability. Experiments are conducted on a newly constructed dataset combining HRSID and SSDD. Ablation experiment results demonstrate that, compared to the baseline, MC-ASFF-ShipYOLO achieves improvements of 1.39% in precision, 2.63% in recall, 2.28% in AP50, and 3.04% in AP, indicating a significant enhancement in overall detection performance. Furthermore, comparative experiments show that our method outperforms mainstream models. Even under high-confidence thresholds, MC-ASFF-ShipYOLO is capable of predicting more high-quality detection boxes, offering a valuable solution for advancing SAR ship detection technology. Full article
(This article belongs to the Special Issue Recent Advances in Synthetic Aperture Radar (SAR) Remote Sensing)
Show Figures

Figure 1

30 pages, 12300 KiB  
Article
VIOS-Net: A Multi-Task Fusion System for Maritime Surveillance Through Visible and Infrared Imaging
by Junquan Zhan, Jiawen Li, Langtao Wu, Jiahua Sun and Hui Yin
J. Mar. Sci. Eng. 2025, 13(5), 913; https://doi.org/10.3390/jmse13050913 - 6 May 2025
Viewed by 508
Abstract
Automatic ship monitoring models leveraging image recognition have become integral to regulatory applications within maritime management, with multi-source image co-monitoring serving as the primary method for achieving comprehensive, round-the-clock surveillance. Despite their widespread use, the existing models predominantly train each data source independently [...] Read more.
Automatic ship monitoring models leveraging image recognition have become integral to regulatory applications within maritime management, with multi-source image co-monitoring serving as the primary method for achieving comprehensive, round-the-clock surveillance. Despite their widespread use, the existing models predominantly train each data source independently or simultaneously train multiple sources without fully optimizing the integration of similar information. This approach, while capable of all-weather detection, results in the underutilization of data features from related sources and unnecessary repetition in model training, leading to excessive time consumption. To address these inefficiencies, this paper introduces a novel multi-task learning framework designed to enhance the utilization of data features from diverse information sources, thereby reducing training time, lowering costs, and improving recognition accuracy. The proposed model, VIOS-Net, integrates the advantages of both visible and infrared data sources to meet the challenges of all-weather, all-day ship monitoring under complex environmental conditions. VIOS-Net employs a Shared Bottom network architecture, utilizing both shared and specific feature extraction modules at the model’s lower and upper layers, respectively, to optimize the system’s recognition capabilities and maximize data utilization efficiency. The experimental results demonstrate that VIOS-Net achieves an accuracy of 96.20% across both visible and infrared spectral datasets, significantly outperforming the baseline ResNet-34 model, which attained accuracies of only 4.86% and 9.04% in visible and infrared data, respectively. Moreover, VIOS-Net reduces the number of parameters by 48.82% compared to the baseline, achieving optimal performance in multi-spectral ship monitoring. Extensive ablation studies further validate the effectiveness of the individual modules within the proposed framework. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

20 pages, 5808 KiB  
Article
Enhanced YOLOv7 Based on Channel Attention Mechanism for Nearshore Ship Detection
by Qingyun Zhu, Zhen Zhang and Ruizhe Mu
Electronics 2025, 14(9), 1739; https://doi.org/10.3390/electronics14091739 - 24 Apr 2025
Viewed by 507
Abstract
Nearshore ship detection is an important task in marine monitoring, playing a significant role in navigation safety and controlling illegal smuggling. The continuous research and development of Synthetic Aperture Radar (SAR) technology is not only of great importance in military and maritime security [...] Read more.
Nearshore ship detection is an important task in marine monitoring, playing a significant role in navigation safety and controlling illegal smuggling. The continuous research and development of Synthetic Aperture Radar (SAR) technology is not only of great importance in military and maritime security fields but also has great potential in civilian fields, such as disaster emergency response, marine resource monitoring, and environmental protection. Due to the limited sample size of nearshore ship datasets, it is difficult to meet the demand for the large quantity of training data required by existing deep learning algorithms, which limits the recognition accuracy. At the same time, artificial environmental features such as buildings can cause significant interference to SAR imaging, making it more difficult to distinguish ships from the background. Ship target images are greatly affected by speckle noise, posing additional challenges to data-driven recognition methods. Therefore, we utilized a Concurrent Single-Image GAN (ConSinGAN) to generate high-quality synthetic samples for re-labeling and fused them with the dataset extracted from the SAR-Ship dataset for nearshore image extraction and dataset division. Experimental analysis showed that the ship recognition model trained with augmented images had an accuracy increase of 4.66%, a recall rate increase of 3.68%, and an average precision (AP) with Intersection over Union (IoU) at 0.5 increased by 3.24%. Subsequently, an enhanced YOLOv7 algorithm (YOLOv7 + ESE) incorporating channel-wise information fusion was developed based on the YOLOv7 architecture integrated with the Squeeze-and-Excitation (SE) channel attention mechanism. Through comparative experiments, the analytical results demonstrated that the proposed algorithm achieved performance improvements of 0.36% in precision, 0.52% in recall, and 0.65% in average precision (AP@0.5) compared to the baseline model. This optimized architecture enables accurate detection of nearshore ship targets in SAR imagery. Full article
(This article belongs to the Special Issue Intelligent Systems in Industry 4.0)
Show Figures

Figure 1

20 pages, 14434 KiB  
Article
Optimized Marine Target Detection in Remote Sensing Images with Attention Mechanism and Multi-Scale Feature Fusion
by Xiantao Jiang, Tianyi Liu, Tian Song and Qi Cen
Information 2025, 16(4), 332; https://doi.org/10.3390/info16040332 - 21 Apr 2025
Cited by 1 | Viewed by 443
Abstract
With the continuous growth of maritime activities and the shipping trade, the application of maritime target detection in remote sensing images has become increasingly important. However, existing detection methods face numerous challenges, such as small target localization, recognition of targets with large aspect [...] Read more.
With the continuous growth of maritime activities and the shipping trade, the application of maritime target detection in remote sensing images has become increasingly important. However, existing detection methods face numerous challenges, such as small target localization, recognition of targets with large aspect ratios, and high computational demands. In this paper, we propose an improved target detection model, named YOLOv5-ASC, to address the challenges in maritime target detection. The proposed YOLOv5-ASC integrates three core components: an Attention-based Receptive Field Enhancement Module (ARFEM), an optimized SIoU loss function, and a Deformable Convolution Module (C3DCN). These components work together to enhance the model’s performance in detecting complex maritime targets by improving its ability to capture multi-scale features, optimize the localization process, and adapt to the large aspect ratios typical of maritime objects. Experimental results show that, compared to the original YOLOv5 model, YOLOv5-ASC achieves a 4.36 percentage point increase in mAP@0.5 and a 9.87 percentage point improvement in precision, while maintaining computational complexity within a reasonable range. The proposed method not only achieves significant performance improvements on the ShipRSImageNet dataset but also demonstrates strong potential for application in complex maritime remote sensing scenarios. Full article
(This article belongs to the Special Issue Computer Vision for Security Applications)
Show Figures

Figure 1

25 pages, 13401 KiB  
Article
Enhanced U-Net for Underwater Laser Range-Gated Image Restoration: Boosting Underwater Target Recognition
by Peng Liu, Shuaibao Chen, Wei He, Jue Wang, Liangpei Chen, Yuguang Tan, Dong Luo, Wei Chen and Guohua Jiao
J. Mar. Sci. Eng. 2025, 13(4), 803; https://doi.org/10.3390/jmse13040803 - 17 Apr 2025
Viewed by 651
Abstract
Underwater optical imaging plays a crucial role in maritime safety, enabling reliable navigation, efficient search and rescue operations, precise target recognition, and robust military reconnaissance. However, conventional underwater imaging methods often suffer from severe backscattering noise, limited detection range, and reduced image clarity—challenges [...] Read more.
Underwater optical imaging plays a crucial role in maritime safety, enabling reliable navigation, efficient search and rescue operations, precise target recognition, and robust military reconnaissance. However, conventional underwater imaging methods often suffer from severe backscattering noise, limited detection range, and reduced image clarity—challenges that are exacerbated in turbid waters. To address these issues, Underwater Laser Range-Gated Imaging has emerged as a promising solution. By selectively capturing photons within a controlled temporal gate, this technique effectively suppresses backscattering noise-enhancing image clarity, contrast, and detection range. Nevertheless, residual noise within the imaging slice can still degrade image quality, particularly in challenging underwater conditions. In this study, we propose an enhanced U-Net neural network designed to mitigate noise interference in underwater laser range-gated images, improving target recognition performance. Built upon the U-Net architecture with added residual connections, our network combines a VGG16-based perceptual loss with Mean Squared Error (MSE) as the loss function, effectively capturing high-level semantic features while preserving critical target details during reconstruction. Trained on a semi-synthetic grayscale dataset containing synthetically degraded images paired with their reference counterparts, the proposed approach demonstrates improved performance compared to several existing underwater image restoration methods in our experimental evaluations. Through comprehensive qualitative and quantitative evaluations, underwater target detection experiments, and real-world oceanic validations, our method demonstrates significant potential for advancing maritime safety and related applications. Full article
Show Figures

Figure 1

39 pages, 5524 KiB  
Article
Research on Methods for the Recognition of Ship Lights and the Autonomous Determination of the Types of Approaching Vessels
by Xiangyu Gao and Yuelin Zhao
J. Mar. Sci. Eng. 2025, 13(4), 643; https://doi.org/10.3390/jmse13040643 - 24 Mar 2025
Viewed by 568
Abstract
The acquisition of approaching vessels’ information is a critical technological challenge for maritime risk warning and intelligent collision avoidance decision-making. This paper proposes a method for autonomously identifying types of approaching vessels based on an improved YOLOv8 model and ship light features, aiming [...] Read more.
The acquisition of approaching vessels’ information is a critical technological challenge for maritime risk warning and intelligent collision avoidance decision-making. This paper proposes a method for autonomously identifying types of approaching vessels based on an improved YOLOv8 model and ship light features, aiming to infer the propulsion mode, size, movement, and operational nature of the approaching vessels in real-time through the color, quantity, and spatial distribution of lights. Firstly, to address the challenges of the small target characteristics of ship lights and complex environmental interference, an improved YOLOv8 model is developed: The dilation-wise residual (DWR) module is introduced to optimize the feature extraction capability of the C2f structure. The bidirectional feature pyramid network (BiFPN) is adopted to enhance multi-scale feature fusion. A hybrid attention transformer (HAT) is employed to enhance the small target detection capability of the detection head. This framework achieves precise ship light recognition under complex maritime circumstances. Secondly, 23 spatio-semantic feature indicators are established to encode ship light patterns, and a multi-viewing angle dataset is constructed. This dataset covers 36 vessel types under four viewing angles (front, port-side, starboard, and stern viewing angles), including the color, quantity, combinations, and spatial distribution of the ship lights. Finally, a two-stage discriminative model is proposed: ECA-1D-CNN is utilized for the rapid assessment of the viewing angle of the vessel. Deep learning algorithms are dynamically applied for vessel type determination within the assessed viewing angles. Experimental results show that this method achieves high determination accuracy. This paper provides a kind of technical support for intelligent situational awareness and the autonomous collision avoidance of ships. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

19 pages, 5364 KiB  
Article
Enhancing Maritime Domain Awareness Through AI-Enabled Acoustic Buoys for Real-Time Detection and Tracking of Fast-Moving Vessels
by Jeremy Karst, Robert McGurrin, Kimberly Gavin, Joseph Luttrell, William Rippy, Robert Coniglione, Jason McKenna and Ralf Riedel
Sensors 2025, 25(6), 1930; https://doi.org/10.3390/s25061930 - 20 Mar 2025
Viewed by 1314
Abstract
Acoustic target recognition has always played a central role in marine sensing. Traditional signal processing techniques that have been used for target recognition have shown limitations in accuracy, particularly with commodity hardware. To address such limitations, we present the results of our experiments [...] Read more.
Acoustic target recognition has always played a central role in marine sensing. Traditional signal processing techniques that have been used for target recognition have shown limitations in accuracy, particularly with commodity hardware. To address such limitations, we present the results of our experiments to assess the capabilities of AI-enabled acoustic buoys using OpenEar™, a commercial, off-the-shelf, software-defined hydrophone sensor, for detecting and tracking fast-moving vessels. We used a triangular sparse sensor network to investigate techniques necessary to estimate the detection, classification, localization, and tracking of boats transiting through the network. Emphasis was placed on evaluating the sensor’s operational detection range and feasibility of onboard AI for cloud-based data fusion. Results indicated effectiveness for enhancing maritime domain awareness and gaining insight into illegal, unreported, and unregulated activities. Additionally, this study provides a framework for scaling autonomous sensor networks to support persistent maritime surveillance. Full article
Show Figures

Figure 1

32 pages, 4011 KiB  
Article
Enhancing Underwater Acoustic Target Recognition Through Advanced Feature Fusion and Deep Learning
by Yanghong Zhao, Guohao Xie, Haoyu Chen, Mingsong Chen and Li Huang
J. Mar. Sci. Eng. 2025, 13(2), 278; https://doi.org/10.3390/jmse13020278 - 31 Jan 2025
Cited by 1 | Viewed by 1800
Abstract
Underwater Acoustic Target Recognition (UATR) is critical to maritime traffic management and ocean monitoring. However, underwater acoustic analysis is fraught with difficulties. The underwater environment is highly complex, with ambient noise, variable water conditions (such as temperature and salinity), and multi-path propagation of [...] Read more.
Underwater Acoustic Target Recognition (UATR) is critical to maritime traffic management and ocean monitoring. However, underwater acoustic analysis is fraught with difficulties. The underwater environment is highly complex, with ambient noise, variable water conditions (such as temperature and salinity), and multi-path propagation of acoustic signals. These factors make it challenging to accurately acquire and analyze target features. Traditional UATR methods struggle with feature fusion representations and model generalization. This study introduces a novel high-dimensional feature fusion method, CM3F, grounded in signal analysis and brain-like features, and integrates it with the Boundary-Aware Hybrid Transformer Network (BAHTNet), a deep-learning architecture tailored for UATR. BAHTNet comprises CBCARM and XCAT modules, leveraging a Kan network for classification and a large-margin aware focal (LMF) loss function for predictive losses. Experimental results on real-world datasets demonstrate the model’s robust generalization capabilities, achieving 99.8% accuracy on the ShipsEar dataset and 94.57% accuracy on the Deepship dataset. These findings underscore the potential of BAHTNet to significantly improve UATR performance. Full article
(This article belongs to the Special Issue Underwater Target Detection and Recognition)
Show Figures

Figure 1

21 pages, 7780 KiB  
Article
On-Ship Trinocular Stereo Vision: An Experimental Study for Long-Range High-Accuracy Localization of Other Vessels
by Kotaro Yoshihara, Shigehiro Yamamoto and Takeshi Hashimoto
J. Mar. Sci. Eng. 2025, 13(1), 115; https://doi.org/10.3390/jmse13010115 - 10 Jan 2025
Viewed by 1224
Abstract
Recently, several initiatives regarding maritime autonomous surface ships (MASSs) have been implemented worldwide. One of the fundamental technologies for attaining MASSs is the recognition and localization of surrounding ships. Traditional navigational instruments are inadequate for recognizing objects, and the authors investigated the potential [...] Read more.
Recently, several initiatives regarding maritime autonomous surface ships (MASSs) have been implemented worldwide. One of the fundamental technologies for attaining MASSs is the recognition and localization of surrounding ships. Traditional navigational instruments are inadequate for recognizing objects, and the authors investigated the potential of stereo vision. Conventional stereo camera systems are not suitable for localizing very distant objects. One proposed solution is to use an additional camera, thus using three-camera measurements of objects at long distances to reduce positional measurement errors, incorporating time-series averaging and keypoint-based techniques. This study evaluated experimentally the accuracy of measurements using three ship-mounted cameras. The accuracy and precision of stereo measurements depend on the distance between the camera positions, referred to as the baseline length. Conventional stereo cameras are typically used to measure objects at distances of up to 200 times the baseline length. This study indicates that, using trinocular stereo vision, a target ship at distances up to 2500 m, which is 500 times the baseline length, can be measured with an accuracy of approximately 5% of the RMSE. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

28 pages, 43934 KiB  
Article
A Cross-Stage Focused Small Object Detection Network for Unmanned Aerial Vehicle Assisted Maritime Applications
by Gege Ding, Jiayue Liu, Dongsheng Li, Xiaming Fu, Yucheng Zhou, Mingrui Zhang, Wantong Li, Yanjuan Wang, Chunxu Li and Xiongfei Geng
J. Mar. Sci. Eng. 2025, 13(1), 82; https://doi.org/10.3390/jmse13010082 - 5 Jan 2025
Cited by 3 | Viewed by 1567
Abstract
The application potential of unmanned aerial vehicles (UAVs) in marine search and rescue is especially of concern for the ongoing advancement of visual recognition technology and image processing technology. Limited computing resources, insufficient pixel representation for small objects in high-altitude images, and challenging [...] Read more.
The application potential of unmanned aerial vehicles (UAVs) in marine search and rescue is especially of concern for the ongoing advancement of visual recognition technology and image processing technology. Limited computing resources, insufficient pixel representation for small objects in high-altitude images, and challenging visibility conditions hinder UAVs’ target recognition performance in maritime search and rescue operations, highlighting the need for further optimization and enhancement. This study introduces an innovative detection framework, CFSD-UAVNet, designed to boost the accuracy of detecting minor objects within imagery captured from elevated altitudes. To improve the performance of the feature pyramid network (FPN) and path aggregation network (PAN), a newly designed PHead structure was proposed, focusing on better leveraging shallow features. Then, structural pruning was applied to refine the model and enhance its capability in detecting small objects. Moreover, to conserve computational resources, a lightweight CED module was introduced to reduce parameters and conserve the computing resources of the UAV. At the same time, in each detection layer, a lightweight CRE module was integrated, leveraging attention mechanisms and detection heads to enhance precision for small object detection. Finally, to enhance the model’s robustness, WIoUv2 loss function was employed, ensuring a balanced treatment of positive and negative samples. The CFSD-UAVNet model was evaluated on the publicly available SeaDronesSee maritime dataset and compared with other cutting-edge algorithms. The experimental results showed that the CFSD-UAVNet model achieved an mAP@50 of 80.1% with only 1.7 M parameters and a computational cost of 10.2 G, marking a 12.1% improvement over YOLOv8 and a 4.6% increase compared to DETR. The novel CFSD-UAVNet model effectively balances the limitations of scenarios and detection accuracy, demonstrating application potential and value in the field of UAV-assisted maritime search and rescue. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

Back to TopTop