Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (89)

Search Parameters:
Keywords = marine image recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 5889 KiB  
Article
Mobile-YOLO: A Lightweight Object Detection Algorithm for Four Categories of Aquatic Organisms
by Hanyu Jiang, Jing Zhao, Fuyu Ma, Yan Yang and Ruiwen Yi
Fishes 2025, 10(7), 348; https://doi.org/10.3390/fishes10070348 - 14 Jul 2025
Viewed by 164
Abstract
Accurate and rapid aquatic organism recognition is a core technology for fisheries automation and aquatic organism statistical research. However, due to absorption and scattering effects, images of aquatic organisms often suffer from poor contrast and color distortion. Additionally, the clustering behavior of aquatic [...] Read more.
Accurate and rapid aquatic organism recognition is a core technology for fisheries automation and aquatic organism statistical research. However, due to absorption and scattering effects, images of aquatic organisms often suffer from poor contrast and color distortion. Additionally, the clustering behavior of aquatic organisms often leads to occlusion, further complicating the identification task. This study proposes a lightweight object detection model, Mobile-YOLO, for the recognition of four representative aquatic organisms, namely holothurian, echinus, scallop, and starfish. Our model first utilizes the Mobile-Nano backbone network we proposed, which enhances feature perception while maintaining a lightweight design. Then, we propose a lightweight detection head, LDtect, which achieves a balance between lightweight structure and high accuracy. Additionally, we introduce Dysample (dynamic sampling) and HWD (Haar wavelet downsampling) modules, aiming to optimize the feature fusion structure and achieve lightweight goals by improving the processes of upsampling and downsampling. These modules also help compensate for the accuracy loss caused by the lightweight design of LDtect. Compared to the baseline model, our model reduces Params (parameters) by 32.2%, FLOPs (floating point operations) by 28.4%, and weights (model storage size) by 30.8%, while improving FPS (frames per second) by 95.2%. The improvement in mAP (mean average precision) can also lead to better accuracy in practical applications, such as marine species monitoring, conservation efforts, and biodiversity assessment. Furthermore, the model’s accuracy is enhanced, with the mAP increased by 1.6%, demonstrating the advanced nature of our approach. Compared with YOLO (You Only Look Once) series (YOLOv5-12), SSD (Single Shot MultiBox Detector), EfficientDet (Efficient Detection), RetinaNet, and RT-DETR (Real-Time Detection Transformer), our model achieves leading comprehensive performance in terms of both accuracy and lightweight design. The results indicate that our research provides technological support for precise and rapid aquatic organism recognition. Full article
(This article belongs to the Special Issue Technology for Fish and Fishery Monitoring)
Show Figures

Figure 1

23 pages, 5304 KiB  
Article
Improvement and Optimization of Underwater Image Target Detection Accuracy Based on YOLOv8
by Yisong Sun, Wei Chen, Qixin Wang, Tianzhong Fang and Xinyi Liu
Symmetry 2025, 17(7), 1102; https://doi.org/10.3390/sym17071102 - 9 Jul 2025
Viewed by 346
Abstract
The ocean encompasses the majority of the Earth’s surface and harbors substantial energy resources. Nevertheless, the intricate and asymmetrically distributed underwater environment renders existing target detection performance inadequate. This paper presents an enhanced YOLOv8s approach for underwater robot object detection to address issues [...] Read more.
The ocean encompasses the majority of the Earth’s surface and harbors substantial energy resources. Nevertheless, the intricate and asymmetrically distributed underwater environment renders existing target detection performance inadequate. This paper presents an enhanced YOLOv8s approach for underwater robot object detection to address issues of subpar image quality and low recognition accuracy. The precise measures are enumerated as follows: initially, to address the issue of model parameters, we optimized the ninth convolutional layer by substituting certain conventional convolutions with adaptive deformable convolution DCN v4. This modification aims to more effectively capture the deformation and intricate features of underwater targets, while simultaneously decreasing the parameter count and enhancing the model’s ability to manage the deformation challenges presented by underwater images. Furthermore, the Triplet Attention module is implemented to augment the model’s capacity for detecting multi-scale targets. The integration of low-level superficial features with high-level semantic features enhances the feature expression capability. The original CIoU loss function was ultimately substituted with Shape IoU, enhancing the model’s performance. In the underwater robot grasping experiment, the system shows particular robustness in handling radial symmetry in marine organisms and reflection symmetry in artificial structures. The enhanced algorithm attained a mean Average Precision (mAP) of 87.6%, surpassing the original YOLOv8s model by 3.4%, resulting in a marked enhancement of the object detection model’s performance and fulfilling the real-time detection criteria for underwater robots. Full article
Show Figures

Figure 1

17 pages, 3741 KiB  
Article
DeepSeaNet: An Efficient UIE Deep Network
by Jingsheng Li, Yuanbing Ouyang, Hao Wang, Di Wu and Yushan Pan
Electronics 2025, 14(12), 2411; https://doi.org/10.3390/electronics14122411 - 12 Jun 2025
Viewed by 421
Abstract
Underwater image enhancement and object recognition are crucial in multiple fields, like marine biology, archeology, and environmental monitoring, but face severe challenges due to low light, color distortion, and reduced contrast in underwater environments. DeepSeaNet re-evaluates the model guidance strategy from multiple dimensions, [...] Read more.
Underwater image enhancement and object recognition are crucial in multiple fields, like marine biology, archeology, and environmental monitoring, but face severe challenges due to low light, color distortion, and reduced contrast in underwater environments. DeepSeaNet re-evaluates the model guidance strategy from multiple dimensions, enhances color recovery using the MCOLE score, and addresses the problem of inconsistent attenuation across different regions of underwater images by integrating a feature extraction method guided by a global attention mechanism by ViT. Comprehensive tests on diverse underwater datasets show that DeepSeaNet achieves a maximum PSNR of 28.96 dB and an average SSIM of 0.901, representing a 20–40% improvement over baseline methods. These results highlight DeepSeaNet’s superior performance in enhancing image clarity, color richness, and contrast, making it a remarkably effective instrument for underwater image processing and analysis. Full article
Show Figures

Graphical abstract

16 pages, 5226 KiB  
Article
Enhanced Mask R-CNN Incorporating CBAM and Soft-NMS for Identification and Monitoring of Offshore Aquaculture Areas
by Jiajun Zhang, Yonggui Wang, Yaxin Zhang and Yanxin Zhao
Sensors 2025, 25(9), 2792; https://doi.org/10.3390/s25092792 - 29 Apr 2025
Viewed by 504
Abstract
The use of remote sensing images to analyze the change characteristics of large-scale aquaculture areas and monitor aquaculture violations is of great significance for exploring the law of marine aquaculture and assisting the monitoring and standardization of aquaculture areas. In this study, a [...] Read more.
The use of remote sensing images to analyze the change characteristics of large-scale aquaculture areas and monitor aquaculture violations is of great significance for exploring the law of marine aquaculture and assisting the monitoring and standardization of aquaculture areas. In this study, a violation monitoring framework for marine aquaculture areas based on image recognition using an enhanced Mask R-CNN architecture incorporating a convolutional block attention module (CBAM) and soft non-maximum suppression (Soft-NMS) is proposed and applied in Sandu’ao. The results show that the modified Mask R-CNN, when compared to the most basic Mask R-CNN model, exhibits higher accuracy in identifying marine aquaculture areas. The aquaculture patterns in the Xiapu region are characterized by two peak periods of aquaculture area fluctuations, occurring in March and October. Conversely, July marks the month with the smallest aquaculture area in the region and is influenced by factors such as water temperature and aquaculture cycle. Significant changes in the aquaculture area were observed in January, March, June, August, and October, necessitating rigorous monitoring. Furthermore, monitoring and analysis of aquaculture areas have revealed that despite the reduction in illegal aquaculture acreage since 2017 due to the implementation of functional zone planning for marine aquaculture areas, illegal aquaculture activities remain prevalent in prohibited and restricted zones in Xiapu, accounting for a considerable proportion. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

20 pages, 5808 KiB  
Article
Enhanced YOLOv7 Based on Channel Attention Mechanism for Nearshore Ship Detection
by Qingyun Zhu, Zhen Zhang and Ruizhe Mu
Electronics 2025, 14(9), 1739; https://doi.org/10.3390/electronics14091739 - 24 Apr 2025
Viewed by 490
Abstract
Nearshore ship detection is an important task in marine monitoring, playing a significant role in navigation safety and controlling illegal smuggling. The continuous research and development of Synthetic Aperture Radar (SAR) technology is not only of great importance in military and maritime security [...] Read more.
Nearshore ship detection is an important task in marine monitoring, playing a significant role in navigation safety and controlling illegal smuggling. The continuous research and development of Synthetic Aperture Radar (SAR) technology is not only of great importance in military and maritime security fields but also has great potential in civilian fields, such as disaster emergency response, marine resource monitoring, and environmental protection. Due to the limited sample size of nearshore ship datasets, it is difficult to meet the demand for the large quantity of training data required by existing deep learning algorithms, which limits the recognition accuracy. At the same time, artificial environmental features such as buildings can cause significant interference to SAR imaging, making it more difficult to distinguish ships from the background. Ship target images are greatly affected by speckle noise, posing additional challenges to data-driven recognition methods. Therefore, we utilized a Concurrent Single-Image GAN (ConSinGAN) to generate high-quality synthetic samples for re-labeling and fused them with the dataset extracted from the SAR-Ship dataset for nearshore image extraction and dataset division. Experimental analysis showed that the ship recognition model trained with augmented images had an accuracy increase of 4.66%, a recall rate increase of 3.68%, and an average precision (AP) with Intersection over Union (IoU) at 0.5 increased by 3.24%. Subsequently, an enhanced YOLOv7 algorithm (YOLOv7 + ESE) incorporating channel-wise information fusion was developed based on the YOLOv7 architecture integrated with the Squeeze-and-Excitation (SE) channel attention mechanism. Through comparative experiments, the analytical results demonstrated that the proposed algorithm achieved performance improvements of 0.36% in precision, 0.52% in recall, and 0.65% in average precision (AP@0.5) compared to the baseline model. This optimized architecture enables accurate detection of nearshore ship targets in SAR imagery. Full article
(This article belongs to the Special Issue Intelligent Systems in Industry 4.0)
Show Figures

Figure 1

19 pages, 10041 KiB  
Article
Intelligent Detection and Recognition of Marine Plankton by Digital Holography and Deep Learning
by Xianfeng Xu, Weilong Luo, Zhanhong Ren and Xinjiu Song
Sensors 2025, 25(7), 2325; https://doi.org/10.3390/s25072325 - 6 Apr 2025
Viewed by 684
Abstract
The detection, observation, recognition, and statistics of marine plankton are the basis of marine ecological research. In recent years, digital holography has been widely applied to plankton detection and recognition. However, the recording and reconstruction of digital holography require a strictly controlled laboratory [...] Read more.
The detection, observation, recognition, and statistics of marine plankton are the basis of marine ecological research. In recent years, digital holography has been widely applied to plankton detection and recognition. However, the recording and reconstruction of digital holography require a strictly controlled laboratory environment and time-consuming iterative computation, respectively, which impede its application in marine plankton imaging. In this paper, an intelligent method designed with digital holography and deep learning algorithms is proposed to detect and recognize marine plankton (IDRMP). An accurate integrated A-Unet network is established under the principle of deep learning and trained by digital holograms recorded with publicly available plankton datasets. This method can complete the work of reconstructing and recognizing a variety of plankton organisms stably and efficiently by a single hologram, and a system interface of YOLOv5 that can realize the task of the end-to-end detection of plankton by a single frame is provided. The structural similarities of the images reconstructed by IDRMP are all higher than 0.97, and the average accuracy of the detection of four plankton species, namely, Appendicularian, Chaetognath, Echinoderm and Hydromedusae,, reaches 91.0% after using YOLOv5. In optical experiments, typical marine plankton collected from Weifang, China, are employed as samples. For randomly selected samples of Copepods, Tunicates and Polychaetes, the results are ideal and acceptable, and a batch detection function is developed for the learning of the system. Our test and experiment results demonstrate that this method is efficient and accurate for the detection and recognition of numerous plankton within a certain volume of space after they are recorded by digital holography. Full article
(This article belongs to the Special Issue Digital Holography in Optics: Techniques and Applications)
Show Figures

Figure 1

15 pages, 6244 KiB  
Article
Detailed Investigation of Cobalt-Rich Crusts in Complex Seamount Terrains Using the Haima ROV: Integrating Optical Imaging, Sampling, and Acoustic Methods
by Yonghang Li, Huiqiang Yao, Zongheng Chen, Lixing Wang, Haoyi Zhou, Shi Zhang and Bin Zhao
J. Mar. Sci. Eng. 2025, 13(4), 702; https://doi.org/10.3390/jmse13040702 - 1 Apr 2025
Viewed by 574
Abstract
The remotely operated vehicle (ROV), a vital deep-sea platform, offers key advantages, including operational duration via continuous umbilical power, high task adaptability, and zero human risk. It has become indispensable for deep-sea scientific research and marine engineering. To enhance surveys of cobalt-rich crusts [...] Read more.
The remotely operated vehicle (ROV), a vital deep-sea platform, offers key advantages, including operational duration via continuous umbilical power, high task adaptability, and zero human risk. It has become indispensable for deep-sea scientific research and marine engineering. To enhance surveys of cobalt-rich crusts (CRCs) on complex seamount terrains, the 4500-m-class Haima ROV integrates advanced payloads, such as underwater positioning systems, multi-angle cameras, multi-functional manipulators, subsea shallow drilling systems, sediment samplers, and acoustic crust thickness gauges. Coordinated control between deck monitoring and subsea units enables stable multi-task execution within single dives, significantly improving operational efficiency. Survey results from Caiwei Guyot reveal the following: (1) ROV-collected data were highly reliable, with high-definition video mapping CRCs distribution across varied terrains. Captured crust-bearing rocks weighed up to 78 kg, drilled cores reached 110 cm, and acoustic thickness measurements had a 1–2 cm margin of error compared to in situ cores; (2) Video and cores analysis showed summit platforms (3–5° slopes) dominated by tabular crusts with gravel-type counterparts, summit margins (5–10° slopes) hosting gravel crusts partially covered by sediment, and steep slopes (12–15° slopes) exhibiting mixed crust types under sediment coverage. Thicker crusts clustered at summit margins (14 and 15 cm, respectively) compared to thinner crusts on platforms and slopes (10 and 7 cm, respectively). The Haima ROV successfully investigated CRC resources in complex terrains, laying the groundwork for seamount crust resource evaluations. Future advancements will focus on high-precision navigation and control, high-resolution crust thickness measurement, optical imaging optimization, and AI-enhanced image recognition. Full article
Show Figures

Figure 1

19 pages, 28456 KiB  
Article
YOLO-SG: Seafloor Topography Unit Recognition and Segmentation Algorithm Based on Lightweight Upsampling Operator and Attention Mechanisms
by Yifan Jiang, Ziyin Wu, Fanlin Yang, Dineng Zhao, Xiaoming Qin, Mingwei Wang and Qiang Wang
J. Mar. Sci. Eng. 2025, 13(3), 583; https://doi.org/10.3390/jmse13030583 - 16 Mar 2025
Cited by 1 | Viewed by 753
Abstract
The recognition and segmentation of seafloor topography play a crucial role in marine science research and engineering applications. However, traditional methods for seafloor topography recognition and segmentation face several issues, such as poor capability in analyzing complex terrains and limited generalization ability. To [...] Read more.
The recognition and segmentation of seafloor topography play a crucial role in marine science research and engineering applications. However, traditional methods for seafloor topography recognition and segmentation face several issues, such as poor capability in analyzing complex terrains and limited generalization ability. To address these challenges, this study introduces the SG-MKD dataset (Submarine Geomorphology Dataset—Seamounts, Sea Knolls, Submarine Depressions) and proposes YOLO-SG (You Only Look Once—Submarine Geomorphology), an algorithm for seafloor topographic unit recognition and segmentation that leverages a lightweight upsampling operator and attention mechanisms. The SG-MKD dataset provides instance segmentation annotations for three types of seafloor topographic units—seamounts, sea knolls, and submarine depressions—across a total of 419 images. YOLO-SG is an optimized version of the YOLOv8l-Segment model, incorporating a convolutional block attention module in the backbone network to enhance feature extraction. Additionally, it integrates a lightweight, general upsampling operator to create a new feature fusion network, thereby improving the model’s ability to fuse and represent features. Experimental results demonstrate that YOLO-SG significantly outperforms the original YOLOv8l-Segment, with a 14.7% increase in mean average precision. Furthermore, inference experiments conducted across various research areas highlight the model’s strong generalization capability. Full article
Show Figures

Figure 1

18 pages, 35678 KiB  
Article
Novelty Recognition: Fish Species Classification via Open-Set Recognition
by Manuel Córdova, Ricardo da Silva Torres, Aloysius van Helmond and Gert Kootstra
Sensors 2025, 25(5), 1570; https://doi.org/10.3390/s25051570 - 4 Mar 2025
Viewed by 1020
Abstract
To support the sustainable use of marine resources, regulations have been proposed to reduce fish discards focusing on the registration of all listed species. To comply with such regulations, computer vision methods have been developed. Nevertheless, current approaches are constrained by their closed-set [...] Read more.
To support the sustainable use of marine resources, regulations have been proposed to reduce fish discards focusing on the registration of all listed species. To comply with such regulations, computer vision methods have been developed. Nevertheless, current approaches are constrained by their closed-set nature, where they are designed only to recognize fish species that were present during training. In the real world, however, samples of unknown fish species may appear in different fishing regions or seasons, requiring fish classification to be treated as an open-set problem. This work focuses on the assessment of open-set recognition to automate the registration process of fish. The state-of-the-art Multiple Gaussian Prototype Learning (MGPL) was compared with the simple yet powerful Open-Set Nearest Neighbor (OSNN) and the Probability of Inclusion Support Vector Machine (PISVM). For the experiments, the Fish Detection and Weight Estimation dataset, containing images of 2216 fish instances from nine species, was used. Experimental results demonstrated that OSNN and PISVM outperformed MGPL in both recognizing known and unknown species. OSNN achieved the best results when classifying samples as either one of the known species or as an unknown species with an F1-macro of 0.79±0.05 and an AUROC score of 0.92±0.01 surpassing PISVM by 0.05 and 0.03, respectively. Full article
Show Figures

Figure 1

13 pages, 6501 KiB  
Article
Recognition of Underwater Engineering Structures Using CNN Models and Data Expansion on Side-Scan Sonar Images
by Xing Du, Yongfu Sun, Yupeng Song, Lifeng Dong, Changfei Tao and Dong Wang
J. Mar. Sci. Eng. 2025, 13(3), 424; https://doi.org/10.3390/jmse13030424 - 25 Feb 2025
Viewed by 758
Abstract
Side-scan sonar (SSS) is a critical tool in marine geophysical exploration, enabling the detection of seabed structures and geological phenomena. However, the manual interpretation of SSS images is time-consuming and relies heavily on expertise, limiting its efficiency and scalability. This study addresses these [...] Read more.
Side-scan sonar (SSS) is a critical tool in marine geophysical exploration, enabling the detection of seabed structures and geological phenomena. However, the manual interpretation of SSS images is time-consuming and relies heavily on expertise, limiting its efficiency and scalability. This study addresses these challenges by employing deep learning techniques for the automatic recognition of SSS images and introducing Marine-PULSE, a specialized dataset focusing on underwater engineering structures. The dataset refines previous classifications by distinguishing four categories of objects: pipeline or cable, underwater residual mound, seabed surface, and engineering platform. A convolutional neural network (CNN) model based on GoogleNet architecture, combined with transfer learning, was applied to assess classification accuracy and the impact of data expansion. The results demonstrate a test accuracy exceeding 92%, with data expansion improving small-sample model performance by over 7%. Notably, mutual influence effects were observed between categories, with similar features enhancing classification accuracy and distinct features causing inhibitory effects. These findings highlight the importance of balanced datasets and effective data expansion strategies in overcoming data scarcity. This work establishes a robust framework for SSS image recognition, advancing applications in marine geophysical exploration and underwater object detection. Full article
(This article belongs to the Special Issue Marine Geohazards: Characterization to Prediction)
Show Figures

Figure 1

20 pages, 8476 KiB  
Article
AquaPile-YOLO: Pioneering Underwater Pile Foundation Detection with Forward-Looking Sonar Image Processing
by Zhongwei Xu, Rui Wang, Tianyu Cao, Wenbo Guo, Bo Shi and Qiqi Ge
Remote Sens. 2025, 17(3), 360; https://doi.org/10.3390/rs17030360 - 22 Jan 2025
Cited by 3 | Viewed by 950
Abstract
Underwater pile foundation detection is crucial for environmental monitoring and marine engineering. Traditional methods for detecting underwater pile foundations are labor-intensive and inefficient. Deep learning-based image processing has revolutionized detection, enabling identification through sonar imagery analysis. This study proposes an innovative methodology, named [...] Read more.
Underwater pile foundation detection is crucial for environmental monitoring and marine engineering. Traditional methods for detecting underwater pile foundations are labor-intensive and inefficient. Deep learning-based image processing has revolutionized detection, enabling identification through sonar imagery analysis. This study proposes an innovative methodology, named the AquaPile-YOLO algorithm, for underwater pile foundation detection. Our approach significantly enhances detection accuracy and robustness by integrating multi-scale feature fusion, improved attention mechanisms, and advanced data augmentation techniques. Trained on 4000 sonar images, the model excels in delineating pile structures and effectively identifying underwater targets. Experimental data show that the model can achieve good target identification results in similar experimental scenarios, with a 96.89% accuracy rate for underwater target recognition. Full article
(This article belongs to the Special Issue Artificial Intelligence for Ocean Remote Sensing)
Show Figures

Figure 1

26 pages, 5245 KiB  
Article
An Imaging Method for Marine Targets in Corner Reflector Jamming Scenario Based on Time–Frequency Analysis and Modified Clean Technique
by Changhong Chen, Wenkang Liu, Yuexin Gao, Lei Cui, Quan Chen, Jixiang Fu and Mengdao Xing
Remote Sens. 2025, 17(2), 310; https://doi.org/10.3390/rs17020310 - 16 Jan 2025
Viewed by 794
Abstract
In the corner reflector jamming scenario, the ship target and the corner reflector array have different degrees of defocusing in the synthetic aperture radar (SAR) image due to their complex motions, which is unfavorable to the subsequent target recognition. In this manuscript, we [...] Read more.
In the corner reflector jamming scenario, the ship target and the corner reflector array have different degrees of defocusing in the synthetic aperture radar (SAR) image due to their complex motions, which is unfavorable to the subsequent target recognition. In this manuscript, we propose an imaging method for marine targets based on time–frequency analysis with the modified Clean technique. Firstly, the motion models of the ship target and the corner reflector array are established, and the characteristics of their Doppler parameter distribution are analyzed. Then, the Chirp Rate–Quadratic Chirp Rate Distribution (CR-QCRD) algorithm is utilized to estimate the Doppler parameters. To address the challenges posed by the aggregated scattering points of the ship target and the overlapping Doppler histories of the corner reflector array, the Clean technique is modified by short-time Fourier transform (STFT) filtering and amplitude–phase distortion correction using fractional Fourier transform (FrFT) filtering. This modification aims to improve the accuracy and efficiency of extracting scattering point components. Thirdly, in response to the poor universality of the traditional Clean iterative termination condition, the kurtosis of the residual signal spectrum amplitude is adopted as the new iterative termination condition. Compared with the existing imaging methods, the proposed method can adapt to the different Doppler distribution characteristics of the ship target and the corner reflector array, thus realizing better robustness in obtaining a well-focused target image. Finally, simulation experiments verify the effectiveness of the algorithm. Full article
Show Figures

Figure 1

10 pages, 827 KiB  
Technical Note
A Novel and Automated Approach to Detect Sea- and Land-Based Aquaculture Facilities
by Maxim Veroli, Marco Martinoli, Arianna Martini, Riccardo Napolitano, Domitilla Pulcini, Nicolò Tonachella and Fabrizio Capoccioni
AgriEngineering 2025, 7(1), 11; https://doi.org/10.3390/agriengineering7010011 - 6 Jan 2025
Cited by 1 | Viewed by 826
Abstract
Aquaculture is a globally widespread practice and the world’s fastest-growing food sector and requires technological advances to both increase productivity and minimize environmental impacts. Monitoring the sector is one of the priorities of state governments, international organizations, such as the Food and Agriculture [...] Read more.
Aquaculture is a globally widespread practice and the world’s fastest-growing food sector and requires technological advances to both increase productivity and minimize environmental impacts. Monitoring the sector is one of the priorities of state governments, international organizations, such as the Food and Agriculture Organization of the United States (FAO), and the European Commission. Data collection in aquaculture, particularly information on the location, number, and size of production facilities, is challenging due to the time required, the extent of the area to be monitored, the frequent changes in farming infrastructures and licenses, and the lack of automated tools. Such information is usually obtained through direct communications (e.g., phone calls and e-mails) with aquaculture producers and is rarely confirmed with on-site measurements. This study describes an innovative and automated method to obtain data on the number and placement of structures for marine and freshwater finfish farming through a YOLOv4 model trained on high-resolution images. High-resolution images were extracted from Google Maps to test their use with the YOLO model for the identification and geolocation of both land (raceways used in salmonids farming) and sea-based (floating sea cages used in seabream, seabass, and meagre farming) aquaculture systems in Italy. An overall accuracy of approximately 85% of correct object recognition of the target class was achieved. Model accuracy was tested with a dataset that includes images from Tuscany (Italy), where all these farm typologies are represented. The results demonstrate that the approach proposed can identify, characterize, and geolocate sea- and land-based aquaculture structures without performing any post-processing procedure, by directly applying customized deep learning and artificial intelligence algorithms. Full article
(This article belongs to the Special Issue The Future of Artificial Intelligence in Agriculture)
Show Figures

Graphical abstract

28 pages, 43934 KiB  
Article
A Cross-Stage Focused Small Object Detection Network for Unmanned Aerial Vehicle Assisted Maritime Applications
by Gege Ding, Jiayue Liu, Dongsheng Li, Xiaming Fu, Yucheng Zhou, Mingrui Zhang, Wantong Li, Yanjuan Wang, Chunxu Li and Xiongfei Geng
J. Mar. Sci. Eng. 2025, 13(1), 82; https://doi.org/10.3390/jmse13010082 - 5 Jan 2025
Cited by 3 | Viewed by 1541
Abstract
The application potential of unmanned aerial vehicles (UAVs) in marine search and rescue is especially of concern for the ongoing advancement of visual recognition technology and image processing technology. Limited computing resources, insufficient pixel representation for small objects in high-altitude images, and challenging [...] Read more.
The application potential of unmanned aerial vehicles (UAVs) in marine search and rescue is especially of concern for the ongoing advancement of visual recognition technology and image processing technology. Limited computing resources, insufficient pixel representation for small objects in high-altitude images, and challenging visibility conditions hinder UAVs’ target recognition performance in maritime search and rescue operations, highlighting the need for further optimization and enhancement. This study introduces an innovative detection framework, CFSD-UAVNet, designed to boost the accuracy of detecting minor objects within imagery captured from elevated altitudes. To improve the performance of the feature pyramid network (FPN) and path aggregation network (PAN), a newly designed PHead structure was proposed, focusing on better leveraging shallow features. Then, structural pruning was applied to refine the model and enhance its capability in detecting small objects. Moreover, to conserve computational resources, a lightweight CED module was introduced to reduce parameters and conserve the computing resources of the UAV. At the same time, in each detection layer, a lightweight CRE module was integrated, leveraging attention mechanisms and detection heads to enhance precision for small object detection. Finally, to enhance the model’s robustness, WIoUv2 loss function was employed, ensuring a balanced treatment of positive and negative samples. The CFSD-UAVNet model was evaluated on the publicly available SeaDronesSee maritime dataset and compared with other cutting-edge algorithms. The experimental results showed that the CFSD-UAVNet model achieved an mAP@50 of 80.1% with only 1.7 M parameters and a computational cost of 10.2 G, marking a 12.1% improvement over YOLOv8 and a 4.6% increase compared to DETR. The novel CFSD-UAVNet model effectively balances the limitations of scenarios and detection accuracy, demonstrating application potential and value in the field of UAV-assisted maritime search and rescue. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

32 pages, 6380 KiB  
Article
Application and Analysis of the MFF-YOLOv7 Model in Underwater Sonar Image Target Detection
by Kun Zheng, Haoshan Liang, Hongwei Zhao, Zhe Chen, Guohao Xie, Liguo Li, Jinghua Lu and Zhangda Long
J. Mar. Sci. Eng. 2024, 12(12), 2326; https://doi.org/10.3390/jmse12122326 - 18 Dec 2024
Cited by 3 | Viewed by 1375
Abstract
The need for precise identification of underwater sonar image targets is growing in areas such as marine resource exploitation, subsea construction, and ocean ecosystem surveillance. Nevertheless, conventional image recognition algorithms encounter several obstacles, including intricate underwater settings, poor-quality sonar image data, and limited [...] Read more.
The need for precise identification of underwater sonar image targets is growing in areas such as marine resource exploitation, subsea construction, and ocean ecosystem surveillance. Nevertheless, conventional image recognition algorithms encounter several obstacles, including intricate underwater settings, poor-quality sonar image data, and limited sample quantities, which hinder accurate identification. This study seeks to improve underwater sonar image target recognition capabilities by employing deep learning techniques and developing the Multi-Gradient Feature Fusion YOLOv7 model (MFF-YOLOv7) to address these challenges. This model incorporates the Multi-Scale Information Fusion Module (MIFM) as a replacement for YOLOv7’s SPPCSPC, substitutes the Conv of CBS following ELAN with RFAConv, and integrates the SCSA mechanism at three junctions where the backbone links to the head, enhancing target recognition accuracy. Trials were conducted using datasets like URPC, SCTD, and UATD, encompassing comparative studies of attention mechanisms, ablation tests, and evaluations against other leading algorithms. The findings indicate that the MFF-YOLOv7 model substantially surpasses other models across various metrics, demonstrates superior underwater target detection capabilities, exhibits enhanced generalization potential, and offers a more dependable and precise solution for underwater target identification. Full article
(This article belongs to the Special Issue Application of Deep Learning in Underwater Image Processing)
Show Figures

Figure 1

Back to TopTop