Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (149)

Search Parameters:
Keywords = single shot multibox detection

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 8796 KB  
Article
Subgrade Distress Detection in GPR Radargrams Using an Improved YOLOv11 Model
by Mingzhou Bai, Qun Ma, Hongyu Liu and Zilun Zhang
Sustainability 2026, 18(3), 1273; https://doi.org/10.3390/su18031273 - 27 Jan 2026
Viewed by 411
Abstract
This study compares three detectors—Single Shot MultiBox Detector (SSD), Faster Region-based Convolutional Neural Network (Faster R-CNN), and Only Look Once v11 (YOLOv11)—for detecting subgrade distress in GPR radargrams. SSD converges fastest but shows weaker detection performance, while Faster R-CNN achieves higher localization accuracy [...] Read more.
This study compares three detectors—Single Shot MultiBox Detector (SSD), Faster Region-based Convolutional Neural Network (Faster R-CNN), and Only Look Once v11 (YOLOv11)—for detecting subgrade distress in GPR radargrams. SSD converges fastest but shows weaker detection performance, while Faster R-CNN achieves higher localization accuracy at the cost of slower convergence. YOLOv11 offers the best overall performance. To push YOLOv11 further, we introduce three enhancements: a Multi-Scale Edge Enhancement Module (MEEM), a Multi-Feature Multi-Scale Attention (MFMSA) mechanism, and a hybrid configuration that combines both. On a representative dataset, YOLOv11_MEEM yields a 0.2 percentage-point increase in precision with a 0.2 percentage-point decrease in recall and a 0.3 percentage-point gain in mean Average Precision@0.5:0.95, indicating improved generalization and efficiency. YOLOv11_MFMSA achieves precision comparable to MEEM but suffers a substantial recall drop and slower inference. The hybrid YOLOv11_MEEM+MFMSA underperforms on key metrics due to gradient conflicts. MEEM reduces electromagnetic interference through dynamic edge enhancement, preserving real-time performance and robust generalization. Overall, MEEM-enhanced YOLOv11 is suitable for real-time subgrade distress detection in GPR radargrams. The research findings can offer technical support for the intelligent detection of subgrade engineering while also promoting the resilient development and sustainable operation and maintenance of urban infrastructure. Full article
Show Figures

Figure 1

22 pages, 8610 KB  
Article
A Lightweight Degradation-Aware Framework for Robust Object Detection in Adverse Weather
by Seungun Park, Jiakang Kuai, Hyunsu Kim, Hyunseong Ko, ChanSung Jung and Yunsik Son
Electronics 2026, 15(1), 146; https://doi.org/10.3390/electronics15010146 - 29 Dec 2025
Cited by 1 | Viewed by 839
Abstract
Object detection in adverse weather remains challenging due to the simultaneous degradation of visibility, structural boundaries, and semantic consistency. Existing restoration-driven or multi-branch detection approaches often fail to recover task-relevant features or introduce substantial computational overhead. To address this problem, DLC-SSD, a lightweight [...] Read more.
Object detection in adverse weather remains challenging due to the simultaneous degradation of visibility, structural boundaries, and semantic consistency. Existing restoration-driven or multi-branch detection approaches often fail to recover task-relevant features or introduce substantial computational overhead. To address this problem, DLC-SSD, a lightweight degradation-aware framework for detecting robust objects in adverse weather environments, is proposed. The framework integrates image enhancement and feature refinement into a single detection pipeline and adopts a hierarchical strategy in which global and local degradations are corrected at the image level, structural cues are reinforced in shallow high-resolution features, and semantic representations are refined in deep layers to suppress weather-induced noise. These components are jointly optimized end-to-end with the single-shot multibox detection (SSD) backbone. In rain, fog, and low-light conditions, DLC-SSD demonstrated more stable performance than conventional detectors and maintained a quasi-real-time inference speed, confirming its practicality in intelligent monitoring and autonomous driving environments. Full article
Show Figures

Figure 1

14 pages, 1787 KB  
Article
HE-DMDeception: Adversarial Attack Network for 3D Object Detection Based on Human Eye and Deep Learning Model Deception
by Pin Zhang, Yawen Liu, Heng Liu, Yichao Teng, Jiazheng Ni, Zhuansun Xiaobo and Jiajia Wang
Information 2025, 16(10), 867; https://doi.org/10.3390/info16100867 - 7 Oct 2025
Viewed by 1010
Abstract
This paper presents HE-DMDeception, a novel adversarial attack network that integrates human visual deception with deep model deception to enhance the security of 3D object detection. Existing patch-based and camouflage methods can mislead deep learning models but struggle to generate visually imperceptible, high-quality [...] Read more.
This paper presents HE-DMDeception, a novel adversarial attack network that integrates human visual deception with deep model deception to enhance the security of 3D object detection. Existing patch-based and camouflage methods can mislead deep learning models but struggle to generate visually imperceptible, high-quality textures. Our framework employs a CycleGAN-based camouflage network to generate highly camouflaged background textures, while a dedicated deception module disrupts non-maximum suppression (NMS) and attention mechanisms through optimized constraints that balance attack efficacy and visual fidelity. To overcome the scarcity of annotated vehicle data, an image segmentation module based on the pre-trained Segment Anything (SAM) model is introduced, leveraging a two-stage training strategy combining semi-supervised self-training and supervised fine-tuning. Experimental results show that the minimum P@0.5 values (50%, 55%, 20%, 25%, 25%) were achieved by HE-DMDeception across You Only Look Once version 8 (YOLOv8), Real-Time Detection Transformer (RT-DETR), Fast Region-based Convolutional Neural Network (Faster-RCNN), Single Shot MultiBox Detector (SSD), and MaskRegion-based Convolutional Neural Network (Mask RCNN) detection models, while maintaining high visual consistency with the original camouflage. These findings demonstrate the robustness and practicality of HE-DMDeception, offering new insights into 3D object detection adversarial attacks. Full article
Show Figures

Figure 1

19 pages, 5844 KB  
Article
Cloud Particle Detection in 2D-S Imaging Data via an Adaptive Anchor SSD Model
by Shuo Liu, Dingkun Yang and Luhong Fan
Atmosphere 2025, 16(8), 985; https://doi.org/10.3390/atmos16080985 - 19 Aug 2025
Viewed by 953
Abstract
The airborne 2D-S optical array probe has worked for more than ten years and has collected a large number of cloud particle images. However, existing detection methods cannot detect cloud particles with high precision due to the size differences of cloud particles and [...] Read more.
The airborne 2D-S optical array probe has worked for more than ten years and has collected a large number of cloud particle images. However, existing detection methods cannot detect cloud particles with high precision due to the size differences of cloud particles and the occurrence of particle fragmentation during imaging. So, this paper proposes a novel cloud particle detection method. The key innovation is an adaptive anchor SSD module, which overcomes existing limitations by generating anchor points that adaptively align with cloud particle size distributions. Firstly, morphological transformations generate multi-scale image information through repeated dilation and erosion operations, while removing irrelevant artifacts and fragmented particles for data cleaning. After that, the method generates geometric and mass centers across multiple scales and dynamically merges these centers to form adaptive anchor points. Finally, a detection module integrates a modified SSD with a ResNet-50 backbone for accurate bounding box predictions. Experimental results show that the proposed method achieves an mAP of 0.934 and a recall of 0.905 on the test set, demonstrating its effectiveness and reliability for cloud particle detection using the 2D-S probe. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Atmospheric Sciences)
Show Figures

Figure 1

21 pages, 5889 KB  
Article
Mobile-YOLO: A Lightweight Object Detection Algorithm for Four Categories of Aquatic Organisms
by Hanyu Jiang, Jing Zhao, Fuyu Ma, Yan Yang and Ruiwen Yi
Fishes 2025, 10(7), 348; https://doi.org/10.3390/fishes10070348 - 14 Jul 2025
Cited by 5 | Viewed by 3715
Abstract
Accurate and rapid aquatic organism recognition is a core technology for fisheries automation and aquatic organism statistical research. However, due to absorption and scattering effects, images of aquatic organisms often suffer from poor contrast and color distortion. Additionally, the clustering behavior of aquatic [...] Read more.
Accurate and rapid aquatic organism recognition is a core technology for fisheries automation and aquatic organism statistical research. However, due to absorption and scattering effects, images of aquatic organisms often suffer from poor contrast and color distortion. Additionally, the clustering behavior of aquatic organisms often leads to occlusion, further complicating the identification task. This study proposes a lightweight object detection model, Mobile-YOLO, for the recognition of four representative aquatic organisms, namely holothurian, echinus, scallop, and starfish. Our model first utilizes the Mobile-Nano backbone network we proposed, which enhances feature perception while maintaining a lightweight design. Then, we propose a lightweight detection head, LDtect, which achieves a balance between lightweight structure and high accuracy. Additionally, we introduce Dysample (dynamic sampling) and HWD (Haar wavelet downsampling) modules, aiming to optimize the feature fusion structure and achieve lightweight goals by improving the processes of upsampling and downsampling. These modules also help compensate for the accuracy loss caused by the lightweight design of LDtect. Compared to the baseline model, our model reduces Params (parameters) by 32.2%, FLOPs (floating point operations) by 28.4%, and weights (model storage size) by 30.8%, while improving FPS (frames per second) by 95.2%. The improvement in mAP (mean average precision) can also lead to better accuracy in practical applications, such as marine species monitoring, conservation efforts, and biodiversity assessment. Furthermore, the model’s accuracy is enhanced, with the mAP increased by 1.6%, demonstrating the advanced nature of our approach. Compared with YOLO (You Only Look Once) series (YOLOv5-12), SSD (Single Shot MultiBox Detector), EfficientDet (Efficient Detection), RetinaNet, and RT-DETR (Real-Time Detection Transformer), our model achieves leading comprehensive performance in terms of both accuracy and lightweight design. The results indicate that our research provides technological support for precise and rapid aquatic organism recognition. Full article
(This article belongs to the Special Issue Technology for Fish and Fishery Monitoring)
Show Figures

Figure 1

14 pages, 6120 KB  
Article
Drones and Deep Learning for Detecting Fish Carcasses During Fish Kills
by Edna G. Fernandez-Figueroa, Stephanie R. Rogers and Dinesh Neupane
Drones 2025, 9(7), 482; https://doi.org/10.3390/drones9070482 - 8 Jul 2025
Cited by 2 | Viewed by 1825
Abstract
Fish kills are sudden mass mortalities that occur in freshwater and marine systems worldwide. Fish kill surveys are essential for assessing the ecological and economic impacts of fish kill events, but are often labor-intensive, time-consuming, and spatially limited. This study aims to address [...] Read more.
Fish kills are sudden mass mortalities that occur in freshwater and marine systems worldwide. Fish kill surveys are essential for assessing the ecological and economic impacts of fish kill events, but are often labor-intensive, time-consuming, and spatially limited. This study aims to address these challenges by exploring the application of unoccupied aerial systems (or drones) and deep learning techniques for coastal fish carcass detection. Seven flights were conducted using a DJI Phantom 4 RGB quadcopter to monitor three sites with different substrates (i.e., sand, rock, shored Sargassum). Orthomosaics generated from drone imagery were useful for detecting carcasses washed ashore, but not floating or submerged carcasses. Single shot multibox detection (SSD) with a ResNet50-based model demonstrated high detection accuracy, with a mean average precision (mAP) of 0.77 and a mean average recall (mAR) of 0.81. The model had slightly higher average precision (AP) when detecting large objects (>42.24 cm long, AP = 0.90) compared to small objects (≤14.08 cm long, AP = 0.77) because smaller objects are harder to recognize and require more contextual reasoning. The results suggest a strong potential future application of these tools for rapid fish kill response and automatic enumeration and characterization of fish carcasses. Full article
Show Figures

Figure 1

16 pages, 7005 KB  
Article
Digitization of Medical Device Displays Using Deep Learning Models: A Comparative Study
by Pedro Ferreira, Pedro Lobo, Filipa Reis, João L. Vilaça and Pedro Morais
Appl. Sci. 2025, 15(10), 5436; https://doi.org/10.3390/app15105436 - 13 May 2025
Cited by 1 | Viewed by 1650
Abstract
With the growing number of patients living with chronic conditions, there is an increasing need for efficient systems that can automatically capture and convert medical device readings into digital data, particularly in home-based care settings. However, most home-based medical devices are closed systems [...] Read more.
With the growing number of patients living with chronic conditions, there is an increasing need for efficient systems that can automatically capture and convert medical device readings into digital data, particularly in home-based care settings. However, most home-based medical devices are closed systems that do not support straightforward automatic data export and often require complex connections to access or transmit patient information. Since most of these devices display clinical information on a screen, this research explores how a standard smartphone camera, combined with artificial intelligence, can be used to automatically extract the displayed data in a simple and non-intrusive way. In particular, this study provides a comparative analysis of several You Only Look Once (YOLO) and Single Shot MultiBox Detector (SSD) models to evaluate their effectiveness in detecting and recognizing the readings on medical device displays. In addition to these comparisons, we also explore a hybrid approach that combines the YOLOv8l model for object detection with a Convolutional Neural Network (CNN) for classification. Several iterations of the aforementioned models were tested, using image resolutions of 320 × 320 and 640 × 640. The performance was assessed using metrics such as precision, recall, mean average precision at 0.5 Intersection over Union (mAP@50), and frames per second (FPS). The results show that YOLOv8l (640) achieved the highest mAP@50 of 0.979, but at a lower inference speed (13.20 FPS), while YOLOv8n (320) offered the fastest inference (129.79 FPS) with a reduction in mean average precision (0.786). Combining YOLOv8l with a CNN classifier resulted in a slight reduction in overall accuracy (0.96) when compared to the standalone model (0.98). While the results are promising, the study acknowledges certain limitations, including dataset-specific biases, controlled acquisition settings, and challenges in adapting to real-world scenarios. Nevertheless, the comparative analysis offers valuable insights into the trade-off between inference time and accuracy, helping guide the selection of the most suitable model based on the specific demands of the intended scanning application. Full article
(This article belongs to the Special Issue Innovations in Artificial Neural Network Applications)
Show Figures

Figure 1

18 pages, 3958 KB  
Article
AI-Driven UAV Surveillance for Agricultural Fire Safety
by Akmalbek Abdusalomov, Sabina Umirzakova, Komil Tashev, Nodir Egamberdiev, Guzalxon Belalova, Azizjon Meliboev, Ibragim Atadjanov, Zavqiddin Temirov and Young Im Cho
Fire 2025, 8(4), 142; https://doi.org/10.3390/fire8040142 - 2 Apr 2025
Cited by 9 | Viewed by 2621
Abstract
The increasing frequency and severity of agricultural fires pose significant threats to food security, economic stability, and environmental sustainability. Traditional fire-detection methods, relying on satellite imagery and ground-based sensors, often suffer from delayed response times and high false-positive rates, limiting their effectiveness in [...] Read more.
The increasing frequency and severity of agricultural fires pose significant threats to food security, economic stability, and environmental sustainability. Traditional fire-detection methods, relying on satellite imagery and ground-based sensors, often suffer from delayed response times and high false-positive rates, limiting their effectiveness in mitigating fire-related damages. In this study, we propose an advanced deep learning-based fire-detection framework that integrates the Single-Shot MultiBox Detector (SSD) with the computationally efficient MobileNetV2 architecture. This integration enhances real-time fire- and smoke-detection capabilities while maintaining a lightweight and deployable model suitable for Unmanned Aerial Vehicle (UAV)-based agricultural monitoring. The proposed model was trained and evaluated on a custom dataset comprising diverse fire scenarios, including various environmental conditions and fire intensities. Comprehensive experiments and comparative analyses against state-of-the-art object-detection models, such as You Only Look Once (YOLO), Faster Region-based Convolutional Neural Network (Faster R-CNN), and SSD-based variants, demonstrated the superior performance of our model. The results indicate that our approach achieves a mean Average Precision (mAP) of 97.7%, significantly surpassing conventional models while maintaining a detection speed of 45 frames per second (fps) and requiring only 5.0 GFLOPs of computational power. These characteristics make it particularly suitable for deployment in edge-computing environments, such as UAVs and remote agricultural monitoring systems. Full article
Show Figures

Figure 1

16 pages, 11868 KB  
Article
A Robust YOLOv5 Model with SE Attention and BIFPN for Jishan Jujube Detection in Complex Agricultural Environments
by Hao Chen, Lijun Su, Yiren Tian, Yixin Chai, Gang Hu and Weiyi Mu
Agriculture 2025, 15(6), 665; https://doi.org/10.3390/agriculture15060665 - 20 Mar 2025
Cited by 8 | Viewed by 2561
Abstract
This study presents an improved detection model based on the YOLOv5 (You Only Look Once version 5) framework to enhance the accuracy of Jishan jujube detection in complex natural environments, particularly with varying degrees of occlusion and dense foliage. To improve detection performance, [...] Read more.
This study presents an improved detection model based on the YOLOv5 (You Only Look Once version 5) framework to enhance the accuracy of Jishan jujube detection in complex natural environments, particularly with varying degrees of occlusion and dense foliage. To improve detection performance, we integrate an SE (squeeze-and-excitation) attention module into the backbone network to enhance the model’s ability to focus on target objects while suppressing background noise. Additionally, the original neck network is replaced with a BIFPN (bi-directional feature pyramid network) structure, enabling efficient multiscale feature fusion and improving the extraction of critical features, especially for small and occluded fruits. The experimental results demonstrate that the improved YOLOv5 model achieves a mean average precision (mAP) of 96.5%, outperforming the YOLOv3, YOLOv4, YOLOv5, and SSD (Single-Shot Multibox Detector) models by 7.4%, 9.9%, 2.5%, and 0.8%, respectively. Furthermore, the proposed model improves precision (95.8%) and F1 score (92.4%), reducing false positives and achieving a better balance between precision and recall. These results highlight the model’s effectiveness in addressing missed detections of small and occluded fruits while maintaining higher confidence in predictions. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

14 pages, 7611 KB  
Article
Detection of Apple Trees in Orchard Using Monocular Camera
by Stephanie Nix, Airi Sato, Hirokazu Madokoro, Satoshi Yamamoto, Yo Nishimura and Kazuhito Sato
Agriculture 2025, 15(5), 564; https://doi.org/10.3390/agriculture15050564 - 6 Mar 2025
Cited by 1 | Viewed by 1663
Abstract
This study proposes an object detector for apple trees as a first step in developing agricultural digital twins. An original dataset of orchard images was created and used to train Single Shot MultiBox Detector (SSD) and You Only Look Once (YOLO) models. Performance [...] Read more.
This study proposes an object detector for apple trees as a first step in developing agricultural digital twins. An original dataset of orchard images was created and used to train Single Shot MultiBox Detector (SSD) and You Only Look Once (YOLO) models. Performance was evaluated using mean Average Precision (mAP). YOLO significantly outperformed SSD, achieving 91.3% mAP compared to the SSD’s 46.7%. Results indicate YOLO’s Darknet-53 backbone extracts more complex features suited to tree detection. This work demonstrates the potential of deep learning for automated data collection in smart farming applications. Full article
(This article belongs to the Special Issue Innovations in Precision Farming for Sustainable Agriculture)
Show Figures

Figure 1

27 pages, 10747 KB  
Article
MC-EVM: A Movement-Compensated EVM Algorithm with Face Detection for Remote Pulse Monitoring
by Abdallah Benhamida and Miklos Kozlovszky
Appl. Sci. 2025, 15(3), 1652; https://doi.org/10.3390/app15031652 - 6 Feb 2025
Viewed by 1995
Abstract
Automated tasks, mainly in the biomedical field, help to develop new technics to provide faster solutions for monitoring patients’ health status. For instance, they help to measure different types of human bio-signal, perform fast data analysis, and enable overall patient status monitoring. Eulerian [...] Read more.
Automated tasks, mainly in the biomedical field, help to develop new technics to provide faster solutions for monitoring patients’ health status. For instance, they help to measure different types of human bio-signal, perform fast data analysis, and enable overall patient status monitoring. Eulerian Video Magnification (EVM) can reveal small-scale and hidden changes in real life such as color and motion changes that are used to detect actual pulse. However, due to patient movement during the measurement, the EVM process will result in the wrong estimation of the pulse. In this research, we provide a working prototype for effective artefact elimination using a face movement compensated EVM (MC-EVM) which aims to track the human face as the main Region Of Interest (ROI) and then use EVM to estimate the pulse. Our primary contribution lays on the development and training of two face detection models using TensorFlow Lite: the Single-Shot MultiBox Detector (SSD) and the EfficientDet-Lite0 models that are used based on the computational capabilities of the device in use. By employing one of these models, we can crop the face accurately from the video, which is then processed using EVM to estimate the pulse. MC-EVM showed very promising results and ensured robust pulse measurement by effectively mitigating the impact of patient movement. The results were compared and validated against ground-truth data that were made available online and against pre-existing solutions from the state-of-the-art. Full article
(This article belongs to the Special Issue Monitoring of Human Physiological Signals)
Show Figures

Figure 1

38 pages, 6770 KB  
Article
Evaluation and Selection of Hardware and AI Models for Edge Applications: A Method and A Case Study on UAVs
by Müge Canpolat Şahin and Ayça Kolukısa Tarhan
Appl. Sci. 2025, 15(3), 1026; https://doi.org/10.3390/app15031026 - 21 Jan 2025
Cited by 11 | Viewed by 8487
Abstract
This study proposes a method for selecting suitable edge hardware and Artificial Intelligence (AI) models to be deployed on these edge devices. Edge AI, which enables devices at the network periphery to perform intelligent tasks locally, is rapidly expanding across various domains. However, [...] Read more.
This study proposes a method for selecting suitable edge hardware and Artificial Intelligence (AI) models to be deployed on these edge devices. Edge AI, which enables devices at the network periphery to perform intelligent tasks locally, is rapidly expanding across various domains. However, selecting appropriate edge hardware and AI models is a multi-faceted challenge due to the wide range of available options, diverse application requirements, and the unique constraints of edge environments, such as limited computational power, strict energy constraints, and the need for real-time processing. Ad hoc approaches often lead to non-optimal solutions and inefficiency problems. Considering these issues, we propose a method based on the ISO/IEC 25010:2011 quality standard, integrating Multi-Criteria Decision Analysis (MCDA) techniques to assess both the hardware and software aspects of Edge AI applications systematically. For the proposed method, we conducted an experiment consisting of two stages: In the first stage of the experiment, to show the applicability of the method across different use cases, we tested the method with four scenarios on UAVs, each presenting distinct edge requirements. In the second stage of the experiment, guided by the method’s recommendations for Scenario I, where the STM32H7 series microcontrollers were identified as the suitable hardware and the object detection model with Single Shot Multi-Box Detector (SSD) architecture and MobileNet backbone as the suitable AI model, we developed a TensorFlow Lite model from scratch to enhance the efficiency and versatility of the model for object detection tasks across various categories. This additional TensorFlow Lite model is aimed to show how the proposed method can guide the further development of optimized AI models tailored to the constraints and requirements of specific edge hardware. Full article
Show Figures

Figure 1

21 pages, 4886 KB  
Article
Comparison of CNN-Based Architectures for Detection of Different Object Classes
by Nataliya Bilous, Vladyslav Malko, Marcus Frohme and Alina Nechyporenko
AI 2024, 5(4), 2300-2320; https://doi.org/10.3390/ai5040113 - 11 Nov 2024
Cited by 22 | Viewed by 9327
Abstract
(1) Background: Detecting people and technical objects in various situations, such as natural disasters and warfare, is critical to search and rescue operations and the safety of civilians. A fast and accurate detection of people and equipment can significantly increase the effectiveness of [...] Read more.
(1) Background: Detecting people and technical objects in various situations, such as natural disasters and warfare, is critical to search and rescue operations and the safety of civilians. A fast and accurate detection of people and equipment can significantly increase the effectiveness of search and rescue missions and provide timely assistance to people. Computer vision and deep learning technologies play a key role in detecting the required objects due to their ability to analyze big volumes of visual data in real-time. (2) Methods: The performance of the neural networks such as You Only Look Once (YOLO) v4-v8, Faster R-CNN, Single Shot MultiBox Detector (SSD), and EfficientDet has been analyzed using COCO2017, SARD, SeaDronesSee, and VisDrone2019 datasets. The main metrics for comparison were mAP, Precision, Recall, F1-Score, and the ability of the neural network to work in real-time. (3) Results: The most important metrics for evaluating the efficiency and performance of models for a given task are accuracy (mAP), F1-Score, and processing speed (FPS). These metrics allow us to evaluate both the accuracy of object recognition and the ability to use the models in real-world environments where high processing speed is important. (4) Conclusion: Although different neural networks perform better on certain types of metrics, YOLO outperforms them on all metrics, showing the best results of mAP-0.88, F1-0.88, and FPS-48, so the focus was on these models. Full article
Show Figures

Figure 1

20 pages, 29993 KB  
Article
Research on the Forward Simulation and Intelligent Detection of Defects in Highways Using Ground-Penetrating Radar
by Pengxiang Li, Mingzhou Bai, Xin Li and Chenyang Liu
Appl. Sci. 2024, 14(22), 10183; https://doi.org/10.3390/app142210183 - 6 Nov 2024
Cited by 3 | Viewed by 1754
Abstract
The increasing variety and frequency of subgrade defects in operational highways have led to a rise in road safety incidents. This study employed ground-penetrating radar (GPR) detection and forward simulation to analyze the characteristic patterns of common subgrade defects, such as looseness, voids, [...] Read more.
The increasing variety and frequency of subgrade defects in operational highways have led to a rise in road safety incidents. This study employed ground-penetrating radar (GPR) detection and forward simulation to analyze the characteristic patterns of common subgrade defects, such as looseness, voids, and cavities. Through the integration of instantaneous feature information from different defect patterns with complex signal techniques, the boundary judgment of structural layers and anomalies in GPR images of various subgrade defects was improved. An intelligent recognition platform was established, and a radar image dataset was created and trained to evaluate the recognition performance of the You Only Look Once (YOLO) v3 and Single-Shot Multi-Box Detector (SSD) algorithms. Evaluation metrics such as precision, recall, F1-score, average precision (AP), and mean average precision (mAP) were used to assess the detection efficiency and accuracy for subgrade defect images. The results showed that YOLO v3 achieved an average detection accuracy of 76.69%, while the SSD achieved 75.07%. This study demonstrates that the reliability of the intelligent recognition and classification of highway subgrade defects can be enhanced by using GPR for non-destructive testing. Full article
(This article belongs to the Section Civil Engineering)
Show Figures

Figure 1

28 pages, 27981 KB  
Article
Acoustic Imaging Learning-Based Approaches for Marine Litter Detection and Classification
by Pedro Alves Guedes, Hugo Miguel Silva, Sen Wang, Alfredo Martins, José Almeida and Eduardo Silva
J. Mar. Sci. Eng. 2024, 12(11), 1984; https://doi.org/10.3390/jmse12111984 - 3 Nov 2024
Cited by 4 | Viewed by 3123
Abstract
This paper introduces an advanced acoustic imaging system leveraging multibeam water column data at various frequencies to detect and classify marine litter. This study encompasses (i) the acquisition of test tank data for diverse types of marine litter at multiple acoustic frequencies; (ii) [...] Read more.
This paper introduces an advanced acoustic imaging system leveraging multibeam water column data at various frequencies to detect and classify marine litter. This study encompasses (i) the acquisition of test tank data for diverse types of marine litter at multiple acoustic frequencies; (ii) the creation of a comprehensive acoustic image dataset with meticulous labelling and formatting; (iii) the implementation of sophisticated classification algorithms, namely support vector machine (SVM) and convolutional neural network (CNN), alongside cutting-edge detection algorithms based on transfer learning, including single-shot multibox detector (SSD) and You Only Look once (YOLO), specifically YOLOv8. The findings reveal discrimination between different classes of marine litter across the implemented algorithms for both detection and classification. Furthermore, cross-frequency studies were conducted to assess model generalisation, evaluating the performance of models trained on one acoustic frequency when tested with acoustic images based on different frequencies. This approach underscores the potential of multibeam data in the detection and classification of marine litter in the water column, paving the way for developing novel research methods in real-life environments. Full article
(This article belongs to the Special Issue Applications of Underwater Acoustics in Ocean Engineering)
Show Figures

Figure 1

Back to TopTop