Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,193)

Search Parameters:
Keywords = Yolov5s

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 2839 KiB  
Article
Detection of Maize Pathogenic Fungal Spores Based on Deep Learning
by Yijie Ren, Ying Xu, Huilin Tian, Qian Zhang, Mingxiu Yang, Rongsheng Zhu, Dawei Xin, Qingshan Chen, Qiaorong Wei and Shuang Song
Agriculture 2025, 15(15), 1689; https://doi.org/10.3390/agriculture15151689 - 5 Aug 2025
Abstract
Timely detection of pathogen spores is fundamental to ensuring early intervention and reducing the spread of corn diseases, like northern corn leaf blight, corn head smut, and corn rust. Traditional spore detection methods struggle to identify spore-level targets within complex backgrounds. To improve [...] Read more.
Timely detection of pathogen spores is fundamental to ensuring early intervention and reducing the spread of corn diseases, like northern corn leaf blight, corn head smut, and corn rust. Traditional spore detection methods struggle to identify spore-level targets within complex backgrounds. To improve the recognition accuracy of various maize disease spores, this study introduced the YOLOv8s-SPM model by incorporating the space-to-depth and convolution (SPD-Conv) layers, the Partial Self-Attention (PSA) mechanism, and Minimum Point Distance Intersection over Union (MPDIoU) loss function. First, we combined SPD-Conv layers into the Backbone of the YOLOv8s to enhance recognition performance on small targets and low-resolution images. To improve computational efficiency, the PSA mechanism was incorporated within the Neck layer of the network. Finally, MPDIoU loss function was applied to refine the localization performance of bounding boxes. The results revealed that the YOLOv8s-SPM model achieved 98.9% accuracy on the mixed spore dataset. Relative to the baseline YOLOv8s, the YOLOv8s-SPM model yielded a 1.4% gain in accuracy. The improved model significantly improved spore detection accuracy and demonstrated superior performance in recognizing diverse spore types under complex background conditions. It met the demands for high-precision spore detection and filled a gap in intelligent spore recognition for maize, offering an effective starting point and practical path for future research in this field. Full article
(This article belongs to the Section Crop Protection, Diseases, Pests and Weeds)
Show Figures

Figure 1

21 pages, 4331 KiB  
Article
Research on Lightweight Tracking of Small-Sized UAVs Based on the Improved YOLOv8N-Drone Architecture
by Yongjuan Zhao, Qiang Ma, Guannan Lei, Lijin Wang and Chaozhe Guo
Drones 2025, 9(8), 551; https://doi.org/10.3390/drones9080551 - 5 Aug 2025
Abstract
Traditional unmanned aerial vehicle (UAV) detection and tracking methods have long faced the twin challenges of high cost and poor efficiency. In real-world battlefield environments with complex backgrounds, occlusions, and varying speeds, existing techniques struggle to track small UAVs accurately and stably. To [...] Read more.
Traditional unmanned aerial vehicle (UAV) detection and tracking methods have long faced the twin challenges of high cost and poor efficiency. In real-world battlefield environments with complex backgrounds, occlusions, and varying speeds, existing techniques struggle to track small UAVs accurately and stably. To tackle these issues, this paper presents an enhanced YOLOv8N-Drone-based algorithm for improved target tracking of small UAVs. Firstly, a novel module named C2f-DSFEM (Depthwise-Separable and Sobel Feature Enhancement Module) is designed, integrating Sobel convolution with depthwise separable convolution across layers. Edge detail extraction and multi-scale feature representation are synchronized through a bidirectional feature enhancement mechanism, and the discriminability of target features in complex backgrounds is thus significantly enhanced. For the feature confusion problem, the improved lightweight Context Anchored Attention (CAA) mechanism is integrated into the Neck network, which effectively improves the system’s adaptability to complex scenes. By employing a position-aware weight allocation strategy, this approach enables adaptive suppression of background interference and precise focus on the target region, thereby improving localization accuracy. At the level of loss function optimization, the traditional classification loss is replaced by the focal loss (Focal Loss). This mechanism effectively suppresses the contribution of easy-to-classify samples through a dynamic weight adjustment strategy, while significantly increasing the priority of difficult samples in the training process. The class imbalance that exists between the positive and negative samples is then significantly mitigated. Experimental results show the enhanced YOLOv8 boosts mean average precision (Map@0.5) by 12.3%, hitting 99.2%. In terms of tracking performance, the proposed YOLOv8 N-Drone algorithm achieves a 19.2% improvement in Multiple Object Tracking Accuracy (MOTA) under complex multi-scenario conditions. Additionally, the IDF1 score increases by 6.8%, and the number of ID switches is reduced by 85.2%, indicating significant improvements in both accuracy and stability of UAV tracking. Compared to other mainstream algorithms, the proposed improved method demonstrates significant advantages in tracking performance, offering a more effective and reliable solution for small-target tracking tasks in UAV applications. Full article
Show Figures

Figure 1

23 pages, 23638 KiB  
Article
Enhanced YOLO and Scanning Portal System for Vehicle Component Detection
by Feng Ye, Mingzhe Yuan, Chen Luo, Shuo Li, Duotao Pan, Wenhong Wang, Feidao Cao and Diwen Chen
Sensors 2025, 25(15), 4809; https://doi.org/10.3390/s25154809 - 5 Aug 2025
Abstract
In this paper, a novel online detection system is designed to enhance accuracy and operational efficiency in the outbound logistics of automotive components after production. The system consists of a scanning portal system and an improved YOLOv12-based detection algorithm which captures images of [...] Read more.
In this paper, a novel online detection system is designed to enhance accuracy and operational efficiency in the outbound logistics of automotive components after production. The system consists of a scanning portal system and an improved YOLOv12-based detection algorithm which captures images of automotive parts passing through the scanning portal in real time. By integrating deep learning, the system enables real-time monitoring and identification, thereby preventing misdetections and missed detections of automotive parts, in this way promoting intelligent automotive part recognition and detection. Our system introduces the A2C2f-SA module, which achieves an efficient feature attention mechanism while maintaining a lightweight design. Additionally, Dynamic Space-to-Depth (Dynamic S2D) is employed to improve convolution and replace the stride convolution and pooling layers in the baseline network, helping to mitigate the loss of fine-grained information and enhancing the network’s feature extraction capability. To improve real-time performance, a GFL-MBConv lightweight detection head is proposed. Furthermore, adaptive frequency-aware feature fusion (Adpfreqfusion) is hybridized at the end of the neck network to effectively enhance high-frequency information lost during downsampling, thereby improving the model’s detection accuracy for target objects in complex backgrounds. On-site tests demonstrate that the system achieves a comprehensive accuracy of 97.3% and an average vehicle detection time of 7.59 s, exhibiting not only high precision but also high detection efficiency. These results can make the proposed system highly valuable for applications in the automotive industry. Full article
(This article belongs to the Topic Smart Production in Terms of Industry 4.0 and 5.0)
Show Figures

Figure 1

24 pages, 5644 KiB  
Article
Design and Optimization of Target Detection and 3D Localization Models for Intelligent Muskmelon Pollination Robots
by Huamin Zhao, Shengpeng Xu, Weiqi Yan, Defang Xu, Yongzhuo Zhang, Linjun Jiang, Yabo Zheng, Erkang Zeng and Rui Ren
Horticulturae 2025, 11(8), 905; https://doi.org/10.3390/horticulturae11080905 (registering DOI) - 4 Aug 2025
Abstract
With the expansion of muskmelon cultivation, manual pollination is increasingly inadequate for sustaining industry development. Therefore, the development of automatic pollination robots holds significant importance in improving pollination efficiency and reducing labor dependency. Accurate flower detection and localization is a key technology for [...] Read more.
With the expansion of muskmelon cultivation, manual pollination is increasingly inadequate for sustaining industry development. Therefore, the development of automatic pollination robots holds significant importance in improving pollination efficiency and reducing labor dependency. Accurate flower detection and localization is a key technology for enabling automated pollination robots. In this study, the YOLO-MDL model was developed as an enhancement of YOLOv7 to achieve real-time detection and localization of muskmelon flowers. This approach adds a Coordinate Attention (CA) module to focus on relevant channel information and a Contextual Transformer (CoT) module to leverage contextual relationships among input tokens, enhancing the model’s visual representation. The pollination robot converts the 2D coordinates into 3D coordinates using a depth camera and conducts experiments on real-time detection and localization of muskmelon flowers in a greenhouse. The YOLO-MDL model was deployed in ROS to control a robotic arm for automatic pollination, verifying the accuracy of flower detection and measurement localization errors. The results indicate that the YOLO-MDL model enhances AP and F1 scores by 3.3% and 1.8%, respectively, compared to the original model. It achieves AP and F1 scores of 91.2% and 85.1%, demonstrating a clear advantage in accuracy over other models. In the localization experiments, smaller errors were revealed in all three directions. The RMSE values were 0.36 mm for the X-axis, 1.26 mm for the Y-axis, and 3.87 mm for the Z-axis. The YOLO-MDL model proposed in this study demonstrates strong performance in detecting and localizing muskmelon flowers. Based on this model, the robot can execute more precise automatic pollination and provide technical support for the future deployment of automatic pollination robots in muskmelon cultivation. Full article
Show Figures

Figure 1

19 pages, 2276 KiB  
Article
Segmentation of Stone Slab Cracks Based on an Improved YOLOv8 Algorithm
by Qitao Tian, Runshu Peng and Fuzeng Wang
Appl. Sci. 2025, 15(15), 8610; https://doi.org/10.3390/app15158610 (registering DOI) - 3 Aug 2025
Viewed by 76
Abstract
To tackle the challenges of detecting complex cracks on large stone slabs with noisy textures, this paper presents the first domain-optimized framework for stone slab cracks, an improved semantic segmentation model (YOLOv8-Seg) synergistically integrating U-NetV2, DSConv, and DySample. The network uses the lightweight [...] Read more.
To tackle the challenges of detecting complex cracks on large stone slabs with noisy textures, this paper presents the first domain-optimized framework for stone slab cracks, an improved semantic segmentation model (YOLOv8-Seg) synergistically integrating U-NetV2, DSConv, and DySample. The network uses the lightweight U-NetV2 backbone combined with dynamic feature recalibration and multi-scale refinement to better capture fine crack details. The dynamic up-sampling module (DySample) helps to adaptively reconstruct curved boundaries. In addition, the dynamic snake convolution head (DSConv) improves the model’s ability to follow irregular crack shapes. Experiments on the custom-built ST stone crack dataset show that YOLOv8-Seg achieves an mAP@0.5 of 0.856 and an mAP@0.5–0.95 of 0.479. The model also reaches a mean intersection over union (MIoU) of 79.17%, outperforming both baseline and mainstream segmentation models. Ablation studies confirm the value of each module. Comparative tests and industrial validation demonstrate stable performance across different stone materials and textures and a 30% false-positive reduction in real production environments. Overall, YOLOv8-Seg greatly improves segmentation accuracy and robustness in industrial crack detection on natural stone slabs, offering a strong solution for intelligent visual inspection in real-world applications. Full article
Show Figures

Figure 1

10 pages, 1055 KiB  
Article
Artificial Intelligence and Hysteroscopy: A Multicentric Study on Automated Classification of Pleomorphic Lesions
by Miguel Mascarenhas, Carla Peixoto, Ricardo Freire, Joao Cavaco Gomes, Pedro Cardoso, Inês Castro, Miguel Martins, Francisco Mendes, Joana Mota, Maria João Almeida, Fabiana Silva, Luis Gutierres, Bruno Mendes, João Ferreira, Teresa Mascarenhas and Rosa Zulmira
Cancers 2025, 17(15), 2559; https://doi.org/10.3390/cancers17152559 - 3 Aug 2025
Viewed by 126
Abstract
Background/Objectives: The integration of artificial intelligence (AI) in medical imaging is rapidly advancing, yet its application in gynecologic use remains limited. This proof-of-concept study presents the development and validation of a convolutional neural network (CNN) designed to automatically detect and classify endometrial [...] Read more.
Background/Objectives: The integration of artificial intelligence (AI) in medical imaging is rapidly advancing, yet its application in gynecologic use remains limited. This proof-of-concept study presents the development and validation of a convolutional neural network (CNN) designed to automatically detect and classify endometrial polyps. Methods: A multicenter dataset (n = 3) comprising 65 hysteroscopies was used, yielding 33,239 frames and 37,512 annotated objects. Still frames were extracted from full-length videos and annotated for the presence of histologically confirmed polyps. A YOLOv1-based object detection model was used with a 70–20–10 split for training, validation, and testing. Primary performance metrics included recall, precision, and mean average precision at an intersection over union (IoU) ≥ 0.50 (mAP50). Frame-level classification metrics were also computed to evaluate clinical applicability. Results: The model achieved a recall of 0.96 and precision of 0.95 for polyp detection, with a mAP50 of 0.98. At the frame level, mean recall was 0.75, precision 0.98, and F1 score 0.82, confirming high detection and classification performance. Conclusions: This study presents a CNN trained on multicenter, real-world data that detects and classifies polyps simultaneously with high diagnostic and localization performance, supported by explainable AI features that enhance its clinical integration and technological readiness. Although currently limited to binary classification, this study demonstrates the feasibility and potential of AI to reduce diagnostic subjectivity and inter-observer variability in hysteroscopy. Future work will focus on expanding the model’s capabilities to classify a broader range of endometrial pathologies, enhance generalizability, and validate performance in real-time clinical settings. Full article
Show Figures

Figure 1

17 pages, 3564 KiB  
Article
Comparative Analysis of Conventional and Focused Data Augmentation Methods in Rib Fracture Detection in CT Images
by Mehmet Çağrı Göktekin, Evrim Gül, Feyza Aksu, Yeliz Gül, Metehan Özen, Yusuf Salik, Merve Kesim Önal and Engin Avci
Diagnostics 2025, 15(15), 1938; https://doi.org/10.3390/diagnostics15151938 - 1 Aug 2025
Viewed by 191
Abstract
Background/Objectives: Rib fracture detection holds critical importance in the field of medical image processing. Methods: In this study, two different data augmentation methods, traditional data augmentation (Albumentations) and focused data augmentation (focused augmentation), were compared using computed tomography (CT) images for [...] Read more.
Background/Objectives: Rib fracture detection holds critical importance in the field of medical image processing. Methods: In this study, two different data augmentation methods, traditional data augmentation (Albumentations) and focused data augmentation (focused augmentation), were compared using computed tomography (CT) images for the detection of rib fractures on YOLOv8n, YOLOv8s, and YOLOv8m models. While the traditional data augmentation method applies general transformations to the entire image, the focused data augmentation method performs specific transformations by targeting only the fracture regions. Results: The model performance was evaluated using the Precision, Recall, mAP@50, and mAP@50–95 metrics. The findings revealed that the focused data augmentation method achieved superior performance in certain metrics. Specifically, analysis on the YOLOv8s model showed that the focused data augmentation method increased the mAP@50 value by 2.18%, reaching 0.9412, and improved the recall value for fracture detection by 5.70%, reaching 0.8766. On the other hand, the traditional data augmentation method achieved better results in overall precision metrics with the YOLOv8m model and provided a slight advantage in the mAP@50 value. Conclusions: The study indicates that focused data augmentation can contribute to achieving more reliable and accurate results in medical imaging applications. Full article
Show Figures

Figure 1

21 pages, 33884 KiB  
Article
Rapid Detection and Segmentation of Landslide Hazards in Loess Tableland Areas Using Deep Learning: A Case Study of the 2023 Jishishan Ms 6.2 Earthquake in Gansu, China
by Zhuoli Bai, Lingyun Ji, Hongtao Tang, Jiangtao Qiu, Shuai Kang, Chuanjin Liu and Zongpan Bian
Remote Sens. 2025, 17(15), 2667; https://doi.org/10.3390/rs17152667 - 1 Aug 2025
Viewed by 201
Abstract
Addressing the technical demands for the rapid, precise detection of earthquake-triggered landslides in loess tablelands, this study proposes and validates an innovative methodology integrating enhanced deep learning architectures with large-tile processing strategies, featuring two core advances: (1) a critical enhancement of YOLOv8’s shallow [...] Read more.
Addressing the technical demands for the rapid, precise detection of earthquake-triggered landslides in loess tablelands, this study proposes and validates an innovative methodology integrating enhanced deep learning architectures with large-tile processing strategies, featuring two core advances: (1) a critical enhancement of YOLOv8’s shallow layers via a higher-resolution P2 detection head to boost small-target capture capabilities, and (2) the development of a large-tile segmentation–tile mosaicking workflow to overcome the technical bottlenecks in large-scale high-resolution image processing, ensuring both timeliness and accuracy in loess landslide detection. This study utilized 20 km2 of high-precision UAV imagery acquired after the 2023 Gansu Jishishan Ms 6.2 earthquake as foundational data, applying our methodology to achieve the rapid detection and precise segmentation of landslides in the study area. Validation was conducted through a comparative analysis of high-accuracy 3D models and field investigations. (1) The model achieved simultaneous convergence of all four loss functions within a 500-epoch progressive training strategy, with mAP50(M) = 0.747 and mAP50-95(M) = 0.46, thus validating the superior detection and segmentation capabilities for the Jishishan earthquake-triggered loess landslides. (2) The enhanced algorithm detected 417 landslides with 94.1% recognition accuracy. Landslide areas ranged from 7 × 10−4 km2 to 0.217 km2 (aggregate area: 1.3 km2), indicating small-scale landslide dominance. (3) Morphological characterization and the spatial distribution analysis revealed near-vertical scarps, diverse morphological configurations, and high spatial density clustering in loess tableland landslides. Full article
Show Figures

Figure 1

28 pages, 6624 KiB  
Article
YoloMal-XAI: Interpretable Android Malware Classification Using RGB Images and YOLO11
by Chaymae El Youssofi and Khalid Chougdali
J. Cybersecur. Priv. 2025, 5(3), 52; https://doi.org/10.3390/jcp5030052 - 1 Aug 2025
Viewed by 226
Abstract
As Android malware grows increasingly sophisticated, traditional detection methods struggle to keep pace, creating an urgent need for robust, interpretable, and real-time solutions to safeguard mobile ecosystems. This study introduces YoloMal-XAI, a novel deep learning framework that transforms Android application files into RGB [...] Read more.
As Android malware grows increasingly sophisticated, traditional detection methods struggle to keep pace, creating an urgent need for robust, interpretable, and real-time solutions to safeguard mobile ecosystems. This study introduces YoloMal-XAI, a novel deep learning framework that transforms Android application files into RGB image representations by mapping DEX (Dalvik Executable), Manifest.xml, and Resources.arsc files to distinct color channels. Evaluated on the CICMalDroid2020 dataset using YOLO11 pretrained classification models, YoloMal-XAI achieves 99.87% accuracy in binary classification and 99.56% in multi-class classification (Adware, Banking, Riskware, SMS, and Benign). Compared to ResNet-50, GoogLeNet, and MobileNetV2, YOLO11 offers competitive accuracy with at least 7× faster training over 100 epochs. Against YOLOv8, YOLO11 achieves comparable or superior accuracy while reducing training time by up to 3.5×. Cross-corpus validation using Drebin and CICAndMal2017 further confirms the model’s generalization capability on previously unseen malware. An ablation study highlights the value of integrating DEX, Manifest, and Resources components, with the full RGB configuration consistently delivering the best performance. Explainable AI (XAI) techniques—Grad-CAM, Grad-CAM++, Eigen-CAM, and HiRes-CAM—are employed to interpret model decisions, revealing the DEX segment as the most influential component. These results establish YoloMal-XAI as a scalable, efficient, and interpretable framework for Android malware detection, with strong potential for future deployment on resource-constrained mobile devices. Full article
Show Figures

Figure 1

40 pages, 18923 KiB  
Article
Twin-AI: Intelligent Barrier Eddy Current Separator with Digital Twin and AI Integration
by Shohreh Kia, Johannes B. Mayer, Erik Westphal and Benjamin Leiding
Sensors 2025, 25(15), 4731; https://doi.org/10.3390/s25154731 - 31 Jul 2025
Viewed by 112
Abstract
The current paper presents a comprehensive intelligent system designed to optimize the performance of a barrier eddy current separator (BECS), comprising a conveyor belt, a vibration feeder, and a magnetic drum. This system was trained and validated on real-world industrial data gathered directly [...] Read more.
The current paper presents a comprehensive intelligent system designed to optimize the performance of a barrier eddy current separator (BECS), comprising a conveyor belt, a vibration feeder, and a magnetic drum. This system was trained and validated on real-world industrial data gathered directly from the working separator under 81 different operational scenarios. The intelligent models were used to recommend optimal settings for drum speed, belt speed, vibration intensity, and drum angle, thereby maximizing separation quality and minimizing energy consumption. the smart separation module utilizes YOLOv11n-seg and achieves a mean average precision (mAP) of 0.838 across 7163 industrial instances from aluminum, copper, and plastic materials. For shape classification (sharp vs. smooth), the model reached 91.8% accuracy across 1105 annotated samples. Furthermore, the thermal monitoring unit can detect iron contamination by analyzing temperature anomalies. Scenarios with iron showed a maximum temperature increase of over 20 °C compared to clean materials, with a detection response time of under 2.5 s. The architecture integrates a Digital Twin using Azure Digital Twins to virtually mirror the system, enabling real-time tracking, behavior simulation, and remote updates. A full connection with the PLC has been implemented, allowing the AI-driven system to adjust physical parameters autonomously. This combination of AI, IoT, and digital twin technologies delivers a reliable and scalable solution for enhanced separation quality, improved operational safety, and predictive maintenance in industrial recycling environments. Full article
(This article belongs to the Special Issue Sensors and IoT Technologies for the Smart Industry)
Show Figures

Figure 1

18 pages, 74537 KiB  
Article
SDA-YOLO: Multi-Scale Dynamic Branching and Attention Fusion for Self-Explosion Defect Detection in Insulators
by Zhonghao Yang, Wangping Xu, Nanxing Chen, Yifu Chen, Kaijun Wu, Min Xie, Hong Xu and Enhui Zheng
Electronics 2025, 14(15), 3070; https://doi.org/10.3390/electronics14153070 - 31 Jul 2025
Viewed by 169
Abstract
To enhance the performance of UAVs in detecting insulator self-explosion defects during power inspections, this paper proposes an insulator self-explosion defect recognition algorithm, SDA-YOLO, based on an improved YOLOv11s network. First, the SODL is added to YOLOv11 to fuse shallow features with deeper [...] Read more.
To enhance the performance of UAVs in detecting insulator self-explosion defects during power inspections, this paper proposes an insulator self-explosion defect recognition algorithm, SDA-YOLO, based on an improved YOLOv11s network. First, the SODL is added to YOLOv11 to fuse shallow features with deeper features, thereby improving the model’s focus on small-sized self-explosion defect features. The OBB is also employed to reduce interference from the complex background. Second, the DBB module is incorporated into the C3k2 module in the backbone to extract target features through a multi-branch parallel convolutional structure. Finally, the AIFI module replaces the C2PSA module, effectively directing and aggregating information between channels to improve detection accuracy and inference speed. The experimental results show that the average accuracy of SDA-YOLO reaches 96.0%, which is higher than the YOLOv11s baseline model of 6.6%. While maintaining high accuracy, the inference speed of SDA-YOLO can reach 93.6 frames/s, which achieves the purpose of the real-time detection of insulator faults. Full article
Show Figures

Figure 1

19 pages, 3139 KiB  
Article
Intelligent Recognition and Parameter Estimation of Radar Active Jamming Based on Oriented Object Detection
by Jiawei Lu, Yiduo Guo, Weike Feng, Xiaowei Hu, Jian Gong and Yu Zhang
Remote Sens. 2025, 17(15), 2646; https://doi.org/10.3390/rs17152646 - 30 Jul 2025
Viewed by 178
Abstract
To enhance the perception capability of radar in complex electromagnetic environments, this paper proposes an intelligent jamming recognition and parameter estimation method based on deep learning. The core idea of the method is to reformulate the jamming perception problem as an object detection [...] Read more.
To enhance the perception capability of radar in complex electromagnetic environments, this paper proposes an intelligent jamming recognition and parameter estimation method based on deep learning. The core idea of the method is to reformulate the jamming perception problem as an object detection task in computer vision, and we pioneer the application of oriented object detection to this problem, enabling simultaneous jamming classification and key parameter estimation. This method takes the time–frequency spectrogram of jamming signals as input. First, it employs the oriented object detection network YOLOv8-OBB (You Only Look Once Version 8–oriented bounding box) to identify three types of classic suppression jamming and five types of Interrupted Sampling Repeater Jamming (ISRJ) and outputs the positional information of the jamming in the time–frequency spectrogram. Second, for the five ISRJ types, a post-processing algorithm based on boxes fusion is designed to further extract features for secondary recognition. Finally, by integrating the detection box information and secondary recognition results, parameters of different ISRJ are estimated. In this paper, ablation experiments from the perspective of Non-Maximum Suppression (NMS) are conducted to simulate and compare the OBB method with the traditional horizontal bounding box-based detection approaches, highlighting OBB’s detection superiority in dense jamming scenarios. Experimental results show that, compared with existing jamming detection methods, the proposed method achieves higher detection probabilities under the jamming-to-noise ratio (JNR) ranging from 0 to 20 dB, with correct identification rates exceeding 98.5% for both primary and secondary recognition stages. Moreover, benefiting from the advanced YOLOv8 network, the method exhibits an absolute error of less than 1.85% in estimating six types of jamming parameters, outperforming existing methods in estimation accuracy across different JNR conditions. Full article
(This article belongs to the Special Issue Array and Signal Processing for Radar (Second Edition))
Show Figures

Figure 1

20 pages, 3518 KiB  
Article
YOLO-AWK: A Model for Injurious Bird Detection in Complex Farmland Environments
by Xiang Yang, Yongliang Cheng, Minggang Dong and Xiaolan Xie
Symmetry 2025, 17(8), 1210; https://doi.org/10.3390/sym17081210 - 30 Jul 2025
Viewed by 218
Abstract
Injurious birds pose a significant threat to food production and the agricultural economy. To address the challenges posed by their small size, irregular shape, and frequent occlusion in complex farmland environments, this paper proposes YOLO-AWK, an improved bird detection model based on YOLOv11n. [...] Read more.
Injurious birds pose a significant threat to food production and the agricultural economy. To address the challenges posed by their small size, irregular shape, and frequent occlusion in complex farmland environments, this paper proposes YOLO-AWK, an improved bird detection model based on YOLOv11n. Firstly, to improve the ability of the enhanced model to recognize bird targets in complex backgrounds, we introduce the in-scale feature interaction (AIFI) module to replace the original SPPF module. Secondly, to more accurately localize and identify bird targets of different shapes and sizes, we use WIoUv3 as a new loss function. Thirdly, to remove the noise interference and improve the extraction of bird residual features, we introduce the Kolmogorov–Arnold network (KAN) module. Finally, to improve the model’s detection accuracy for small bird targets, we add a small target detection head. The experimental results show that the detection performance of YOLO-AWK on the farmland bird dataset is significantly improved, and the final precision, recall, mAP@0.5, and mAP@0.5:0.95 reach 93.9%, 91.2%, 95.8%, and 75.3%, respectively, which outperforms the original model by 2.7, 2.3, 1.6, and 3.0 percentage points, respectively. These results demonstrate that the proposed method offers a reliable and efficient technical solution for farmland injurious bird monitoring. Full article
(This article belongs to the Special Issue Symmetry and Its Applications in Image Processing)
Show Figures

Figure 1

15 pages, 4592 KiB  
Article
SSAM_YOLOv5: YOLOv5 Enhancement for Real-Time Detection of Small Road Signs
by Fatima Qanouni, Hakim El Massari, Noreddine Gherabi and Maria El-Badaoui
Digital 2025, 5(3), 30; https://doi.org/10.3390/digital5030030 - 29 Jul 2025
Viewed by 354
Abstract
Many traffic-sign detection systems are available to assist drivers with particular conditions such as small and distant signs, multiple signs on the road, objects similar to signs, and other challenging conditions. Real-time object detection is an indispensable aspect of these detection systems, with [...] Read more.
Many traffic-sign detection systems are available to assist drivers with particular conditions such as small and distant signs, multiple signs on the road, objects similar to signs, and other challenging conditions. Real-time object detection is an indispensable aspect of these detection systems, with detection speed and efficiency being critical parameters. In terms of these parameters, to enhance performance in road-sign detection under diverse conditions, we proposed a comprehensive methodology, SSAM_YOLOv5, to handle feature extraction and small-road-sign detection performance. The method was based on a modified version of YOLOv5s. First, we introduced attention modules into the backbone to focus on the region of interest within video frames; secondly, we replaced the activation function with the SwishT_C activation function to enhance feature extraction and achieve a balance between inference, precision, and mean average precision (mAP@50) rates. Compared to the YOLOv5 baseline, the proposed improvements achieved remarkable increases of 1.4% and 1.9% in mAP@50 on the Tiny LISA and GTSDB datasets, respectively, confirming their effectiveness. Full article
Show Figures

Figure 1

21 pages, 2965 KiB  
Article
Inspection Method Enabled by Lightweight Self-Attention for Multi-Fault Detection in Photovoltaic Modules
by Shufeng Meng and Tianxu Xu
Electronics 2025, 14(15), 3019; https://doi.org/10.3390/electronics14153019 - 29 Jul 2025
Viewed by 237
Abstract
Bird-dropping fouling and hotspot anomalies remain the most prevalent and detrimental defects in utility-scale photovoltaic (PV) plants; their co-occurrence on a single module markedly curbs energy yield and accelerates irreversible cell degradation. However, markedly disparate visual–thermal signatures of the two phenomena impede high-fidelity [...] Read more.
Bird-dropping fouling and hotspot anomalies remain the most prevalent and detrimental defects in utility-scale photovoltaic (PV) plants; their co-occurrence on a single module markedly curbs energy yield and accelerates irreversible cell degradation. However, markedly disparate visual–thermal signatures of the two phenomena impede high-fidelity concurrent detection in existing robotic inspection systems, while stringent onboard compute budgets also preclude the adoption of bulky detectors. To resolve this accuracy–efficiency trade-off for dual-defect detection, we present YOLOv8-SG, a lightweight yet powerful framework engineered for mobile PV inspectors. First, a rigorously curated multi-modal dataset—RGB for stains and long-wave infrared for hotspots—is assembled to enforce robust cross-domain representation learning. Second, the HSV color space is leveraged to disentangle chromatic and luminance cues, thereby stabilizing appearance variations across sensors. Third, a single-head self-attention (SHSA) block is embedded in the backbone to harvest long-range dependencies at negligible parameter cost, while a global context (GC) module is grafted onto the detection head to amplify fine-grained semantic cues. Finally, an auxiliary bounding box refinement term is appended to the loss to hasten convergence and tighten localization. Extensive field experiments demonstrate that YOLOv8-SG attains 86.8% mAP@0.5, surpassing the vanilla YOLOv8 by 2.7 pp while trimming 12.6% of parameters (18.8 MB). Grad-CAM saliency maps corroborate that the model’s attention consistently coincides with defect regions, underscoring its interpretability. The proposed method, therefore, furnishes PV operators with a practical low-latency solution for concurrent bird-dropping and hotspot surveillance. Full article
Show Figures

Figure 1

Back to TopTop