Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (6,091)

Search Parameters:
Keywords = YOLOv8

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 6122 KB  
Article
Automated Detection and Classification of Lunar Linear Tectonic Features Using a Deep Learning Method
by Xiaoyang Liu, Yang Luo, Jianhui Wang, Denggao Qiu, Jianguo Yan, Wensong Zhang and Yaowen Luo
Remote Sens. 2026, 18(9), 1330; https://doi.org/10.3390/rs18091330 (registering DOI) - 26 Apr 2026
Abstract
On the lunar surface, wrinkle ridges, grabens, and lobate scarps represent key tectonic landforms that reflect the evolution of the Moon’s stress field and its tectonic processes. However, these linear structures often exhibit weak textures, low contrast, and large scale variations, making manual [...] Read more.
On the lunar surface, wrinkle ridges, grabens, and lobate scarps represent key tectonic landforms that reflect the evolution of the Moon’s stress field and its tectonic processes. However, these linear structures often exhibit weak textures, low contrast, and large scale variations, making manual interpretation inefficient and subjective. To address this issue, this study introduces an improved YOLOv8 model, termed HL-YOLOv8, for the automated detection of lunar linear features. The model incorporates a multiscale lightweight channel attention (C2f_MLCA) module into the backbone network to enhance the extraction of fine-grained and weak-texture features and integrates a multihead self-attention (C2f_MHSA) module in the feature fusion stage to improve the modelling of long-range spatial dependencies. In addition, the combination of a dual focal loss and a diversified data augmentation strategy effectively mitigates the detection difficulties caused by class imbalance and weak-feature samples. The experimental results obtained using the global LROC-WAC image dataset demonstrate that HL-YOLOv8 significantly outperforms the baseline YOLOv8 and other comparative models in terms of precision, recall, and mAP@0.5. Specifically, the proposed model achieved an average precision of 73.5%, an average recall of 73.1%, and an average mAP@0.5 of 74.6% on the evaluation dataset, showing particularly strong performance in detecting elongated grabens and boundary-blurred lobate scarps. The global distribution maps derived from the model predictions indicate that HL-YOLOv8 can be applied to comprehensively reconstruct the spatial patterns of the three types of linear structures and identify potential new features in high-latitude and geologically complex regions, demonstrating excellent generalizability and robustness. This study provides an efficient and reliable framework for the automated identification and global mapping of lunar linear features and offers a transferable methodological reference for the tectonic interpretation of terrestrial planets. Full article
23 pages, 5919 KB  
Article
Backbone and Feature Fusion Design for YOLOv8-Based Bacterial Microcolony Detection in Microscopy Images
by Malek Rababa, Anas AlSobeh, Namariq Dhahir and Amer AbuGhazaleh
Appl. Sci. 2026, 16(9), 4241; https://doi.org/10.3390/app16094241 (registering DOI) - 26 Apr 2026
Abstract
Foodborne bacterial contamination creates significant public health and economic challenges. In the United States, the CDC estimates that foodborne illness causes approximately 48 million illnesses and 3000 deaths annually. Rapid screening is important because conventional confirmation methods are time- and labor-intensive. Microscopy-based analysis [...] Read more.
Foodborne bacterial contamination creates significant public health and economic challenges. In the United States, the CDC estimates that foodborne illness causes approximately 48 million illnesses and 3000 deaths annually. Rapid screening is important because conventional confirmation methods are time- and labor-intensive. Microscopy-based analysis of early bacterial microcolonies can enable detection within hours rather than days, yet manual inspection is slow, subjective, and impractical at scale. Although deep learning object detectors such as YOLO offer a promising solution, the impact of architectural design choices on microscopy-based bacterial detection has not been systematically characterized under controlled conditions. In this work, we conducted a controlled architectural evaluation of YOLOv8 for detecting bacterial microcolonies in high-resolution microscopy images. We replaced the CSP-Darknet backbone with EfficientNetV2 variants and evaluated three feature fusion designs: no neck, the original PAN-FPN neck, and a NAS-FPN-inspired neck. All experiments were performed under identical conditions on a two-class dataset of Salmonella and E. coli. Our results show that EfficientNetV2 architectures consistently outperform the YOLOv8x baseline, which achieved 0.891 precision, 0.867 recall, and 0.898 mAP@50. The best overall performance was obtained with EfficientNetV2-S and the original YOLOv8 neck, reaching 0.976 precision, 0.968 recall, and 0.987 mAP@50, with comparable performance of 0.986 mAP@50 achieved by EfficientNetV2-S + NAS-FPN. The highest precision was obtained with EfficientNetV2-L + NAS-FPN, reaching 0.978. These findings demonstrate that effective bacterial detection depends on the interaction between backbone capacity and feature fusion design rather than backbone scaling alone. Full article
(This article belongs to the Special Issue Innovative Computer Vision and Deep Learning Applications)
30 pages, 6414 KB  
Article
Research on Distracted and Fatigue-Related Driving Behavior Detection Based on YOLOv12-LAD
by Xiyao Liu, Zhiwei Guan, Qiang Chen and Yi Ren
Electronics 2026, 15(9), 1838; https://doi.org/10.3390/electronics15091838 (registering DOI) - 26 Apr 2026
Abstract
Distracted and fatigue-related driving behaviors are major causes of road traffic accidents, creating an urgent need for reliable driver monitoring systems. Vision-based detection methods have garnered widespread attention due to their low cost of deployment and practical applicability. However, existing lightweight models often [...] Read more.
Distracted and fatigue-related driving behaviors are major causes of road traffic accidents, creating an urgent need for reliable driver monitoring systems. Vision-based detection methods have garnered widespread attention due to their low cost of deployment and practical applicability. However, existing lightweight models often suffer from limited global contextual perception and insufficient preservation of fine details. Motivated by these challenges, this study introduces an improved distracted and fatigue-related driving behavior detection model, YOLOv12-LAD, built on the YOLOv12 architecture. The proposed framework integrates a Large Separable Kernel Attention module (LSKA) to enhance global contextual perception, an Adaptive Downsampling module (ADown) to mitigate information loss during feature compression, and a Dynamic Sampling module (DySample) to enable content-adaptive feature reconstruction and improve multi-scale behavior representation. Experimental results show that YOLOv12-LAD achieved 97.5% precision, 96.3% recall, and 98.4% mAP@50 with only 2.5 million parameters, 6.2 GFLOPs, and an inference speed of 249 FPS. Ablation studies, comparisons with representative models, cross-dataset evaluation, and real-vehicle tests further verify the effectiveness and robustness of the proposed method. The proposed method demonstrates strong performance while maintaining computational efficiency, making it suitable for real-time vision-based driver monitoring applications. Full article
Show Figures

Figure 1

50 pages, 17736 KB  
Article
Swin–YOLOv12: A Hybrid Transformer-Based Deep Learning Approach for Enhanced Real-Time Brain Tumor Detection in MRI Images
by Mubashar Tariq and Kiho Choi
Mathematics 2026, 14(9), 1447; https://doi.org/10.3390/math14091447 (registering DOI) - 25 Apr 2026
Abstract
Brain tumors (BTs) arise from the abnormal growth of cells within brain tissue and may spread rapidly, making them a major cause of mortality worldwide. Early detection of BTs remains highly challenging due to the brain’s complex structure and the heterogeneous nature of [...] Read more.
Brain tumors (BTs) arise from the abnormal growth of cells within brain tissue and may spread rapidly, making them a major cause of mortality worldwide. Early detection of BTs remains highly challenging due to the brain’s complex structure and the heterogeneous nature of tumors. Magnetic Resonance Imaging (MRI) provides detailed information about tumor size, location, and shape, thereby supporting clinical decision-making for treatments such as chemotherapy, radiation therapy, and surgery. Traditional machine learning (ML) approaches mainly rely on manual feature extraction, whereas recent advances in Computer-Aided Diagnosis (CAD) and deep learning (DL) have enabled more accurate detection of small and complex tumor regions. To improve automated tumor detection, we propose a hybrid Swin–YOLO framework that combines the Swin Transformer (ST) with the latest CNN-based YOLOv12 model. In this framework, the Swin Transformer serves as the main backbone for feature extraction, while the Feature Pyramid Network (FPN) and Path Aggregation Network (PANet) are employed in the neck to better capture multi-scale features. For training, we used the publicly available Br35H dataset and applied data augmentation to enhance the model’s robustness and generalization capability. The experimental results show that the proposed framework achieved 99.7% accuracy, 99.4% mAP@50, and 87.2% mAP@50:95. Furthermore, we incorporated Explainable Artificial Intelligence (XAI) techniques, including Grad-CAM and SHAP, to improve the interpretability of the model by visually highlighting the tumor regions that contributed most to the prediction. In addition, we developed NeuroVision AI, a web-based application designed to support faster and more accurate clinical decision-making. Although the proposed model demonstrated strong performance on the dataset, these results should be interpreted within the context of the current experimental setting. Full article
19 pages, 3599 KB  
Article
Automated Pomelo Posture Detection: A Lightweight Deep Learning Solution for Conveyor-Based Fruit Processing
by Qingting Jin, Runqi Yuan, Jiayan Fang, Jing Huang, Jiayu Chen, Shilei Lyu, Zhen Li and Yu Deng
Agriculture 2026, 16(9), 946; https://doi.org/10.3390/agriculture16090946 - 24 Apr 2026
Abstract
In modern intelligent food processing, the unpredictable variability in pomelo orientation on high-speed conveyors poses a significant challenge to automated grading and precision peeling operations. To address this, a deep learning-based method is proposed for the real-time detection of pomelo posture. Firstly, a [...] Read more.
In modern intelligent food processing, the unpredictable variability in pomelo orientation on high-speed conveyors poses a significant challenge to automated grading and precision peeling operations. To address this, a deep learning-based method is proposed for the real-time detection of pomelo posture. Firstly, a pomelo posture dataset was constructed to support model training and validation. Secondly, to balance the extraction of posture features from uniform fruits with the low-power constraints of edge deployment, a domain-specific architectural optimization is presented. Building on the YOLOv8n framework, the proposed model synergistically integrates specialized modules. A lightweight GhostHGNetV2 foundation is utilized to significantly reduce computational redundancy while maintaining the resolution required to detect key anatomical landmarks. To overcome spatial confusion and capture multi-scale global appearance information, a multi-path coordinate attention (MPCA) module is introduced. Furthermore, the SlimNeck architecture and VoVGSCSP module streamline multi-scale feature fusion via one-time aggregation, effectively preventing computational bottlenecks. This design optimizes the computational efficiency of the model while maintaining detection accuracy. Experimental results demonstrate that compared with the baseline YOLOv8n model, the proposed method increased the mAP50 accuracy by 3.67% while reducing parameter count and computational load by 17.5% and 23.3%, respectively. Additionally, it achieved a processing speed of 19.3 FPS on the Jetson Orin Nano 6G edge platform. This research provides a critical technical foundation for the recognition of pomelo posture, enabling subsequent orientation rectification and fostering the development of streamlined, automated pomelo processing lines. Full article
Show Figures

Figure 1

24 pages, 8042 KB  
Article
Ship Target Detection Method Based on Feature Fusion and Bi-Level Routing Attention
by Danfeng Zuo, Liang Qi, Hao Ni, Song Song, Haifeng Li and Xinwen Wang
Symmetry 2026, 18(5), 729; https://doi.org/10.3390/sym18050729 - 24 Apr 2026
Abstract
Ship target detection is a prerequisite for achieving automated monitoring in ship detection systems. To address the challenge of accurately detecting ship targets in complex water environments, this study proposes a ship target detection method based on an improved YOLOv11 framework. To enhance [...] Read more.
Ship target detection is a prerequisite for achieving automated monitoring in ship detection systems. To address the challenge of accurately detecting ship targets in complex water environments, this study proposes a ship target detection method based on an improved YOLOv11 framework. To enhance the model’s ability to perceive and fuse features across multiple scales and in complex backgrounds, an Iterative Attention Feature Fusion (iAFF) module and a Biformer module are integrated at the end of the backbone network. The iAFF module iteratively optimizes multi-scale features through a two-stage attention mechanism, effectively focusing on key target regions, thereby improving the model’s detection capability for small, medium-sized, and occluded ships. The Biformer module leverages its innovative Bi-level Routing Attention (BRA) mechanism to enhance the modeling of global semantic information while reducing computational complexity, mitigating false detections caused by occlusions among ship targets, and consequently improving detection precision. This study employs the Minimum Point Distance Intersection over Union (MPDIoU) loss function, which more comprehensively measures the similarity between predicted and ground-truth bounding boxes by optimizing the distances of their key geometric points, effectively enhancing the accuracy of bounding box regression. Experimental results show that the proposed model achieved 93.96% mAP, 92.93% recall, and 94.97% precision on a self-built ship dataset, surpassing mainstream detection algorithms including YOLOv11 in multiple metrics. The model has only 2.90 M parameters, achieving a good balance between accuracy and efficiency. This provides an accurate and efficient solution for intelligent ship supervision. Full article
(This article belongs to the Section Computer)
32 pages, 18066 KB  
Article
Grapevine Winter Pruning Point Localization Using YOLO-Based Instance Segmentation
by Magdalena Kapłan and Kamil Buczyński
Agriculture 2026, 16(9), 943; https://doi.org/10.3390/agriculture16090943 - 24 Apr 2026
Abstract
Winter pruning is a key management practice in viticulture that directly affects vine architecture, yield balance, and grape quality. At the same time, it is a highly labor-intensive operation, and the selective identification of appropriate cutting locations remains one of the main challenges [...] Read more.
Winter pruning is a key management practice in viticulture that directly affects vine architecture, yield balance, and grape quality. At the same time, it is a highly labor-intensive operation, and the selective identification of appropriate cutting locations remains one of the main challenges limiting the automation of pruning in vineyards. Advances in machine vision provide new opportunities to support the development of robotic pruning systems. The objective of this study was to develop and evaluate a vision-based method for estimating grapevine pruning points and cutting lines using instance segmentation outputs generated by YOLO models. A dataset of 1500 RGB images of dormant grapevines was collected under field conditions in the Nobilis vineyard located in southeastern Poland. Two annotation strategies were implemented to define pruning regions. YOLO-based instance segmentation models were trained and evaluated for detecting cutting-related structures. Based on the predicted segmentation masks, a geometry-based method termed PCAcutSeg-V was developed to estimate class-dependent cutting points and cutting lines using principal component analysis applied to object contours. The results indicate that YOLOv8 and YOLO11 architectures achieved the highest segmentation performance among the evaluated models. The simplified annotation strategy provided more stable geometric inputs for the PCAcutSeg-V method, enabling more reliable estimation of cutting points and cutting lines compared with the extended annotation approach. When combined with the PCAcutSeg-V method, the proposed perception–geometry pipeline achieved high effectiveness in pruning decision estimation. The method was further implemented in a real-time processing pipeline using an RGB camera and an edge computing platform, where it maintained performance consistent with the results obtained from offline image analysis. These findings demonstrate that combining deep learning-based instance segmentation with deterministic geometric reasoning enables accurate and interpretable estimation of grapevine pruning locations and provides a promising foundation for future autonomous pruning systems. Full article
Show Figures

Figure 1

29 pages, 15835 KB  
Article
A Lightweight Detection Model for Peanut Leaf Diseases
by Zongle Xiao, Jie Zhou, Xiaoxiao Li, Wei Ma and Fuchun Sun
Agronomy 2026, 16(9), 864; https://doi.org/10.3390/agronomy16090864 - 24 Apr 2026
Abstract
Peanut leaf disease detection in complex field environments faces two major challenges: distinguishing visually similar symptoms and identifying severely occluded lesions. This study presents YOLOv8-MSDH, a lightweight detection model built upon an improved YOLOv8 framework to address these issues. Four architectural enhancements are [...] Read more.
Peanut leaf disease detection in complex field environments faces two major challenges: distinguishing visually similar symptoms and identifying severely occluded lesions. This study presents YOLOv8-MSDH, a lightweight detection model built upon an improved YOLOv8 framework to address these issues. Four architectural enhancements are introduced. First, the MHSA attention mechanism is integrated to enhance sequential feature dependency modeling and suppress background noise. Second, the Slim-Neck module is adopted for neck reconstruction, which lowers computational cost and facilitates multi-scale feature fusion. Third, the original C2f module is replaced with the C2f-Dual module to further reduce computational load. Last, the HWD downsampling module is incorporated into the backbone to improve the retention of disease-specific features while promoting lightweight design. Evaluated on a peanut leaf disease dataset, YOLOv8-MSDH achieves 90.14% precision, 82.16% recall, 90.71% mAP50 and 72.14% mAP50-95 under complex conditions—surpassing the baseline YOLOv8 by 3.5, 0.7, 1.5 and 2.89 percentage points, respectively. Parameter count and computational complexity are reduced by 12.9% and 20.7%, confirming effective lightweighting. Operating at 509.95 FPS, the model maintains strong real-time performance and exhibits high robustness across varying lighting conditions. Full article
(This article belongs to the Special Issue Digital Twins in Precision Agriculture)
18 pages, 7837 KB  
Article
An In Situ Non-Destructive Detection Method and Device for the Quality of Dried Green Sichuan Pepper Based on the Improved YOLOv11
by Bin Li, Minxi Li, Hongsheng Ren, Chuandong Liu, Guilan Peng and Zhiheng Zeng
Agriculture 2026, 16(9), 940; https://doi.org/10.3390/agriculture16090940 - 24 Apr 2026
Abstract
In response to the subjective issues, inconsistent quality standards, high labor intensity and low sorting efficiency during the drying process of green pepper, an improved YOLOv11 algorithm was proposed for quality detection. A multi-scale edge enhancement module (MEEM) is introduced into the backbone [...] Read more.
In response to the subjective issues, inconsistent quality standards, high labor intensity and low sorting efficiency during the drying process of green pepper, an improved YOLOv11 algorithm was proposed for quality detection. A multi-scale edge enhancement module (MEEM) is introduced into the backbone network, replacing the original basic C3K2 module with C3K2-MEEM to enhance the extraction of detailed features in images of dried green Sichuan pepper and prevent missed detections, false detections, and boundary confusion. The LRSA module is integrated into the 10th layer of the backbone network to improve the clarity of the tumor-like texture of the Sichuan pepper and reduce the influence of impurities, automatically allocating attention based on feature similarity to preserve local information. In the neck layer, the DPCF module is added to the FPN+PAN feature fusion stage to achieve multi-scale feature collaboration, meeting the detection requirements of dried green Sichuan pepper. The results show that the accuracy recall rate, mean average precision, and model size of the improved MLD-YOLOv11 algorithm are 92.1%, 96.6%, 95.6%, and 11.06 MB, respectively. Compared with the training results of the original YOLOv11 model, the average accuracy of the improved model has increased by 2.2 percentage points, and GFLOPs have definitely decreased by 2 G, with parameter reduction of approximately 3.10%. Compared with other mainstream models, the MLD-YOLOv11 model has significant advantages in terms of mean average precision, model size, and floating point operations per second, making it more suitable for industrial applications and providing an efficient, accurate, and lightweight solution for the quality detection of dried green Sichuan pepper. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

31 pages, 6114 KB  
Article
A Multi-Stage YOLOv11-Based Deep Learning Framework for Robust Instance Segmentation and Material Quantification of Mixed Plastic Waste
by Andrew N. Shafik, Mohamed H. Khafagy, Alber S. Aziz and Shereen A. Hussein
Computers 2026, 15(5), 271; https://doi.org/10.3390/computers15050271 - 24 Apr 2026
Abstract
Instance segmentation in heterogeneous waste scenes remains challenging due to object variability, deformable shapes, partial occlusion, and large appearance differences across packaging types. This study presents a YOLOv11-based deep learning framework for mixed plastic waste instance segmentation, developed to connect visual perception with [...] Read more.
Instance segmentation in heterogeneous waste scenes remains challenging due to object variability, deformable shapes, partial occlusion, and large appearance differences across packaging types. This study presents a YOLOv11-based deep learning framework for mixed plastic waste instance segmentation, developed to connect visual perception with reliable material quantification. The framework integrates curated instance-level annotations, strict split isolation, multi-stage optimization, training strategy ablation, and seed-robustness analysis to support reproducible model selection. Experimental results on a held-out test set show that the optimized model achieves a mask mAP@50:95 of 0.9337, indicating strong segmentation performance under heterogeneous waste-scene conditions. To extend the analysis beyond standard vision metrics, the framework incorporates a physics-informed mask-to-mass module that converts predicted masks into class-specific mass estimates using geometric calibration and material priors. Applied to a representative stream of 1253 detected objects, the system estimated a total plastic mass of 15.48 ± 1.08 kg, corresponding to a theoretical H2 potential of 0.41 ± 0.04 kg and a greenhouse-gas avoidance of 34.57 ± 4.15 kg CO2e. Overall, the proposed framework extends waste-scene understanding beyond vision-level assessment toward physically grounded, data-driven decision support for smart material recovery systems. Full article
(This article belongs to the Special Issue Machine Learning: Innovation, Implementation, and Impact)
26 pages, 68213 KB  
Article
LDA-YOLO: A YOLO-Based Rotated Object Detection Method for Remote Sensing with Large Kernel Attention and Deformable Alignment
by Dan Shan, Dadi Cai, Xuan Tong, Yanfeng Li and Dongming Liu
Appl. Sci. 2026, 16(9), 4168; https://doi.org/10.3390/app16094168 - 24 Apr 2026
Viewed by 49
Abstract
Rotated object detection is widely adopted in remote sensing to handle arbitrary object orientations and improve localization accuracy. However, existing methods still suffer from limited global context modeling, degraded feature representation under complex backgrounds, and suboptimal optimization caused by task coupling, which jointly [...] Read more.
Rotated object detection is widely adopted in remote sensing to handle arbitrary object orientations and improve localization accuracy. However, existing methods still suffer from limited global context modeling, degraded feature representation under complex backgrounds, and suboptimal optimization caused by task coupling, which jointly restrict detection performance in challenging scenarios. To address these issues, this paper proposes a novel rotated object detection framework, termed LDA-YOLO, which systematically enhances feature modeling and prediction quality. Specifically, a Large Separable Kernel Attention (LSKA) module is introduced to approximate global spatial interactions through a low-rank separable formulation, enabling effective long-range dependency modeling with linear computational complexity. A Dual-Path Feature Refinement (DPFR) module is designed to improve feature representation by decomposing features into complementary subspaces and performing adaptive fusion to suppress redundancy and noise. In addition, an Angle-Aware Decoupled Head (AADH) is developed to explicitly separate classification, localization, and orientation estimation, thereby reducing inter-task interference and improving optimization stability. The proposed method achieves superior performance compared to existing approaches. Specifically, it improves mAP50 by 1.6% over the baseline YOLOv8n-OBB, while maintaining a lightweight design with significantly reduced computational cost. These results indicate that the proposed framework provides an effective solution for rotated object detection in complex remote sensing scenarios. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

7 pages, 952 KB  
Proceeding Paper
Obstructive Sleep Apnea (OSA) Severity Classification Using Tongue Ultrasound Images and YOLOv8
by Rosezellynda D. Regular and Cyrel O. Manlises
Eng. Proc. 2026, 134(1), 80; https://doi.org/10.3390/engproc2026134080 (registering DOI) - 23 Apr 2026
Viewed by 71
Abstract
Obstructive sleep apnea (OSA) is a widely known sleep disorder that leads to serious health problems and complications. The standard diagnosis method of OSA is polysomnography. However, the process is time-intensive, expensive, and not readily accessible. Machine learning (ML) has been increasingly applied [...] Read more.
Obstructive sleep apnea (OSA) is a widely known sleep disorder that leads to serious health problems and complications. The standard diagnosis method of OSA is polysomnography. However, the process is time-intensive, expensive, and not readily accessible. Machine learning (ML) has been increasingly applied in various medical imaging modalities; however, there is still a lack of research on applying ML to ultrasound imaging for OSA classification. Previous studies on ML applications in medical imaging adopt X-rays, Computed Tomography, and Magnetic Resonance Imaging, leaving ultrasound as an underexplored area. Using the You-Only-Look-Once version 8 algorithm and static tongue ultrasound images, we classified OSA severity: normal, mild, moderate, and severe. A total of 280 ultrasound images were augmented to 838 images using brightness scaling, which enhanced the training process of the model. The system was tested on 60 images, achieving an overall classification accuracy of 85%. The results demonstrate the possibility and potential of using machine learning and ultrasound imaging for classifying the severity of OSA, suggesting potential assistance to clinicians in diagnosing and intervening in this condition. Full article
Show Figures

Figure 1

19 pages, 20662 KB  
Article
YOLO-MSG: A Lightweight and Real-Time Photovoltaic Defect Detection Algorithm for Edge Computing
by Jingdong Zhu, Xu Qian, Liangliang Wang, Chong Yin, Tao Wang, Zhanpeng Xu, Zhenqin Yao and Ban Wang
Energies 2026, 19(9), 2043; https://doi.org/10.3390/en19092043 - 23 Apr 2026
Viewed by 154
Abstract
Photovoltaic (PV) power stations are pivotal for the renewable energy transition, yet their operational efficiency is often compromised by defects such as surface dust accumulation and cracks. Traditional manual inspections are labor-intensive and subjective, while conventional monitoring methods struggle with environmental interference. This [...] Read more.
Photovoltaic (PV) power stations are pivotal for the renewable energy transition, yet their operational efficiency is often compromised by defects such as surface dust accumulation and cracks. Traditional manual inspections are labor-intensive and subjective, while conventional monitoring methods struggle with environmental interference. This study proposes YOLO-MSG, a lightweight framework specifically designed for the automated detection of PV module defects during system operation, including normal panels as well as defective conditions such as dusty and cracked panels. The methodology integrates a Multi-Scale Grouped Convolution (MSGC) module for enhanced feature extraction and a Group-Stem Decoupled Head (GSD-Head) to reduce parameter redundancy. Furthermore, a joint optimization strategy involving LAMP and logits-based knowledge distillation is employed to facilitate edge deployment. Experimental results on a specialized PV defect dataset demonstrate that YOLO-MSG achieves a superior balance between detection accuracy and computational cost. Compared to state-of-the-art models like YOLO11 and YOLOv12, YOLO-MSG significantly reduces GFLOPs and parameter count while maintaining highly competitive mean Average Precision (mAP), with improvements of 1.35% in mAP and 2.37% in mAP50-95 over the baseline models. Specifically, the model achieves an average inference speed of 90.30 FPS on the NVIDIA Jetson AGX platform. These findings confirm the algorithm’s industrial viability, providing a robust and efficient solution for the real-time automated maintenance of photovoltaic infrastructures. Full article
Show Figures

Figure 1

26 pages, 8883 KB  
Article
Strip Steel Defect Detection Algorithm Integrating Dynamic Convolution and Attention
by Changchun Shao, Zhijie Chen and Jianjun Meng
Electronics 2026, 15(9), 1796; https://doi.org/10.3390/electronics15091796 - 23 Apr 2026
Viewed by 80
Abstract
To address the issues of low accuracy, high false positives, and missed detections in hot-rolled strip steel surface defect inspection, this paper proposes an improved detection model named DFEM-NET based on YOLOv8n. First, an efficient feature extraction module (DSC2f) based on Dynamic Snake [...] Read more.
To address the issues of low accuracy, high false positives, and missed detections in hot-rolled strip steel surface defect inspection, this paper proposes an improved detection model named DFEM-NET based on YOLOv8n. First, an efficient feature extraction module (DSC2f) based on Dynamic Snake Convolution is designed to enhance the model’s capability in capturing features of irregular and elongated defects. Second, a Feature Pyramid Shared Convolution module (FPSC) is constructed to expand the model’s receptive field and effectively suppress interference from complex backgrounds. Third, an Enhanced Feature Correction (EFC) strategy is adopted during the feature fusion stage to help the model better learn the detailed features of small defect targets. Finally, a Multi-Scale Attention Aggregation module (MSAA) is introduced before the detection head, enabling the network to focus on critical feature information and thereby comprehensively improve detection accuracy for target defects. Experimental results demonstrate that, compared to the baseline model YOLOv8n, DFEM-NET achieves a detection accuracy (mAP@0.5) of 83.5%, representing an increase of 4.8%; a recall rate of 76.4%, an increase of 3.3%; and a precision of 84.7%, an increase of 3.1%, without a significant increase in model complexity. Furthermore, generalization experiments conducted on the GC10-DET dataset confirm that the proposed algorithm exhibits exceptional generalization capability. Full article
25 pages, 14230 KB  
Article
EP-YOLO: An Enhanced Lightweight Model for Micro-Pest Detection in Agricultural Light-Trap Environments
by Yuyang Tang, Jiaxuan Wang, Wenxi Sheng and Jilong Bian
Sensors 2026, 26(9), 2607; https://doi.org/10.3390/s26092607 - 23 Apr 2026
Viewed by 112
Abstract
As food security gains increasing attention, automated pest monitoring is crucial for agricultural early warning systems. However, in practical light-trap capturing sensors, the extremely small scale of pests and complex background interference, such as unexpected reflection and occlusions, severely undermine the performance of [...] Read more.
As food security gains increasing attention, automated pest monitoring is crucial for agricultural early warning systems. However, in practical light-trap capturing sensors, the extremely small scale of pests and complex background interference, such as unexpected reflection and occlusions, severely undermine the performance of existing models, resulting in frequent missed and false detections. To deal with these challenges, this study proposes EP-YOLO, an enhanced lightweight detection architecture based on YOLOv8n. Specifically, to retain the spatial pixels of micro-targets during downsampling and isolate pest features while eliminating background noise without compromising channel information, the Spatial-to-Depth Convolution (SPD) module and the Efficient Multi-Scale Attention (EMA) module are introduced. We evaluate our model through experiments on Pest24, a dataset consisting of 24 tiny pest categories. The results demonstrate that EP-YOLO achieves a mAP@50 and mAP@50:95 of 70.5% and 47.3%, respectively, improving upon the baseline by 1.1% and 1.9%. Furthermore, EP-YOLO achieves a significant improvement in detecting certain extremely small pests. For example, Rice planthopper and Plutella xylostella show improvements of 8.4% and 3.1%, respectively, compared to the baseline. In conclusion, the physical limitations of detecting tiny pests are successfully overcome by EP-YOLO, providing a robust and deployable design for real-time agricultural monitoring systems. Full article
(This article belongs to the Section Smart Agriculture)
Back to TopTop