Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (98)

Search Parameters:
Keywords = SIOU loss

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 7643 KiB  
Article
Automated Detection of Micro-Scale Porosity Defects in Reflective Metal Parts via Deep Learning and Polarization Imaging
by Haozhe Li, Xing Peng, Bo Wang, Feng Shi, Yu Xia, Shucheng Li, Chong Shan and Shiqing Li
Nanomaterials 2025, 15(11), 795; https://doi.org/10.3390/nano15110795 - 25 May 2025
Viewed by 479
Abstract
Aiming at the key technology of defect detection in precision additive manufacturing of highly reflective metal materials, this study proposes an enhanced SCK-YOLOV5 framework, which combines polarization imaging and deep learning methods to significantly improve the intelligent identification ability of small metal micro [...] Read more.
Aiming at the key technology of defect detection in precision additive manufacturing of highly reflective metal materials, this study proposes an enhanced SCK-YOLOV5 framework, which combines polarization imaging and deep learning methods to significantly improve the intelligent identification ability of small metal micro and nano defects. This framework introduces the SNWD (Selective Network with attention for Defect and Weathering Degradation) Loss function, which combines the SIOU Angle Loss with the NWD distribution sensing characteristics. It is specially designed for automatic positioning and identification of micrometer hole defects. At the same time, we employ global space construction with a dual-attention mechanism and multi-scale feature refining technique with selection kernel convolution to extract multi-scale defect information from highly reflective surfaces stably. Combined with the polarization imaging preprocessing and the comparison of enhancement defects under high reflectivity, the experimental results show that the proposed method significantly improves the precision, recall rate, and mAP50 index compared with the YOLOv5 baseline (increased by 0.5%, 1.2%, and 1.8%, respectively). It is the first time that this improvement has been achieved among the existing methods based on the YOLO framework. It creates a new paradigm for intelligent defect detection in additive manufacturing of high-precision metal materials and provides more reliable technical support for quality control in industrial manufacturing. Full article
(This article belongs to the Section Nanofabrication and Nanomanufacturing)
Show Figures

Figure 1

21 pages, 7065 KiB  
Article
Lightweight UAV Detection Method Based on IASL-YOLO
by Huaiyu Yang, Bo Liang, Song Feng, Ji Jiang, Ao Fang and Chunyun Li
Drones 2025, 9(5), 325; https://doi.org/10.3390/drones9050325 - 23 Apr 2025
Cited by 1 | Viewed by 856
Abstract
The widespread application of drone technology has raised security concerns, as unauthorized drones can lead to illegal intrusions and privacy breaches. Traditional detection methods often fall short in balancing performance and lightweight design, making them unsuitable for resource-constrained scenarios. To address this, we [...] Read more.
The widespread application of drone technology has raised security concerns, as unauthorized drones can lead to illegal intrusions and privacy breaches. Traditional detection methods often fall short in balancing performance and lightweight design, making them unsuitable for resource-constrained scenarios. To address this, we propose the IASL-YOLO algorithm, which optimizes the YOLOv8s model to enhance detection accuracy and lightweight efficiency. First, we design the CFE-AFPN network to streamline the architecture while boosting feature fusion capabilities across non-adjacent layers. Second, we introduce the SIoU loss function to address the orientation mismatch issue between predicted and ground truth bounding boxes. Finally, we employ the LAMP pruning algorithm to compress the model. Experimental results on the Anti-UAV dataset show that the improved model achieves a 2.9% increase in Precision, a 6.8% increase in Recall, and 3.9% and 3.8% improvements in mAP50 and mAP50-95, respectively. Additionally, the model size is reduced by 75%, the parameter count by 78%, and computational workload by 30%. Compared to mainstream algorithms, IASL-YOLO demonstrates significant advantages in both performance and lightweight design, offering an efficient solution for drone detection tasks. Full article
(This article belongs to the Special Issue UAV Detection, Classification, and Tracking)
Show Figures

Figure 1

22 pages, 2567 KiB  
Article
FA-YOLO: A Pedestrian Detection Algorithm with Feature Enhancement and Adaptive Sparse Self-Attention
by Hang Sui, Huiyan Han, Yuzhu Cui, Menglong Yang and Binwei Pei
Electronics 2025, 14(9), 1713; https://doi.org/10.3390/electronics14091713 - 23 Apr 2025
Viewed by 835
Abstract
Pedestrian detection technology refers to identifying pedestrians within the field of view and is widely used in smart cities, public safety surveillance, and other scenarios. However, in real-world complex scenes, challenges such as high pedestrian density, occlusion, and low lighting conditions lead to [...] Read more.
Pedestrian detection technology refers to identifying pedestrians within the field of view and is widely used in smart cities, public safety surveillance, and other scenarios. However, in real-world complex scenes, challenges such as high pedestrian density, occlusion, and low lighting conditions lead to blurred image boundaries, which significantly impact accuracy of pedestrian detection. To address these challenges, we propose a novel pedestrian detection algorithm, FA-YOLO. First, to address issues of limited effective information extraction in backbone network and insufficient feature map representation, we propose a feature enhancement module (FEM) that integrates both global and local features of the feature map, thereby enhancing the network’s feature representation capability. Then, to reduce redundant information and improve adaptability to complex scenes, an adaptive sparse self-attention (ASSA) module is designed to suppress noise interactions in irrelevant regions and eliminate feature redundancy across both spatial and channel dimensions. Finally, to further enhance the model’s focus on target features, we propose cross stage partial with adaptive sparse self-attention (C3ASSA), which improves overall detection performance by reinforcing the importance of target features during the final detection stage. Additionally, a scalable intersection over union (SIoU) loss function is introduced to address the vector angle differences between predicted and ground-truth bounding boxes. Extensive experiments on the WiderPerson and RTTS datasets demonstrate that FA-YOLO achieves State-of-the-Art performance, with a precision improvement of 3.5% on the WiderPerson and 3.0% on RTTS compared to YOLOv11. Full article
(This article belongs to the Special Issue Applications of Computer Vision, 3rd Edition)
Show Figures

Figure 1

20 pages, 14434 KiB  
Article
Optimized Marine Target Detection in Remote Sensing Images with Attention Mechanism and Multi-Scale Feature Fusion
by Xiantao Jiang, Tianyi Liu, Tian Song and Qi Cen
Information 2025, 16(4), 332; https://doi.org/10.3390/info16040332 - 21 Apr 2025
Cited by 1 | Viewed by 463
Abstract
With the continuous growth of maritime activities and the shipping trade, the application of maritime target detection in remote sensing images has become increasingly important. However, existing detection methods face numerous challenges, such as small target localization, recognition of targets with large aspect [...] Read more.
With the continuous growth of maritime activities and the shipping trade, the application of maritime target detection in remote sensing images has become increasingly important. However, existing detection methods face numerous challenges, such as small target localization, recognition of targets with large aspect ratios, and high computational demands. In this paper, we propose an improved target detection model, named YOLOv5-ASC, to address the challenges in maritime target detection. The proposed YOLOv5-ASC integrates three core components: an Attention-based Receptive Field Enhancement Module (ARFEM), an optimized SIoU loss function, and a Deformable Convolution Module (C3DCN). These components work together to enhance the model’s performance in detecting complex maritime targets by improving its ability to capture multi-scale features, optimize the localization process, and adapt to the large aspect ratios typical of maritime objects. Experimental results show that, compared to the original YOLOv5 model, YOLOv5-ASC achieves a 4.36 percentage point increase in mAP@0.5 and a 9.87 percentage point improvement in precision, while maintaining computational complexity within a reasonable range. The proposed method not only achieves significant performance improvements on the ShipRSImageNet dataset but also demonstrates strong potential for application in complex maritime remote sensing scenarios. Full article
(This article belongs to the Special Issue Computer Vision for Security Applications)
Show Figures

Figure 1

23 pages, 12090 KiB  
Article
Smart Car Damage Assessment Using Enhanced YOLO Algorithm and Image Processing Techniques
by Muhammad Remzy Syah Ramazhan, Alhadi Bustamam and Rinaldi Anwar Buyung
Information 2025, 16(3), 211; https://doi.org/10.3390/info16030211 - 10 Mar 2025
Viewed by 1702
Abstract
Conventional inspections in car damage assessments depend on visual judgments by human inspectors, which are labor-intensive and prone to fraudulent practices through manipulating damages. Recent advancements in artificial intelligence have given rise to a state-of-the-art object detection algorithm, the You Only Look Once [...] Read more.
Conventional inspections in car damage assessments depend on visual judgments by human inspectors, which are labor-intensive and prone to fraudulent practices through manipulating damages. Recent advancements in artificial intelligence have given rise to a state-of-the-art object detection algorithm, the You Only Look Once algorithm (YOLO), that sets a new standard in smart and automated damage assessment. This study proposes an enhanced YOLOv9 network tailored to detect six types of car damage. The enhancements include the convolutional block attention module (CBAM), applied to the backbone layer to enhance the model’s ability to focus on key damaged regions, and the SCYLLA-IoU (SIoU) loss function, introduced for bounding box regression. To be able to assess the damage severity comprehensively, we propose a novel formula named damage severity index (DSI) for quantifying damage severity directly from images, integrating multiple factors such as the number of detected damages, the ratio of damage to the image size, object detection confidence, and the type of damage. Experimental results on the CarDD dataset show that the proposed model outperforms state-of-the-art YOLO algorithms by 1.75% and that the proposed DSI demonstrates intuitive assessment of damage severity with numbers, aiding repair decisions. Full article
(This article belongs to the Special Issue Information Processing in Multimedia Applications)
Show Figures

Figure 1

20 pages, 5206 KiB  
Article
An Improved YOLOv8-Based Method for Detecting Pests and Diseases on Cucumber Leaves in Natural Backgrounds
by Jiacong Xie, Xingliu Xie, Wu Xie and Qianxin Xie
Sensors 2025, 25(5), 1551; https://doi.org/10.3390/s25051551 - 2 Mar 2025
Viewed by 1162
Abstract
The accurate detection and identification of pests and diseases on cucumber leaves is a prerequisite for scientifically controlling such issues. To address the limited detection accuracy of existing models in complex and diverse natural backgrounds, this study proposes an improved deep learning network [...] Read more.
The accurate detection and identification of pests and diseases on cucumber leaves is a prerequisite for scientifically controlling such issues. To address the limited detection accuracy of existing models in complex and diverse natural backgrounds, this study proposes an improved deep learning network model based on YOLOv8, named SEDCN-YOLOv8. First, the deformable convolution network DCNv2 (Deformable Convolution Network version 2) is introduced, replacing the original C2f module with an improved C2f_DCNv2 module in the backbone feature extraction network’s final C2f block. This enhances the model’s ability to recognize multi-scale, deformable leaf shapes and disease characteristics. Second, a Separated and Enhancement Attention Module (SEAM) is integrated to construct an improved detection head, Detect_SEAM, which strengthens the learning of critical features in pest and disease channels. This module also captures the relationship between occluded and non-occluded leaves, thereby improving the recognition of diseased leaves that are partially obscured. Finally, the original CIOU loss function of YOLOv8 is replaced with the Focaler-SIOU loss function. The experimental results demonstrate that the SEDCN-YOLOv8 network achieves a mean average precision (mAP) of 75.1% for mAP50 and 53.1% for mAP50-95 on a cucumber pest and disease dataset, representing improvements of 1.8 and 1.5 percentage points, respectively, over the original YOLOv8 model. The new model exhibits superior detection accuracy and generalization capabilities, with a model size of 6 MB and a detection speed of 400 frames per second, fully meeting the requirements for industrial deployment and real-time detection. Therefore, the SEDCN-YOLOv8 network model demonstrates broad applicability and can be effectively used in large-scale real-world scenarios for cucumber leaf pest and disease detection. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

22 pages, 9277 KiB  
Article
LRNTRM-YOLO: Research on Real-Time Recognition of Non-Tobacco-Related Materials
by Chunjie Zhang, Lijun Yun, Chenggui Yang, Zaiqing Chen and Feiyan Cheng
Agronomy 2025, 15(2), 489; https://doi.org/10.3390/agronomy15020489 - 18 Feb 2025
Cited by 2 | Viewed by 998
Abstract
The presence of non-tobacco-related materials can significantly compromise the quality of tobacco. To accurately detect non-tobacco-related materials, this study introduces a lightweight and real-time detection model derived from the YOLOv11 framework, named LRNTRM-YOLO. Initially, due to the sub-optimal accuracy in detecting diminutive non-tobacco-related [...] Read more.
The presence of non-tobacco-related materials can significantly compromise the quality of tobacco. To accurately detect non-tobacco-related materials, this study introduces a lightweight and real-time detection model derived from the YOLOv11 framework, named LRNTRM-YOLO. Initially, due to the sub-optimal accuracy in detecting diminutive non-tobacco-related materials, the model was augmented by incorporating an additional layer dedicated to enhancing the detection of small targets, thereby improving the overall accuracy. Furthermore, an attention mechanism was incorporated into the backbone network to focus on the features of the detection targets, thereby improving the detection efficacy of the model. Simultaneously, for the introduction of the SIoU loss function, the angular vector between the bounding box regressions was utilized to define the loss function, thus improving the training efficiency of the model. Following these enhancements, a channel pruning technique was employed to streamline the network, which not only reduced the parameter count but also expedited the inference process, yielding a more compact model for non-tobacco-related material detection. The experimental results on the NTRM dataset indicate that the LRNTRM-YOLO model achieved a mean average precision (mAP) of 92.9%, surpassing the baseline model by a margin of 4.8%. Additionally, there was a 68.3% reduction in the parameters and a 15.9% decrease in floating-point operations compared to the baseline model. Comparative analysis with prominent models confirmed the superiority of the proposed model in terms of its lightweight architecture, high accuracy, and real-time capabilities, thereby offering an innovative and practical solution for detecting non-tobacco-related materials in the future. Full article
(This article belongs to the Special Issue Robotics and Automation in Farming)
Show Figures

Figure 1

18 pages, 3271 KiB  
Article
GES-YOLO: A Light-Weight and Efficient Method for Conveyor Belt Deviation Detection in Mining Environments
by Hongwei Wang, Ziming Kou and Yandong Wang
Machines 2025, 13(2), 126; https://doi.org/10.3390/machines13020126 - 8 Feb 2025
Viewed by 979
Abstract
Conveyor belt deviation is one of the most common failures in belt conveyors. To address issues such as the high computational complexity, large number of parameters, long inference time, and difficulty in feature extraction of existing conveyor belt deviation detection models, we propose [...] Read more.
Conveyor belt deviation is one of the most common failures in belt conveyors. To address issues such as the high computational complexity, large number of parameters, long inference time, and difficulty in feature extraction of existing conveyor belt deviation detection models, we propose a GES-YOLO algorithm for detecting deviation in mining belt conveyors, based on an improved YOLOv8s model. The core of this algorithm is to enhance the model’s ability to extract features in complex scenarios, thereby improving the detection efficiency. Specifically, to improve real-time detection capabilities, we introduce the Groupwise Separable Convolution (GSConv) module. Additionally, by analyzing scene features, we remove the large object detection layer, which enhances the detection speed while maintaining the feature extraction capability. Furthermore, to strengthen feature perception under low-light conditions, we introduce the Efficient Multi-Scale Attention Mechanism (EMA), allowing the model to obtain more robust features. Finally, to improve the detection capability for small objects such as conveyor rollers, we introduce the Scaled Intersection over Union (SIoU) loss function, enabling the algorithm to sensitively detect rollers and provide a precise localization for deviation detection. The experimental results show that the GES-YOLO significantly improves the detection performance in complex environments such as high-noise and low-illumination conditions in coal mines. Compared to the baseline YOLOv8s model, GES-YOLO’s mAP@0.5 and mAP@0.5:0.95 increase by 1.5% and 2.3%, respectively, while the model’s parameter count and computational complexity decrease by 38.2% and 10.5%, respectively. The Frames Per Second (FPS) of the average detection speed reaches 63.62. This demonstrates that GES-YOLO achieves a good balance between detection accuracy and inference speed, with excellent accuracy, robustness, and industrial application potential. Full article
(This article belongs to the Section Robotics, Mechatronics and Intelligent Machines)
Show Figures

Figure 1

26 pages, 7145 KiB  
Article
An Efficient and Lightweight Surface Defect Detection Method for Micro-Motor Commutators in Complex Industrial Scenarios Based on the CLS-YOLO Network
by Qipeng Chen, Qiaoqiao Xiong, Haisong Huang and Saihong Tang
Electronics 2025, 14(3), 505; https://doi.org/10.3390/electronics14030505 - 26 Jan 2025
Cited by 2 | Viewed by 1417
Abstract
Existing surface defect detection methods for micro-motor commutators suffer from low detection accuracy, poor real-time performance, and high false detection and missed detection rates for small targets. To address these issues, this paper proposes a high-performance and robust commutator surface defect detection model [...] Read more.
Existing surface defect detection methods for micro-motor commutators suffer from low detection accuracy, poor real-time performance, and high false detection and missed detection rates for small targets. To address these issues, this paper proposes a high-performance and robust commutator surface defect detection model (CLS-YOLO), using YOLOv11-n as the baseline model. First, a lightweight Cross-Scale Feature Fusion Module (CCFM) is introduced to integrate features from different scales, enhancing the model’s adaptability to scale variations and ability to detect small objects. This approach reduces model parameters and improves detection speed without compromising detection accuracy. Second, a Large Separable Kernel Attention (LSKA) module is incorporated into the detection head to strengthen feature understanding and capture, reducing interference from complex surface patterns on the commutator and significantly improving adaptability to various target types. Finally, to address issues related to the center point location, aspect ratio, angle, and sample imbalance in bounding boxes, SIoU Loss replaces the CIoU Loss in the original network, overcoming limitations of the original loss function and enhancing overall detection performance. Model performance was evaluated and compared on a commutator surface defect detection dataset, with additional experiments designed to verify the model’s effectiveness and feasibility. Experimental results show that, compared to YOLOv11-n, the CLS-YOLO model achieves a 2.08% improvement in mAP@0.5. This demonstrates that CLS-YOLO can accurately detect large defect targets while maintaining accuracy for tiny defects. Additionally, CLS-YOLO outperforms most YOLO-series models, including YOLOv8-n and YOLOv10-n. The model’s parameter count is only 1.860 million, lower than YOLOv11-n, with a detection speed increase of 8.34%, making it suitable for deployment on resource-limited terminal devices in complex industrial scenarios. Full article
(This article belongs to the Special Issue Convolutional Neural Networks and Vision Applications, 3rd Edition)
Show Figures

Figure 1

20 pages, 4889 KiB  
Article
Lightweight Transmission Line Outbreak Target Obstacle Detection Incorporating ACmix
by Junbo Hao, Guangying Yan, Lidong Wang, Honglan Pei, Xu Xiao and Baifu Zhang
Processes 2025, 13(1), 271; https://doi.org/10.3390/pr13010271 - 18 Jan 2025
Cited by 1 | Viewed by 939
Abstract
To address challenges such as the frequent misdetection of targets, missed detections of multiple targets, high computational demands, and poor real-time detection performance in the video surveillance of external breakage obstacles on transmission lines, we propose a lightweight target detection algorithm incorporating the [...] Read more.
To address challenges such as the frequent misdetection of targets, missed detections of multiple targets, high computational demands, and poor real-time detection performance in the video surveillance of external breakage obstacles on transmission lines, we propose a lightweight target detection algorithm incorporating the ACmix mechanism. First, the ShuffleNetv2 backbone network is used to reduce the model parameters and improve the detection speed. Next, the ACmix attention mechanism is integrated into the Neck layer to suppress irrelevant information, mitigate the impact of complex backgrounds on feature extraction, and enhance the network’s ability to detect small external breakage targets. Additionally, we introduce the PC-ELAN module to replace the ELAN-W module, reducing redundant feature extraction in the Neck network, lowering the model parameters, and boosting the detection efficiency. Finally, we adopt the SIoU loss function for bounding box regression, which enhances the model stability and convergence speed due to its smoothing characteristics. The experimental results show that the proposed algorithm achieves an mAP of 92.7%, which is 3% higher than the baseline network. The number of model parameters and the computational complexity are reduced by 32.3% and 44.9%, respectively, while the detection speed is improved by 3.5%. These results demonstrate that the proposed method significantly enhances the detection performance. Full article
(This article belongs to the Section AI-Enabled Process Engineering)
Show Figures

Figure 1

21 pages, 4678 KiB  
Article
TBF-YOLOv8n: A Lightweight Tea Bud Detection Model Based on YOLOv8n Improvements
by Wenhui Fang and Weizhen Chen
Sensors 2025, 25(2), 547; https://doi.org/10.3390/s25020547 - 18 Jan 2025
Cited by 2 | Viewed by 1006
Abstract
Tea bud localization detection not only ensures tea quality, improves picking efficiency, and advances intelligent harvesting, but also fosters tea industry upgrades and enhances economic benefits. To solve the problem of the high computational complexity of deep learning detection models, we developed the [...] Read more.
Tea bud localization detection not only ensures tea quality, improves picking efficiency, and advances intelligent harvesting, but also fosters tea industry upgrades and enhances economic benefits. To solve the problem of the high computational complexity of deep learning detection models, we developed the Tea Bud DSCF-YOLOv8n (TBF-YOLOv8n)lightweight detection model. Improvement of the Cross Stage Partial Bottleneck Module with Two Convolutions(C2f) module via efficient Distributed Shift Convolution (DSConv) yields the C2f module with DSConv(DSCf)module, which reduces the model’s size. Additionally, the coordinate attention (CA) mechanism is incorporated to mitigate interference from irrelevant factors, thereby improving mean accuracy. Furthermore, the SIOU_Loss (SCYLLA-IOU_Loss) function and the Dynamic Sample(DySample)up-sampling operator are implemented to accelerate convergence and enhance both average precision and detection accuracy. The experimental results show that compared to the YOLOv8n model, the TBF-YOLOv8n model has a 3.7% increase in accuracy, a 1.1% increase in average accuracy, a 44.4% reduction in gigabit floating point operations (GFLOPs), and a 13.4% reduction in the total number of parameters included in the model. In comparison experiments with a variety of lightweight detection models, the TBF-YOLOv8n still performs well in terms of detection accuracy while remaining more lightweight. In conclusion, the TBF-YOLOv8n model achieves a commendable balance between efficiency and precision, offering valuable insights for advancing intelligent tea bud harvesting technologies. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

16 pages, 4947 KiB  
Article
FE-YOLO: An Efficient Deep Learning Model Based on Feature-Enhanced YOLOv7 for Microalgae Identification and Detection
by Gege Ding, Yuhang Shi, Zhenquan Liu, Yanjuan Wang, Zhixuan Yao, Dan Zhou, Xuexiu Zhu and Yiqin Li
Biomimetics 2025, 10(1), 62; https://doi.org/10.3390/biomimetics10010062 - 16 Jan 2025
Viewed by 1983
Abstract
The identification and detection of microalgae are essential for the development and utilization of microalgae resources. Traditional methods for microalgae identification and detection have many limitations. Herein, a Feature-Enhanced YOLOv7 (FE-YOLO) model for microalgae cell identification and detection is proposed. Firstly, the feature [...] Read more.
The identification and detection of microalgae are essential for the development and utilization of microalgae resources. Traditional methods for microalgae identification and detection have many limitations. Herein, a Feature-Enhanced YOLOv7 (FE-YOLO) model for microalgae cell identification and detection is proposed. Firstly, the feature extraction capability was enhanced by integrating the CAGS (Coordinate Attention Group Shuffle Convolution) attention module into the Neck section. Secondly, the SIoU (SCYLLA-IoU) algorithm was employed to replace the CIoU (Complete IoU) loss function in the original model, addressing the issues of unstable convergence. Finally, we captured and constructed a microalgae dataset containing 6300 images of seven species of microalgae, addressing the issue of a lack of microalgae cell datasets. Compared to the YOLOv7 model, the proposed method shows greatly improved average Precision, Recall, mAP@50, and mAP@95; our proposed algorithm achieved increases of 9.6%, 1.9%, 9.7%, and 6.9%, respectively. In addition, the average detection time of a single image was 0.0455 s, marking a 9.2% improvement. Full article
(This article belongs to the Section Bioinspired Sensorics, Information Processing and Control)
Show Figures

Figure 1

19 pages, 12769 KiB  
Article
YOLOv8n-CA: Improved YOLOv8n Model for Tomato Fruit Recognition at Different Stages of Ripeness
by Xin Gao, Jieyuan Ding, Ruihong Zhang and Xiaobo Xi
Agronomy 2025, 15(1), 188; https://doi.org/10.3390/agronomy15010188 - 14 Jan 2025
Cited by 4 | Viewed by 1266
Abstract
This study addresses the challenges of tomato maturity recognition in natural environments, such as occlusion caused by branches and leaves, and the difficulty in detecting stacked fruits. To overcome these issues, we propose a novel YOLOv8n-CA method for tomato maturity recognition, which defines [...] Read more.
This study addresses the challenges of tomato maturity recognition in natural environments, such as occlusion caused by branches and leaves, and the difficulty in detecting stacked fruits. To overcome these issues, we propose a novel YOLOv8n-CA method for tomato maturity recognition, which defines four maturity stages: unripe, turning color, turning ripe, and fully ripe. The model is based on the YOLOv8n architecture, incorporating the coordinate attention (CA) mechanism into the backbone network to enhance the model’s ability to capture and express features of the tomato fruits. Additionally, the C2f-FN structure was utilized in both the backbone and neck networks to strengthen the model’s capacity to extract maturity-related features. The CARAFE up-sampling operator was integrated to expand the receptive field for improved feature fusion. Finally, the SIoU loss function was used to solve the problem of insufficient CIoU of the original loss function. Experimental results showed that the YOLOv8n-CA model had a parameter count of only 2.45 × 106, computational complexity of 6.9 GFLOPs, and a weight file size of just 4.90 MB. The model achieved a mean average precision (mAP) of 97.3%. Compared to the YOLOv8n model, it reduced the model size slightly while improving accuracy by 1.3 percentage points. When compared to seven other models—Faster R-CNN, YOLOv3s, YOLOv5s, YOLOv5m, YOLOv7, YOLOv8n, YOLOv10s, and YOLOv11n—the YOLOv8n-CA model was the smallest in size and demonstrated superior detection performance. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

22 pages, 18812 KiB  
Article
Enhancing Thyroid Nodule Detection in Ultrasound Images: A Novel YOLOv8 Architecture with a C2fA Module and Optimized Loss Functions
by Shidan Wang, Zi-An Zhao, Yuze Chen, Ye-Jiao Mao and James Chung-Wai Cheung
Technologies 2025, 13(1), 28; https://doi.org/10.3390/technologies13010028 - 9 Jan 2025
Cited by 2 | Viewed by 2689
Abstract
Thyroid-related diseases, particularly thyroid cancer, are rising globally, emphasizing the critical need for the early detection and accurate screening of thyroid nodules. Ultrasound imaging has inherent limitations—high noise, low contrast, and blurred boundaries—that make manual interpretation subjective and error-prone. To address these challenges, [...] Read more.
Thyroid-related diseases, particularly thyroid cancer, are rising globally, emphasizing the critical need for the early detection and accurate screening of thyroid nodules. Ultrasound imaging has inherent limitations—high noise, low contrast, and blurred boundaries—that make manual interpretation subjective and error-prone. To address these challenges, YOLO-Thyroid, an improved model for the automatic detection of thyroid nodules in ultrasound images, is presented herein. Building upon the YOLOv8 architecture, YOLO-Thyroid introduces the C2fA module—an extension of C2f that incorporates Coordinate Attention (CA)—to enhance feature extraction. Additionally, loss functions were incorporated, including class-weighted binary cross-entropy to alleviate class imbalance and SCYLLA-IoU (SIoU) to improve localization accuracy during boundary regression. A publicly available thyroid ultrasound image dataset was optimized using format conversion and data augmentation. The experimental results demonstrate that YOLO-Thyroid outperforms mainstream object detection models across multiple metrics, achieving a higher detection precision of 54%. The recall, calculated based on the detection of nodules containing at least one feature suspected of being malignant, reaches 58.2%, while the model maintains a lightweight structure. The proposed method significantly advances ultrasound nodule detection, providing an effective and practical solution for enhancing diagnostic accuracy in medical imaging. Full article
Show Figures

Figure 1

18 pages, 4989 KiB  
Article
SSHP-YOLO: A High Precision Printed Circuit Board (PCB) Defect Detection Algorithm with a Small Sample
by Jianxin Wang, Lingcheng Ma, Zixin Li, Yuan Cao and Hao Zhang
Electronics 2025, 14(2), 217; https://doi.org/10.3390/electronics14020217 - 7 Jan 2025
Cited by 2 | Viewed by 1222
Abstract
In the domain of printed circuit board (PCB) defect detection, challenges such as missed detections and false positives remain prevalent. To address these challenges, we propose a small-sample, high-precision PCB defect detection algorithm, called SSHP-YOLO. The proposed method incorporates an ELAN-C module that [...] Read more.
In the domain of printed circuit board (PCB) defect detection, challenges such as missed detections and false positives remain prevalent. To address these challenges, we propose a small-sample, high-precision PCB defect detection algorithm, called SSHP-YOLO. The proposed method incorporates an ELAN-C module that merges the convolutional block attention module (CBAM) with the efficient layer aggregation network (ELAN), thereby enhancing the model’s focus on defect features and improving the detection of minute defect details. Furthermore, we introduce the ASPPCSPC structure, which extracts multi-scale features using pyramid pooling combined with dilated convolutions while maintaining the resolution of feature maps. This design improves the detection accuracy and robustness, thereby enhancing the algorithm’s generalization ability. Additionally, we employ the SIoU loss function to optimize the regression between the predicted and ground-truth bounding boxes, thus improving the localization accuracy of minute defects. The experimental results show that SSHP-YOLO achieves a recall rate that is 11.84% higher than traditional YOLOv7, with a mean average precision (mAP) of 97.80%. This leads to a substantial improvement in the detection accuracy, effectively mitigating issues related to missed and false detections in PCB defect detection tasks. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

Back to TopTop