Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,606)

Search Parameters:
Keywords = YOLOv3 algorithm

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 4782 KB  
Article
Cattle Farming Activity Monitoring Using Advanced Deep Learning Approach
by Muhammad Asim, Bareera Anam, Muhammad Nadeem Ali and Byung-Seo Kim
Sensors 2026, 26(3), 785; https://doi.org/10.3390/s26030785 (registering DOI) - 24 Jan 2026
Abstract
Technological advancements have significantly improved cattle farming, particularly in sensor-based activity monitoring for health management, estrus detection, and overall herd supervision. However, such a sensor-based monitoring framework often illustrates several issues, such as high cost, animal discomfort, and susceptibility to false measurement. This [...] Read more.
Technological advancements have significantly improved cattle farming, particularly in sensor-based activity monitoring for health management, estrus detection, and overall herd supervision. However, such a sensor-based monitoring framework often illustrates several issues, such as high cost, animal discomfort, and susceptibility to false measurement. This study introduces a vision-based cattle activity monitoring approach deployed in a commercial Nestlé dairy farm, specifically one that is estrus-focused, where overhead cameras capture unconstrained herd behavior under variable lighting, occlusions, and crowding. A custom dataset of 2956 Images are collected and then annotated into four fine-grained behaviors—standing, lying, grazing, and estrus—enabling detailed analysis beyond coarse activity categories commonly used in prior livestock monitoring studies. Furthermore, computer vision-based deep learning algorithms are deployed on this dataset to classify the aforementioned classes. A comparative analysis of YOLOv8 and YOLOv9 is provided, which clearly illustrates that YOLOv8-L achieved a mAP of 91.11%, whereas YOLOv9-E achieved a mAP of 90.23%. Full article
(This article belongs to the Special Issue Smart Decision Systems for Digital Farming: 2nd Edition)
Show Figures

Figure 1

20 pages, 49658 KB  
Article
Dead Chicken Identification Method Based on a Spatial-Temporal Graph Convolution Network
by Jikang Yang, Chuang Ma, Haikun Zheng, Zhenlong Wu, Xiaohuan Chao, Cheng Fang and Boyi Xiao
Animals 2026, 16(3), 368; https://doi.org/10.3390/ani16030368 - 23 Jan 2026
Abstract
In intensive cage rearing systems, accurate dead hen detection remains difficult due to complex environments, severe occlusion, and the high visual similarity between dead hens and live hens in a prone posture. To address these issues, this study proposes a dead hen identification [...] Read more.
In intensive cage rearing systems, accurate dead hen detection remains difficult due to complex environments, severe occlusion, and the high visual similarity between dead hens and live hens in a prone posture. To address these issues, this study proposes a dead hen identification method based on a Spatial-Temporal Graph Convolutional Network (STGCN). Unlike conventional static image-based approaches, the proposed method introduces temporal information to enable dynamic spatial-temporal modeling of hen health states. First, a multimodal fusion algorithm is applied to visible light and thermal infrared images to strengthen multimodal feature representation. Then, an improved YOLOv7-Pose algorithm is used to extract the skeletal keypoints of individual hens, and the ByteTrack algorithm is employed for multi-object tracking. Based on these results, spatial-temporal graph-structured data of hens are constructed by integrating spatial and temporal dimensions. Finally, a spatial-temporal graph convolution model is used to identify dead hens by learning spatial-temporal dependency features from skeleton sequences. Experimental results show that the improved YOLOv7-Pose model achieves an average precision (AP) of 92.8% in keypoint detection. Based on the constructed spatial-temporal graph data, the dead hen identification model reaches an overall classification accuracy of 99.0%, with an accuracy of 98.9% for the dead hen category. These results demonstrate that the proposed method effectively reduces interference caused by feeder occlusion and ambiguous visual features. By using dynamic spatial-temporal information, the method substantially improves robustness and accuracy of dead hen detection in complex cage rearing environments, providing a new technical route for intelligent monitoring of poultry health status. Full article
(This article belongs to the Special Issue Welfare and Behavior of Laying Hens)
Show Figures

Figure 1

26 pages, 1165 KB  
Article
Improved Dual-Module YOLOv8 Algorithm for Building Crack Detection
by Xinyu Zuo, Ahmed D. D. Almutairi, Muneer K. K. Saeed and Yiqing Dai
Buildings 2026, 16(2), 461; https://doi.org/10.3390/buildings16020461 - 22 Jan 2026
Viewed by 15
Abstract
Building cracks are significant indicators of structural integrity. Conventional fracture detection methodologies, however, are characterized by extended durations, significant labor requirements, and limitations in both precision and operational effectiveness. Findings are also subject to subjective and technical constraints inherent in manual assessments. To [...] Read more.
Building cracks are significant indicators of structural integrity. Conventional fracture detection methodologies, however, are characterized by extended durations, significant labor requirements, and limitations in both precision and operational effectiveness. Findings are also subject to subjective and technical constraints inherent in manual assessments. To overcome these challenges, this paper introduces an enhanced YOLOv8-based methodology for developing a building crack detection system, thereby achieving high precision, operational efficiency, and cost-effectiveness. Initially, classified and segmented datasets of building fractures were obtained from field photography, online image aggregation, and open-source databases, thereby providing the basis for training the experimental model. Subsequently, the Swin Transformer window multi-head self-attention mechanism was implemented to augment small-object recognition capabilities and reduce computational demands, thereby enabling the development of an enhanced image segmentation module. Utilizing the U-Net’s segmentation capabilities, a rotated split method was implemented to quantify fracture width and derive geometric parameters from the segmented crack regions. In order to evaluate the effectiveness of the model, two experiments were conducted: one to demonstrate the performance of the classification category and the other to show the capabilities of the segmentation category. The result is that the proposed model has high accuracy and efficiency in the frac detection task. This approach effectively enhances fracture detection in structural safety evaluations of these buildings, providing technical support for relevant management decisions. Full article
(This article belongs to the Special Issue Automation and Intelligence in the Construction Industry)
24 pages, 2902 KB  
Article
Research on Prolonged Violation Behavior Recognition in Construction Sites Based on Artificial Intelligence
by Kai Yu, Zhenyue Wang, Lujie Zhou, Xuesong Yang, Zhaoxiang Mu and Tianyu Wang
Symmetry 2026, 18(1), 204; https://doi.org/10.3390/sym18010204 - 22 Jan 2026
Viewed by 9
Abstract
Prolonged violation behavior is characterized by sustained temporal presence, slow action changes, and similarity to normal behavior. Due to the complex construction environment, intelligent recognition algorithms face significant challenges. This paper proposes an improved YOLOv8-based model, DGEA-YOLOv8, to address these issues, using “playing [...] Read more.
Prolonged violation behavior is characterized by sustained temporal presence, slow action changes, and similarity to normal behavior. Due to the complex construction environment, intelligent recognition algorithms face significant challenges. This paper proposes an improved YOLOv8-based model, DGEA-YOLOv8, to address these issues, using “playing with mobile phones” as a case study. The model integrates the DCNv3 module in the backbone to enhance behavior deformation adaptability and the GELAN module to improve lightweight performance and global perception in resource-limited environments. An ECA attention mechanism is added to enhance small target detection, while the ASPP module boosts multi-scale perception. ByteTrack is incorporated for continuous tracking of prolonged violation behavior in construction scenarios. Experimental results show that DGEA-YOLOv8 achieves 94.5% mAP50, a 2.95% improvement over the YOLOv8s baseline, with better data capture rates and lower ID change rates compared to algorithms like Deepsort and Strongsort. A construction-specific dataset of over 3000 images verifies the model’s effectiveness. From the perspective of data symmetry, the proposed model demonstrates strong capability in addressing asymmetric feature distributions and behavioral imbalance inherent in prolonged violations, restoring spatiotemporal consistency in detection. In conclusion, DGEA-YOLOv8 provides a precise, efficient, and adaptive solution for recognizing prolonged violation behaviors in construction sites. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

15 pages, 4315 KB  
Article
Deep Learning for Real-Time Detection of Brassicogethes aeneus in Oilseed Rape Using the YOLOv4 Architecture
by Ziemowit Malecha, Kajetan Ożarowski, Rafał Siemasz, Maciej Chorowski, Krzysztof Tomczuk, Bernadeta Strochalska and Anna Wondołowska-Grabowska
Appl. Sci. 2026, 16(2), 1075; https://doi.org/10.3390/app16021075 - 21 Jan 2026
Viewed by 71
Abstract
The growing global population and increasing food demand highlight the need for sustainable agricultural practices that balance productivity with environmental protection. Traditional blanket pesticide spraying leads to overuse of chemicals, environmental pollution, and biodiversity loss. This study aims to develop an innovative approach [...] Read more.
The growing global population and increasing food demand highlight the need for sustainable agricultural practices that balance productivity with environmental protection. Traditional blanket pesticide spraying leads to overuse of chemicals, environmental pollution, and biodiversity loss. This study aims to develop an innovative approach to precision pest management using mobile computing, computer vision, and deep learning techniques. A mobile measurement platform equipped with cameras and an onboard computer was designed to collect real-time field data and detect pest infestations. The system uses an advanced object detection algorithm based on the YOLOv4 architecture, trained on a custom dataset of rapeseed pest images. Modifications were made to enhance detection accuracy, especially for small objects. Field tests demonstrated the system’s ability to identify and count pests, such as the pollen beetle (Brassicogethes aeneus), in rapeseed crops. The collected data, combined with GPS information, generated pest density maps, which can guide site-specific pesticide applications. The results show that the proposed method achieved a mean average precision (mAP) of 83.7% on the test dataset. Field measurements conducted during the traversal of rapeseed fields enabled the creation of density maps illustrating the distribution of pollen beetles. Based on these maps, the potential for pesticide savings was demonstrated, and the migration dynamics of pollen beetle were discussed. Full article
Show Figures

Figure 1

24 pages, 7972 KB  
Article
YOLO-MCS: A Lightweight Loquat Object Detection Algorithm in Orchard Environments
by Wei Zhou, Leina Gao, Fuchun Sun and Yuechao Bian
Agriculture 2026, 16(2), 262; https://doi.org/10.3390/agriculture16020262 - 21 Jan 2026
Viewed by 52
Abstract
To address the challenges faced by loquat detection algorithms in orchard settings—including complex backgrounds, severe branch and leaf occlusion, and inaccurate identification of densely clustered fruits—which lead to high computational complexity, insufficient real-time performance, and limited recognition accuracy, this study proposed a lightweight [...] Read more.
To address the challenges faced by loquat detection algorithms in orchard settings—including complex backgrounds, severe branch and leaf occlusion, and inaccurate identification of densely clustered fruits—which lead to high computational complexity, insufficient real-time performance, and limited recognition accuracy, this study proposed a lightweight detection model based on the YOLO-MCS architecture. First, to address fruit occlusion by branches and leaves, the backbone network adopts the lightweight EfficientNet-b0 architecture. Leveraging its composite model scaling feature, this significantly reduces computational costs while balancing speed and accuracy. Second, to deal with inaccurate recognition of densely clustered fruits, the C2f module is enhanced. Spatial Channel Reconstruction Convolution (SCConv) optimizes and reconstructs the bottleneck structure of the C2f module, accelerating inference while improving the model’s multi-scale feature extraction capabilities. Finally, to overcome interference from complex natural backgrounds in loquat fruit detection, this study introduces the SimAm module during the initial detection phase. Its feature recalibration strategy enhances the model’s ability to focus on target regions. According to the experimental results, the improved YOLO-MCS model outperformed the original YOLOv8 model in terms of Precision (P) and mean Average Precision (mAP) by 1.3% and 2.2%, respectively. Additionally, the model reduced GFLOPs computation by 34.1% and Params by 43.3%. Furthermore, in tests under complex weather conditions and with interference factors such as leaf occlusion, branch occlusion, and fruit mutual occlusion, the YOLO-MCS model demonstrated significant robustness, achieving mAP of 89.9% in the loquat recognition task. The exceptional performance serves as a robust technical base on the development and research of intelligent systems for harvesting loquats. Full article
Show Figures

Figure 1

17 pages, 5027 KB  
Article
Symmetry-Enhanced YOLOv8s Algorithm for Small-Target Detection in UAV Aerial Photography
by Zhiyi Zhou, Chengyun Wei, Lubin Wang and Qiang Yu
Symmetry 2026, 18(1), 197; https://doi.org/10.3390/sym18010197 - 20 Jan 2026
Viewed by 150
Abstract
In order to solve the problems of small-target detection in UAV aerial photography, such as small scale, blurred features and complex background interference, this article proposes the ACS-YOLOv8s method to optimize the YOLOv8s network: notably, most small man-made targets in UAV aerial scenes [...] Read more.
In order to solve the problems of small-target detection in UAV aerial photography, such as small scale, blurred features and complex background interference, this article proposes the ACS-YOLOv8s method to optimize the YOLOv8s network: notably, most small man-made targets in UAV aerial scenes (e.g., small vehicles, micro-drones) inherently possess symmetry, a key geometric attribute that can significantly enhance the discriminability of blurred or incomplete target features, and thus symmetry-aware mechanisms are integrated into the aforementioned improved modules to further boost detection performance. The backbone network introduces an adaptive feature enhancement module, the edge and detail representation of small targets is enhanced by dynamically modulating the receptive field with deformable attention while also capturing symmetric contour features to strengthen the perception of target geometric structures; a cascaded multi-receptive field module is embedded at the end of the trunk to integrate multi-scale features in a hierarchical manner to take into account both expressive ability and computational efficiency with a focus on fusing symmetric multi-scale features to optimize feature representation; the neck is integrated with a spatially adaptive feature modulation network to achieve dynamic weighting of cross-layer features and detail fidelity and, meanwhile, models symmetric feature dependencies across channels to reduce the loss of discriminative information. Experimental results based on the VisDrone2019 data set show that ACS-YOLOv8s is superior to the baseline model in precision, recall, and mAP indicators, with mAP50 increased by 2.8% to 41.6% and mAP50:90 increased by 1.9% to 25.0%, verifying its effectiveness and robustness in small-target detection in complex drone aerial-photography scenarios. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

22 pages, 5115 KB  
Article
Intelligent Detection Method of Defects in High-Rise Building Facades Using Infrared Thermography
by Daiming Liu, Yongqiang Jin, Yuan Yang, Zhenyang Xiao, Zeming Zhao, Changling Gao and Dingcheng Zhang
Sensors 2026, 26(2), 694; https://doi.org/10.3390/s26020694 - 20 Jan 2026
Viewed by 214
Abstract
High-rise building facades are prone to defects due to prolonged exposure to complex environments. Infrared detection, as a commonly employed method for facade defect inspection, often results in low accuracy owing to abundant interferences and blurred defect boundaries. In this work, an intelligent [...] Read more.
High-rise building facades are prone to defects due to prolonged exposure to complex environments. Infrared detection, as a commonly employed method for facade defect inspection, often results in low accuracy owing to abundant interferences and blurred defect boundaries. In this work, an intelligent defect detection method for high-rise building facades is proposed. In the first stage of the proposed method, a segmentation model based on DeepLabV3+ is proposed to remove interferences in infrared images using masks. The model incorporates a Post-Decoder Dual-Branch Boundary Refinement Module, which is subdivided into a boundary feature optimization branch and a boundary-guided attention branch. Sub-pixel-level contour refinement and boundary-adaptive weighting are hence achieved to mitigate edge blurring induced by thermal diffusion and to enhance the perception of slender cracks and cavity edges. A triple constraint mechanism is also introduced, combining cross-entropy, multi-scale Dice, and boundary-aware losses to address class imbalance and enhance segmentation performance for small targets. Furthermore, superpixel linear iterative clustering (SLIC) is utilized to enforce regional consistency, hence improving the smoothness and robustness of predictions. In the second stage of the proposed method, a defect detection model based on YOLOV11 is proposed to process masked infrared images for detecting hollow, seepage, cracks and detachment. This work validates the proposed method using 180 infrared images collected via unmanned aerial vehicles. The experimental results demonstrate that the proposed method achieves a detection precision of 89.7%, an mAP@0.5 of 87.9%, and a 57.8 mAP@50-95. surpassing other algorithms and confirming its effectiveness and superiority. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

24 pages, 69667 KB  
Article
YOLO-ELS: A Lightweight Cherry Tomato Maturity Detection Algorithm
by Zhimin Tong, Yu Zhou, Changhao Li, Changqing Cai and Lihong Rong
Appl. Sci. 2026, 16(2), 1043; https://doi.org/10.3390/app16021043 - 20 Jan 2026
Viewed by 63
Abstract
Within the domain of intelligent picking robotics, fruit recognition and positioning are essential. Challenging conditions such as varying light, occlusion, and limited edge-computing power compromise fruit maturity detection. To tackle these issues, this paper proposes a lightweight algorithm YOLO-ELS based on YOLOv8n. Specifically, [...] Read more.
Within the domain of intelligent picking robotics, fruit recognition and positioning are essential. Challenging conditions such as varying light, occlusion, and limited edge-computing power compromise fruit maturity detection. To tackle these issues, this paper proposes a lightweight algorithm YOLO-ELS based on YOLOv8n. Specifically, we reconstruct the backbone by replacing the bottlenecks in the C2f structure with Edge-Information-Enhanced Modules (EIEM) to prioritize morphological cues and filter background redundancy. Furthermore, a Large Separable Kernel Attention (LSKA) mechanism is integrated into the SPPF layer to expand the effective receptive field for multi-scale targets. To mitigate occlusion-induced errors, a Spatially Enhanced Attention Module (SEAM) is incorporated into the decoupled detection head to enhance feature responses in obscured regions. Finally, the Inner-GIoU loss is adopted to refine bounding box regression and accelerate convergence. Experimental results demonstrate that compared to the YOLOv8n baseline, the proposed YOLO-ELS achieves a 14.8% reduction in GFLOPs and a 2.3% decrease in parameters, while attaining a precision, recall, and mAP@50% of 92.7%, 83.9%, and 92.0%, respectively. When compared with mainstream models such as DETR, Faster-RCNN, SSD, TOOD, YOLOv5s, and YOLO11n, the mAP@50% is improved by 7.0%, 4.7%, 11.4%, 8.6%, 3.1%, and 3.2%. Deployment tests on the NVIDIA Jetson Orin Nano Super edge platform yield an inference latency of 25.2 ms and a detection speed of 28.2 FPS, successfully meeting the real-time operational requirements of automated harvesting systems. These findings confirm that YOLO-ELS effectively balances high detection accuracy with lightweight architecture, providing a robust technical foundation for intelligent fruit picking in resource-constrained greenhouse environments. Full article
(This article belongs to the Section Agricultural Science and Technology)
Show Figures

Figure 1

18 pages, 4205 KB  
Article
Research on Field Weed Target Detection Algorithm Based on Deep Learning
by Ziyang Chen, Le Wu, Zhenhong Jia, Jiajia Wang, Gang Zhou and Zhensen Zhang
Sensors 2026, 26(2), 677; https://doi.org/10.3390/s26020677 - 20 Jan 2026
Viewed by 108
Abstract
Weed detection algorithms based on deep learning are considered crucial for smart agriculture, with the YOLO series algorithms being widely adopted due to their efficiency. However, existing YOLO algorithms struggle to maintain high accuracy, while low parameter requirements and computational efficiency are achieved [...] Read more.
Weed detection algorithms based on deep learning are considered crucial for smart agriculture, with the YOLO series algorithms being widely adopted due to their efficiency. However, existing YOLO algorithms struggle to maintain high accuracy, while low parameter requirements and computational efficiency are achieved when weeds with occlusion or overlap are detected. To address this challenge, a target detection algorithm called SSS-YOLO based on YOLOv9t is proposed in this paper. First, the SCB (Spatial Channel Conv Block) module is introduced, in which large kernel convolution is employed to capture long-range dependencies, occluded weed regions are bypassed by being associated with unobstructed areas, and features of unobstructed regions are enhanced through inter-channel relationships. Second, the SPPF EGAS (Spatial Pyramid Pooling Fast Edge Gaussian Aggregation Super) module is proposed, where multi-scale max pooling is utilized to extract hierarchical contextual features, large receptive fields are leveraged to acquire background information around occluded objects, and features of weed regions obscured by crops are inferred. Finally, the EMSN (Efficient Multi-Scale Spatial-Feedforward Network) module is developed, through which semantic information of occluded regions is reconstructed by contextual reasoning and background vegetation interference is effectively suppressed while visible regional details are preserved. To validate the performance of this method, experiments are conducted on both our self-built dataset and the publicly available Cotton WeedDet12 dataset. The results demonstrate that compared to existing algorithms, significant performance improvements are achieved by the proposed method. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

18 pages, 14158 KB  
Article
Vision-Based Perception and Execution Decision-Making for Fruit Picking Robots Using Generative AI Models
by Yunhe Zhou, Chunjiang Yu, Jiaming Zhang, Yuanhang Liu, Jiangming Kan, Xiangjun Zou, Kang Zhang, Hanyan Liang, Sheng Zhang and Fengyun Wu
Machines 2026, 14(1), 117; https://doi.org/10.3390/machines14010117 - 19 Jan 2026
Viewed by 116
Abstract
At present, fruit picking mainly relies on manual operation. Taking the litchi (litchi chinensis Sonn.)-picking robot as an example, visual perception is often affected by illumination variations, low recognition accuracy, complex maturity judgment, and occlusion, which lead to inaccurate fruit localization. This study [...] Read more.
At present, fruit picking mainly relies on manual operation. Taking the litchi (litchi chinensis Sonn.)-picking robot as an example, visual perception is often affected by illumination variations, low recognition accuracy, complex maturity judgment, and occlusion, which lead to inaccurate fruit localization. This study aims to establish an embodied perception mechanism based on “perception-reasoning-execution” to enhance the visual perception and decision-making capability of the robot in complex orchard environments. First, a Y-LitchiC instance segmentation method is proposed to achieve high-precision segmentation of litchi clusters. Second, a generative artificial intelligence model is introduced to intelligently assess fruit maturity and occlusion, providing auxiliary support for automatic picking. Based on the auxiliary judgments provided by the generative AI model, two types of dynamic harvesting decisions are formulated for subsequent operations. For unoccluded main fruit-bearing branches, a skeleton thinning algorithm is applied within the segmented region to extract the skeleton line, and the midpoint of the skeleton is used to perform the first type of localization and harvesting decision. In contrast, for main fruit-bearing branches occluded by leaves, threshold-based segmentation combined with maximum connected component extraction is employed to obtain the target region, followed by skeleton thinning, thereby completing the second type of dynamic picking decision. Experimental results show that the Y-LitchiC model improves the mean average precision (mAP) by 1.6% compared with the YOLOv11s-seg model, achieving higher accuracy in litchi cluster segmentation and recognition. The generative artificial intelligence model provides higher-level reasoning and decision-making capabilities for automatic picking. Overall, the proposed embodied perception mechanism and dynamic picking strategies effectively enhance the autonomous perception and decision-making of the picking robot in complex orchard environments, providing a reliable theoretical basis and technical support for accurate fruit localization and precision picking. Full article
(This article belongs to the Special Issue Control Engineering and Artificial Intelligence)
Show Figures

Figure 1

25 pages, 12600 KB  
Article
Underwater Object Recovery Using a Hybrid-Controlled ROV with Deep Learning-Based Perception
by Inés Pérez-Edo, Salvador López-Barajas, Raúl Marín-Prades and Pedro J. Sanz
J. Mar. Sci. Eng. 2026, 14(2), 198; https://doi.org/10.3390/jmse14020198 - 18 Jan 2026
Viewed by 295
Abstract
The deployment of large remotely operated vehicles (ROVs) or autonomous underwater vehicles (AUVs) typically requires support vessels, crane systems, and specialized personnel, resulting in increased logistical complexity and operational costs. In this context, lightweight and modular underwater robots have emerged as a cost-effective [...] Read more.
The deployment of large remotely operated vehicles (ROVs) or autonomous underwater vehicles (AUVs) typically requires support vessels, crane systems, and specialized personnel, resulting in increased logistical complexity and operational costs. In this context, lightweight and modular underwater robots have emerged as a cost-effective alternative, capable of reaching significant depths and performing tasks traditionally associated with larger platforms. This article presents a system architecture for recovering a known object using a hybrid-controlled ROV, integrating autonomous perception, high-level interaction, and low-level control. The proposed architecture includes a perception module that estimates the object pose using a Perspective-n-Point (PnP) algorithm, combining object segmentation from a YOLOv11-seg network with 2D keypoints obtained from a YOLOv11-pose model. In addition, a Natural Language ROS Agent is incorporated to enable high-level command interaction between the operator and the robot. These modules interact with low-level controllers that regulate the vehicle degrees of freedom and with autonomous behaviors such as target approach and grasping. The proposed system is evaluated through simulation and experimental tank trials, including object recovery experiments conducted in a 12 × 8 × 5 m test tank at CIRTESU, as well as perception validation in simulated, tank, and harbor scenarios. The results demonstrate successful recovery of a black box using a BlueROV2 platform, showing that architectures of this type can effectively support operators in underwater intervention tasks, reducing operational risk, deployment complexity, and mission costs. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

16 pages, 4801 KB  
Article
Welding Seam Recognition and Trajectory Planning Based on Deep Learning in Electron Beam Welding
by Hao Yang, Congjin Zuo, Haiying Xu and Xiaofei Xu
Sensors 2026, 26(2), 641; https://doi.org/10.3390/s26020641 - 18 Jan 2026
Viewed by 202
Abstract
To address challenges in weld recognition during vacuum electron beam welding caused by dark environments and metal reflections, this study proposes an improved hybrid algorithm combining YOLOv11-seg with adaptive Canny edge detection. By incorporating the UFO-ViT attention mechanism and optimizing the network architecture [...] Read more.
To address challenges in weld recognition during vacuum electron beam welding caused by dark environments and metal reflections, this study proposes an improved hybrid algorithm combining YOLOv11-seg with adaptive Canny edge detection. By incorporating the UFO-ViT attention mechanism and optimizing the network architecture with the EIoU loss function, along with adaptive threshold setting for the Canny operator using the Otsu method, the recognition performance under complex conditions is significantly enhanced. Experimental results demonstrate that the optimized model achieves an average precision (mAP) of 77.4%, representing a 9-percentage-point improvement over the baseline YOLOv11-seg. The system operates at 20 frames per second (FPS), meeting real-time requirements, with the generated welding trajectories showing an average length deviation of less than 3 mm from actual welds. This approach provides an effective pre-weld visual guidance solution, which is a critical step towards the automation of electron beam welding. Full article
Show Figures

Figure 1

16 pages, 998 KB  
Article
Architecture Design of a Convolutional Neural Network Accelerator for Heterogeneous Computing Based on a Fused Systolic Array
by Yang Zong, Zhenhao Ma, Jian Ren, Yu Cao, Meng Li and Bin Liu
Sensors 2026, 26(2), 628; https://doi.org/10.3390/s26020628 - 16 Jan 2026
Viewed by 205
Abstract
Convolutional Neural Networks (CNNs) generally suffer from excessive computational overhead, high resource consumption, and complex network structures, which severely restrict the deployment on microprocessor chips. Existing related accelerators only have an energy efficiency ratio of 2.32–6.5925 GOPs/W, making it difficult to meet the [...] Read more.
Convolutional Neural Networks (CNNs) generally suffer from excessive computational overhead, high resource consumption, and complex network structures, which severely restrict the deployment on microprocessor chips. Existing related accelerators only have an energy efficiency ratio of 2.32–6.5925 GOPs/W, making it difficult to meet the low-power requirements of embedded application scenarios. To address these issues, this paper proposes a low-power and high-energy-efficiency CNN accelerator architecture based on a central processing unit (CPU) and an Application-Specific Integrated Circuit (ASIC) heterogeneous computing architecture, adopting an operator-fused systolic array algorithm with the YOLOv5n target detection network as the application benchmark. It integrates a 2D systolic array with Conv-BN fusion technology to achieve deep operator fusion of convolution, batch normalization and activation functions; optimizes the RISC-V core to reduce resource usage; and adopts a locking mechanism and a prefetching strategy for the asynchronous platform to ensure operational stability. Experiments on the Nexys Video development board show that the architecture achieves 20.6 GFLOPs of computational performance, 1.96 W of power consumption, and 10.46 GOPs/W of energy efficiency ratio, which is 58–350% higher than existing mainstream accelerators, thus demonstrating excellent potential for embedded deployment. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

28 pages, 3390 KB  
Article
SDC-YOLOv8: An Improved Algorithm for Road Defect Detection Through Attention-Enhanced Feature Learning and Adaptive Feature Reconstruction
by Hao Yang, Yulong Song, Yue Liang, Enhao Tang and Danyang Cao
Sensors 2026, 26(2), 609; https://doi.org/10.3390/s26020609 - 16 Jan 2026
Viewed by 251
Abstract
Road defect detection is essential for timely road damage repair and traffic safety assurance. However, existing object detection algorithms suffer from insufficient accuracy in detecting small road surface defects and are prone to missed detections and false alarms under complex lighting and background [...] Read more.
Road defect detection is essential for timely road damage repair and traffic safety assurance. However, existing object detection algorithms suffer from insufficient accuracy in detecting small road surface defects and are prone to missed detections and false alarms under complex lighting and background conditions. To address these challenges, this study proposes SDC-YOLOv8, an improved YOLOv8-based algorithm for road defect detection that employs attention-enhanced feature learning and adaptive feature reconstruction. The model incorporates three key innovations: (1) an SPPF-LSKA module that integrates Fast Spatial Pyramid Pooling with Large Separable Kernel Attention to enhance multi-scale feature representation and irregular defect modeling capabilities; (2) DySample dynamic upsampling that replaces conventional interpolation methods for adaptive feature reconstruction with reduced computational cost; and (3) a Coordinate Attention module strategically inserted to improve spatial localization accuracy under complex conditions. Comprehensive experiments on a public pothole dataset demonstrate that SDC-YOLOv8 achieves 78.0% mAP@0.5, 81.0% Precision, and 70.7% Recall while maintaining real-time performance at 85 FPS. Compared to the baseline YOLOv8n model, the proposed method improves mAP@0.5 by 2.0 percentage points, Precision by 3.3 percentage points, and Recall by 1.8 percentage points, yielding an F1 score of 75.5%. These results demonstrate that SDC-YOLOv8 effectively enhances small-target detection accuracy while preserving real-time processing capability, offering a practical and efficient solution for intelligent road defect detection applications. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

Back to TopTop