Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (342)

Search Parameters:
Keywords = industrial visual inspection

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 3064 KB  
Article
Non-Destructive Detection of Elasmopalpus lignosellus Infestation in Fresh Asparagus Using VIS–NIR Hyperspectral Imaging and Machine Learning
by André Rodríguez-León, Jimy Oblitas, Jhonsson Luis Quevedo-Olaya, William Vera, Grimaldo Wilfredo Quispe-Santivañez and Rebeca Salvador-Reyes
Foods 2026, 15(2), 355; https://doi.org/10.3390/foods15020355 - 19 Jan 2026
Viewed by 101
Abstract
The early detection of internal damage caused by Elasmopalpus lignosellus in fresh asparagus constitutes a challenge for the agro-export industry due to the limited sensitivity of traditional visual inspection. This study evaluated the potential of VIS–NIR hyperspectral imaging (390–1036 nm) combined with machine-learning [...] Read more.
The early detection of internal damage caused by Elasmopalpus lignosellus in fresh asparagus constitutes a challenge for the agro-export industry due to the limited sensitivity of traditional visual inspection. This study evaluated the potential of VIS–NIR hyperspectral imaging (390–1036 nm) combined with machine-learning models to discriminate between infested (PB) and sound (SB) asparagus spears. A balanced dataset of 900 samples was acquired, and preprocessing was performed using Savitzky–Golay and SNV. Four classifiers (SVM, MLP, Elastic Net, and XGBoost) were compared. The optimized SVM model achieved the best results (CV Accuracy = 0.9889; AUC = 0.9997). The spectrum was reduced to 60 bands while LOBO and RFE were used to maintain high performance. In external validation (n = 3000), the model achieved an accuracy of 97.9% and an AUC of 0.9976. The results demonstrate the viability of implementing non-destructive systems based on VIS–NIR to improve the quality control of asparagus destined for export. Full article
(This article belongs to the Section Food Analytical Methods)
Show Figures

Figure 1

20 pages, 8043 KB  
Article
Development of a Cost-Effective UUV Localisation System Integrable with Aquaculture Infrastructure
by Thein Than Tun, Loulin Huang and Mark Anthony Preece
J. Mar. Sci. Eng. 2026, 14(2), 115; https://doi.org/10.3390/jmse14020115 - 7 Jan 2026
Viewed by 217
Abstract
In many aquaculture farms, Unmanned Underwater Vehicles (UUVs) are being deployed to perform dangerous and time-consuming repetitive tasks (e.g., fish net-pen visual inspection) on behalf of or in collaboration with farm operators. Mostly, they are remotely operated, and one of the main barriers [...] Read more.
In many aquaculture farms, Unmanned Underwater Vehicles (UUVs) are being deployed to perform dangerous and time-consuming repetitive tasks (e.g., fish net-pen visual inspection) on behalf of or in collaboration with farm operators. Mostly, they are remotely operated, and one of the main barriers to deploying them autonomously is the UUV localisation. Specifically, the cost of the localisation sensor suite, sensor reliability in constrained operational workspace and return on investment (ROI) for the huge initial investment on the UUV and its localisation hinder the R&D work and adoption of the autonomous UUV deployment on an industrial scale. The proposed system, which leverages the AprilTag (a fiducial marker used as a frame of reference) detection, provides cost-effective UUV localisation for the initial trials of autonomous UUV deployment, requiring only minor modifications to the aquaculture infrastructure. With such a cost-effective approach, UUV R&D engineers can demonstrate and validate the advantages and challenges of autonomous UUV deployment to farm operators, policymakers, and governing authorities to make informed decision-making for the future large-scale adoption of autonomous UUVs in aquaculture. Initial validation of the proposed cost-effective localisation system indicates that centimetre-level accuracy can be achieved with a single monocular camera and only 10 AprilTags, without requiring physical measurements, in a 115.46 m3 laboratory workspace under various lighting conditions. Full article
(This article belongs to the Special Issue Infrastructure for Offshore Aquaculture Farms)
Show Figures

Figure 1

30 pages, 5831 KB  
Systematic Review
A Systematic Literature Review of Augmented Reality’s Development in Construction
by José Marinho, Filipe Sá, João Durães, Inácio Fonseca and Nuno Cid Martins
Electronics 2026, 15(1), 225; https://doi.org/10.3390/electronics15010225 - 3 Jan 2026
Viewed by 307
Abstract
Augmented reality (AR) has emerged as a transformative technology, allowing users to engage with digital content overlaid on the physical world. In the construction industry, AR shows significant potential to enhance visualization, collaboration, training, and safety across the project lifecycle. This paper presents [...] Read more.
Augmented reality (AR) has emerged as a transformative technology, allowing users to engage with digital content overlaid on the physical world. In the construction industry, AR shows significant potential to enhance visualization, collaboration, training, and safety across the project lifecycle. This paper presents a systematic literature review (SLR) of 136 publications on the use of AR in construction published between 2019 and 2025, focusing on architectures, technologies, trends, and challenges. The review identifies the main architectures (cloud, hybrid, and local) and examines how AR is combined with Building Information Modeling (BIM) systems, digital twins, the Internet of Things (IoT), and Unmanned Aerial Vehicles (UAVs). Key application trends are identified and discussed, including on-site visualization, inspection and monitoring, immersive training, hazard detection, and remote collaboration. Challenges and constraints to the adoption of AR in construction are highlighted and examined such as hardware limitations, usability and ergonomics issues, interoperability with existing systems, high costs, and resistance to organizational change. By systematizing existing approaches and mapping both opportunities and barriers, this review provides a comprehensive reference for researchers, practitioners, and policy makers aiming to accelerate AR adoption in the construction sector. Full article
Show Figures

Figure 1

14 pages, 1512 KB  
Article
YOLO-LA: Prototype-Based Vision–Language Alignment for Silicon Wafer Defect Pattern Detection
by Ziyue Wang, Yichen Yang, Jianning Chu, Yikai Zang, Zhongdi She, Weikang Fang and Ruoxin Wang
Micromachines 2026, 17(1), 67; https://doi.org/10.3390/mi17010067 - 31 Dec 2025
Viewed by 547
Abstract
With the rapid development of semiconductor manufacturing technology, methods to effectively control the production process, reduce variation in the manufacturing process, and improve the yield rate represent important competitive factors for wafer factories. Wafer bin maps, a method for characterizing wafer defect patterns, [...] Read more.
With the rapid development of semiconductor manufacturing technology, methods to effectively control the production process, reduce variation in the manufacturing process, and improve the yield rate represent important competitive factors for wafer factories. Wafer bin maps, a method for characterizing wafer defect patterns, provide valuable information for engineers to quickly identify potential root causes through accurate pattern recognition. Vision-based deep learning approaches rely on visual patterns to achieve robust performance. However, they rarely exploit the rich semantic information embedded in defect descriptions, limiting interpretability and generalization. To address this gap, we propose YOLO-LA, a lightweight prototype-based vision–language alignment framework that integrates a pretrained frozen YOLO backbone with a frozen text encoder to enhance wafer defect recognition. A learnable projection head is introduced to map visual features into a shared embedding space, enabling classification through cosine similarity Experimental results on the WM-811K dataset demonstrate that YOLO-LA consistently improves classification accuracy across different backbones while introducing minimal additional parameters. In particular, YOLOv12 achieves the fastest speed while maintaining competitive accuracy, whereas YOLOv10 benefits most from semantic prototype alignment. The proposed framework is lightweight and suitable for real-time industrial wafer inspection systems. Full article
(This article belongs to the Special Issue Future Trends in Ultra-Precision Machining)
Show Figures

Figure 1

15 pages, 3365 KB  
Article
Lightweight YOLO-Based Online Inspection Architecture for Cup Rupture Detection in the Strip Steel Welding Process
by Yong Qin and Shuai Zhao
Machines 2026, 14(1), 40; https://doi.org/10.3390/machines14010040 - 29 Dec 2025
Viewed by 238
Abstract
Cup rupture failures in strip steel welds can lead to strip breakage, resulting in unplanned downtime of high-speed continuous rolling mills and scrap steel losses. Manual visual inspection suffers from a high false positive rate and cannot meet the production cycle time requirements. [...] Read more.
Cup rupture failures in strip steel welds can lead to strip breakage, resulting in unplanned downtime of high-speed continuous rolling mills and scrap steel losses. Manual visual inspection suffers from a high false positive rate and cannot meet the production cycle time requirements. This paper proposes a lightweight online cup rupture visual inspection method based on an improved YOLOv10 algorithm. The backbone feature extraction network is replaced with ShuffleNetV2 to reduce the model’s parameter count and computational complexity. An ECA attention mechanism is incorporated into the backbone network to enhance the model’s focus on cup rupture micro-cracks. A Slim-Neck design is adopted, utilizing a dual optimization with GSConv and VoV-GSCSP, significantly improving the balance between real-time performance and accuracy. Based on the results, the optimized model achieves a precision of 98.8% and a recall of 99.2%, with a mean average precision (mAP) of 99.5%—an improvement of 0.2 percentage points over the baseline. The model has a computational load of 4.4 GFLOPs and a compact size of only 3.24 MB, approximately half that of the original model. On embedded devices, it achieves a real-time inference speed of 122 FPS, which is about 2.5, 11, and 1.8 times faster than SSD, Faster R-CNN, and YOLOv10n, respectively. Therefore, the lightweight model based on the improved YOLOv10 not only enhances detection accuracy but also significantly reduces computational cost and model size, enabling efficient real-time cup rupture detection in industrial production environments on embedded platforms. Full article
(This article belongs to the Section Advanced Manufacturing)
Show Figures

Figure 1

31 pages, 10819 KB  
Article
Research on High-Precision Localization Method of Curved Surface Feature Points Based on RGB-D Data Fusion
by Enguo Wang, Rui Zou and Chengzhi Su
Sensors 2026, 26(1), 137; https://doi.org/10.3390/s26010137 - 25 Dec 2025
Viewed by 292
Abstract
Although RGB images contain rich details, they lack 3D depth information. Depth data, while providing spatial positioning, is often affected by noise and suffers from sparsity or missing data at key feature points, leading to low accuracy and high computational complexity in traditional [...] Read more.
Although RGB images contain rich details, they lack 3D depth information. Depth data, while providing spatial positioning, is often affected by noise and suffers from sparsity or missing data at key feature points, leading to low accuracy and high computational complexity in traditional visual localization. To address this, this paper proposes a high-precision, sub-pixel-level localization method for workpiece feature points based on RGB-D data fusion. The method specifically targets two types of localization objects: planar corner keypoints and sharp-corner keypoints. It employs the YOLOv10 model combined with a Background Misdetection Filtering Module (BMFM) to classify and identify feature points in RGB images. An improved Prewitt operator (using 5 × 5 convolution kernels in 8 directions) and sub-pixel refinement techniques are utilized to enhance 2D localization accuracy. The 2D feature boundaries are then mapped into 3D point cloud space based on camera extrinsic parameters. After coarse error detection in the point cloud and local quadric surface fitting, 3D localization is achieved by intersecting spatial rays with the fitted surfaces. Experimental results demonstrate that the proposed method achieves a mean absolute error (MAE) of 0.17 mm for localizing flat, free-form, and grooved components, with a maximum error of less than 0.22 mm, meeting the requirements of high-precision industrial applications such as precision manufacturing and quality inspection. Full article
(This article belongs to the Section Navigation and Positioning)
Show Figures

Figure 1

14 pages, 17578 KB  
Article
A Two-Stage High-Precision Recognition and Localization Framework for Key Components on Industrial PCBs
by Li Wang, Liu Ouyang, Huiying Weng, Xiang Chen, Anna Wang and Kexin Zhang
Mathematics 2026, 14(1), 4; https://doi.org/10.3390/math14010004 - 19 Dec 2025
Viewed by 240
Abstract
Precise recognition and localization of electronic components on printed circuit boards (PCBs) are crucial for industrial automation tasks, including robotic disassembly, high-precision assembly, and quality inspection. However, strong visual interference from silkscreen characters, copper traces, solder pads, and densely packed small components often [...] Read more.
Precise recognition and localization of electronic components on printed circuit boards (PCBs) are crucial for industrial automation tasks, including robotic disassembly, high-precision assembly, and quality inspection. However, strong visual interference from silkscreen characters, copper traces, solder pads, and densely packed small components often degrades the accuracy of deep learning-based detectors, particularly under complex industrial imaging conditions. This paper presents a two-stage, coarse-to-fine PCB component localization framework based on an optimized YOLOv11 architecture and a sub-pixel geometric refinement module. The proposed method enhances the backbone with a Convolutional Block Attention Module (CBAM) to suppress background noise and strengthen discriminative features. It also integrates a tiny-object detection branch and a weighted Bi-directional Feature Pyramid Network (BiFPN) for more effective multi-scale feature fusion, and it employs a customized hybrid loss with vertex-offset supervision to enable pose-aware bounding box regression. In the second stage, the coarse predictions guide contour-based sub-pixel fitting using template geometry to achieve industrial-grade precision. Experiments show significant improvements over baseline YOLOv11, particularly for small and densely arranged components, indicating that the proposed approach meets the stringent requirements of industrial robotic disassembly. Full article
(This article belongs to the Special Issue Complex Process Modeling and Control Based on AI Technology)
Show Figures

Figure 1

23 pages, 7391 KB  
Article
TSE-YOLO: A Model for Tomato Ripeness Segmentation
by Liangquan Jia, Xinhui Yuan, Ze Chen, Tao Wang, Lu Gao, Guosong Gu, Xuechun Wang and Yang Wang
Agriculture 2026, 16(1), 8; https://doi.org/10.3390/agriculture16010008 - 19 Dec 2025
Viewed by 458
Abstract
Accurate and efficient tomato ripeness estimation is crucial for robotic harvesting and supply chain grading in smart agriculture. However, manual visual inspection is subjective, slow and difficult to scale, while existing vision models often struggle with cluttered field backgrounds, small targets and limited [...] Read more.
Accurate and efficient tomato ripeness estimation is crucial for robotic harvesting and supply chain grading in smart agriculture. However, manual visual inspection is subjective, slow and difficult to scale, while existing vision models often struggle with cluttered field backgrounds, small targets and limited throughput. To overcome these limitations, we introduce TSE-YOLO, an improved real-time detector tailored for tomato ripeness estimation with joint detection and segmentation. In the TSE-YOLO model, three key enhancements are introduced. The C2PSA module is improved with ConvGLU, adapted from TransNeXt, to strengthen feature extraction within tomato regions. A novel segmentation head is designed to accelerate ripeness-aware segmentation and improve recall. Additionally, the C3k2 module is augmented with partial and frequency-dynamic convolutions, enhancing feature representation under complex planting conditions. These components enable precise instance-level localization and pixel-wise segmentation of tomatoes at three ripeness stages: verde, semi-ripe (semi-maduro), and ripe. Experiments on a self-constructed tomato ripeness dataset demonstrate that TSE-YOLO achieves 92.5% mAP@0.5 for detection and 92.2% mAP@0.5 for segmentation with only 9.8 GFLOPs. Deployed on Android via Ncnn Convolutional Neural Network (NCNN), the model runs at 30 fps on Dimensity 9300, offering a practical solution for automated tomato harvesting and grading that accelerates smart agriculture’s industrial adoption. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

28 pages, 20201 KB  
Systematic Review
Extended Realityin Construction 4.0: A Systematic Review of Applications, Implementation Barriers, and Research Trends
by Jose Gornall, Alvaro Peña, Hernan Pinto, Jorge Rojas, Fabiano Correa and Jose García
Appl. Sci. 2026, 16(1), 9; https://doi.org/10.3390/app16010009 - 19 Dec 2025
Viewed by 325
Abstract
Extended reality (XR) is increasingly used to address productivity, communication, and safety challenges in the construction industry, but large-scale adoption within Construction 4.0 remains limited. The existing reviews rarely provide an integrated perspective that jointly examines XR applications, underlying technology stacks, and the [...] Read more.
Extended reality (XR) is increasingly used to address productivity, communication, and safety challenges in the construction industry, but large-scale adoption within Construction 4.0 remains limited. The existing reviews rarely provide an integrated perspective that jointly examines XR applications, underlying technology stacks, and the barriers that constrain implementation. This study fills that gap by combining a PRISMA-compliant systematic review with a bibliometric analysis of 76 journal articles published between 2019 and 2024. The review maps XR usage in construction, which XR modes, devices, and graphics engines are most prevalent, and which barriers hinder deployment in real projects. Design visualization and coordination, immersive training, and remote assistance or inspection emerge as the dominant application areas. Augmented reality (AR) and virtual reality (VR) lead the technology landscape, with Microsoft HoloLens and Meta Quest as the most frequently reported head-mounted displays and Unity as the main graphics engine. Implementation barriers are categorized into five groups—technological, organizational, economic, infrastructural, and methodological—with interoperability issues, hardware performance limitations, and the lack of standardized BIM-to-XR workflows being particularly recurrent. The review contributes to the Construction 4.0 agenda by providing a consolidated map of XR applications, technologies, and barriers, and by highlighting enablers such as open data schemas and competency-based training programs. Future research should validate AI-assisted, bidirectional BIM–XR workflows in real projects, report cost–benefit metrics, and advance interoperability standards that integrate XR into broader Construction 4.0 strategies. Full article
Show Figures

Figure 1

27 pages, 1475 KB  
Article
Operationalizing the R4VR-Framework: Safe Human-in-the-Loop Machine Learning for Image Recognition
by Julius Wiggerthale and Christoph Reich
Processes 2025, 13(12), 4086; https://doi.org/10.3390/pr13124086 - 18 Dec 2025
Viewed by 483
Abstract
Visual inspection is a crucial quality assurance process across many manufacturing industries. While many companies now employ machine learning-based systems, they face a significant challenge, particularly in safety-critical domains. The outcomes of these systems are often complex and difficult to comprehend, making them [...] Read more.
Visual inspection is a crucial quality assurance process across many manufacturing industries. While many companies now employ machine learning-based systems, they face a significant challenge, particularly in safety-critical domains. The outcomes of these systems are often complex and difficult to comprehend, making them less reliable and trustworthy. To address this challenge, we build on our previously proposed R4VR-framework and provide practical, step-by-step guidelines that enable the safe and efficient implementation of machine learning in visual inspection tasks, even when starting from scratch. The framework leverages three complementary safety mechanisms—uncertainty detection, explainability, and model diversity—to enhance both accuracy and system safety while minimizing manual effort. Using the example of steel surface inspection, we demonstrate how a self-accelerating process of data collection where model performance improves while manual effort decreases progressively can arise. Based on that, we create a system with various safety mechanisms where less than 0.1% of images are classified wrongly and remain undetected. We provide concrete recommendations and an open-source code base to facilitate reproducibility and adaptation to diverse industrial contexts. Full article
Show Figures

Figure 1

32 pages, 14384 KB  
Article
CSPC-BRS: An Enhanced Real-Time Multi-Target Detection and Tracking Algorithm for Complex Open Channels
by Wei Li, Xianpeng Zhu, Aghaous Hayat, Hu Yuan and Xiaojiang Yang
Electronics 2025, 14(24), 4942; https://doi.org/10.3390/electronics14244942 - 16 Dec 2025
Viewed by 242
Abstract
Ensuring worker safety compliance and secure cargo transportation in complex port environments is critical for modern logistics hubs. However, conventional supervision methods, including manual inspection and passive video monitoring, suffer from limited coverage, poor real-time responsiveness, and low robustness under frequent occlusion, scale [...] Read more.
Ensuring worker safety compliance and secure cargo transportation in complex port environments is critical for modern logistics hubs. However, conventional supervision methods, including manual inspection and passive video monitoring, suffer from limited coverage, poor real-time responsiveness, and low robustness under frequent occlusion, scale variation, and cross-camera transitions, leading to unstable target association and missed risk events. To address these challenges, this paper proposes CSPC-BRS, a real-time multi-object detection and tracking framework for open-channel port scenarios. CSPC (Coordinated Spatial Perception Cascade) enhances the YOLOv8 backbone by integrating CASAM, SPPELAN-DW, and CACC modules to improve feature representation under cluttered backgrounds and degraded visual conditions. Meanwhile, BRS (Bounding Box Reduction Strategy) mitigates scale distortion during tracking, and a Multi-Dimensional Re-identification Scoring (MDRS) mechanism fuses six perceptual features—color, texture, shape, motion, size, and time—to achieve stable cross-camera identity consistency. Experimental results demonstrate that CSPC-BRS outperforms the YOLOv8-n baseline by improving the mAP@0.5:0.95 by 9.6% while achieving a real-time speed of 132.63 FPS. Furthermore, in practical deployment, it reduces the false capture rate by an average of 59.7% compared to the YOLOv8 + Bot-SORT tracker. These results confirm that CSPC-BRS effectively balances detection accuracy and computational efficiency, providing a practical and deployable solution for intelligent safety monitoring in complex industrial logistics environments. Full article
Show Figures

Figure 1

41 pages, 8287 KB  
Article
Smart Image-Based Deep Learning System for Automated Quality Grading of Phalaenopsis Seedlings in Outsourced Production
by Hong-Dar Lin, Zheng-Yuan Zhang and Chou-Hsien Lin
Sensors 2025, 25(24), 7502; https://doi.org/10.3390/s25247502 - 10 Dec 2025
Viewed by 492
Abstract
Phalaenopsis orchids are one of Taiwan’s key floral export products, and maintaining consistent quality is crucial for international competitiveness. To improve production efficiency, many orchid farms outsource the early flask seedling stage to contract growers, who raise the plants to the 2.5-inch potted [...] Read more.
Phalaenopsis orchids are one of Taiwan’s key floral export products, and maintaining consistent quality is crucial for international competitiveness. To improve production efficiency, many orchid farms outsource the early flask seedling stage to contract growers, who raise the plants to the 2.5-inch potted seedling stage before returning them for further greenhouse cultivation. Traditionally, the quality of these outsourced seedlings is evaluated manually by inspectors who visually detect defects and assign quality grades based on experience, a process that is time-consuming and subjective. This study introduces a smart image-based deep learning system for automatic quality grading of Phalaenopsis potted seedlings, combining computer vision, deep learning, and machine learning techniques to replace manual inspection. The system uses YOLOv8 and YOLOv10 models for defect and root detection, along with SVM and Random Forest classifiers for defect counting and grading. It employs a dual-view imaging approach, utilizing top-view RGB-D images to capture spatial leaf structures and multi-angle side-view RGB images to assess leaf and root conditions. Two grading strategies are developed: a three-stage hierarchical method that offers interpretable diagnostic results and a direct grading method for fast, end-to-end quality prediction. Performance comparisons and ablation studies show that using RGB-D top-view images and optimal viewing-angle combinations significantly improve grading accuracy. The system achieves F1-scores of 84.44% (three-stage) and 90.44% (direct), demonstrating high reliability and strong potential for automated quality assessment and export inspection in the orchid industry. Full article
(This article belongs to the Special Issue Sensing and Imaging for Defect Detection: 2nd Edition)
Show Figures

Figure 1

19 pages, 2362 KB  
Article
TCQI-YOLOv5: A Terminal Crimping Quality Defect Detection Network
by Yingjuan Yu, Dawei Ren and Lingwei Meng
Sensors 2025, 25(24), 7498; https://doi.org/10.3390/s25247498 - 10 Dec 2025
Viewed by 431
Abstract
With the rapid development of the automotive industry, terminals—as critical components of wiring harnesses—play a pivotal role in ensuring the reliability and stability of signal transmission. At present, terminal crimping quality inspection (TCQI) primarily relies on manual visual examination, which suffers from low [...] Read more.
With the rapid development of the automotive industry, terminals—as critical components of wiring harnesses—play a pivotal role in ensuring the reliability and stability of signal transmission. At present, terminal crimping quality inspection (TCQI) primarily relies on manual visual examination, which suffers from low efficiency, high labor intensity, and susceptibility to missed detections. To address these challenges, this study proposes an improved YOLOv5-based model, TCQI-YOLOv5, designed to achieve efficient and accurate automatic detection of terminal crimping quality. In the feature extraction module, the model integrates the C2f structure, FasterNet module, and Efficient Multi-scale Attention (EMA) attention mechanism, enhancing its capability to identify small targets and subtle defects. Moreover, the SIOU loss function is employed to replace the traditional IOU, thereby improving the localization accuracy of predicted bounding boxes. Experimental results demonstrate that TCQI-YOLOv5 significantly improves recognition ccuracy for difficult-to-detect defects such as shallow insulation crimps, achieving a mean average precision (mAP) of 98.3%, outperforming comparative models. Furthermore, the detection speed meets the requirements of real-time industrial applications, indicating strong potential for practical deployment. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

22 pages, 8773 KB  
Article
Reconfigurable Multispectral Imaging System Design and Implementation with FPGA Control
by Shuyang Chen, Min Huang, Wenbin Ge, Guangming Wang, Xiangning Lu, Yixin Zhao, Jinlin Chen, Lulu Qian and Zhanchao Wang
Appl. Sci. 2025, 15(24), 12951; https://doi.org/10.3390/app152412951 - 8 Dec 2025
Viewed by 581
Abstract
Multispectral imaging plays an important role in fields such as environmental monitoring and industrial inspection. To meet the demands for high spatial resolution, portability, and multi-scenario use, this study presents a reconfigurable 2 × 3 multispectral camera-array imaging system. The system features a [...] Read more.
Multispectral imaging plays an important role in fields such as environmental monitoring and industrial inspection. To meet the demands for high spatial resolution, portability, and multi-scenario use, this study presents a reconfigurable 2 × 3 multispectral camera-array imaging system. The system features a modular architecture, allowing for the flexible exchange of lenses and narrowband filters. Each camera node is equipped with an FPGA that performs real-time sensor control and data preprocessing. A companion host program, based on the GigE Vision protocol, was developed for synchronous control, multi-channel real-time visualization, and unified parameter configuration. End-to-end performance verification confirmed stable, lossless, and synchronous acquisition from all six 3072 × 2048-pixel resolution channels. Following field alignment, the 16 mm lens achieves an effective 4.7 MP spatial resolution. Spectral profile measurements further confirm that the system exhibits favorable spectral response characteristics. The proposed framework provides a high-resolution and flexible solution for portable multispectral imaging. Full article
(This article belongs to the Section Optics and Lasers)
Show Figures

Figure 1

28 pages, 5016 KB  
Article
A Lightweight Improved YOLOv8-Based Method for Rebar Intersection Detection
by Rui Wang, Fangjun Shi, Yini She, Li Zhang, Kaifeng Lin, Longshun Fu and Jingkun Shi
Appl. Sci. 2025, 15(24), 12898; https://doi.org/10.3390/app152412898 - 7 Dec 2025
Viewed by 451
Abstract
As industrialized construction and smart building continue to advance, rebar-tying robots place higher demands on the real-time and accurate recognition of rebar intersections and their tying status. Existing deep learning-based detection methods generally rely on heavy backbone networks and complex feature-fusion structures, making [...] Read more.
As industrialized construction and smart building continue to advance, rebar-tying robots place higher demands on the real-time and accurate recognition of rebar intersections and their tying status. Existing deep learning-based detection methods generally rely on heavy backbone networks and complex feature-fusion structures, making it difficult to deploy them efficiently on resource-constrained mobile robots and edge devices, and there is also a lack of dedicated datasets for rebar intersections. In this study, 12,000 rebar mesh images were collected and annotated from two indoor scenes and one outdoor scene to construct a rebar-intersection dataset that supports both object detection and instance segmentation, enabling simultaneous learning of intersection locations and tying status. On this basis, a lightweight improved YOLOv8-based method for rebar intersection detection and segmentation is proposed. The original backbone is replaced with ShuffleNetV2, and a C2f_Dual residual module is introduced in the neck; the same improvements are further transferred to YOLOv8-seg to form a unified lightweight detection–segmentation framework for joint prediction of intersection locations and tying status. Experimental results show that, compared with the original YOLOv8L and several mainstream detectors, the proposed model achieves comparable or superior performance in terms of mAP@50, precision and recall, while reducing model size and computational cost by 51.2% and 58.1%, respectively, and significantly improving inference speed. The improved YOLOv8-seg also achieves satisfactory contour alignment and regional consistency for rebar regions and intersection masks. Owing to its combination of high accuracy and low resource consumption, the proposed method is well suited for deployment on edge-computing devices used in rebar-tying robots and construction quality inspection, providing an effective visual perception solution for intelligent construction. Full article
(This article belongs to the Special Issue Advances in Smart Construction and Intelligent Buildings)
Show Figures

Figure 1

Back to TopTop