Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (219)

Search Parameters:
Keywords = ORB features

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 2873 KB  
Article
An Online Calibration Method for UAV Electro-Optical Pod Zoom Cameras Based on IMU-Vision Fusion
by Weiming Zhu, Zhangsong Shi, Huihui Xu, Qingping Hu, Wenjian Ying and Fan Gui
Drones 2026, 10(3), 224; https://doi.org/10.3390/drones10030224 - 22 Mar 2026
Viewed by 307
Abstract
To address the calibration challenge caused by the nonlinear variation in intrinsic parameters during continuous camera zooming in UAV electro-optical pods, this paper proposes an online calibration method based on IMU-visual fusion. Traditional offline calibration cannot adapt to dynamic scenarios, while existing self-calibration [...] Read more.
To address the calibration challenge caused by the nonlinear variation in intrinsic parameters during continuous camera zooming in UAV electro-optical pods, this paper proposes an online calibration method based on IMU-visual fusion. Traditional offline calibration cannot adapt to dynamic scenarios, while existing self-calibration methods suffer from slow convergence and insufficient robustness. The proposed method aims to achieve real-time and accurate estimation of camera intrinsic parameters during zooming. Specifically, we first construct a unified state estimation framework that encodes the internal and external parameters of the camera and the 3D positions of scene feature points into a high-dimensional state vector, then establish a camera motion model based on IMU data, construct a visual observation model by combining the pinhole camera and second-order radial distortion model to establish a nonlinear mapping from 3D feature points to 2D pixel coordinates, and adopt an improved ORB algorithm for feature extraction and LK optical flow method to achieve high-precision cross-frame feature matching to enhance the stability of visual observation. Most importantly, we design a tight-coupling fusion strategy based on the Extended Kalman Filter (EKF) prediction-update iteration mechanism, which fuses IMU high-frequency motion constraints and visual geometric constraints in real time to suppress parameter drift induced by focal length changes. Finally, we recursively solve the state vector to complete the online dynamic estimation of intrinsic parameters. Monte Carlo simulation experiments and real UAV flight experiments confirm that the method has both high estimation accuracy and strong environmental adaptability, can meet the high-precision calibration needs of UAVs in dynamic scenarios, and provides reliable technical support for accurate target positioning. Full article
Show Figures

Figure 1

18 pages, 2493 KB  
Article
Improved Kernel Correlation Filtering Algorithm Integrating Scale Adaptation and Occlusion Redetection
by Tianbo Liu, Yuya Wang, Hong Sun and Shuai Yuan
Appl. Sci. 2026, 16(6), 2843; https://doi.org/10.3390/app16062843 - 16 Mar 2026
Viewed by 236
Abstract
To address the limitations of the Kernelized Correlation Filter (KCF) in handling scale variation and occlusion during visual tracking, this paper proposes a scale-adaptive and occlusion-robust KCF-based tracking method. The proposed approach integrates the Histogram of Oriented Gradients (HOGs) and Color Name (CN) [...] Read more.
To address the limitations of the Kernelized Correlation Filter (KCF) in handling scale variation and occlusion during visual tracking, this paper proposes a scale-adaptive and occlusion-robust KCF-based tracking method. The proposed approach integrates the Histogram of Oriented Gradients (HOGs) and Color Name (CN) features to fully exploit pixel-level information, thereby improving the accuracy of target localization. On this basis, a sub-region-based scale adaptation mechanism is introduced. Specifically, the target is partitioned into multiple sub-regions, and the KCF classifier is applied to each sub-region to estimate its center position. The relative displacement among these sub-region centers is then utilized to estimate target scale variation, enabling adaptive scale tracking. In addition, an occlusion-aware mechanism is designed to enhance robustness under occlusion. During tracking, occlusion detection is performed, and once occlusion is detected, template updating is suspended. Oriented FAST and Rotated BRIEF (ORB) features extracted from the template are subsequently matched with features from subsequent frames to re-acquire the target. Experimental results on the OTB2013 and OTB2015 benchmarks demonstrate that the proposed method achieves competitive precision and success rates compared with the baseline KCF and other representative trackers, while satisfying real-time tracking requirements using only CPU resources, indicating its practical applicability in resource-constrained environments. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

30 pages, 29830 KB  
Article
From Hematoxylin and Eosin to Masson’s Trichrome: A Comprehensive Framework for Virtual Stain Transformation in Chronic Liver Disease Diagnosis
by Hossam Magdy Balaha, Khadiga M. Ali, Ali Mahmoud, Ahmed Aboudessouki, Mohamed T. Azam, Guruprasad A. Giridharan, Dibson Gondim and Ayman El-Baz
Diagnostics 2026, 16(5), 764; https://doi.org/10.3390/diagnostics16050764 - 4 Mar 2026
Viewed by 544
Abstract
Background/Objectives: Virtual histological staining offers a rapid, cost-effective alternative to physical reprocessing but faces challenges related to spatial misalignment and staining heterogeneity between Hematoxylin and Eosin (H&E) and Masson’s Trichrome (MT) domains. This study develops a robust framework for H&E-to-MT virtual staining [...] Read more.
Background/Objectives: Virtual histological staining offers a rapid, cost-effective alternative to physical reprocessing but faces challenges related to spatial misalignment and staining heterogeneity between Hematoxylin and Eosin (H&E) and Masson’s Trichrome (MT) domains. This study develops a robust framework for H&E-to-MT virtual staining to enable accurate fibrosis assessment without additional tissue consumption. Methods: We propose a transformer-based generative adversarial network (TbGAN) supported by a multi-stage alignment pipeline (SIFT (scale-invariant feature transform) coarse alignment, ORB/homography patch registration, and B-spline free-form deformation) and a weighted fusion mechanism combining four configuration outputs (O/10/3, O/3/10, R/10/3, and R/3/10). The framework was validated on 27 whole-slide images (>100,000 aligned patches) through 24 independent experiments. Results: The fused approach achieved state-of-the-art performance: MI = 0.9815 ± 0.0934, SSIM = 0.7474 ± 0.0597, NCC = 0.9320 ± 0.0220, and CS = 0.9946 ± 0.0014. Statistical analysis confirmed enhanced stability through narrower interquartile ranges, fewer outliers, and tighter 95% confidence intervals compared to individual configurations. Qualitative assessment demonstrated preserved collagen morphology critical for fibrosis staging. Conclusions: Our framework provides a reliable, IRB-compliant solution for virtual MT staining that maintains high structural fidelity suitable for diagnostic support. It enables resource-efficient fibrosis quantification and supports integration into clinical digital pathology workflows without patient-specific recalibration. Full article
Show Figures

Figure 1

20 pages, 4751 KB  
Article
Improving ORB-SLAM3 Accuracy in Dynamic Scenes with YOLO11 Segmentation
by Renata Raffaine Villegas, Anselmo Rafael Cukla, Gabriel Alejandro Tarnowski, Guillermo Mudry, Sergio Omar Lapczuk, Ely Carneiro de Paiva and Daniel Fernando Tello Gamarra
Sensors 2026, 26(5), 1487; https://doi.org/10.3390/s26051487 - 27 Feb 2026
Viewed by 585
Abstract
Traditional Visual SLAM systems, like ORB-SLAM3, often lose accuracy in dynamic environments. This work presents YOLO11-ORB-SLAM3, an enhancement to ORB-SLAM3 for dynamic scenarios, which integrates a YOLO11-based instance segmentation module to detect and exclude dynamic features from the tracking process. The system is [...] Read more.
Traditional Visual SLAM systems, like ORB-SLAM3, often lose accuracy in dynamic environments. This work presents YOLO11-ORB-SLAM3, an enhancement to ORB-SLAM3 for dynamic scenarios, which integrates a YOLO11-based instance segmentation module to detect and exclude dynamic features from the tracking process. The system is designed to work with stereo and RGB-D cameras, and its performance was evaluated on challenging dynamic sequences of the public TUM RGB-D dataset, and also through real-world experiments on a mobile robot using a stereo camera to highlight its robustness and viability for real robotic applications. Experimental results demonstrate that the proposed system outperforms the original ORB-SLAM3, reducing the error by 93% in the public TUM dataset while preserving computational efficiency. Full article
Show Figures

Figure 1

27 pages, 6570 KB  
Article
LiDAR–Inertial–Visual Odometry Based on Elastic Registration and Dynamic Feature Removal
by Qiang Ma, Fuhong Qin, Peng Xiao, Meng Wei, Sihong Chen, Wenbo Xu, Xingrui Yue, Ruicheng Xu and Zheng He
Electronics 2026, 15(4), 741; https://doi.org/10.3390/electronics15040741 - 9 Feb 2026
Viewed by 509
Abstract
Simultaneous Localization and Mapping (SLAM) is a fundamental capability for autonomous robots. However, in highly dynamic scenes, conventional SLAM systems often suffer from degraded accuracy due to LiDAR motion distortion and interference from moving objects. To address these challenges, this paper proposes a [...] Read more.
Simultaneous Localization and Mapping (SLAM) is a fundamental capability for autonomous robots. However, in highly dynamic scenes, conventional SLAM systems often suffer from degraded accuracy due to LiDAR motion distortion and interference from moving objects. To address these challenges, this paper proposes a LiDAR–Inertial–Visual odometry framework based on elastic registration and dynamic feature removal, with the aim of enhancing system robustness through detailed algorithmic supplements. In the LiDAR odometry module, an elastic registration-based de-skewing method is introduced by modeling second-order motion, enabling accurate point cloud correction under non-uniform motion. In the visual odometry module, a multi-strategy dynamic feature suppression mechanism is developed, combining IMU-assisted motion consistency verification with a lightweight YOLOv5-based detection network to effectively filter out dynamic interference with low computational overhead. Furthermore, depth information for visual key points is recovered using LiDAR assistance to enable tightly coupled pose estimation. Extensive experiments on the TUM and M2DGR datasets demonstrate that the proposed method achieves a 96.3% reduction in absolute trajectory error (ATE) compared with ORB-SLAM2 in highly dynamic scenarios. Real-world deployment on an embedded computing device further confirms the framework’s real-time performance and practical applicability in complex environments. Full article
Show Figures

Figure 1

20 pages, 1202 KB  
Article
Adaptive ORB Accelerator on FPGA: High Throughput, Power Consumption, and More Efficient Vision for UAVs
by Hussam Rostum and József Vásárhelyi
Signals 2026, 7(1), 13; https://doi.org/10.3390/signals7010013 - 2 Feb 2026
Viewed by 863
Abstract
Feature extraction and description are fundamental components of visual perception systems used in applications such as visual odometry, Simultaneous Localization and Mapping (SLAM), and autonomous navigation. In resource-constrained platforms, such as Unmanned Aerial Vehicles (UAVs), achieving real-time hardware acceleration on Field-Programmable Gate Arrays [...] Read more.
Feature extraction and description are fundamental components of visual perception systems used in applications such as visual odometry, Simultaneous Localization and Mapping (SLAM), and autonomous navigation. In resource-constrained platforms, such as Unmanned Aerial Vehicles (UAVs), achieving real-time hardware acceleration on Field-Programmable Gate Arrays (FPGAs) is challenging. This work demonstrates an FPGA-based implementation of an adaptive ORB (Oriented FAST and Rotated BRIEF) feature extraction pipeline designed for high-throughput and energy-efficient embedded vision. The proposed architecture is a completely new design for the main algorithmic blocks of ORB, including the FAST (Features from Accelerated Segment Test) feature detector, Gaussian image filtering, moment computation, and descriptor generation. Adaptive mechanisms are introduced to dynamically adjust thresholds and filtering behavior, improving robustness under varying illumination conditions. The design is developed using a High-Level Synthesis (HLS) approach, where all processing modules are implemented as reusable hardware IP cores and integrated at the system level. The architecture is deployed and evaluated on two FPGA platforms, PYNQ-Z2 and KRIA KR260, and its performance is compared against CPU and GPU implementations using a dedicated C++ testbench based on OpenCV. Experimental results demonstrate significant improvements in throughput and energy efficiency while maintaining stable and scalable performance, making the proposed solution suitable for real-time embedded vision applications on UAVs and similar platforms. Notably, the FPGA implementation increases DSP utilization from 11% to 29% compared to the previous designs implemented by other researchers, effectively offloading computational tasks from general purpose logic (LUTs and FFs), reducing LUT usage by 6% and FF usage by 13%, while maintaining overall design stability, scalability, and acceptable thermal margins at 2.387 W. This work establishes a robust foundation for integrating the optimized ORB pipeline into larger drone systems and opens the door for future system-level enhancements. Full article
Show Figures

Figure 1

22 pages, 7096 KB  
Article
An Improved ORB-KNN-Ratio Test Algorithm for Robust Underwater Image Stitching on Low-Cost Robotic Platforms
by Guanhua Yi, Tianxiang Zhang, Yunfei Chen and Dapeng Yu
J. Mar. Sci. Eng. 2026, 14(2), 218; https://doi.org/10.3390/jmse14020218 - 21 Jan 2026
Viewed by 427
Abstract
Underwater optical images often exhibit severe color distortion, weak texture, and uneven illumination due to light absorption and scattering in water. These issues result in unstable feature detection and inaccurate image registration. To address these challenges, this paper proposes an underwater image stitching [...] Read more.
Underwater optical images often exhibit severe color distortion, weak texture, and uneven illumination due to light absorption and scattering in water. These issues result in unstable feature detection and inaccurate image registration. To address these challenges, this paper proposes an underwater image stitching method that integrates ORB (Oriented FAST and Rotated BRIEF) feature extraction with a fixed-ratio constraint matching strategy. First, lightweight color and contrast enhancement techniques are employed to restore color balance and improve local texture visibility. Then, ORB descriptors are extracted and matched via a KNN (K-Nearest Neighbors) nearest-neighbor search, and Lowe’s ratio test is applied to eliminate false matches caused by weak texture similarity. Finally, the geometric transformation between image frames is estimated by incorporating robust optimization, ensuring stable homography computation. Experimental results on real underwater datasets show that the proposed method significantly improves stitching continuity and structural consistency, achieving 40–120% improvements in SSIM (Structural Similarity Index) and PSNR (peak signal-to-noise ratio) over conventional Harris–ORB + KNN, SIFT (scale-invariant feature transform) + BF (brute force), SIFT + KNN, and AKAZE (accelerated KAZE) + BF methods while maintaining processing times within one second. These results indicate that the proposed method is well-suited for real-time underwater environment perception and panoramic mapping on low-cost, micro-sized underwater robotic platforms. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

20 pages, 8055 KB  
Article
Research on an Underwater Visual Enhancement Method Based on Adaptive Parameter Optimization in a Multi-Operator Framework
by Zhiyong Yang, Shengze Yang, Yuxuan Fu and Hao Jiang
Sensors 2026, 26(2), 668; https://doi.org/10.3390/s26020668 - 19 Jan 2026
Viewed by 337
Abstract
Underwater images often suffer from luminance attenuation, structural degradation, and color distortion due to light absorption and scattering in water. The variations in illumination and color distribution across different water bodies further increase the uncertainty of these degradations, making traditional enhancement methods that [...] Read more.
Underwater images often suffer from luminance attenuation, structural degradation, and color distortion due to light absorption and scattering in water. The variations in illumination and color distribution across different water bodies further increase the uncertainty of these degradations, making traditional enhancement methods that rely on fixed parameters, such as underwater dark channel prior (UDCP) and histogram equalization (HE), unstable in such scenarios. To address these challenges, this paper proposes a multi-operator underwater image enhancement framework with adaptive parameter optimization. To achieve luminance compensation, structural detail enhancement, and color restoration, a collaborative enhancement pipeline was constructed using contrast-limited adaptive histogram equalization (CLAHE) with highlight protection, texture-gated and threshold-constrained unsharp masking (USM), and mild saturation compensation. Building upon this pipeline, an adaptive multi-operator parameter optimization strategy was developed, where a unified scoring function jointly considers feature gains, geometric consistency of feature matches, image quality metrics, and latency constraints to dynamically adjust the CLAHE clip limit, USM gain, and Gaussian scale under varying water conditions. Subjective visual comparisons and quantitative experiments were conducted on several public underwater datasets. Compared with conventional enhancement methods, the proposed approach achieved superior structural clarity and natural color appearance on the EUVP and UIEB datasets, and obtained higher quality metrics on the RUIE dataset (Average Gradient (AG) = 0.5922, Underwater Image Quality Measure (UIQM) = 2.095). On the UVE38K dataset, the proposed adaptive optimization method improved the oriented FAST and rotated BRIEF (ORB) feature counts by 12.5%, inlier matches by 9.3%, and UIQM by 3.9% over the fixed-parameter baseline, while the adjacent-frame matching visualization and stability metrics such as inlier ratio further verified the geometric consistency and temporal stability of the enhanced features. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

19 pages, 38545 KB  
Article
Improving Dynamic Visual SLAM in Robotic Environments via Angle-Based Optical Flow Analysis
by Sedat Dikici and Fikret Arı
Electronics 2026, 15(1), 223; https://doi.org/10.3390/electronics15010223 - 3 Jan 2026
Viewed by 636
Abstract
Dynamic objects present a major challenge for visual simultaneous localization and mapping (Visual SLAM), as feature measurements originating from moving regions can corrupt camera pose estimation and lead to inaccurate maps. In this paper, we propose a lightweight, semantic-free front-end enhancement for ORB-SLAM [...] Read more.
Dynamic objects present a major challenge for visual simultaneous localization and mapping (Visual SLAM), as feature measurements originating from moving regions can corrupt camera pose estimation and lead to inaccurate maps. In this paper, we propose a lightweight, semantic-free front-end enhancement for ORB-SLAM that detects and suppresses dynamic features using optical flow geometry. The key idea is to estimate a global motion direction point (MDP) from optical flow vectors and to classify feature points based on their angular consistency with the camera-induced motion field. Unlike magnitude-based flow filtering, the proposed strategy exploits the geometric consistency of optical flow with respect to a motion direction point, providing robustness not only to depth variation and camera speed changes but also to different camera motion patterns, including pure translation and pure rotation. The method is integrated into the ORB-SLAM front-end without modifying the back-end optimization or cost function. Experiments on public dynamic-scene datasets demonstrate that the proposed approach reduces absolute trajectory error by up to approximately 45% compared to baseline ORB-SLAM, while maintaining real-time performance on a CPU-only platform. These results indicate that reliable dynamic feature suppression can be achieved without semantic priors or deep learning models. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

24 pages, 14385 KB  
Article
LDFE-SLAM: Light-Aware Deep Front-End for Robust Visual SLAM Under Challenging Illumination
by Cong Liu, You Wang, Weichao Luo and Yanhong Peng
Machines 2026, 14(1), 44; https://doi.org/10.3390/machines14010044 - 29 Dec 2025
Viewed by 980
Abstract
Visual SLAM systems face significant performance degradation under dynamic lighting conditions, where traditional feature extraction methods suffer from reduced keypoint detection and unstable matching. This paper presents LDFE-SLAM, a novel visual SLAM framework that addresses illumination challenges through a Light-Aware Deep Front-End (LDFE) [...] Read more.
Visual SLAM systems face significant performance degradation under dynamic lighting conditions, where traditional feature extraction methods suffer from reduced keypoint detection and unstable matching. This paper presents LDFE-SLAM, a novel visual SLAM framework that addresses illumination challenges through a Light-Aware Deep Front-End (LDFE) architecture. Our key insight is that low-light degradation in SLAM is fundamentally a geometric feature distribution problem rather than merely a visibility issue. The proposed system integrates three synergistic components: (1) an illumination-adaptive enhancement module based on EnlightenGAN with geometric consistency loss that restores gradient structures for downstream feature extraction, (2) SuperPoint-based deep feature detection that provides illumination-invariant keypoints, and (3) LightGlue attention-based matching that filters enhancement-induced noise while maintaining geometric consistency. Through systematic evaluation of five method configurations (M1–M5), we demonstrate that enhancement, deep features, and learned matching must be co-designed rather than independently optimized. Experiments on EuRoC and TUM sequences under synthetic illumination degradation show that LDFE-SLAM maintains stable localization accuracy (∼1.2 m ATE) across all brightness levels, while baseline methods degrade significantly (up to 3.7 m). Our method operates normally down to severe lighting conditions (30% ambient brightness and 20–50 lux—equivalent to underground parking or night-time streetlight illumination), representing a 4–6× lower illumination threshold compared to ORB-SLAM3 (200–300 lux minimum). Under severe (25% brightness) conditions, our method achieves a 62% tracking success rate, compared to 12% for ORB-SLAM3, with keypoint detection remaining above the critical 100-point threshold, even under extreme degradation. Full article
Show Figures

Figure 1

23 pages, 3212 KB  
Article
AKAZE-GMS-PROSAC: A New Progressive Framework for Matching Dynamic Characteristics of Flotation Foam
by Zhen Peng, Zhihong Jiang, Pengcheng Zhu, Gaipin Cai and Xiaoyan Luo
J. Imaging 2026, 12(1), 7; https://doi.org/10.3390/jimaging12010007 - 25 Dec 2025
Viewed by 370
Abstract
The dynamic characteristics of flotation foam, such as velocity and breakage rate, are critical factors that influence mineral separation efficiency. However, challenges inherent in foam images, including weak textures, severe deformations, and motion blur, present significant technical hurdles for dynamic monitoring. These issues [...] Read more.
The dynamic characteristics of flotation foam, such as velocity and breakage rate, are critical factors that influence mineral separation efficiency. However, challenges inherent in foam images, including weak textures, severe deformations, and motion blur, present significant technical hurdles for dynamic monitoring. These issues lead to a fundamental conflict between the efficiency and accuracy of traditional feature matching algorithms. This paper introduces a novel progressive framework for dynamic feature matching in flotation foam images, termed “stable extraction, efficient coarse screening, and precise matching.” This framework first employs the Accelerated-KAZE (AKAZE) algorithm to extract robust, scale- and rotation-invariant feature points from a non-linear scale-space, effectively addressing the challenge of weak textures. Subsequently, it innovatively incorporates the Grid-based Motion Statistics (GMS) algorithm to perform efficient coarse screening based on motion consistency, rapidly filtering out a large number of obvious mismatches. Finally, the Progressive Sample and Consensus (PROSAC) algorithm is used for precise matching, eliminating the remaining subtle mismatches through progressive sampling and geometric constraints. This framework enables the precise analysis of dynamic foam characteristics, including displacement, velocity, and breakage rate (enhanced by a robust “foam lifetime” mechanism). Comparative experimental results demonstrate that, compared to ORB-GMS-RANSAC (with a Mean Absolute Error, MAE of 1.20 pixels and a Mean Relative Error, MRE of 9.10%) and ORB-RANSAC (MAE: 3.53 pixels, MRE: 27.36%), the proposed framework achieves significantly lower error rates (MAE: 0.23 pixels, MRE: 2.13%). It exhibits exceptional stability and accuracy, particularly in complex scenarios involving low texture and minor displacements. This research provides a high-precision, high-robustness technical solution for the dynamic monitoring and intelligent control of the flotation process. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

24 pages, 22793 KB  
Article
GL-VSLAM: A General Lightweight Visual SLAM Approach for RGB-D and Stereo Cameras
by Xu Li, Tuanjie Li, Yulin Zhang, Ziang Li, Lixiang Ban and Yuming Ning
Sensors 2025, 25(24), 7467; https://doi.org/10.3390/s25247467 - 8 Dec 2025
Cited by 1 | Viewed by 847
Abstract
Feature-based indirect SLAM is more robust than direct SLAM; however, feature extraction and descriptor computation are time-consuming. In this paper, we propose GL-VSLAM, a general lightweight visual SLAM approach designed for RGB-D and stereo cameras. GL-VSLAM utilizes sparse optical flow matching based on [...] Read more.
Feature-based indirect SLAM is more robust than direct SLAM; however, feature extraction and descriptor computation are time-consuming. In this paper, we propose GL-VSLAM, a general lightweight visual SLAM approach designed for RGB-D and stereo cameras. GL-VSLAM utilizes sparse optical flow matching based on uniform motion model prediction to establish keypoint correspondences between consecutive frames, rather than relying on descriptor-based feature matching, thereby achieving high real-time performance. To enhance positioning accuracy, we adopt a coarse-to-fine strategy for pose estimation in two stages. In the first stage, the initial camera pose is estimated using RANSAC PnP based on robust keypoint correspondences from sparse optical flow. In the second stage, the camera pose is further refined by minimizing the reprojection error. Keypoints and descriptors are extracted from keyframes for backend optimization and loop closure detection. We evaluate our system on the TUM and KITTI datasets, as well as in a real-world environment, and compare it with several state-of-the-art methods. Experimental results demonstrate that our method achieves comparable positioning accuracy, while its efficiency is up to twice that of ORB-SLAM2. Full article
Show Figures

Figure 1

22 pages, 5092 KB  
Article
Fault Diagnosis Method for Excitation Dry-Type Transformer Based on Multi-Channel Vibration Signal and Visual Feature Fusion
by Yang Liu, Mingtao Yu, Jingang Wang, Peng Bao, Weiguo Zu, Yinglong Deng, Shiyi Chen, Lijiang Ma, Pengcheng Zhao and Jinyao Dou
Sensors 2025, 25(24), 7460; https://doi.org/10.3390/s25247460 - 8 Dec 2025
Cited by 1 | Viewed by 751
Abstract
To address the limitations of existing fault diagnosis methods for excitation dry-type transformers, such as inadequate utilization of multi-axis vibration data, low recognition accuracy under complex operational conditions, and limited computational efficiency, this paper presents a lightweight fault diagnosis approach based on the [...] Read more.
To address the limitations of existing fault diagnosis methods for excitation dry-type transformers, such as inadequate utilization of multi-axis vibration data, low recognition accuracy under complex operational conditions, and limited computational efficiency, this paper presents a lightweight fault diagnosis approach based on the fusion of multi-channel vibration signals and visual features. Initially, a multi-physics field coupling simulation model of the excitation dry-type transformer is developed. Vibration data collected from field-installed three-axis sensors are combined to generate typical fault samples, including normal operation, winding looseness, core looseness, and winding eccentricity. Due to the high dimensionality of vibration signals, the Symmetrized Dot Pattern (ISDP) method is extended to aggregate and map time- and frequency-domain information from the x-, y-, and z-axes into a two-dimensional feature map. To optimize the inter-class separability and intra-class consistency of the map, Particle Swarm Optimization (PSO) is employed to adaptively adjust the angle gain factor (η) and time delay coefficient (t). Keypoint descriptors are then extracted from the map using the Oriented FAST and Rotated BRIEF (ORB) feature extraction operator, which improves computational efficiency while maintaining sensitivity to local details. Finally, an efficient fault classification model is constructed using an Adaptive Boosting Support Vector Machine (Adaboost-SVM) to achieve robust fault mode recognition across multiple operating conditions. Experimental results demonstrate that the proposed method achieves a fault diagnosis accuracy of 94.00%, outperforming signal-to-image techniques such as Gramian Angular Field (GAF), Recurrence Plot (RP), and Markov Transition Field (MTF), as well as deep learning models based on Convolutional Neural Networks (CNN) in both training and testing time. Additionally, the method exhibits superior stability and robustness in repeated trials. This approach is well-suited for online monitoring and rapid diagnosis in resource-constrained environments, offering significant engineering value in enhancing the operational safety and reliability of excitation dry-type transformers. Full article
(This article belongs to the Collection Sensors and Sensing Technology for Industry 4.0)
Show Figures

Figure 1

16 pages, 87659 KB  
Article
UAV-TIRVis: A Benchmark Dataset for Thermal–Visible Image Registration from Aerial Platforms
by Costin-Emanuel Vasile, Călin Bîră and Radu Hobincu
J. Imaging 2025, 11(12), 432; https://doi.org/10.3390/jimaging11120432 - 4 Dec 2025
Viewed by 1271
Abstract
Registering UAV-based thermal and visible images is a challenging task due to differences in appearance across spectra and the lack of public benchmarks. To address this issue, we introduce UAV-TIRVis, a dataset consisting of 80 accurately and manually registered UAV-based thermal (640 × [...] Read more.
Registering UAV-based thermal and visible images is a challenging task due to differences in appearance across spectra and the lack of public benchmarks. To address this issue, we introduce UAV-TIRVis, a dataset consisting of 80 accurately and manually registered UAV-based thermal (640 × 512) and visible (4K) image pairs, captured across diverse environments. We benchmark our dataset using well-known registration methods, including feature-based (ORB, SURF, SIFT, KAZE), correlation-based, and intensity-based methods, as well as a custom, heuristic intensity-based method. We evaluate the performance of these methods using four metrics: RMSE, PSNR, SSIM, and NCC, averaged per scenario and across the entire dataset. The results show that conventional methods often fail to generalize across scenes, yielding <0.6 NCC on average, whereas the heuristic method shows that it is possible to achieve 0.77 SSIM and 0.82 NCC, highlighting the difficulty of cross-spectral UAV alignment and the need for further research to improve optimization in existing registration methods. Full article
Show Figures

Figure 1

16 pages, 1863 KB  
Article
Superpoint Network-Based Video Stabilization Technology for Mine Rescue Robots
by Shuqi Wang, Zhaowenbo Zhu and Yikai Jiang
Appl. Sci. 2025, 15(22), 12322; https://doi.org/10.3390/app152212322 - 20 Nov 2025
Viewed by 556
Abstract
Mine rescue robots operate in extremely adverse subterranean environments, where the acquired video data are frequently affected by severe jitter and motion distortion. Such instability leads to the loss of critical visual information, thereby reducing the reliability of rescue decision-making. To address this [...] Read more.
Mine rescue robots operate in extremely adverse subterranean environments, where the acquired video data are frequently affected by severe jitter and motion distortion. Such instability leads to the loss of critical visual information, thereby reducing the reliability of rescue decision-making. To address this issue, a dual-channel visual stabilization framework based on the SuperPoint network is proposed, extending the traditional ORB descriptor framework. Here, dual-channel refers to two configurable and mutually exclusive feature extraction paths—an ORB-based path and a SuperPoint-based path—that can be flexibly switched according to scene conditions and computational requirements, rather than operating simultaneously on the same frame. The subsequent stabilization pipeline remains unified and consistent across both modes. The method employs an optimized detector head that integrates deep feature extraction, non-maximum suppression, and boundary filtering to enable precise estimation of inter-frame motion. When combined with smoothing filters, the approach effectively attenuates vibrations induced by irregular terrain and dynamic operational conditions. Experimental evaluations conducted across diverse scenarios demonstrate that the proposed algorithm achieves an average improvement of 27.91% in Peak Signal-to-Noise Ratio (PSNR), a 55.04% reduction in Mean Squared Error (MSE), and more than a twofold increase in the Structural Similarity Index (SSIM) relative to pre-stabilized sequences. Moreover, runtime analysis indicates that the algorithm can operate in near-real-time, supporting its practical deployment on embedded mine rescue robot platforms.These results verify the algorithm’s robustness and applicability in environments requiring high visual stability and image fidelity, providing a reliable foundation for enhanced visual perception and autonomous decision-making in complex disaster scenarios. Full article
Show Figures

Figure 1

Back to TopTop