Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (487)

Search Parameters:
Keywords = stereo mapping

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
31 pages, 162558 KB  
Article
SAOF: A Semantic-Aware Optical Flow Framework for Fine-Grained Disparity Estimation in High-Resolution Satellite Stereo Images
by Dingkai Wang, Feng Wang, Jingyi Cao, Niangang Jiao, Yuming Xiang, Enze Zhu, Jingxing Zhu and Hongjian You
Remote Sens. 2025, 17(24), 4017; https://doi.org/10.3390/rs17244017 - 12 Dec 2025
Viewed by 164
Abstract
Disparity estimation from high-resolution satellite stereo images is critical for 3D reconstruction but remains challenging due to large disparities, complex structures, and textureless regions. To address this, we propose a Semantic-Aware Optical Flow (SAOF) framework for fine-grained disparity estimation, which enhances optical flow-based [...] Read more.
Disparity estimation from high-resolution satellite stereo images is critical for 3D reconstruction but remains challenging due to large disparities, complex structures, and textureless regions. To address this, we propose a Semantic-Aware Optical Flow (SAOF) framework for fine-grained disparity estimation, which enhances optical flow-based via a multi-level optimization incorporating sub-top pyramid re-PatchMatch, scale-adaptive matching windows, and multi-feature cost refinement. For improving the spatial consistency of the resulting disparity map, SAMgeo-Reg is utilized to produce semantic prototypes, which are used to build guidance embeddings for integration into the optical flow estimation process. Experiments on the US3D dataset demonstrate that SAOF outperforms state-of-the-art methods across challenging scenarios. It achieves an average endpoint error (EPE) of 1.317 and a D1 error of 9.09%. Full article
Show Figures

Figure 1

20 pages, 14182 KB  
Article
Automated 3D Phenotyping of Maize Plants: Stereo Matching Guided by Deep Learning
by Juan Zapata-Londoño, Juan Botero-Valencia, Ítalo A. Torres, Erick Reyes-Vera and Ruber Hernández-García
Agriculture 2025, 15(24), 2573; https://doi.org/10.3390/agriculture15242573 - 12 Dec 2025
Viewed by 186
Abstract
Automated three-dimensional plant phenotyping is an essential tool for non-destructive analysis of plant growth and structure. This paper presents a low-cost system based on stereo vision for depth estimation and morphological characterization of maize plants. The system incorporates an automatic detection stage for [...] Read more.
Automated three-dimensional plant phenotyping is an essential tool for non-destructive analysis of plant growth and structure. This paper presents a low-cost system based on stereo vision for depth estimation and morphological characterization of maize plants. The system incorporates an automatic detection stage for the object of interest using deep learning techniques to delimit the region of interest (ROI) corresponding to the plant. The Semi-Global Block Matching (SGBM) algorithm is applied to the detected region to compute the disparity map and generate a partial three-dimensional representation of the plant structure. The ROI delimitation restricts the disparity calculation to the plant area, reducing processing of the background and optimizing computational resource use. The deep learning-based detection stage maintains stable foliage identification even under varying lighting conditions and shadowing, ensuring consistent depth data across different experimental conditions. Overall, the proposed system integrates detection and disparity estimation into an efficient processing flow, providing an accessible alternative for automated three-dimensional phenotyping in agricultural environments. Full article
(This article belongs to the Special Issue Field Phenotyping for Precise Crop Management)
Show Figures

Figure 1

20 pages, 5083 KB  
Article
MDR–SLAM: Robust 3D Mapping in Low-Texture Scenes with a Decoupled Approach and Temporal Filtering
by Kailin Zhang and Letao Zhou
Electronics 2025, 14(24), 4864; https://doi.org/10.3390/electronics14244864 - 10 Dec 2025
Viewed by 193
Abstract
Realizing real-time dense 3D reconstruction on resource-limited mobile platforms remains a significant challenge, particularly in low-texture environments that demand robust multi-frame fusion to resolve matching ambiguities. However, the inherent tight coupling of pose estimation and mapping in traditional monolithic SLAM architectures imposes a [...] Read more.
Realizing real-time dense 3D reconstruction on resource-limited mobile platforms remains a significant challenge, particularly in low-texture environments that demand robust multi-frame fusion to resolve matching ambiguities. However, the inherent tight coupling of pose estimation and mapping in traditional monolithic SLAM architectures imposes a severe restriction on integrating high-complexity fusion algorithms without compromising tracking stability. To overcome these limitations, this paper proposes MDR–SLAM, a modular and fully decoupled stereo framework. The system features a novel keyframe-driven temporal filter that synergizes efficient ELAS stereo matching with Kalman filtering to effectively accumulate geometric constraints, thereby enhancing reconstruction density in textureless areas. Furthermore, a confidence-based fusion backend is employed to incrementally maintain global map consistency and filter outliers. Quantitative evaluation on the NUFR-M3F indoor dataset demonstrates the effectiveness of the proposed method: compared to the standard single-frame baseline, MDR–SLAM reduces map RMSE by 83.3% (to 0.012 m) and global trajectory drift by 55.6%, while significantly improving map completeness. The system operates entirely on CPU resources with a stable 4.7 Hz mapping frequency, verifying its suitability for embedded mobile robotics. Full article
(This article belongs to the Special Issue Recent Advance of Auto Navigation in Indoor Scenarios)
Show Figures

Figure 1

35 pages, 1152 KB  
Review
Why “Where” Matters as Much as “How Much”: Single-Cell and Spatial Transcriptomics in Plants
by Kinga Moskal, Marta Puchta-Jasińska, Paulina Bolc, Adrian Motor, Rafał Frankowski, Aleksandra Pietrusińska-Radzio, Anna Rucińska, Karolina Tomiczak and Maja Boczkowska
Int. J. Mol. Sci. 2025, 26(24), 11819; https://doi.org/10.3390/ijms262411819 - 7 Dec 2025
Viewed by 230
Abstract
Plant tissues exhibit a layered architecture that makes spatial context decisive for interpreting transcriptional changes. This review explains why the location of gene expression is as important as its magnitude and synthesizes advances uniting single-cell/nucleus RNA-seq with spatial transcriptomics in plants. Surveyed topics [...] Read more.
Plant tissues exhibit a layered architecture that makes spatial context decisive for interpreting transcriptional changes. This review explains why the location of gene expression is as important as its magnitude and synthesizes advances uniting single-cell/nucleus RNA-seq with spatial transcriptomics in plants. Surveyed topics include platform selection and material preparation; plant-specific sample processing and quality control; integration with epigenomic assays such as single-nucleus Assay for Transposase-Accessible Chromatin using sequencing (ATAC) and Multiome; and computational workflows for label transfer, deconvolution, spatial embedding, and neighborhood-aware cell–cell communication. Protoplast-based single-cell RNA sequencing (scRNA-seq) enables high-resolution profiling but introduces dissociation artifacts and cell-type biases, whereas ingle-nucleus RNA sequencing (snRNA-seq) improves the representation of recalcitrant lineages and reduces stress signatures while remaining compatible with multiomics profiling. Practical guidance is provided for mitigating ambient RNA, interpreting organellar and intronic metrics, identifying doublets, and harmonizing batches across chemistries and studies. Spatial platforms (Visium HD, Stereo-seq, bead arrays) and targeted imaging (Single-molecule fluorescence in situ hybridization (smFISH), Hairpin-chain-reaction FISH (HCR-FISH), Multiplexed Error-Robust Fluorescence In Situ Hybridization (MERFISH) ) are contrasted with plant-specific adaptations and integration pipelines that anchor dissociated profiles in anatomical coordinates. Recent atlases in Arabidopsis, soybean, and maize illustrate how cell identities, chromatin accessibility, and spatial niches reveal developmental trajectories and stress responses jointly. A roadmap is outlined for moving from atlases to interventions by deriving gene regulatory networks, prioritizing cis-regulatory targets, and validating perturbations with spatial readouts in crops. Together, these principles support a transition from descriptive maps to mechanism-informed, low-pleiotropy engineering of agronomic traits. Full article
(This article belongs to the Special Issue Plant Physiology and Molecular Nutrition: 2nd Edition)
22 pages, 5368 KB  
Article
Training Data for Stereo Matching Algorithms Based on Neural Networks and a Method for Data Evaluation
by Adam L. Kaczmarek
Appl. Sci. 2025, 15(23), 12663; https://doi.org/10.3390/app152312663 - 29 Nov 2025
Viewed by 443
Abstract
This paper addresses the problem of selecting data for learning stereo matching algorithms. The paper presents an overview of currently available learning datasets, including synthetic data and data from real environments. Stereo matching algorithms based on neural networks require high quality learning data [...] Read more.
This paper addresses the problem of selecting data for learning stereo matching algorithms. The paper presents an overview of currently available learning datasets, including synthetic data and data from real environments. Stereo matching algorithms based on neural networks require high quality learning data in order to provide expected results, which have a form of disparity maps. There are hundreds of stereo matching algorithms for processing images from stereo cameras. However, a significant problem with this 3D data processing technology is that there is a relatively low availability of learning data with dense ground truth from real environments. The paper also introduces an evaluation method for estimating the quality of input data. The method considers features such as the quality of calibration, the density of ground truth, and the occurrence of occluded areas. The research presented in this paper is aimed at developing stereo matching methods applicable not only to benchmarks of this kind of algorithms but also to applications in real environments. Full article
(This article belongs to the Section Robotics and Automation)
Show Figures

Figure 1

20 pages, 15632 KB  
Article
Investigating an Earthquake Surface Rupture Along the Kumysh Fault (Eastern Tianshan, Central Asia) from High-Resolution Topographic Data
by Jiahui Han, Haiyun Bi, Wenjun Zheng, Hui Qiu, Fuer Yang, Xinyuan Chen and Jiaoyan Yang
Remote Sens. 2025, 17(23), 3847; https://doi.org/10.3390/rs17233847 - 27 Nov 2025
Viewed by 227
Abstract
As direct geomorphic evidence and records of earthquakes on the surface, coseismic surface ruptures have long been a key focus in earthquake research. However, compared with strike-slip and normal faults, studies on reverse-fault surface ruptures remain relatively scarce. In this study, surface rupture [...] Read more.
As direct geomorphic evidence and records of earthquakes on the surface, coseismic surface ruptures have long been a key focus in earthquake research. However, compared with strike-slip and normal faults, studies on reverse-fault surface ruptures remain relatively scarce. In this study, surface rupture characteristics of the most recent earthquake on the Kumysh thrust fault in eastern Tianshan were investigated using high-resolution topographic data, including 0.5 m- and 5 cm-resolution Digital Elevation Models (DEMs) generated from the WorldView-2 satellite stereo image pairs and Unmanned Aerial Vehicle (UAV) images, respectively. We carefully mapped the spatial geometry of the surface rupture and measured 120 vertical displacements along the rupture strike. Using the moving-window method and statistical analysis, both moving-mean and moving-maximum coseismic displacement curves were obtained for the entire rupture zone. Results show that the most recent rupture on the Kumysh Fault extends ~25 km with an overall NWW strike, exhibits complex spatial geometry, and can be subdivided into five secondary segments, which are discontinuously distributed in arcuate shapes across both piedmont alluvial fans and mountain fronts. Reverse fault scarps dominate the rupture pattern. The along-strike coseismic displacements generally form three asymmetric triangles, with an average displacement of 0.9–1.1 m and a maximum displacement of 2.8–3.2 m, yielding an estimated earthquake magnitude of Mw 6.6–6.7. This study not only highlights the strong potential of high-resolution remote sensing data for investigating surface earthquake ruptures, but also provides an additional example to the relatively underexplored reverse-fault surface ruptures. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

14 pages, 1310 KB  
Article
Stereo-GS: Online 3D Gaussian Splatting Mapping Using Stereo Depth Estimation
by Junkyu Park, Byeonggwon Lee, Sanggi Lee and Soohwan Song
Electronics 2025, 14(22), 4436; https://doi.org/10.3390/electronics14224436 - 14 Nov 2025
Viewed by 1297
Abstract
We present Stereo-GS, a real-time system for online 3D Gaussian Splatting (3DGS) that reconstructs photorealistic 3D scenes from streaming stereo pairs. Unlike prior offline 3DGS methods that require dense multi-view input or precomputed depth, Stereo-GS estimates metrically accurate depth maps directly from rectified [...] Read more.
We present Stereo-GS, a real-time system for online 3D Gaussian Splatting (3DGS) that reconstructs photorealistic 3D scenes from streaming stereo pairs. Unlike prior offline 3DGS methods that require dense multi-view input or precomputed depth, Stereo-GS estimates metrically accurate depth maps directly from rectified stereo geometry, enabling progressive, globally consistent reconstruction. The frontend combines a stereo implementation of DROID-SLAM for robust tracking and keyframe selection with FoundationStereo, a generalizable stereo network that needs no scene-specific fine-tuning. A two-stage filtering pipeline improves depth reliability by removing outliers using a variance-based refinement filter followed by a multi-view consistency check. In the backend, we selectively initialize new Gaussians in under-represented regions flagged by low PSNR during rendering and continuously optimize them via differentiable rendering. To maintain global coherence with minimal overhead, we apply a lightweight rigid alignment after periodic bundle adjustment. On EuRoC and TartanAir, Stereo-GS attains state-of-the-art performance, improving average PSNR by 0.22 dB and 2.45 dB over the best baseline, respectively. Together with superior visual quality, these results show that Stereo-GS delivers high-fidelity, geometrically accurate 3D reconstructions suitable for real-time robotics, navigation, and immersive AR/VR applications. Full article
(This article belongs to the Special Issue Real-Time Computer Vision)
Show Figures

Figure 1

8 pages, 2554 KB  
Proceeding Paper
Optimal Sensor Placement for Autonomous Formula Student Vehicles: A Field-of-View Analysis of Dual LIDAR and Stereo Camera Configurations
by Máté Kapocsi, László Illés Orova and Zoltán Pusztai
Eng. Proc. 2025, 113(1), 27; https://doi.org/10.3390/engproc2025113027 - 31 Oct 2025
Viewed by 873
Abstract
The optimal configuration of perception systems in autonomous vehicles is essential for accurate environmental sensing, precise navigation, and overall operational safety. In Formula Student Driverless (FSD) vehicles, sensor placement is particularly challenging due to the compact design constraints and the highly dynamic nature [...] Read more.
The optimal configuration of perception systems in autonomous vehicles is essential for accurate environmental sensing, precise navigation, and overall operational safety. In Formula Student Driverless (FSD) vehicles, sensor placement is particularly challenging due to the compact design constraints and the highly dynamic nature of the racing environment. This study investigates the positioning and configuration of two LIDAR sensors and a stereo camera on an FSD race car, focusing on field-of-view coverage, sensing redundancy, and sensor fusion potential. To achieve a comprehensive evaluation, measurements are conducted exclusively in a simulation environment, where field-of-view maps are generated, detection ranges are analyzed, and perception reliability is assessed under various conditions. The results provide insights into the optimal sensor arrangement that minimizes blind spots, maximizes sensing accuracy, and enhances the efficiency of the autonomous vehicle’s perception architecture. Full article
(This article belongs to the Proceedings of The Sustainable Mobility and Transportation Symposium 2025)
Show Figures

Figure 1

18 pages, 3754 KB  
Article
Hardware Implementation of Improved Oriented FAST and Rotated BRIEF-Simultaneous Localization and Mapping Version 2
by Ji-Long He, Ying-Hua Chen, Wenny Ramadha Putri, Chung-I. Huang, Ming-Hsiang Su, Kuo-Chen Li, Jian-Hong Wang, Shih-Lun Chen, Yung-Hui Li and Jia-Ching Wang
Sensors 2025, 25(20), 6404; https://doi.org/10.3390/s25206404 - 17 Oct 2025
Viewed by 975
Abstract
The field of autonomous driving has seen continuous advances, yet achieving higher levels of automation in real-world applications remains challenging. A critical requirement for autonomous navigation is accurate map construction, particularly in novel and unstructured environments. In recent years, Simultaneous Localization and Mapping [...] Read more.
The field of autonomous driving has seen continuous advances, yet achieving higher levels of automation in real-world applications remains challenging. A critical requirement for autonomous navigation is accurate map construction, particularly in novel and unstructured environments. In recent years, Simultaneous Localization and Mapping (SLAM) has evolved to support diverse sensor modalities, with some implementations incorporating machine learning to improve performance. However, these approaches often demand substantial computational resources. The key challenge lies in achieving efficiency within resource-constrained environments while minimizing errors that could degrade downstream tasks. This paper presents an enhanced ORB-SLAM2 (Oriented FAST and Rotated BRIEF Simultaneous Localization and Mapping, version 2) algorithm implemented on a Raspberry Pi 3 (ARM A53 CPU) to improve mapping performance under limited computational resources. ORB-SLAM2 comprises four main stages: Tracking, Local Mapping, Loop Closing, and Full Bundle Adjustment (BA). The proposed improvements include employing a more efficient feature descriptor to increase stereo feature-matching rates and optimizing loop-closing parameters to reduce accumulated errors. Experimental results demonstrate that the proposed system achieves notable improvements on the Raspberry Pi 3 platform. For monocular SLAM, RMSE is reduced by 18.11%, mean error by 22.97%, median error by 29.41%, and maximum error by 17.18%. For stereo SLAM, RMSE decreases by 0.30% and mean error by 0.38%. Furthermore, the ROS topic frequency stabilizes at 10 Hz, with quad-core CPU utilization averaging approximately 90%. These results indicate that the system satisfies real-time requirements while maintaining a balanced trade-off between accuracy and computational efficiency under resource constraints. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

22 pages, 17900 KB  
Article
Custom Material Scanning System for PBR Texture Acquisition: Hardware Design and Digitisation Workflow
by Lunan Wu, Federico Morosi and Giandomenico Caruso
Appl. Sci. 2025, 15(20), 10911; https://doi.org/10.3390/app152010911 - 11 Oct 2025
Viewed by 1079
Abstract
Real-time rendering is increasingly used in augmented and virtual reality (AR/VR), interactive design, and product visualisation, where materials must prioritise efficiency and consistency rather than the extreme accuracy required in offline rendering. In parallel, the growing demand for personalised and customised products has [...] Read more.
Real-time rendering is increasingly used in augmented and virtual reality (AR/VR), interactive design, and product visualisation, where materials must prioritise efficiency and consistency rather than the extreme accuracy required in offline rendering. In parallel, the growing demand for personalised and customised products has created a need for digital materials that can be generated in-house without relying on expensive commercial systems. To address these requirements, this paper presents a low-cost digitisation workflow based on photometric stereo. The system integrates a custom-built scanner with cross-polarised illumination, automated multi-light image acquisition, a dual-stage colour calibration process, and a node-based reconstruction pipeline that produces albedo and normal maps. A reproducible evaluation methodology is also introduced, combining perceptual colour-difference analysis using the CIEDE2000 (ΔE00) metric with angular-error assessment of normal maps on known-geometry samples. By openly providing the workflow, bill of materials, and implementation details, this work delivers a practical and replicable solution for reliable material capture in real-time rendering and product customisation scenarios. Full article
Show Figures

Figure 1

25 pages, 18016 KB  
Article
Joint Modeling of Pixel-Wise Visibility and Fog Structure for Real-World Scene Understanding
by Jiayu Wu, Jiaheng Li, Jianqiang Wang, Xuezhe Xu, Sidan Du and Yang Li
Atmosphere 2025, 16(10), 1161; https://doi.org/10.3390/atmos16101161 - 4 Oct 2025
Viewed by 559
Abstract
Reduced visibility caused by foggy weather has a significant impact on transportation systems and driving safety, leading to increased accident risks and decreased operational efficiency. Traditional methods rely on expensive physical instruments, limiting their scalability. To address this challenge in a cost-effective manner, [...] Read more.
Reduced visibility caused by foggy weather has a significant impact on transportation systems and driving safety, leading to increased accident risks and decreased operational efficiency. Traditional methods rely on expensive physical instruments, limiting their scalability. To address this challenge in a cost-effective manner, we propose a two-stage network for visibility estimation from stereo image inputs. The first stage computes scene depth via stereo matching, while the second stage fuses depth and texture information to estimate metric-scale visibility. Our method produces pixel-wise visibility maps through a physically constrained, progressive supervision strategy, providing rich spatial visibility distributions beyond a single global value. Moreover, it enables the detection of patchy fog, allowing a more comprehensive understanding of complex atmospheric conditions. To facilitate training and evaluation, we propose an automatic fog-aware data generation pipeline that incorporates both synthetically rendered foggy images and real-world captures. Furthermore, we construct a large-scale dataset encompassing diverse scenarios. Extensive experiments demonstrate that our method achieves state-of-the-art performance in both visibility estimation and patchy fog detection. Full article
(This article belongs to the Section Atmospheric Techniques, Instruments, and Modeling)
Show Figures

Figure 1

20 pages, 74841 KB  
Article
Autonomous Concrete Crack Monitoring Using a Mobile Robot with a 2-DoF Manipulator and Stereo Vision Sensors
by Seola Yang, Daeik Jang, Jonghyeok Kim and Haemin Jeon
Sensors 2025, 25(19), 6121; https://doi.org/10.3390/s25196121 - 3 Oct 2025
Cited by 1 | Viewed by 1162
Abstract
Crack monitoring in concrete structures is essential to maintaining structural integrity. Therefore, this paper proposes a mobile ground robot equipped with a 2-DoF manipulator and stereo vision sensors for autonomous crack monitoring and mapping. To facilitate crack detection over large areas, a 2-DoF [...] Read more.
Crack monitoring in concrete structures is essential to maintaining structural integrity. Therefore, this paper proposes a mobile ground robot equipped with a 2-DoF manipulator and stereo vision sensors for autonomous crack monitoring and mapping. To facilitate crack detection over large areas, a 2-DoF motorized manipulator providing linear and rotational motions, with a stereo vision sensor mounted on the end effector, was deployed. In combination with a manual rotation plate, this configuration enhances accessibility and expands the field of view for crack monitoring. Another stereo vision sensor, mounted at the front of the robot, was used to acquire point cloud data of the surrounding environment, enabling tasks such as SLAM (simultaneous localization and mapping), path planning and following, and obstacle avoidance. Cracks are detected and segmented using the deep learning algorithms YOLO (You Only Look Once) v6-s and SFNet (Semantic Flow Network), respectively. To enhance the performance of crack segmentation, synthetic image generation and preprocessing techniques, including cropping and scaling, were applied. The dimensions of cracks are calculated using point clouds filtered with the median absolute deviation method. To validate the performance of the proposed crack-monitoring and mapping method with the robot system, indoor experimental tests were performed. The experimental results confirmed that, in cases of divided imaging, the crack propagation direction was predicted, enabling robotic manipulation and division-point calculation. Subsequently, total crack length and width were calculated by combining reconstructed 3D point clouds from multiple frames, with a maximum relative error of 1%. Full article
Show Figures

Figure 1

19 pages, 5861 KB  
Article
Topological Signal Processing from Stereo Visual SLAM
by Eleonora Di Salvo, Tommaso Latino, Maria Sanzone, Alessia Trozzo and Stefania Colonnese
Sensors 2025, 25(19), 6103; https://doi.org/10.3390/s25196103 - 3 Oct 2025
Viewed by 574
Abstract
Topological signal processing is emerging alongside Graph Signal Processing (GSP) in various applications, incorporating higher-order connectivity structures—such as faces—in addition to nodes and edges, for enriched connectivity modeling. Rich point clouds acquired by multi-camera systems in Visual Simultaneous Localization and Mapping (V-SLAM) are [...] Read more.
Topological signal processing is emerging alongside Graph Signal Processing (GSP) in various applications, incorporating higher-order connectivity structures—such as faces—in addition to nodes and edges, for enriched connectivity modeling. Rich point clouds acquired by multi-camera systems in Visual Simultaneous Localization and Mapping (V-SLAM) are typically processed using graph-based methods. In this work, we introduce a topological signal processing (TSP) framework that integrates texture information extracted from V-SLAM; we refer to this framework as TSP-SLAM. We show how TSP-SLAM enables the extension of graph-based point cloud processing to more advanced topological signal processing techniques. We demonstrate, on real stereo data, that TSP-SLAM enables a richer point cloud representation by associating signals not only with vertices but also with edges and faces of the mesh computed from the point cloud. Numerical results show that TSP-SLAM supports the design of topological filtering algorithms by exploiting the mapping between the 3D mesh faces, edges and vertices and their 2D image projections. These findings confirm the potential of TSP-SLAM for topological signal processing of point cloud data acquired in challenging V-SLAM environments. Full article
(This article belongs to the Special Issue Stereo Vision Sensing and Image Processing)
Show Figures

Figure 1

31 pages, 25818 KB  
Article
FishKP-YOLOv11: An Automatic Estimation Model for Fish Size and Mass in Complex Underwater Environments
by Jinfeng Wang, Zhipeng Cheng, Mingrun Lin, Renyou Yang and Qiong Huang
Animals 2025, 15(19), 2862; https://doi.org/10.3390/ani15192862 - 30 Sep 2025
Viewed by 1183
Abstract
The size and mass of fish are crucial parameters in aquaculture management. However, existing research primarily focuses on conducting fish size and mass estimation under ideal conditions, which limits its application in actual aquaculture scenarios with complex water quality and fluctuating lighting. A [...] Read more.
The size and mass of fish are crucial parameters in aquaculture management. However, existing research primarily focuses on conducting fish size and mass estimation under ideal conditions, which limits its application in actual aquaculture scenarios with complex water quality and fluctuating lighting. A non-contact size and mass measurement framework is proposed for complex underwater environments, which integrates the improved FishKP-YOLOv11 module based on YOLOv11, stereo vision technology, and a Random Forest model. This framework fuses the detected 2D key points with binocular stereo technology to reconstruct the 3D key point coordinates. Fish size is computed based on these 3D key points, and a Random Forest model establishes a mapping relationship between size and mass. For validating the performance of the framework, a self-constructed grass carp dataset for key point detection is established. The experimental results indicate that the mean average precision (mAP) of FishKP-YOLOv11 surpasses that of diverse versions of YOLOv5–YOLOv12. The mean absolute errors (MAEs) for length and width estimations are 0.35 cm and 0.10 cm, respectively. The MAE for mass estimations is 2.7 g. Therefore, the proposed framework is well suited for application in actual breeding environments. Full article
Show Figures

Figure 1

22 pages, 23570 KB  
Article
Bundled-Images Based Geo-Positioning Method for Satellite Images Without Using Ground Control Points
by Zhenling Ma, Yuan Chen, Xu Zhong, Hong Xie, Yanlin Liu, Zhengjie Wang and Peng Shi
Remote Sens. 2025, 17(19), 3289; https://doi.org/10.3390/rs17193289 - 25 Sep 2025
Viewed by 471
Abstract
Bundle adjustment without Ground Control Points (GCPs) using stereo remote sensing images represents a reliable and efficient approach for realizing the demand for regional and global mapping. This paper proposes a bundled-images based geo-positioning method that leverages a Kalman filter to effectively integrate [...] Read more.
Bundle adjustment without Ground Control Points (GCPs) using stereo remote sensing images represents a reliable and efficient approach for realizing the demand for regional and global mapping. This paper proposes a bundled-images based geo-positioning method that leverages a Kalman filter to effectively integrate new image observations with their corresponding historical bundled images. Under the assumption that the noise follows a Gaussian distribution, a linear mean square estimator is employed to orient the new images. The historical bundled images can be updated with posterior covariance information to maintain consistent accuracy with the newly oriented images. This method employs recursive computation to dynamically orient the new images, ensuring consistent accuracy across all the historical and new images. To validate the proposed method, extensive experiments were carried out using two satellite datasets comprising both homologous (IKONOS) and heterogeneous (TH-1 and ZY-3) sources. The experiment results reveal that without using GCPs, the proposed method can meet 1:50,000 mapping standards with heterogeneous TH-1 and ZY-3 datasets and 1:10,000 mapping accuracy requirements with homologous IKONOS datasets. These experiments indicate that as the bundled images expand further, the image quantity growth no longer results in substantial improvements in precision, suggesting the presence of an accuracy ceiling. The final positioning accuracy is predominantly influenced by the initial bundled image quality. Experimental evidence suggests that when using the proposed method, the bundled image sets should exhibit superior precision compared to subsequently new images. In future research, we will expand the coverage to regional or global scales. Full article
Show Figures

Figure 1

Back to TopTop