Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (127)

Search Parameters:
Keywords = semantic point cloud reconstruction

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 5467 KB  
Article
Automated Dimension Recognition and BIM Modeling of Frame Structures Based on 3D Point Clouds
by Fengyu Zhang, Jinyang Liu, Peizhen Li, Lin Chen and Qingsong Xiong
Electronics 2026, 15(2), 293; https://doi.org/10.3390/electronics15020293 - 9 Jan 2026
Abstract
Building information models (BIMs) serve as a foundational tool for digital management of existing structures. Traditional methods suffer from low automation and heavy reliance on manual intervention. This paper proposes an automated method for structural component dimension recognition and BIM modeling based on [...] Read more.
Building information models (BIMs) serve as a foundational tool for digital management of existing structures. Traditional methods suffer from low automation and heavy reliance on manual intervention. This paper proposes an automated method for structural component dimension recognition and BIM modeling based on 3D point cloud data. The proposed methodology follows a three-step workflow. First, the raw point cloud is semantically segmented using the PointNet++ deep learning network, and individual structural components are effectively isolated using the Fast Euclidean Clustering (FEC) algorithm. Second, the principal axis of each component is determined through Principal Component Analysis, and the Random Sample Consensus (RANSAC) algorithm is applied to fit the boundary lines of the projected cross-sections, enabling the automated extraction of geometric dimensions. Finally, an automated script maps the extracted geometric parameters to standard IFC entities to generate the BIM model. The experimental results demonstrate that the average dimensional error for beams and columns is within 3 mm, with the exception of specific occluded components. This study realizes the efficient transformation from point cloud data to BIM models through an automated workflow, providing reliable technical support for the digital reconstruction of existing buildings. Full article
Show Figures

Figure 1

22 pages, 17762 KB  
Article
Highway Reconstruction Through Fine-Grained Semantic Segmentation of Mobile Laser Scanning Data
by Yuyu Chen, Zhou Yang, Huijing Zhang and Jinhu Wang
Sensors 2026, 26(1), 40; https://doi.org/10.3390/s26010040 - 20 Dec 2025
Viewed by 333
Abstract
The highway is a crucial component of modern transportation systems, and its efficient management is essential for ensuring safety and facilitating communication. The automatic understanding and reconstruction of highway environments are therefore pivotal for advanced traffic management and intelligent transportation systems. This work [...] Read more.
The highway is a crucial component of modern transportation systems, and its efficient management is essential for ensuring safety and facilitating communication. The automatic understanding and reconstruction of highway environments are therefore pivotal for advanced traffic management and intelligent transportation systems. This work introduces a methodology for the fine-grained semantic segmentation and reconstruction of highway environments using dense 3D point cloud data acquired via mobile laser scanning. First, a multi-scale, object-based data augmentation and down-sampling method is introduced to address the issue of training sample imbalance. Subsequently, a deep learning approach utilizing the KPConv convolutional network is proposed to achieve fine-grained semantic segmentation. The segmentation results are then used to reconstruct a 3D model of the highway environment. The methodology is validated on a 32 km stretch of highway, achieving semantic segmentation across 27 categories of environmental features. When evaluated against a manually annotated ground truth, the results exhibit a mean Intersection over Union (mIoU) of 87.27%. These findings demonstrate that the proposed methodology is effective for fine-grained semantic segmentation and instance-level reconstruction of highways in practical scenarios. Full article
(This article belongs to the Special Issue Application of LiDAR Remote Sensing and Mapping)
Show Figures

Figure 1

33 pages, 9178 KB  
Article
Automated Image-to-BIM Using Neural Radiance Fields and Vision-Language Semantic Modeling
by Mohammad H. Mehraban, Shayan Mirzabeigi, Mudan Wang, Rui Liu and Samad M. E. Sepasgozar
Buildings 2025, 15(24), 4549; https://doi.org/10.3390/buildings15244549 - 16 Dec 2025
Viewed by 567
Abstract
This study introduces a novel, automated image-to-BIM (Building Information Modeling) workflow designed to generate semantically rich and geometrically useful BIM models directly from RGB images. Conventional scan-to-BIM often relies on specialized, costly, and time-intensive equipment, specifically if LiDAR is used to generate point [...] Read more.
This study introduces a novel, automated image-to-BIM (Building Information Modeling) workflow designed to generate semantically rich and geometrically useful BIM models directly from RGB images. Conventional scan-to-BIM often relies on specialized, costly, and time-intensive equipment, specifically if LiDAR is used to generate point clouds (PCs). Typical workflows are followed by a separate post-processing step for semantic segmentation recently performed by deep learning models on the generated PCs. Instead, the proposed method integrates vision language object detection (YOLOv8x-World v2) and vision based segmentation (SAM 2.1) with Neural Radiance Fields (NeRF) 3D reconstruction to generate segmented, color-labeled PCs directly from images. The key novelty lies in bypassing post-processing on PCs by embedding semantic information at the pixel level in images, preserving it through reconstruction, and encoding it into the resulting color labeled PC, which allows building elements to be directly identified and geometrically extracted based on color labels. Extracted geometry is serialized into a JSON format and imported into Revit to automate BIM creation for walls, windows, and doors. Experimental validation on BIM models generated from Unmanned Aerial Vehicle (UAV)-based exterior datasets and standard camera-based interior datasets demonstrated high accuracy in detecting windows and doors. Spatial evaluations yielded up to 0.994 precision and 0.992 Intersection over Union (IoU). NeRF and Gaussian Splatting models, Nerfacto, Instant-NGP, and Splatfacto, were assessed. Nerfacto produced the most structured PCs suitable for geometry extraction and Splatfacto achieved the highest image reconstruction quality. The proposed method removes dependency on terrestrial surveying tools and separate segmentation processes on PCs. It provides a low-cost and scalable solution for generating BIM models in aging or undocumented buildings and supports practical applications such as renovation, digital twin, and facility management. Full article
(This article belongs to the Special Issue Artificial Intelligence in Architecture and Interior Design)
Show Figures

Graphical abstract

25 pages, 22959 KB  
Article
A Semi-Automatic Framework for Dry Beach Extraction in Tailings Ponds Using Photogrammetry and Deep Learning
by Bei Cao, Yinsheng Wang, Yani Li, Xudong Zhu, Zicheng Yang, Xinlong Liu and Guangyin Lu
Remote Sens. 2025, 17(24), 4022; https://doi.org/10.3390/rs17244022 - 13 Dec 2025
Viewed by 309
Abstract
The spatial characteristics of the dry beach in tailings ponds are critical indicators for the safety assessment of tailings dams. This study presents a method for dry beach extraction that combines deep learning-based semantic segmentation with 3D reconstruction, overcoming the limitations of 2D [...] Read more.
The spatial characteristics of the dry beach in tailings ponds are critical indicators for the safety assessment of tailings dams. This study presents a method for dry beach extraction that combines deep learning-based semantic segmentation with 3D reconstruction, overcoming the limitations of 2D methods in spatial analysis. The workflow includes four steps: (1) High-resolution 3D point clouds are reconstructed from UAV images, and the projection matrix of each image is derived to link 2D pixels with 3D points. (2) AlexNet and GoogLeNet are employed to extract image features and automatically select images containing the dry beach boundary. (3) A DeepLabv3+ network is trained on manually labeled samples to perform semantic segmentation of the dry beach, with a lightweight incremental training strategy for enhanced adaptability. (4) Boundary pixels are detected and back-projected into 3D space to generate consistent point cloud boundaries. The method was validated on two-phase UAV datasets from a tailings pond in Yunnan Province, China. In phase I, the model achieved high segmentation performance, with a mean Accuracy and IoU of approximately 0.95 and a BF of 0.8267. When applied to phase II without retraining, the model maintained stable performance on dam boundaries, while slight performance degradation was observed on hillside and water boundaries. The 3D back-projection converted 2D boundary pixels into 3D coordinates, enabling the extraction of dry beach point clouds and supporting reliable dry beach length monitoring and deposition morphology analysis. Full article
(This article belongs to the Section Engineering Remote Sensing)
Show Figures

Graphical abstract

18 pages, 5130 KB  
Article
Efficient Hierarchical Spatial Indexing for Managing Remote Sensing Data Streams Using the PL-2000 Map-Sheet System
by Mariusz Zygmunt and Marta Róg
Appl. Sci. 2025, 15(24), 12915; https://doi.org/10.3390/app152412915 - 8 Dec 2025
Viewed by 290
Abstract
Efficient spatial indexing is critical for processing large-scale remote sensing datasets (e.g., LiDAR point clouds, orthophotos, hyperspectral imagery). We present a bidirectional, hierarchical index based on the Polish PL-2000 coordinate reference system for (1) direct computation of a map-sheet identifier from metric coordinates [...] Read more.
Efficient spatial indexing is critical for processing large-scale remote sensing datasets (e.g., LiDAR point clouds, orthophotos, hyperspectral imagery). We present a bidirectional, hierarchical index based on the Polish PL-2000 coordinate reference system for (1) direct computation of a map-sheet identifier from metric coordinates (forward encoder) and (2) reconstruction of the sheet extent from the identifier alone (inverse decoder). By replacing geometric point-in-polygon tests with closed-form arithmetic, the method achieves constant-time assignment O(1), eliminates boundary-geometry loading, and enables multi-scale aggregation via simple code truncation. Unlike global spatial indices (e.g., H3, S2), a CRS-native, aligned with cartographic map sheets in PL-2000 implementation, removes reprojection overhead and preserves the legal sheet semantics, enabling the direct use of deterministic O(1) numeric keys for remote-sensing data and Polish archives. We detail the algorithms, formalize their complexity and boundary rules across all PL-2000 zones, and analyze memory trade-offs, including a compact 26-bit packing of numeric keys for nationwide single-table indexing. We also discuss integration patterns with the OGC Tile Matrix Set (TMS), ETL pipelines, and GeoAI workflows, showing how bidirectional indexing accelerates ingest, training and inference, and national-scale visualization. Although demonstrated for PL-2000, the approach is transferable to other national coordinate reference systems, illustrating how statutory map-sheet identification schemes can be transformed into high-performance indices for modern remote sensing and AI data pipelines. Full article
Show Figures

Figure 1

36 pages, 14822 KB  
Article
Deep Learning for Unsupervised 3D Shape Representation with Superquadrics
by Mahmoud Eltaher and Michael Breuß
AI 2025, 6(12), 317; https://doi.org/10.3390/ai6120317 - 4 Dec 2025
Viewed by 731
Abstract
The representation of 3D shapes from point clouds remains a fundamental challenge in computer vision. A common approach decomposes 3D objects into interpretable geometric primitives, enabling compact, structured, and efficient representations. Building upon prior frameworks, this study introduces an enhanced unsupervised deep learning [...] Read more.
The representation of 3D shapes from point clouds remains a fundamental challenge in computer vision. A common approach decomposes 3D objects into interpretable geometric primitives, enabling compact, structured, and efficient representations. Building upon prior frameworks, this study introduces an enhanced unsupervised deep learning approach for 3D shape representation using superquadrics. The proposed framework fits a set of superquadric primitives to 3D objects through a fully integrated, differentiable pipeline that enables efficient optimization and parameter learning, directly extracting geometric structure from 3D point clouds without requiring ground-truth segmentation labels. This work introduces three key advancements that substantially improve representation quality, interpretability, and evaluation rigor: (1) A uniform sampling strategy that enhances training stability compared with random sampling used in earlier models; (2) An overlapping loss that penalizes intersections between primitives, reducing redundancy and improving reconstruction coherence; and (3) A novel evaluation framework comprising Primitive Accuracy, Structural Accuracy, and Overlapping Percentage metrics. This new metric design transitions from point-based to structure-aware assessment, enabling fairer and more interpretable comparison across primitive-based models. Comprehensive evaluations on benchmark 3D shape datasets demonstrate that the proposed modifications yield coherent, compact, and semantically consistent shape representations, establishing a robust foundation for interpretable and quantitative evaluation in primitive-based 3D reconstruction. Full article
(This article belongs to the Section AI Systems: Theory and Applications)
Show Figures

Figure 1

27 pages, 5618 KB  
Article
Real-Time Semantic Reconstruction and Semantically Constrained Path Planning for Agricultural Robots in Greenhouses
by Tianrui Quan, Junjie Luo, Shuxin Xie, Xuesong Ren and Yubin Miao
Agronomy 2025, 15(12), 2696; https://doi.org/10.3390/agronomy15122696 - 23 Nov 2025
Viewed by 711
Abstract
To address perception and navigation challenges in precision agriculture caused by GPS signal loss and weakly structured environments in greenhouses, this study proposes an integrated framework for real-time semantic reconstruction and path planning. This framework comprises three core components: First, it introduces a [...] Read more.
To address perception and navigation challenges in precision agriculture caused by GPS signal loss and weakly structured environments in greenhouses, this study proposes an integrated framework for real-time semantic reconstruction and path planning. This framework comprises three core components: First, it introduces a semantic segmentation method tailored for greenhouse environments, enhancing recognition accuracy of key navigable areas such as furrows. Second, it designs a visual-semantic fusion SLAM point cloud reconstruction algorithm and proposes a semantic point cloud rasterization method. Finally, it develops a semantic-constrained A* path planning algorithm adapted for semantic maps. We collected a segmentation dataset (1083 images, 4 classes) and a reconstruction dataset from greenhouses in Shanghai. Experiments demonstrate that the segmentation algorithm achieves 95.44% accuracy and 87.93% mIoU, with a 3.9% improvement in furrow category recognition accuracy. The reconstructed point cloud exhibits an average relative error of 7.37% on furrows. In practical greenhouse validation, single-frame point cloud fusion took approximately 0.35 s, while path planning was completed in under 1 s. Feasible paths avoiding crops were successfully generated across three structurally distinct greenhouses. Results demonstrate that this framework can stably and in real-time accomplish semantic mapping and path planning, providing effective technical support for digital agriculture. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

35 pages, 20479 KB  
Article
Comprehensive Forensic Tool for Crime Scene and Traffic Accident 3D Reconstruction
by Alejandra Ospina-Bohórquez, Esteban Ruiz de Oña, Roy Yali, Emmanouil Patsiouras, Katerina Margariti and Diego González-Aguilera
Algorithms 2025, 18(11), 707; https://doi.org/10.3390/a18110707 - 7 Nov 2025
Viewed by 1532
Abstract
This article presents a comprehensive forensic tool for crime scene and traffic accident investigations, integrating advanced 3D reconstruction and semantic and dynamic analyses; the tool facilitates the accurate documentation and preservation of crime scenes through photogrammetric techniques, producing detailed 3D models based on [...] Read more.
This article presents a comprehensive forensic tool for crime scene and traffic accident investigations, integrating advanced 3D reconstruction and semantic and dynamic analyses; the tool facilitates the accurate documentation and preservation of crime scenes through photogrammetric techniques, producing detailed 3D models based on images or video captured under specified protocols. The system includes modules for semantic analysis, enabling object detection and classification in 3D point clouds and 2D images. By employing machine learning methods such as the Random Forest model for point cloud classification and the YOLOv8 architecture for object detection, the tool enhances the accuracy and reliability of forensic analysis. Furthermore, a dynamic analysis module supports ballistic trajectory calculations for crime scene investigations and the vehicle impact speed estimation using the Equivalent Barrier Speed (EBS) model for traffic accidents. These capabilities are integrated into a single, user-friendly platform offering significant improvements over existing forensic tools, which often focus on singular tasks and require expertise. This tool provides a robust, accessible solution for law enforcement agencies, enabling more efficient and precise forensic investigations across different scenarios. Full article
(This article belongs to the Special Issue Modern Algorithms for Image Processing and Computer Vision)
Show Figures

Figure 1

25 pages, 18310 KB  
Article
A Multimodal Fusion Method for Weld Seam Extraction Under Arc Light and Fume Interference
by Lei Cai and Han Zhao
J. Manuf. Mater. Process. 2025, 9(11), 350; https://doi.org/10.3390/jmmp9110350 - 26 Oct 2025
Viewed by 1190
Abstract
During the Gas Metal Arc Welding (GMAW) process, intense arc light and dense fumes cause local overexposure in RGB images and data loss in point clouds, which severely compromises the extraction accuracy of circular closed-curve weld seams. To address this challenge, this paper [...] Read more.
During the Gas Metal Arc Welding (GMAW) process, intense arc light and dense fumes cause local overexposure in RGB images and data loss in point clouds, which severely compromises the extraction accuracy of circular closed-curve weld seams. To address this challenge, this paper proposes a multimodal fusion method for weld seam extraction under arc light and fume interference. The method begins by constructing a weld seam edge feature extraction (WSEF) module based on a synergistic fusion network, which achieves precise localization of the weld contour by coupling image arc light-removal and semantic segmentation tasks. Subsequently, an image-to-point cloud mapping-guided Local Point Cloud Feature extraction (LPCF) module was designed, incorporating the Shuffle Attention mechanism to enhance robustness against noise and occlusion. Building upon this, a cross-modal attention-driven multimodal feature fusion (MFF) module integrates 2D edge features with 3D structural information to generate a spatially consistent and detail-rich fused point cloud. Finally, a hierarchical trajectory reconstruction and smoothing method is employed to achieve high-precision reconstruction of the closed weld seam path. The experimental results demonstrate that under severe arc light and fume interference, the proposed method achieves a Root Mean Square Error below 0.6 mm, a maximum error not exceeding 1.2 mm, and a processing time under 5 s. Its performance significantly surpasses that of existing methods, showcasing excellent accuracy and robustness. Full article
Show Figures

Figure 1

29 pages, 13306 KB  
Article
Building Outline Extraction via Topology-Aware Loop Parsing and Parallel Constraint from Airborne LiDAR
by Ke Liu, Hongchao Ma, Li Li, Shixin Huang, Liang Zhang, Xiaoli Liang and Zhan Cai
Remote Sens. 2025, 17(20), 3498; https://doi.org/10.3390/rs17203498 - 21 Oct 2025
Viewed by 675
Abstract
Building outlines are important vector data for various applications, but due to the uneven point density and complex building structures, extracting satisfactory building outlines from airborne light detection and ranging point cloud data poses significant challenges. Thus, a building outline extraction method based [...] Read more.
Building outlines are important vector data for various applications, but due to the uneven point density and complex building structures, extracting satisfactory building outlines from airborne light detection and ranging point cloud data poses significant challenges. Thus, a building outline extraction method based on topology-aware loop parsing and parallel constraint is proposed. First, constrained Delaunay triangulation (DT) is used to organize scattered projected building points, and initial boundary points and edges are extracted based on the constrained DT. Subsequently, accurate semantic boundary points are obtained by parsing the topology-aware loops searched from an undirected graph. Building dominant directions are estimated through angle normalization, merging, and perpendicular pairing. Finally, outlines are regularized using the parallel constraint-based method, which simultaneously considers the fitness between the dominant direction and boundary points, and the length of line segments. Experiments on five datasets, including three datasets provided by ISPRS and two datasets with high-density point clouds and complex building structures, verify that the proposed method can extract sequential and semantic boundary points, with over 97.88% correctness. Additionally, the regularized outlines are attractive, and most line segments are parallel or perpendicular. The RMSE, PoLiS, and RCC metrics are better than 0.94 m, 0.84 m, and 0.69 m, respectively. The extracted building outlines can be used for building three-dimensional (3D) reconstruction. Full article
Show Figures

Figure 1

26 pages, 12151 KB  
Article
Toward Automatic 3D Model Reconstruction of Building Curtain Walls from UAV Images Based on NeRF and Deep Learning
by Zeyu Li, Qian Wang, Hongzhe Yue and Xiang Nie
Remote Sens. 2025, 17(19), 3368; https://doi.org/10.3390/rs17193368 - 5 Oct 2025
Cited by 2 | Viewed by 1314
Abstract
The Automated Building Information Modeling (BIM) reconstruction of existing building curtain walls is crucial for promoting digital Operation and Maintenance (O&M). However, existing 3D reconstruction technologies are mainly designed for general architectural scenes, and there is currently a lack of research specifically focused [...] Read more.
The Automated Building Information Modeling (BIM) reconstruction of existing building curtain walls is crucial for promoting digital Operation and Maintenance (O&M). However, existing 3D reconstruction technologies are mainly designed for general architectural scenes, and there is currently a lack of research specifically focused on the BIM reconstruction of curtain walls. This study proposes a BIM reconstruction method from unmanned aerial vehicle (UAV) images based on neural radiance field (NeRF) and deep learning-based semantic segmentation. The proposed method compensates for the lack of semantic information in traditional NeRF methods and could fill the gap in the automatic reconstruction of semantic models for curtain walls. A comprehensive high-rise building is selected as a case study to validate the proposed method. The results show that the overall accuracy (OA) for semantic segmentation of curtain wall point clouds is 71.8%, and the overall dimensional error of the reconstructed BIM model is less than 0.1m, indicating high modeling accuracy. Additionally, this study compares the proposed method with photogrammetry-based reconstruction and traditional semantic segmentation methods to further validate its effectiveness. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

20 pages, 74841 KB  
Article
Autonomous Concrete Crack Monitoring Using a Mobile Robot with a 2-DoF Manipulator and Stereo Vision Sensors
by Seola Yang, Daeik Jang, Jonghyeok Kim and Haemin Jeon
Sensors 2025, 25(19), 6121; https://doi.org/10.3390/s25196121 - 3 Oct 2025
Cited by 1 | Viewed by 1300
Abstract
Crack monitoring in concrete structures is essential to maintaining structural integrity. Therefore, this paper proposes a mobile ground robot equipped with a 2-DoF manipulator and stereo vision sensors for autonomous crack monitoring and mapping. To facilitate crack detection over large areas, a 2-DoF [...] Read more.
Crack monitoring in concrete structures is essential to maintaining structural integrity. Therefore, this paper proposes a mobile ground robot equipped with a 2-DoF manipulator and stereo vision sensors for autonomous crack monitoring and mapping. To facilitate crack detection over large areas, a 2-DoF motorized manipulator providing linear and rotational motions, with a stereo vision sensor mounted on the end effector, was deployed. In combination with a manual rotation plate, this configuration enhances accessibility and expands the field of view for crack monitoring. Another stereo vision sensor, mounted at the front of the robot, was used to acquire point cloud data of the surrounding environment, enabling tasks such as SLAM (simultaneous localization and mapping), path planning and following, and obstacle avoidance. Cracks are detected and segmented using the deep learning algorithms YOLO (You Only Look Once) v6-s and SFNet (Semantic Flow Network), respectively. To enhance the performance of crack segmentation, synthetic image generation and preprocessing techniques, including cropping and scaling, were applied. The dimensions of cracks are calculated using point clouds filtered with the median absolute deviation method. To validate the performance of the proposed crack-monitoring and mapping method with the robot system, indoor experimental tests were performed. The experimental results confirmed that, in cases of divided imaging, the crack propagation direction was predicted, enabling robotic manipulation and division-point calculation. Subsequently, total crack length and width were calculated by combining reconstructed 3D point clouds from multiple frames, with a maximum relative error of 1%. Full article
Show Figures

Figure 1

25 pages, 10178 KB  
Article
Super-Resolution Point Cloud Completion for Large-Scale Missing Data in Cotton Leaves
by Hui Geng, Zhiben Yin, Mingdeng Shi, Junzhang Pan and Chunjing Si
Agriculture 2025, 15(18), 1989; https://doi.org/10.3390/agriculture15181989 - 22 Sep 2025
Viewed by 718
Abstract
Point cloud completion for cotton leaves is critical for accurately reconstructing complete shapes from sparse and significantly incomplete data. Traditional methods typically assume small missing ratios (≤25%), which limits their effectiveness for morphologically complex cotton leaves with severe sparsity (50–75%), large geometric distortions, [...] Read more.
Point cloud completion for cotton leaves is critical for accurately reconstructing complete shapes from sparse and significantly incomplete data. Traditional methods typically assume small missing ratios (≤25%), which limits their effectiveness for morphologically complex cotton leaves with severe sparsity (50–75%), large geometric distortions, and extensive point loss. To overcome these challenges, we introduce an end-to-end neural network that combines PF-Net and PointNet++ to effectively reconstruct dense, uniform point clouds from incomplete inputs. The model initially uses a multiresolution encoder to extract multiscale features from locally incomplete point clouds at different resolutions. By capturing both low-level and high-level attributes, these features significantly enhance the network’s ability to represent semantic content and geometric structure. Next, a point pyramid decoder generates missing point clouds hierarchically from layers at different depths, effectively reconstructing the fine details of the original structure. PointNet++ is then used to fuse and reshape the incomplete input point clouds with the generated missing points, yielding a fully reconstructed and uniformly distributed point cloud. To ensure effective task completion at different training stages, a loss function freezing strategy is employed, optimizing the network’s performance throughout the training process. Experimental evaluation on the cotton leaf dataset demonstrated that the proposed model outperformed PF-Net, reducing the Chamfer distance by 80.15% and the Earth Mover distance by 54.35%. These improvements underscore the model’s robustness in reconstructing sparse point clouds for precise agricultural phenotyping. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

16 pages, 2720 KB  
Article
Multi-Trait Phenotypic Extraction and Fresh Weight Estimation of Greenhouse Lettuce Based on Inspection Robot
by Xiaodong Zhang, Xiangyu Han, Yixue Zhang, Lian Hu and Tiezhu Li
Agriculture 2025, 15(18), 1929; https://doi.org/10.3390/agriculture15181929 - 11 Sep 2025
Viewed by 920
Abstract
In situ detection of growth information in greenhouse crops is crucial for germplasm resource optimization and intelligent greenhouse management. To address the limitations of poor flexibility and low automation in traditional phenotyping platforms, this study developed a controlled environment inspection robot. By means [...] Read more.
In situ detection of growth information in greenhouse crops is crucial for germplasm resource optimization and intelligent greenhouse management. To address the limitations of poor flexibility and low automation in traditional phenotyping platforms, this study developed a controlled environment inspection robot. By means of a SCARA robotic arm equipped with an information acquisition device consisting of an RGB camera, a depth camera, and an infrared thermal imager, high-throughput and in situ acquisition of lettuce phenotypic information can be achieved. Through semantic segmentation and point cloud reconstruction, 12 phenotypic parameters, such as lettuce plant height and crown width, were extracted from the acquired images as inputs for three machine learning models to predict fresh weight. By analyzing the training results, a Backpropagation Neural Network (BPNN) with an added feature dimension-increasing module (DE-BP) was proposed, achieving improved prediction accuracy. The R2 values for plant height, crown width, and fresh weight predictions were 0.85, 0.93, and 0.84, respectively, with RMSE values of 7 mm, 6 mm, and 8 g, respectively. This study achieved in situ, high-throughput acquisition of lettuce phenotypic information under controlled environmental conditions, providing a lightweight solution for crop phenotypic information analysis algorithms tailored for inspection tasks. Full article
Show Figures

Figure 1

23 pages, 4627 KB  
Article
Dynamic SLAM Dense Point Cloud Map by Fusion of Semantic Information and Bayesian Moving Probability
by Qing An, Shao Li, Yanglu Wan, Wei Xuan, Chao Chen, Bufan Zhao and Xijiang Chen
Sensors 2025, 25(17), 5304; https://doi.org/10.3390/s25175304 - 26 Aug 2025
Viewed by 1442
Abstract
Most existing Simultaneous Localization and Mapping (SLAM) systems rely on the assumption of static environments to achieve reliable and efficient mapping. However, such methods often suffer from degraded localization accuracy and mapping consistency in dynamic settings, as they lack explicit mechanisms to distinguish [...] Read more.
Most existing Simultaneous Localization and Mapping (SLAM) systems rely on the assumption of static environments to achieve reliable and efficient mapping. However, such methods often suffer from degraded localization accuracy and mapping consistency in dynamic settings, as they lack explicit mechanisms to distinguish between static and dynamic elements. To overcome this limitation, we present BMP-SLAM, a vision-based SLAM approach that integrates semantic segmentation and Bayesian motion estimation to robustly handle dynamic indoor scenes. To enable real-time dynamic object detection, we integrate YOLOv5, a semantic segmentation network that identifies and localizes dynamic regions within the environment, into a dedicated dynamic target detection thread. Simultaneously, the data association Bayesian mobile probability proposed in this paper effectively eliminates dynamic feature points and successfully reduces the impact of dynamic targets in the environment on the SLAM system. To enhance complex indoor robotic navigation, the proposed system integrates semantic keyframe information with dynamic object detection outputs to reconstruct high-fidelity 3D point cloud maps of indoor environments. The evaluation conducted on the TUM RGB-D dataset indicates that the performance of BMP-SLAM is superior to that of ORB-SLAM3, with the trajectory tracking accuracy improved by 96.35%. Comparative evaluations demonstrate that the proposed system achieves superior performance in dynamic environments, exhibiting both lower trajectory drift and enhanced positioning precision relative to state-of-the-art dynamic SLAM methods. Full article
(This article belongs to the Special Issue Indoor Localization Technologies and Applications)
Show Figures

Figure 1

Back to TopTop