Next Article in Journal
An Anti-Mainlobe Suppression Jamming Method Based on Improved Blind Source Separation Using Variational Mode Decomposition and Wavelet Packet Decomposition
Previous Article in Journal
Energy-Efficient Multi-Agent Deep Reinforcement Learning Task Offloading and Resource Allocation for UAV Edge Computing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review of Optical-Based Three-Dimensional Reconstruction and Multi-Source Fusion for Plant Phenotyping

College of Information Science and Engineering, Shanxi Agricultural University, Jinzhong 030800, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(11), 3401; https://doi.org/10.3390/s25113401
Submission received: 28 March 2025 / Revised: 23 May 2025 / Accepted: 25 May 2025 / Published: 28 May 2025
(This article belongs to the Section Smart Agriculture)

Abstract

:
In the context of the booming development of precision agriculture and plant phenotyping, plant 3D reconstruction technology has become a research hotspot, with widespread applications in plant growth monitoring, pest and disease detection, and smart agricultural equipment. Given the complex geometric and textural characteristics of plants, traditional 2D image analysis methods are difficult to meet the modeling requirements, highlighting the growing importance of 3D reconstruction technology. This paper reviews active vision techniques (such as structured light, time-of-flight, and laser scanning methods), passive vision techniques (such as stereo vision and structure from motion), and deep learning-based 3D reconstruction methods (such as NeRF, CNN, and 3DGS). These technologies enhance crop analysis accuracy from multiple perspectives, provide strong support for agricultural production, and significantly promote the development of the field of plant research.

1. Introduction

With the continuous advancement of precision agriculture and plant phenotyping research, plant 3D reconstruction technology has become an essential tool in crop growth monitoring, pest and disease detection, and the development of intelligent agricultural equipment [1,2,3]. This review synthesizes research primarily from the last 10 years, emphasizing the significant role of 3D reconstruction in accurately modeling plant morphology, improving the analysis of plant growth status, enabling high-throughput phenotypic trait extraction, and providing a solid scientific foundation for crop variety improvement and precision management [4,5,6]. However, due to the complex geometric morphology and intricate textural features of plants, traditional 2D image analysis methods are insufficient for accurate modeling, which makes the application of 3D reconstruction technology in plant research especially critical [7,8].
Over the past decade, 3D reconstruction techniques have made significant strides. From active vision methods such as structured light, time-of-flight (ToF), and laser scanning, to passive vision techniques like stereo vision and Structure from Motion (SfM), and more recently, advanced deep learning-based methods such as Neural Radiance Fields (NeRF), Convolutional Neural Networks (CNNs), and 3D Gaussian Splatting (3DGS), continuous innovation has opened new frontiers for plant research [9,10,11]. These methods enable the capture of crop morphological features from various angles and at multiple levels, providing comprehensive details ranging from macroscopic plant architecture to microscopic leaf texture and venation. This versatility makes them invaluable tools for precise crop growth analysis, early pest and disease prevention, and optimizing agricultural yield [12,13].
A growing trend in the field is the exploration of multi-source fusion for 3D reconstruction, which combines data from various sensors and integrates 3D models with plant growth physical models. This fusion enhances the accuracy and completeness of the reconstructed models and addresses challenges such as occlusion, wind-induced disturbances, and growth variability [14]. For instance, by combining depth sensors with optical sensors, and integrating these with physiological data, more detailed and reliable plant models can be obtained. Moreover, the application of high-speed imaging systems and event cameras has shown promise in reconstructing dynamic plant scenes, further advancing the real-time capabilities of 3D plant reconstruction technologies.
This review aims to systematically organize the current state and development trends of 3D reconstruction technologies in plant research. We will discuss both active and passive vision techniques as well as deep learning-driven methods, elaborating on their specific applications, advantages, and challenges in plant research. The goal is to provide a thorough overview and offer practical guidance for future research, helping to advance the field and enhance the precision and efficiency of agricultural management. By also considering the potential of multi-source data fusion and its role in improving model accuracy and adaptability, this paper outlines a comprehensive path forward for integrating 3D reconstruction with plant phenotyping in precision agriculture. Table 1 summarizes the technology evolution and application of 3D reconstruction methods.

2. Three-Dimensional Reconstruction Technology

Three-dimensional (3D) reconstruction technology refers to the use of specialized equipment or the processing of two-dimensional (2D) image data to create precise 3D models of real-world objects or scenes [15,16]. This technology enables comprehensive capture of the geometric morphology and spatial structure of plants, making it highly valuable in agricultural research, ecological monitoring, and plant biology. Based on data acquisition methods, 3D reconstruction can be categorized into active 3D reconstruction and passive 3D reconstruction [17]. Additionally, according to algorithmic principles, there are deep learning-based 3D reconstruction techniques [18].

2.1. Active 3D Reconstruction Techniques

Active 3D reconstruction technique refers to the use of actively emitted signals (e.g., light, laser, infrared, etc.) to measure the distance between the surface of the object and the sensor in the 3D reconstruction process, so as to obtain the 3D information of the object, which plays an important role in the field of plant research [19].

2.1.1. Structured Light Method

The advantage of the structured light method is that it can achieve high-precision 3D information extraction in a large field of view, especially suitable for complex surfaces with rich details [20]. The core principle is based on optical triangulation, where the depth information of an object’s surface is deduced by the known relationship between the relative positions of the light source and the camera, and by measuring the offsets of the light stripes in the camera’s image plane [21]. In practice, a laser transmitter emits a laser beam, which is modulated by the optical system into a striped beam and projected onto the surface of the object [22]. The camera then captures the image of the light pattern on the object’s surface, and subsequently calculates the precise coordinates of the object surface in 3D space according to the calibration equation and the degree of distortion of the light stripes [23].
As it relies on active light source illumination, the structured light method is highly resistant to ambient light interference, can work stably under natural light environments, and has good real-time performance, making it suitable for real-time 3D measurement of dynamic scenes [24].
With its high accuracy and good real-time performance, the structured light method has been widely applied in the field of plant 3D reconstruction, especially in measuring the geometric morphology of plant leaves, the surface texture of fruits, and the overall 3D structure of plants [25]. In agricultural research, the structured light method can be used to monitor plant growth conditions, analyze leaf curl, and evaluate the surface quality of fruits. However, traditional measurement methods, such as 2D image processing or laser scanning, can easily fail to accurately extract the 3D information of an object when the surface features are not prominent (e.g., smooth fruit surfaces or unclear textures) [26]. Chen Xujia et al. proposed a solution to the inability to effectively measure objects due to inconspicuous surface features or occlusion problems in traditional methods. By projecting light stripes of known patterns onto the object’s surface, the deformation of structured light provides obvious depth information, enabling the extraction of 3D morphology from fruit surfaces even in the absence of clear surface features. Furthermore, the relative measurement error of the standard part dimensions through 3D reconstruction is within 3.32%. The deformation index of apples has an R2 of 0.97, RMSE of 0.755 mm, and mean absolute percentage error (MAPE) of 7.23%. For spherical-like fruits, the R2 values for volume and maximum diameter are 0.99 and 0.92, with RMSE values of 6.015 cubic centimeters and 1.823 mm, and MAPE values of 1.946% and 1.859%, respectively [27]. This method effectively overcomes the measurement difficulties caused by the lack of clear texture in traditional methods.

2.1.2. Time-of-Flight (TOF)

Time-of-Flight (TOF) is a 3D imaging technique based on measuring the time-of-flight of an optical signal to determine the distance. The technique calculates the distance traveled by a light wave by transmitting a light pulse and accurately measuring the time elapsed from transmission to reception of the reflected light pulse [28]. TOF technology has been widely used in many fields due to its easy operation, fast response speed, and low cost. In particular, in the field of plant 3D reconstruction, TOF can effectively measure a wide range of plant structural information [29].
A TOF camera system typically consists of a transmitter (such as a laser or LED light) and a receiver (such as a camera). The transmitter emits short pulses of light waves, which are reflected back after striking the target object, and the receiver captures these reflected light pulses, as shown in Figure 1 [30]. By accurately measuring the round-trip time of the light waves and based on the principle of the constant speed of light, the TOF camera can calculate the distance from the object to the camera using the formula
d = c t 2 ,
where d represents the distance, c represents the speed of light, and t is the round-trip time of the light wave [31].
In plant science research and agriculture, TOF cameras have been successfully used for 3D modeling of plants, ranging from large-scale trees to small plants [32]. These data are critical for analyzing plant growth adaptations, monitoring pests and diseases, and evaluating crop yields [14]. TOF technology can be applied for accurate measurements of tree height, branch distribution, and fruit count in orchards, assisting agricultural practitioners in optimizing plant cultivation and management plans [33]. Manuel Vázquez-Arellano et al. proposed a 3D reconstruction method for maize plants based on Time-of-Flight (TOF) cameras. They employed the Iterative Closest Point (ICP) algorithm for point cloud registration and used the Random Sample Consensus (RANSAC) algorithm to remove soil points, thus improving the accuracy of the plant point cloud. The method achieves reconstruction by capturing high-resolution 3D images of the maize plants, stitching these point clouds together, and segmenting plant points from soil points. By comparing the positions of the maize seedlings in the generated point cloud with ground-truth measurements, the average deviation of the reconstruction results was 3.4 cm, with a standard deviation of ±1.3 cm [28].
One of the main advantages of TOF technology is that it can perform high-precision measurements under various lighting conditions, especially when dealing with large-scale scenes. However, in environments with highly reflective or extremely dark surfaces, these conditions can affect the effective reception of the light pulses [34].

2.1.3. Laser Scanning Method

The laser scanning method is an active vision 3D reconstruction technique based on laser ranging technology to obtain high-precision 3D point cloud data from the surface of an object [35]. This method uses a laser scanner to emit a laser beam to the target object and measures the time or phase shift of the reflected laser beam returning to the scanner, thus calculating the exact position of each point on the object’s surface [36]. This technology is particularly suitable for accurate mapping of large-scale environments and complex terrains, including forest surveys, farmland analysis, and plant structure studies [37].
A laser scanning system typically consists of a laser transmitter, a receiver, a ranging system, and a computer system [38]. The laser transmitter emits a focused laser beam onto the surface of the target object, and the receiver records the intensity and time information of the reflected laser signal [39]. By precisely measuring these parameters, the system can calculate the distance between the laser and the object and determine its exact position in 3D space. To further enhance measurement accuracy and data reliability, advanced laser scanning systems are typically equipped with Global Navigation Satellite Systems (GNSS) and Inertial Measurement Units (IMUs) to ensure precise positioning of the data over large-scale operations [40].
In plant science research and precision agriculture, laser scanning technology plays a key role in recording and analyzing plant growth [41]. For instance, laser scanning can accurately measure tree height, branch distribution, and canopy structure. In forest management, these 3D data can be used to estimate timber volume, monitor forest health, and optimize forestry management strategies. Laser scanning is also used for crop growth monitoring. By analyzing the 3D structure of crops, it is possible to assess growth conditions and predict yields [42]. Leonard Freißmuth et al. proposed an online tree reconstruction and forest inventory system based on a mobile robotic platform, which integrates LiDAR-Inertial Odometry (LIO), a graph-based SLAM system, and a Voronoi heuristic clustering algorithm. The researchers developed a “Tree Manager” module that enables real-time tree segmentation, reconstruction, and extraction of tree features such as Diameter at Breast Height (DBH) and trunk curvature. Experiments were conducted in coniferous, broadleaf, and mixed forests in the UK, Switzerland, and Finland to validate the method’s accuracy. The results showed a tree detection recall rate of 98.1%, with an average RMSE of 1.93 cm for DBH measurements, a bias of 0.65 cm, and a standard deviation of 1.81 cm. The average RMSE for trunk curvature was 8.05 cm [43]. The system produces complete results without requiring post-processing and demonstrates real-time reconstruction capability, achieving a good balance between accuracy and efficiency. The point cloud generated by the laser scanning method for 3D reconstruction is shown in Figure 2.
The main advantage of the laser scanning method is its high accuracy and ability to cover large areas. This technology is not affected by lighting conditions, making it effective for use at night or in low-light environments [44]. However, laser scanning can be problematic when encountering transparent or highly reflective surfaces, as these surfaces can interfere with the laser reflection, compromising data accuracy [45]. Additionally, the high cost of laser scanning equipment and its expensive maintenance remain major constraints to its widespread application [46].

2.1.4. Wavelength Selection in Active 3D Reconstruction

In active 3D reconstruction, the choice of emitted wavelength is a critical factor influencing system performance, particularly when capturing plant surfaces with varying reflectance and translucency. Structured light and Time-of-Flight (TOF) systems commonly operate in the visible spectrum (e.g., red, green, blue) or the near-infrared range (NIR). Among these, NIR wavelengths—typically around 850 nm or 940 nm—are widely preferred for plant-related applications due to their lower sensitivity to ambient light fluctuations and improved penetration into plant tissues. These wavelengths correspond well with the high reflectance exhibited by healthy foliage in the NIR region, enabling more reliable surface detection and depth estimation [47]. Moreover, NIR light reduces the influence of plant pigments, enhancing structural contrast even in dense canopies.
For targeted agricultural tasks, avoiding wavelengths with strong absorption by water or pigments can further improve the signal-to-noise ratio. For instance, bypassing chlorophyll absorption bands (such as around 670 nm) helps ensure stronger signal return for depth sensing. In dynamic lighting environments—such as greenhouses or open fields—narrow-band infrared sources combined with optical filtering can greatly enhance system stability and reconstruction quality [48]. Ultimately, tailoring wavelength selection to the spectral properties of plant materials is vital for achieving high-precision reconstruction, particularly when resolving fine structures like leaves, stems, and fruit surfaces.

2.2. Passive Three-Dimensional Reconstruction Techniques

Passive 3D reconstruction techniques refer to methods for recovering the 3D structure of an object by analyzing image sequences captured from different viewpoints, without relying on an external emission source, but rather using natural light or existing ambient lighting conditions [49,50]. These techniques offer unique advantages in plant research.

2.2.1. Stereo Vision Method

The stereo vision method is a passive vision 3D reconstruction technique that simulates how the human eye acquires depth information to reconstruct 3D structures. The basic principle involves analyzing two or more images of the same scene captured from different angles and deducing the depth information of each point in the scene by accurately calculating the parallax [51]. The core idea is to obtain the 3D structural information of the object by utilizing the correspondence between images captured from different viewpoints.
A stereo vision system typically consists of two or more cameras that capture the same scene with a fixed baseline (i.e., the distance between the cameras), as shown in Figure 3. The images captured by each camera provide different views of the scene, and the parallax (the difference in the position of the same scene point in different images) becomes the key to calculating depth information. Using the principle of triangulation, the distance from a point on the object’s surface to the camera can be calculated from the parallax, completing the 3D reconstruction [52].
In plant science research and agricultural applications, stereo vision technology is widely used across various fields such as plant growth monitoring, disease identification, and crop phenotypic analysis [53]. By analyzing plant images taken from different angles, researchers can measure key parameters such as plant growth rates, leaf size, and fruit volume, which are critical for crop breeding and health management [54]. For instance, Sébastien Dandrifosse et al. used the stereo vision method to measure the morphological characteristics of winter wheat, including Leaf Area Index (LAI), Mean Leaf Angle (MLA), Leaf Angle Distribution (LAD), and canopy height. The experiment was conducted in a winter wheat field in Lonzée, Belgium, where two cameras were used to capture overhead images. Image segmentation was performed using both RGB and HSV color spaces, and the Support Vector Machine (SVM) algorithm was employed for soil and plant segmentation to calculate the plant’s morphological characteristics. By adjusting the camera and image resolution during the experiment, the team improved the accuracy of image matching and depth computation, ultimately achieving 99.8% accuracy in segmenting soil, leaves, and spikes. In comparison, the stereo vision method achieved 97.1% accuracy when measuring canopy height, with an RMSE of 0.016 m. The estimation error for the Leaf Area Index (LAI) was RMSE = 0.37 [55].
The primary advantages of the stereo vision method are its relatively simple equipment, low cost, and independence from specific lighting conditions, making it ideal for use in natural environments [56]. Additionally, this technique allows for rapid 3D data collection without physical contact with the plant, thus avoiding any potential physical damage to the plant [57]. However, the stereo vision method also has some limitations. For example, it requires high precision in image alignment, which can lead to errors when dealing with occlusions or scenes with repetitive textures. Moreover, parallax calculation requires advanced algorithms, often involving complex image processing techniques to ensure the accuracy of the data [58].

2.2.2. Structure from Motion (SfM)

Structure from Motion (SfM) is a passive visual 3D reconstruction technique that analyzes a series of images taken from different viewpoints to recover the camera trajectory and reconstruct the 3D structure of a scene. It integrates photogrammetry and computer vision techniques [59], enabling accurate 3D reconstruction without the need for specialized equipment.
One of the key advantages of SfM technology is its low hardware requirements, typically needing only one or more standard cameras to operate [60]. The workflow consists primarily of feature detection, feature matching, motion estimation, triangulation, and bundle adjustment. By tracking and matching common feature points in continuous or semi-continuous image sequences, the SfM algorithm estimates the positions of these feature points in 3D space, as well as the positions and orientations of the cameras during image capture, ultimately generating a dense 3D point cloud model of the scene [61].
In botany and agricultural research, SfM is widely used for 3D modeling across various plant growth stages, from seedlings to maturity [62]. These 3D models not only help researchers analyze plant growth dynamics and morphological changes, but can also be used to assess crop health and predict yields [63]. By analyzing a 3D model of a crop, agricultural experts can more accurately identify early signs of diseases and optimize fertilization and irrigation strategies based on these data, thereby improving crop productivity.
The main advantages of SfM are its flexibility and low cost. As the technology utilizes existing common camera equipment without the need for additional specialized devices, it is particularly well-suited for research environments with limited resources. Additionally, SfM can handle large-scale scenes and generate high-resolution 3D data [64]. However, SfM may experience errors in feature matching when dealing with scenes that have sparse feature points or repetitive textures, which can lead to a decrease in reconstruction accuracy. Furthermore, SfM requires high image quality; factors such as image blur or improper exposure can negatively impact the final 3D reconstruction results. Cao Xin et al. proposed a 3D reconstruction method for field-grown soybean plants that combines Structure from Motion (SfM) with Instant Neural Graphics Primitives (Instant-NGP), referred to as SfM-INGP, aiming to address the high equipment costs and long reconstruction times of traditional field reconstruction approaches. They used consumer-grade smartphones to capture multi-view videos, applied motion detection for adaptive frame extraction to generate image sequences, recovered camera poses using SfM, and performed efficient 3D reconstruction via Instant-NGP. Field experiments were conducted on 56 soybean plants, with 12 plants used for method validation. The results showed that the SfM-INGP method achieved an average reconstruction time of only 2.82 min, representing reductions of 90.7% and 99.4% compared to MVS and NeRF, respectively. The average Peak Signal-to-Noise Ratio (PSNR) reached 24.47 dB, which was 15.4% and 9.3% higher than MVS and NeRF, respectively. The reconstructed Mean Squared Error (MSE) was only 0.15, significantly lower than that of MVS (0.46) and NeRF (0.37) [65].

2.3. Deep Learning-Based 3D Reconstruction Technology

Deep learning-based 3D reconstruction techniques are methods that utilize deep neural networks to learn and infer 3D structural information from 2D images.

2.3.1. Neural Radiance Field (NeRF)

Neural Radiance Field (NeRF) is a deep learning-based implicit representation method for 3D scenes, capable of reconstructing highly realistic 3D environments from dense and comprehensively covered multi-view images [66]. NeRF employs a deep neural network to model the color and volume density of each 3D point in the scene and utilizes a differentiable volumetric rendering algorithm to integrate along the camera’s light rays, generating photo-realistic renderings from novel viewpoints. The core innovation of NeRF lies in its high-frequency positional encoding, which enhances detail representation and enables the implicit modeling of complex geometry and appearance [67].
The NeRF model implicitly represents a 3D scene as a five-dimensional radiance field using a multilayer perceptron (MLP). It takes high-frequency encoded 3D coordinates and 2D viewing directions as input, and outputs the volume density and color at each point. Based on known camera parameters, 3D points are hierarchically sampled along light rays, and the final image is synthesized through differentiable volumetric rendering. The MLP is trained by minimizing the mean squared error (MSE) between the rendered and ground truth images, enabling high-fidelity reconstruction of complex geometries and lighting conditions [66].
NeRF has shown great potential in 3D plant reconstruction, particularly in handling complex geometries and fine textures. It has been applied to capture detailed morphological features such as rice panicles and seedling vegetables, providing new tools for agricultural research [68]. This technique is well-suited for automated, accurate plant phenotyping, supporting optimized planting strategies and crop improvement. Zhu Lei et al. proposed a NeRF-based method for 3D modeling and phenotypic extraction of seedling-stage vegetables using smartphone-captured multi-view RGB images. The method reconstructs seedlings like pepper, tomato, and strawberry, followed by point cloud processing—such as line fitting, region growing, and triangulation—to extract traits including plant height, stem diameter, and leaf area. Compared to the SfM-MVS method, NeRF achieved higher reconstruction quality and improved efficiency by an average of 700%. For various plants, point-to-point RMSE ranged from 0.128 to 0.395 cm. For pepper seedlings, R2 values for plant height and stem diameter were 0.971 and 0.907, with RMSEs of 0.86 cm and 0.017 cm. Leaf area estimation yielded R2 values of 0.909–0.935 and RMSEs of 0.75–3.22 cm2, confirming the method’s accuracy and efficiency in seedling phenotyping [69].
The main advantage of NeRF is its ability to reconstruct highly detailed 3D scenes from limited data, without the need for a prior scene model or complex calibration process [70]. However, the major challenges of NeRF include its high demand for computational resources, especially during the training and rendering phases, and its long processing times, which limit its use in real-time applications. In addition, although NeRF performs well in static scenes, it still faces difficulties when dealing with dynamic and large-scale scenes [66].

2.3.2. 3D Reconstruction Method Based on Convolutional Neural Networks (CNNs)

Convolutional Neural Networks (CNNs), a fundamental deep learning architecture, are widely used in image analysis and have proven effective in 3D reconstruction tasks. By extracting hierarchical features from 2D images through multiple convolutional layers, CNNs can interpret complex visual information and infer 3D structures. As network depth increases, features evolve from low-level edges and textures to high-level representations, enabling accurate reconstruction beyond traditional image classification [71].
In the field of 3D plant reconstruction, CNNs are capable of predicting 3D models of plants from a single image or a small number of images. For example, researchers can use CNNs to analyze the morphology of plant leaves, automatically identifying and quantifying the size, shape, and disease conditions of the leaves [72]. CNNs are also used to monitor plant growth dynamics from time-series images, providing accurate 3D data to support plant growth processes and enabling better decision-making, such as optimizing irrigation and fertilization strategies.
The key advantage of CNNs in 3D reconstruction lies in their high degree of automation and sensitivity to complex visual features, making them well-suited for information-rich images. However, their performance heavily depends on the quantity and quality of training data. They also struggle with occlusion, background noise, and demand high computational resources and training time, which limits practical efficiency [73]. To address structure loss due to occlusion in cotton canopies, Yang Li et al. proposed a CNN-based 3D reconstruction method combining a Cascade Leaf Segmentation and Completion Network (CLSCN) with a Fragmental Leaf Point-cloud Reconstruction Algorithm (FLPRA). The method integrates an Instance Segmentation Network (ISN) and a Generative Adversarial Network (GAN) to reconstruct occluded leaves using RGB-D images and point clouds captured bya RealSense L515 camera (Intel Corporation, Santa Clara, CA, USA). The training set included 2720 images with over 12,000 labeled leaves. Experiments showed the CLSCN network achieved an average mIoU of 87.22% under varying occlusion types, while the GAN module exceeded 94% accuracy in leaf completion with Fréchet Inception Distance (FID) scores below 35. The final reconstructed canopy model reached an accuracy of 82.70%, demonstrating improved completeness and reliability [74].

2.3.3. 3D Gaussian Splatting (3DGS)

3D Gaussian Splatting (3DGS) is an advanced 3D reconstruction method that uses 3D Gaussian distributions to represent and reconstruct scenes. This technology combines the mathematical properties of Gaussian distributions with modern computer graphics and deep learning optimization techniques, enabling efficient and precise reconstruction of complex 3D structures [75].
3DGS represents objects in a 3D scene using a set of 3D Gaussian distributions, each of which has four main attributes: position, shape and size, color, and transparency. The position is represented by the mean of the Gaussian distribution, describing its location in 3D space; shape and size are represented through the covariance matrix; color is represented by spherical harmonic (SH) coefficients, which describe the color and texture of the object’s surface; transparency is controlled by the opacity value (α). Through these Gaussian distributions, 3DGS can efficiently describe and render objects in the scene, especially when reconstructing 3D models from 2D images. During the rendering process, initial Gaussian distributions are first derived from images or sparse point cloud data; then, each Gaussian is projected onto the 2D image plane through differentiable rendering, and its color and transparency are calculated. Optimization is performed through back-propagation. Using optimization algorithms such as gradient descent, the parameters of the Gaussian distributions are optimized by comparing the differences between the real and reconstructed images (e.g., L1 loss or SSIM loss), thus achieving high-quality 3D reconstruction [76].
In the field of agriculture, 3D Gaussian Splatting (3DGS) can be applied to crop growth monitoring, plant phenotype analysis, and precision agriculture. By utilizing 3DGS to reconstruct 3D models of crops from 2D images taken from multiple perspectives, researchers can accurately extract various phenotypic features of plants, such as plant height, leaf area, branching distribution, and fruit count [77]. These high-precision 3D models help farmers and researchers better understand the growth state, health condition, and environmental adaptability of crops. Compared with traditional sensors (e.g., LiDAR), 3DGS overcomes the limitations in cost and precision, offering a more economical and efficient solution, especially for large-scale crop monitoring and automated agricultural systems. Lizhi Jiang and others used 3D Gaussian Splatting (3DGS) to successfully reconstruct high-fidelity 3D models of cotton plants and performed detailed 3D cotton boll segmentation. The research team captured images of cotton plants using a smartphone, generated sparse point clouds through photogrammetry, and optimized them into 3DGS models. They developed the Cotton3DGaussians workflow, which performs boll segmentation through multi-view images from four perspectives and removes redundant bolls through cross-view clustering. The study also compared the performance of YOLOv11x and the Segment Anything Model (SAM) in generating 2D instance masks. The results showed that YOLOv11x had an F1 score 5.9% higher than SAM. Using the 3DGS model, the researchers successfully estimated phenotypic features such as boll count, volume, plant height, and canopy width. Compared with LiDAR ground truth data, the model’s mean absolute percentage error (MAPE) for boll count was 9.23%, while the errors for canopy width, plant height, and boll volume were 3.66%, 2.38%, and 8.17%, respectively, demonstrating high accuracy [78].
Table 2 summarizes the above 3D reconstruction methods.
Table 3 summarizes the application examples of the above 3D reconstruction.

3. Exploration of Multi-Technology Fusion for 3D Reconstruction of Plants

3.1. Multi-Sensor Fusion Techniques

3.1.1. Sensor Types and Functions

Optical Sensors: Including RGB cameras and multispectral cameras. RGB cameras are used to capture the color information of plants, while multispectral cameras record the spectral characteristics of different bands (e.g., near-infrared and mid-infrared), which can support the precise discrimination of various plant parts such as leaves, stems, and fruits [80].
Depth Sensors: For example, depth cameras such as Kinect V2 can acquire 3D point cloud data of plants, providing highly accurate geometric morphological information. These data are particularly crucial for addressing the occlusion problem and can help reconstruct the complete structure of the plant [81].
Thermal Imaging Sensors: These sensors are used to generate images of the thermal distribution of plants, reflecting the temperature status. Such data are valuable for studying the physiological state and health levels of plants [82].

3.1.2. Fusion Methods

Data Fusion: This involves the direct integration of raw data from different sensors at the data layer. For instance, fusing RGB images with depth images to generate multispectral 3D images. This method effectively addresses the occlusion problem, while improving the integrity and accuracy of the data [83].
Feature Fusion: At the feature layer, features are extracted from the data of each sensor and then fused. For example, infrared image features are extracted by Convolutional Neural Networks (CNNs), and visible image features are processed using Transformer networks, followed by their unified integration. This method fully utilizes the complementary advantages of different data modalities and enhances feature expression [84].
Decision Fusion: At the decision layer, the outputs from different sensors are integrated. For example, by matching and integrating the target detection results generated by multiple sensors, more reliable and comprehensive target information is produced. This approach significantly enhances system robustness and reduces false positives and missed detections [85].

3.1.3. Advantages and Challenges of Multi-Sensor Fusion Techniques

(1)
Advantages
Enhanced Accuracy and Completeness of 3D Reconstruction: Multi-sensor fusion provides a rich set of information, significantly improving the effectiveness of 3D reconstruction of plants. For instance, fusing RGB images with depth images can effectively solve the occlusion problem and generate complete point cloud models. Additionally, the physiological state information from thermal imaging further enhances the descriptive power of the models [86].
Improved System Robustness: The complementary nature of data from different sensors plays a key role in fusion. By integrating data from LiDAR and cameras, more accurate target detection results can be achieved, reducing the likelihood of missed detections and false positives.
(2)
Challenges
Calibration: Data from different sensors must be accurately aligned both temporally and spatially to ensure consistency and accuracy. For example, data from LiDAR and cameras need to be aligned to the same coordinate system [87]. This step is crucial for ensuring the accuracy of subsequent data processing.
Registration: Due to differences in data formats and sampling rates among sensors, appropriate registration is required. For example, data from millimeter-wave radar and LiDAR need to be aligned both temporally and spatially. To improve the efficiency and robustness of registration, a method for coarse registration of single-plant 3D structural point clouds was proposed, which improves registration efficiency by pre-processing point clouds and introducing initial pose constraints [88].
Weight Allocation: During fusion, it is necessary to allocate weights to each sensor’s data based on its reliability. The weighted average method is commonly used to assign weights according to the reliability of each sensor’s data. Recursive estimation algorithms, such as Kalman filtering, can also be applied to fuse uncertain data and dynamically adjust weights for ambiguity and conflicts, thereby improving fusion accuracy [89].
Despite continuous technological advancements, the widespread adoption of 3D reconstruction tools remains limited, particularly in small-scale agricultural settings. Key barriers include the high cost of specialized sensors, substantial computational requirements for running deep learning models, and the lack of accessible training resources for agricultural practitioners. Addressing these challenges will require the development of affordable hardware solutions, the deployment of edge computing technologies, and the design of user-friendly interfaces to lower the threshold for adoption.

3.2. Physical Modeling Combined with 3D Reconstruction

3.2.1. Plant Physical Modeling

(1)
Growth Model
The growth process of plants is influenced by a variety of factors, including light, temperature, water, and nutrients. The Logistic growth equation is a commonly used mathematical model that effectively controls the growth rate of plants, as well as the allocation of light and resources in each iteration. This equation automatically adjusts the final height of the tree and simulates the growth process of a single tree or multiple trees under the influence of competition [90].
(2)
Mechanical Model
The mechanical properties of plants are mainly manifested in their structural stability and deformation behavior. The mass-spring system is a commonly used mechanical model that can effectively simulate the morphology and dynamic changes of plant leaves. For example, motion capture technology can be used to obtain motion data of corn leaves, construct the mass–spring model, and calculate the model parameters using the least squares method to simulate the dynamic behavior of corn leaves [91].

3.2.2. Combined Method and Effect

(1)
Combination Method
Data-driven and Physical Model Combination: This method combines 3D real data with plant growth rules, using actual data to guide the process-based modeling of plant growth. For example, Lei Yi et al. [92] proposed a data-driven tree modeling method, assuming that trees have multi-level representations, including main trunks, branches, thin branches, and leaves. By scanning unilateral data using a 3D laser, and combining it with plant growth rules, the missing parts of the branches and trunks are automatically filled using a mechanism based on grammatical rules.
Deep Learning and Physical Modeling Combined: In this approach, deep learning techniques are used to complete incomplete point cloud data, which are then combined with physical models for 3D reconstruction. Haibo Chen et al. [93] proposed a 3D reconstruction method for plants based on a deep camera and deep learning point cloud completion. This method uses deep learning techniques to complement incomplete leaf point cloud data captured by RGB-D cameras under occlusion conditions.
(2)
Effectiveness Enhancement
Biological Plausibility: 3D reconstruction methods that integrate physical models significantly improve the biological plausibility of the reconstruction results. For example, by combining data-driven methods with physical modeling, the reconstructed model not only achieves high accuracy (with minimal error compared to real data) but also has higher completeness, aligning with botanical growth patterns [94].
Completeness and Accuracy: The combination of deep learning and physical modeling shows significant advantages in handling occlusion and missing data issues. Experimental results demonstrate that this method has made notable progress in enhancing the completeness of plant point cloud data, especially in the presence of multiple occlusions or high missing data percentages. It also shows more uniform completeness and high effectiveness in leaf area estimation [93].

3.3. Dynamic Scene Reconstruction Techniques

3.3.1. Impact of Dynamic Factors and Response Strategies

(1)
Wind Impact and Coping Strategies
The impact of wind causes plants to oscillate in space, leading to continuous changes in their form and position. This dynamic change is especially significant at high wind speeds. This requires the reconstruction system to continuously capture and compensate for plant movement to avoid reconstruction errors. To address this challenge, high-speed photography techniques can be employed to reduce motion blur by providing higher temporal resolution data, thus improving reconstruction accuracy in dynamic scenes [95]. Additionally, Event Cameras (ECs), due to their ultra-high temporal accuracy and ability to record light changes in real time without relying on traditional frame rate systems, are particularly suitable for capturing the rapid dynamic changes in plants under the influence of wind and other dynamic factors [96].
(2)
Impact of Plant Growth and Motion and Coping Strategies
During plant growth, branches, leaves, and other parts undergo continuous morphological changes due to factors such as light and gravity. These slow dynamic changes can affect the accuracy of 3D reconstruction. To cope with this challenge, plant images and depth data can be collected at regular intervals to capture dynamic changes at different growth stages [97]. In addition, combining data from different time scales and gradually updating the 3D model can reflect the plant’s growth process more accurately, thereby improving the accuracy and biological plausibility of the reconstruction [26].

3.3.2. Applications of High-Speed Cameras, Event Cameras, and Other Equipment

(1)
High-Speed Camera Equipment
High-speed camera equipment provides high frame-rate image sequences that can capture the rapid dynamic changes in plants, especially the oscillation of leaves and branches caused by wind. By recording high frame-rate images, these devices reduce motion blur and provide clearer, more accurate image data, significantly improving the accuracy of subsequent 3D reconstruction [98].
(2)
Event Cameras
Event Cameras detect brightness changes through independent pixels and can capture real-time information on plant movement, generating a stream of events with temporal information. These streaming data record dynamic changes at high temporal resolution, making it ideal for dynamic scene reconstruction. Event cameras are particularly suited for capturing rapid dynamic changes in plants during wind or growth, excelling in both high accuracy and real-time performance [99].
(3)
Algorithm Applications
Deep learning-based algorithms, such as Convolutional Neural Networks (CNNs), can be used to process event stream data, extract key features, and perform 3D reconstruction. For example, the Pix2Vox framework is capable of handling both single- and multi-view 3D reconstruction, generating higher-quality 3D models in dynamic scenes through the use of encoder, decoder, and refiner modules [100].

3.3.3. Time Dimension Information Processing

Capturing dynamic changes is at the core of dynamic scene reconstruction and accurately capturing temporal information is crucial under the influence of plant growth and environmental factors such as wind. Event cameras, by working independently with each pixel and only recording information when brightness changes, effectively avoid the loss of crucial information caused by motion blur in traditional cameras [96]. Their ultra-high temporal resolution enables real-time capture of dynamic plant changes, providing critical temporal information for 3D reconstruction. The EventPS method, through the use of event cameras, efficiently captures light changes and performs real-time surface normal estimation, accurately capturing the dynamic morphology of the plant surface over time, and solving the motion blur and temporal delay issues present in traditional methods [101].
In 3D reconstruction, processing time information is essential, especially in dynamic scenes where plant morphology and structure change over time. Temporal information can serve as the fourth dimension (4D), which, when combined with 3D models, generates dynamic point cloud data. Taking the DymSLAM system as an example, it can reconstruct a 4D point cloud of a dynamic scene through multi-motion segmentation and optimization techniques, handling dynamic changes in moving targets and further improving the accuracy and biological plausibility of the reconstructed model (DymSLAM: 4D Dynamic Scene Reconstruction Based on Geometrical Motion Segmentation). Another deep learning-based technology, NDR (Neural-Dynamic Reconstruction), can generate more accurate dynamic scene geometries by capturing time-varying information and addressing plant growth and wind movement factors in dynamic environments, resulting in 3D models that are closer to reality [79].

4. Future Perspectives

4.1. Technology Convergence and Innovative Applications

Future developments in 3D reconstruction are expected to move toward deeper integration with emerging technologies such as mixed reality (MR), augmented reality (AR), and virtual simulation platforms. These advances will enable more intuitive visualization and interaction for applications including crop monitoring, spatial analysis, and agricultural decision-making. In parallel, the maturation of unmanned aerial vehicles (UAVs), mobile robotic systems, and ground-based autonomous platforms will facilitate more flexible and real-time acquisition of plant structural data under varying field conditions. The lightweight deployment of reconstruction algorithms on embedded and edge computing devices will further enhance their applicability in environments with limited computational resources.

4.2. Intelligence and Automation

Three-dimensional reconstruction is poised to serve as a foundational component of intelligent agricultural systems. When combined with deep learning models, vision transformers, or large pre-trained vision-language architectures such as CLIP and SAM, 3D data can significantly enhance the robustness of feature extraction, semantic segmentation, and instance-level phenotyping across diverse crop species and growth stages [102]. Looking ahead, autonomous field robots equipped with these capabilities may dynamically perform complex tasks such as disease detection, growth monitoring, and adaptive resource application. Moreover, integrating 3D models with AI-driven forecasting tools holds promise for early anomaly detection and more informed agronomic management strategies.

4.3. Popularization and Application of Precision Agriculture

As precision agriculture continues to expand, 3D reconstruction will play an increasingly pivotal role in supporting site-specific operations such as precision planting, variable-rate fertilization, and targeted micro-irrigation [4]. These applications demand highly accurate spatial information at both canopy and sub-canopy levels. High-resolution reconstruction techniques—such as Neural Radiance Fields (NeRF) and 3D Gaussian Splatting (3DGS)—enable fine-scale morphological analysis that complements traditional 2D imaging and remote sensing approaches. When linked with physiological measurements or plant growth models, reconstructed 3D representations can support more efficient resource allocation, higher crop yields, and reduced environmental impact.

4.4. Data Integration and Decision Support Through Cross-Domain Fusion

Future 3D reconstruction frameworks are likely to be increasingly integrated with big data platforms, cloud computing, and geographic information systems (GISs), enabling scalable storage, processing, and spatiotemporal visualization. By dynamically fusing data from satellites and UAV-based remote sensing systems, multi-resolution and multi-modal crop monitoring pipelines can be established. Additionally, coupling 3D reconstructions with plant growth simulation and agro-ecological models will facilitate 4D modeling—capturing temporal changes in plant structure and physiology. These advancements will enhance agricultural decision support systems by providing actionable insights for land use optimization, stress diagnostics, and yield forecasting [103].

5. Summary

Plant 3D reconstruction has become a pivotal technology in supporting precision agriculture and plant phenotyping, with expanding applications in crop growth assessment, intelligent pest and disease detection, and the automation of farm operations. This review presents a comprehensive overview of three principal technological paradigms—active vision, passive vision, and deep learning—alongside widely adopted methods including structured light, time-of-flight (ToF), laser scanning, stereo vision, Structure from Motion (SfM), Neural Radiance Field (NeRF), Convolutional Neural Networks (CNNs), and 3D Gaussian Splatting (3DGS). Each paradigm offers distinct advantages: active vision provides high accuracy and resilience to lighting variability; passive vision stands out for its flexibility and cost efficiency; while deep learning-based approaches demonstrate strong capabilities in modeling complex structures and extracting fine-grained features, positioning them as a growing force in the development of high-fidelity reconstruction technologies.
In real-world agricultural settings, the choice of reconstruction method must be tailored to specific application scenarios. Structured light and ToF systems are particularly effective in controlled environments such as automated greenhouses, where their precision and stability can be fully leveraged. SfM and stereo vision are more suitable for open-field conditions, offering a balance between affordability and adaptability to natural illumination. Deep learning approaches, such as NeRF and 3DGS, are especially well suited for capturing intricate morphological details in static or semi-controlled environments. Meanwhile, LiDAR continues to be the preferred option for large-scale mapping tasks in forestry and orchard management, where dense canopy reconstruction is essential. Key factors—including deployment scale, ambient lighting, sensor mobility, and the required reconstruction resolution—must be carefully considered when selecting a technique.
Beyond the classification of technologies, this paper emphasizes the rising importance of multi-source data fusion. This includes integrating multiple sensors, coupling with plant physiological models, and designing reconstruction pipelines that account for dynamic variables such as wind interference and growth variability. Emerging technologies—such as event-based cameras and high-speed imaging systems—offer new potential for rapid, temporally resolved 3D capture. Looking forward, future research should focus on improving model robustness in complex and variable agricultural environments, enabling real-time processing on embedded platforms, and advancing unified systems that combine 3D modeling with intelligent analysis to support data-driven decision-making in smart agriculture.

Funding

This research was funded by the Agricultural Production Process Digital Twin Platform Construction and Display project, grant number TYGC25-42. The APC was funded by Shanxi Agricultural University.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study, in the collection, analyses, or interpretation of data, in the writing of the manuscript, or in the decision to publish the results.

References

  1. Ju, J.; Zhang, M.; Zhang, Y.; Chen, Q.; Gao, Y.; Yu, Y.; Wu, Z.; Hu, Y.; Liu, X.; Song, J.; et al. Development of Lettuce Growth Monitoring Model Based on Three-Dimensional Reconstruction Technology. Agronomy 2024, 15, 29. [Google Scholar] [CrossRef]
  2. Arshad, M.A.; Jubery, T.; Afful, J.; Jignasu, A.; Balu, A.; Ganapathysubramanian, B.; Sarkar, S.; Krishnamurthy, A. Evaluating Neural Radiance Fields for 3D Plant Geometry Reconstruction in Field Conditions. Plant Phenomics 2024, 6, 0235. [Google Scholar] [CrossRef]
  3. Yang, D.; Yang, H.; Liu, D.; Wang, X. Research on automatic 3D reconstruction of plant phenotype based on Multi-View images. Comput. Electron. Agric. 2024, 220, 108866. [Google Scholar] [CrossRef]
  4. Gu, W.; Wen, W.; Wu, S.; Zheng, C.; Lu, X.; Chang, W.; Xiao, P.; Guo, X. 3D reconstruction of wheat plants by integrating point cloud data and virtual design optimization. Agriculture 2024, 14, 391. [Google Scholar] [CrossRef]
  5. Zhang, H.; Wang, L.; Jin, X.; Bian, L.; Ge, Y. High-throughput phenotyping of plant leaf morphological, physiological, and biochemical traits on multiple scales using optical sensing. Crop J. 2023, 11, 1303–1318. [Google Scholar] [CrossRef]
  6. Miao, Y.; Wang, L.; Peng, C.; Li, H.; Li, X.; Zhang, M. Banana plant counting and morphological parameters measurement based on terrestrial laser scanning. Plant Methods 2022, 18, 66. [Google Scholar] [CrossRef]
  7. Stausberg, L.; Jost, B.; Klingbeil, L.; Kuhlmann, H. A 3D Surface Reconstruction Pipeline for Plant Phenotyping. Remote Sens. 2024, 16, 4720. [Google Scholar] [CrossRef]
  8. Guan, H.; Liu, M.; Ma, X.; Yu, S. Three-dimensional reconstruction of soybean canopies using multisource imaging for phenotyping analysis. Remote Sens. 2018, 10, 1206. [Google Scholar] [CrossRef]
  9. Hu, K.; Ying, W.; Pan, Y.; Kang, H.; Chen, C. High-fidelity 3D reconstruction of plants using Neural Radiance Fields. Comput. Electron. Agric. 2024, 220, 108848. [Google Scholar] [CrossRef]
  10. Chen, H.; Liu, S.; Wang, L.; Wang, C.; Xiang, X.; Zhao, Y.; Lan, Y. Automated 3D Reconstruction and Verification of Single Crop Plants Based on Kinect V3. J. Agric. Eng. 2022, 38, 215–223. [Google Scholar]
  11. Gao, J.; Chen, L.; Shen, Q.; Cao, X.; Yao, Y. Progress in Differentiable Rendering Based on 3D Gaussian Sputtering Technology. Prog. Laser Optoelectron. 2024, 61, 1–23. [Google Scholar] [CrossRef]
  12. Cui, B.; Tao, W.; Zhao, H. High-precision 3D reconstruction for small-to-medium-sized objects utilizing line-structured light scanning: A review. Remote Sens. 2021, 13, 4457. [Google Scholar] [CrossRef]
  13. Zhu, J.; Zhai, R.; Ren, H.; Xie, K.; Du, A.; He, X.; Cui, C.; Wang, Y.; Ye, J.; Wang, J.; et al. Crops3D: A diverse 3D crop dataset for realistic perception and segmentation toward agricultural applications. Sci. Data 2024, 11, 1438. [Google Scholar] [CrossRef]
  14. Yadav, N.; Sidana, N. Precision Agriculture Technologies: Analysing the Use of Advanced Technologies, Such as Drones, Sensors, and GPS, in Precision Agriculture for Optimizing Resource Management, Crop Monitoring, and Yield Prediction. J. Adv. Zool. 2023, 44, 255. [Google Scholar]
  15. Choy, C.B.; Xu, D.; Gwak, J.; Chen, K.; Savarese, S. 3d-r2n2: A unified approach for single and multi-view 3d object reconstruction. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Part VIII 14, pp. 628–644. [Google Scholar] [CrossRef]
  16. Xiao, S.; Fei, S.; Ye, Y.; Xu, D.; Xie, Z.; Bi, K.; Guo, Y.; Li, B.; Zhang, R.; Ma, Y. 3D reconstruction and characterization of cotton bolls in situ based on UAV technology. ISPRS J. Photogramm. Remote Sens. 2024, 209, 101–116. [Google Scholar] [CrossRef]
  17. Harandi, N.; Vandenberghe, B.; Vankerschaver, J.; Depuydt, S.; Van Messem, A. How to make sense of 3D representations for plant phenotyping: A compendium of processing and analysis techniques. Plant Methods 2023, 19, 60. [Google Scholar] [CrossRef] [PubMed]
  18. Nasiri, A.; Omid, M.; Taheri-Garavand, A.; Jafari, A. Deep learning-based precision agriculture through weed recognition in sugar beet fields. Sustain. Comput. Inform. Syst. 2022, 35, 100759. [Google Scholar] [CrossRef]
  19. Jiang, Y.; Li, C.; Paterson, A.H. High throughput phenotyping of cotton plant height using depth images under field conditions. Comput. Electron. Agric. 2016, 130, 57–68. [Google Scholar] [CrossRef]
  20. Yang, Y.; Wang, X.; Ren, M. Data Fusion Method for Photometric-Structured Light Measurement on Metal Surfaces Based on Multilayer Perceptron. Acta Opt. Sin. 2023, 43, 200–211. [Google Scholar] [CrossRef]
  21. Meza, J.; Vargas, R.; Romero, L.A.; Zhang, S.; Marrugo, A.G. What is the best triangulation approach for a structured light system? In Proceedings of the Dimensional Optical Metrology and Inspection for Practical Applications IX, Online, 27 April–9 May 2020; Volume 11397, pp. 65–74. [Google Scholar] [CrossRef]
  22. Huang, H.; Liu, G.; Deng, L.; Song, T.; Qin, F. Multi-line laser 3D reconstruction method based on spatial quadric surface and geometric estimation. Sci. Rep. 2024, 14, 23589. [Google Scholar] [CrossRef]
  23. Geng, J. Structured-light 3D surface imaging: A tutorial. Adv. Opt. Photonics 2011, 3, 128–160. [Google Scholar] [CrossRef]
  24. Zhou, W.; Jia, Y.; Fan, L.; Fan, G.; Lu, F. A MEMS-based real-time structured light 3-D measuring architecture on FPGA. J. Real-Time Image Process. 2024, 21, 98. [Google Scholar] [CrossRef]
  25. Feng, Q.; Cheng, W.; Yang, Q.; Xun, Y.; Wang, X. Research on Recognition and Localization of Overlapping Tomato Fruits Based on Line Structured Light Vision. J. China Agric. Univ. 2015, 20, 100–106. [Google Scholar]
  26. Lv, H. Research on 3D Reconstruction and Segmentation of Plants Based on Multi-View Images. Master’s Thesis, Chongqing Jiaotong University, Chongqing, China, 2022. [Google Scholar]
  27. Chen, X.; Peng, Y.; Yu, S.; Hu, D. Measurement of Shape Index of Spherical Fruits Based on Binocular Structured Light 3D Reconstruction. Trans. Chin. Soc. Agric. Eng. 2024, 40, 187–194. [Google Scholar]
  28. Vázquez-Arellano, M.; Reiser, D.; Paraforos, D.S.; Garrido-Izard, M.; Burce, M.E.C.; Griepentrog, H.W. 3-D reconstruction of maize plants using a time-of-flight camera. Comput. Electron. Agric. 2018, 145, 235–247. [Google Scholar] [CrossRef]
  29. Si, X. Application of Depth Image Processing in Precision Agriculture. Master’s Thesis, University of Science and Technology of China, Hefei, China, 2017. [Google Scholar]
  30. Foix, S.; Alenya, G.; Torras, C. Lock-in time-of-flight (ToF) cameras: A survey. IEEE Sens. J. 2011, 11, 1917–1926. [Google Scholar] [CrossRef]
  31. Hansard, M.; Lee, S.; Choi, O.; Horaud, R.P. Time-of-Flight Cameras: Principles, Methods and Applications; Springer Science & Business Media: London, UK, 2012. [Google Scholar]
  32. Kazmi, W.; Foix, S.; Alenyà, G.; Andersen, H.J. Indoor and outdoor depth imaging of leaves with time-of-flight and stereo vision sensors: Analysis and comparison. ISPRS J. Photogramm. Remote Sens. 2014, 88, 128–146. [Google Scholar] [CrossRef]
  33. Xia, C.; Shi, Y.; Yin, W. 3D Plant Point Cloud Data Acquisition and Denoising Using TOF Depth Sensing. Trans. Chin. Soc. Agric. Eng. 2018, 34, 168–174. [Google Scholar]
  34. Yang, C.; Kang, J.; Eom, D.S. Enhancing ToF Sensor Precision Using 3D Models and Simulation for Vision Inspection in Industrial Mobile Robots. Appl. Sci. 2024, 14, 4595. [Google Scholar] [CrossRef]
  35. Xu, A.; Rao, L.; Fan, G.; Chen, N. Fast and high accuracy 3D point cloud registration for automatic reconstruction from laser scanning data. IEEE Access 2023, 11, 42497–42509. [Google Scholar] [CrossRef]
  36. Ebrahim, M.A.B. 3D laser scanners’ techniques overview. Int. J. Sci. Res. 2015, 4, 323–331. [Google Scholar]
  37. Farhan, S.M.; Yin, J.; Chen, Z.; Memon, M.S. A comprehensive review of LiDAR applications in crop management for precision agriculture. Sensors 2024, 24, 5409. [Google Scholar] [CrossRef] [PubMed]
  38. Muralikrishnan, B. Performance evaluation of terrestrial laser scanners—A review. Meas. Sci. Technol. 2021, 32, 072001. [Google Scholar] [CrossRef]
  39. Suresh, A.; Mathew, A.; Dhanish, P. Factors influencing the measurement using 3D laser scanner: A designed experimental study. J. Mech. Sci. Technol. 2024, 38, 5605–5615. [Google Scholar] [CrossRef]
  40. Lau, L.; Quan, Y.; Wan, J.; Zhou, N.; Wen, C.; Qian, N.; Jing, F. An autonomous ultra-wide band-based attitude and position determination technique for indoor mobile laser scanning. ISPRS Int. J. Geo-Inf. 2018, 7, 155. [Google Scholar] [CrossRef]
  41. Cao, L.; Coops, N.C.; Sun, Y.; Ruan, H.; Wang, G.; Dai, J.; She, G. Estimating canopy structure and biomass in bamboo forests using airborne LiDAR data. ISPRS J. Photogramm. Remote Sens. 2019, 148, 114–129. [Google Scholar] [CrossRef]
  42. Micheletto, M.J.; Chesñevar, C.I.; Santos, R. Methods and applications of 3D ground crop analysis using LiDAR technology: A survey. Sensors 2023, 23, 7212. [Google Scholar] [CrossRef]
  43. Freißmuth, L.; Mattamala, M.; Chebrolu, N.; Schaefer, S.; Leutenegger, S.; Fallon, M. Online tree reconstruction and forest inventory on a mobile robotic system. In Proceedings of the 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Abu Dhabi, United Arab Emirates, 14–18 October 2024; pp. 11765–11772. [Google Scholar] [CrossRef]
  44. Debnath, S.; Paul, M.; Debnath, T. Applications of LiDAR in agriculture and future research directions. J. Imaging 2023, 9, 57. [Google Scholar] [CrossRef]
  45. Li, Y.; Zhao, X.; Schwertfeger, S. Detection and Utilization of Reflections in LiDAR Scans through Plane Optimization and Plane SLAM. Sensors 2024, 24, 4794. [Google Scholar] [CrossRef]
  46. Bleier, M.; Nüchter, A. Low-cost 3d laser scanning in air orwater using self-calibrating structured light. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 105–112. [Google Scholar] [CrossRef]
  47. Xie, D.; Lü, C.; Zu, M.; Cheng, H. Research Progress on Simulated Materials of Visible–Near Infrared Reflectance Spectra of Green Vegetation. Spectrosc. Spectr. Anal. 2021, 41, 1032–1038. [Google Scholar]
  48. Reddy, P.; Guthridge, K.M.; Panozzo, J.; Ludlow, E.J.; Spangenberg, G.C.; Rochfort, S.J. Near-infrared hyperspectral imaging pipelines for pasture seed quality evaluation: An overview. Sensors 2022, 22, 1981. [Google Scholar] [CrossRef] [PubMed]
  49. Golla, T.; Kneiphof, T.; Kuhlmann, H.; Weinmann, M.; Klein, R. Temporal upsampling of point cloud sequences by optimal transport for plant growth visualization. Comput. Graph. Forum 2020, 39, 167–179. [Google Scholar] [CrossRef]
  50. Ando, R.; Ozasa, Y.; Guo, W. Robust surface reconstruction of plant leaves from 3D point clouds. Plant Phenomics 2021. [Google Scholar] [CrossRef] [PubMed]
  51. Zhou, L.; Wu, G.; Zuo, Y.; Chen, X.; Hu, H. A comprehensive review of vision-based 3d reconstruction methods. Sensors 2024, 24, 2314. [Google Scholar] [CrossRef]
  52. Lin, H.Y.; Tsai, C.L.; Tran, V.L. Depth measurement based on stereo vision with integrated camera rotation. IEEE Trans. Instrum. Meas. 2021, 70, 1–10. [Google Scholar] [CrossRef]
  53. Hou, C.; Xu, J.; Tang, Y.; Zhuang, J.; Tan, Z.; Chen, W.; Wei, S.; Huang, H.; Fang, M. Detection and localization of citrus picking points based on binocular vision. Precis. Agric. 2024, 25, 2321–2355. [Google Scholar] [CrossRef]
  54. Tang, Y.; Zhou, H.; Wang, H.; Zhang, Y. Fruit detection and positioning technology for a Camellia oleifera C. Abel orchard based on improved YOLOv4-tiny model and binocular stereo vision. Expert Syst. Appl. 2023, 211, 118573. [Google Scholar] [CrossRef]
  55. Dandrifosse, S.; Bouvry, A.; Leemans, V.; Dumont, B.; Mercatoris, B. Imaging wheat canopy through stereo vision: Overcoming the challenges of the laboratory to field transition for morphological features extraction. Front. Plant Sci. 2020, 11, 96. [Google Scholar] [CrossRef]
  56. Ge, L.; Yang, Z.; Sun, Z.; Zhang, G.; Zhang, M.; Zhang, K.; Zhang, C.; Tan, Y.; Li, W. A method for broccoli seedling recognition in natural environment based on binocular stereo vision and Gaussian mixture model. Sensors 2019, 19, 1132. [Google Scholar] [CrossRef]
  57. Yang, X.; Zhong, J.; Lin, K.; Wu, J.; Chen, J.; Si, H. Progress in Binocular Stereo Vision Technology and Its Application in Smart Agriculture. Trans. Chin. Soc. Agric. Eng. 2025, 41, 27–39. [Google Scholar] [CrossRef]
  58. Bai, Y.; Zhang, B.; Xu, N.; Zhou, J.; Shi, J.; Diao, Z. Vision-based navigation and guidance for agricultural autonomous vehicles and robots: A review. Comput. Electron. Agric. 2023, 205, 107584. [Google Scholar] [CrossRef]
  59. Liang, X.; Zhou, F.; Chen, H.; Liang, B.; Xu, X.; Yang, W. 3D Reconstruction and Trait Extraction of Maize Plants Based on Motion Recovery Structure. J. Agric. Mach. 2020, 51, 209–219. [Google Scholar]
  60. Gao, L.; Zhao, Y.; Han, J.; Liu, H. Research on multi-view 3D reconstruction technology based on SFM. Sensors 2022, 22, 4366. [Google Scholar] [CrossRef] [PubMed]
  61. Zhao, Y.; He, S. Improvement of Feature Point Detection and Matching Methods in Incremental SFM. Laser J. 2020, 41, 59–66. [Google Scholar] [CrossRef]
  62. Wei, K.; Liu, S.; Chen, Q.; Huang, S.; Zhong, M.; Zhang, J.; Sun, H.; Wu, K.; Fan, S.; Ye, Z.; et al. Fast Multi-View 3D reconstruction of seedlings based on automatic viewpoint planning. Comput. Electron. Agric. 2024, 218, 108708. [Google Scholar] [CrossRef]
  63. Arief, M.A.A.; Nugroho, A.P.; Putro, A.W.; Sutiarso, L.; Cho, B.K.; Okayasu, T. Development and Application of a Low-Cost 3-Dimensional (3D) Reconstruction System Based on the Structure from Motion (SfM) Approach for Plant Phenotyping. J. Biosyst. Eng. 2024, 49, 326–336. [Google Scholar] [CrossRef]
  64. Brandolini, F.; Patrucco, G. Structure-from-motion (SFM) photogrammetry as a non-invasive methodology to digitalize historical documents: A highly flexible and low-cost approach? Heritage 2019, 2, 2124–2136. [Google Scholar] [CrossRef]
  65. Xin, C.; Xujia, Q.; Xiaogang, X. 3D Reconstruction Method for Field-grown Soybean Plants Based on SfM and Instant-NGP. Trans. Chin. Soc. Agric. Eng. 2025, 41, 171–180. [Google Scholar]
  66. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2021, 65, 99–106. [Google Scholar] [CrossRef]
  67. Gao, K.; Gao, Y.; He, H.; Lu, D.; Xu, L.; Li, J. Nerf: Neural radiance field in 3d vision, a comprehensive review. arXiv 2022, arXiv:2210.00379. [Google Scholar]
  68. Wu, S.; Hu, C.; Tian, B.; Huang, Y.; Yang, S.; Li, S.; Xu, S. A 3D reconstruction platform for complex plants using OB-NeRF. Front. Plant Sci. 2025, 16, 1449626. [Google Scholar] [CrossRef] [PubMed]
  69. Zhu, L.; Jiang, W.; Sun, B.; Chai, M.; Li, S.; Ding, Y. 3D Modeling and Phenotypic Parameter Acquisition of Seedling Vegetables Based on Neural Radiance Fields. J. Agric. Mach. 2024, 55, 184–192. [Google Scholar] [CrossRef]
  70. Martin-Brualla, R.; Radwan, N.; Sajjadi, M.S.; Barron, J.T.; Dosovitskiy, A.; Duckworth, D. Nerf in the wild: Neural radiance fields for unconstrained photo collections. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7210–7219. [Google Scholar] [CrossRef]
  71. Flynn, J.; Neulander, I.; Philbin, J.; Snavely, N. Deepstereo: Learning to predict new views from the world’s imagery. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5515–5524. [Google Scholar] [CrossRef]
  72. Wang, J.; Zhen, Z.; Zhao, Y.; Ma, Y.; Zhao, Y. 3D-CNN with Multi-Scale Fusion for Tree Crown Segmentation and Species Classification. Remote Sens. 2024, 16, 4544. [Google Scholar] [CrossRef]
  73. Zhang, J.; Zheng, C. 3D Reconstruction Network from Single Image Based on Multi-scale CNN-RNN. Appl. Res. Comput. 2020, 37, 3487–3491. [Google Scholar] [CrossRef]
  74. Li, Y.; Si, S.; Liu, X.; Zou, L.; Wu, W.; Liu, X.; Zhang, L. Three-dimensional reconstruction of cotton plant with internal canopy occluded structure recovery. Comput. Electron. Agric. 2023, 215, 108370. [Google Scholar] [CrossRef]
  75. Hou, L.J.; Shen, Y.S.; Liu, X.C.; Chen, X.M.; Jia, R.C.; Fan, G.W.; Shen, C. 3D Gaussian Sputtering Super-Resolution Visual Scene Construction Algorithm. China Test 2024, 50, 13–20. [Google Scholar]
  76. Kerbl, B.; Kopanas, G.; Leimkühler, T.; Drettakis, G. 3d gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 2023, 42, 139. [Google Scholar] [CrossRef]
  77. Chen, J.; Li, Q.; Jiang, D. From Images to Loci: Applying 3D Deep Learning to Enable Multivariate and Multitemporal Digital Phenotyping and Mapping the Genetics Underlying Nitrogen Use Efficiency in Wheat. Plant Phenomics 2024, 6, 0270. [Google Scholar] [CrossRef]
  78. Jiang, L.; Sun, J.; Chee, P.W.; Li, C.; Fu, L. Cotton3DGaussians: Multiview 3D Gaussian Splatting for boll mapping and plant architecture analysis. Comput. Electron. Agric. 2025, 234, 110293. [Google Scholar] [CrossRef]
  79. Wang, H.; Wang, J.; Agapito, L. Morpheus: Neural dynamic 360deg surface reconstruction from monocular rgb-d video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 20965–20976. [Google Scholar] [CrossRef]
  80. Li, W.; Zhu, D.; Wang, Q. A single view leaf reconstruction method based on the fusion of ResNet and differentiable render in plant growth digital twin system. Comput. Electron. Agric. 2022, 193, 106712. [Google Scholar] [CrossRef]
  81. Wang, L.; Hu, Y.; Jiang, H.; Shi, W.; Ni, X. Monitor geomatical information of plant by reconstruction 3D model based on Kinect V2. In Proceedings of the 2018 ASABE Annual International Meeting. American Society of Agricultural and Biological Engineers, Detroit, MI, USA, 29 July–1 August 2018; p. 1. [Google Scholar] [CrossRef]
  82. Atencia Payares, L.K.; Gomez-del Campo, M.; Tarquis, A.M.; García, M. Thermal imaging from UAS for estimating crop water status in a Merlot vineyard in semi-arid conditions. Irrig. Sci. 2025, 43, 87–103. [Google Scholar] [CrossRef]
  83. Yu, S.; Liu, X.; Tan, Q.; Wang, Z.; Zhang, B. Sensors, systems and algorithms of 3D reconstruction for smart agriculture and precision farming: A review. Comput. Electron. Agric. 2024, 224, 109229. [Google Scholar] [CrossRef]
  84. Si, K.; Niu, C. Triple Multimodal Image Fusion Algorithm Based on Hybrid Differential Convolution and Efficient Visual Transformer Network. Infrared Laser Eng. 2024, 53, 331–345. [Google Scholar] [CrossRef]
  85. Tran, X.T.; Do, T.; Pal, N.R.; Jung, T.P.; Lin, C.T. Multimodal fusion for anticipating human decision performance. Sci. Rep. 2024, 14, 13217. [Google Scholar] [CrossRef]
  86. Xiang, S.; Zeng, Z.; Jiang, J.; Zhang, D.; Liu, N. An Efficient Dense Reconstruction Algorithm from LiDAR and Monocular Camera. Symmetry 2024, 16, 1496. [Google Scholar] [CrossRef]
  87. Cai, H.; Pang, W.; Chen, X.; Wang, Y.; Liang, H. A novel calibration board and experiments for 3D LiDAR and camera calibration. Sensors 2020, 20, 1130. [Google Scholar] [CrossRef]
  88. Lu, J.; Lan, Y.; Wu, Z.; Liang, X.; Chang, H.; Deng, X.; Wu, Z.; Tang, Y. Optimization of ICP Point Cloud Registration for Plant 3D Modeling. Trans. Chin. Soc. Agric. Eng. 2022, 38, 183–191. [Google Scholar]
  89. Wang, M.; Liu, H.; Tang, H.; Zhang, M.; Shen, X. IFNet: Data-driven multisensor estimate fusion with unknown correlation in sensor measurement noises. Inf. Fusion 2025, 115, 102750. [Google Scholar] [CrossRef]
  90. Sepaskhah, A.R.; Fahandezh-Saadi, S.; Zand-Parsa, S. Logistic model application for prediction of maize yield under water and nitrogen management. Agric. Water Manag. 2011, 99, 51–57. [Google Scholar] [CrossRef]
  91. Boxiang, X.; Hong, Q. Kinematic Analysis Method for Physical Modeling and Parameter Determination of Corn Leaf. J. Comput.-Aided Des. Comput. Graph. 2019, 31, 718–725. [Google Scholar]
  92. Yi, L.; Li, H.; Guo, J.; Deussen, O.; Zhang, X. Tree growth modelling constrained by growth equations. Comput. Graph. Forum 2018, 37, 239–253. [Google Scholar] [CrossRef]
  93. Chen, H.; Liu, S.; Wang, C.; Wang, C.; Gong, K.; Li, Y.; Lan, Y. Point cloud completion of plant leaves under occlusion conditions based on deep learning. Plant Phenomics 2023, 5, 0117. [Google Scholar] [CrossRef]
  94. Vinodkumar, P.K.; Karabulut, D.; Avots, E.; Ozcinar, C.; Anbarjafari, G. Deep learning for 3d reconstruction, augmentation, and registration: A review paper. Entropy 2024, 26, 235. [Google Scholar] [CrossRef]
  95. Burgess, A.J.; Retkute, R.; Preston, S.P.; Jensen, O.E.; Pound, M.P.; Pridmore, T.P.; Murchie, E.H. The 4-dimensional plant: Effects of wind-induced canopy movement on light fluctuations and photosynthesis. Front. Plant Sci. 2016, 7, 1392. [Google Scholar] [CrossRef]
  96. Yu, B.; Ren, J.; Han, J.; Wang, F.; Liang, J.; Shi, B. EventPS: Real-time photometric stereo using an event camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 9602–9611. [Google Scholar] [CrossRef]
  97. Zhu, R.; Sun, K.; Yan, Z.; Yan, X.; Yu, J.; Shi, J.; Hu, Z.; Jiang, H.; Xin, D.; Zhang, Z.; et al. Analysing the phenotype development of soybean plants using low-cost 3D reconstruction. Sci. Rep. 2020, 10, 7055. [Google Scholar] [CrossRef]
  98. Zhu, C.; Gao, Z.; Xue, W.; Tu, H.; Zhang, Q. High-speed deformation measurement with event-based cameras. Exp. Mech. 2023, 63, 987–994. [Google Scholar] [CrossRef]
  99. Nkrumah, I.; Moshrefizadeh, M.; Tahri, O.; Blasch, E.; Palaniappan, K.; AliAkbarpour, H. EC-WAMI: Event camera-based pose optimization in remote sensing and wide-area motion imagery. Sensors 2024, 24, 7493. [Google Scholar] [CrossRef] [PubMed]
  100. Xie, H.; Yao, H.; Sun, X.; Zhou, S.; Zhang, S. Pix2vox: Context-aware 3d reconstruction from single and multi-view images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2690–2698. [Google Scholar] [CrossRef]
  101. Elms, E.; Latif, Y.; Park, T.H.; Chin, T.J. Event-based structure-from-orbit. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 19541–19550. [Google Scholar] [CrossRef]
  102. Logeshwaran, J.; Srivastava, D.; Kumar, K.S.; Rex, M.J.; Al-Rasheed, A.; Getahun, M.; Soufiene, B.O. Improving crop production using an agro-deep learning framework in precision agriculture. BMC Bioinform. 2024, 25, 341. [Google Scholar] [CrossRef]
  103. Wang, Y.D.; Zhou, J.; Sun, J.W.; Wang, K.; Jiang, Z.Z.; Zhang, Z. Research on the Remote Operation Visualization System for Orchard Robots Based on Augmented Reality. Trans. Chin. Soc. Agric. Mach. 2023, 54, 22–31+78. [Google Scholar]
Figure 1. Schematic of a TOF system.
Figure 1. Schematic of a TOF system.
Sensors 25 03401 g001
Figure 2. The point cloud generated by the laser scanning method for 3D reconstruction.
Figure 2. The point cloud generated by the laser scanning method for 3D reconstruction.
Sensors 25 03401 g002
Figure 3. Stereoscopic vision method diagram.
Figure 3. Stereoscopic vision method diagram.
Sensors 25 03401 g003
Table 1. Technology evolution and application of 3D reconstruction methods.
Table 1. Technology evolution and application of 3D reconstruction methods.
TechnologyDevelopment PeriodMaturity/ApplicationKey Features and Applications
Structured LightOriginating in the 1980s, algorithm optimization and system implementation in the 1990s, and gradually maturing after the 2010sHigh Precision, Widely Adopted in Controlled EnvironmentsUsed for precise plant morphology modeling in controlled environments (e.g., greenhouses).
Laser ScanningOriginating in the 1960s, and with the development of computer technology since the 2000s, it has been widely applied in various domainsMature in Large-Scale ApplicationsApplied in large-scale environments (e.g., forestry, orchard), offering high precision and detailed canopy modeling.
ToF (Time-of-Flight)Started in the 1990s, entered commercial applications in the 2000s, and the development of high-performance compact sensors after 2010 has driven technological advancementsEmerging in Real-Time ApplicationsApplied in fast dynamic reconstruction for precision agriculture, with strong real-time capabilities.
Stereo VisionFrom the development of basic 3D reconstruction using stereo cameras in the 1980s, to the introduction of optimization algorithms in the 1990s, and significant improvements in matching accuracy and application performance in the 2010sCost-Effective, Well-EstablishedUsed in open-field crop monitoring under natural light, offering low-cost and flexible solutions.
SfM (Structure from Motion)Developed starting from the 1980s, optimization algorithms were introduced in the 1990s, and matching accuracy and application performance were significantly improved in the 2010sMature, Widely Used in Large-Scale Field StudiesApplied to large-scale plant phenotyping through multi-view image analysis, cost-effective and adaptable.
NeRF (Neural Radiance Fields)First proposed in 2020, widely applied in multiple fields, and still continuously developing and optimizingHigh Fidelity, Emerging in Controlled and Static EnvironmentsSuitable for reconstructing complex plant geometries and fine textures in static or controlled environments with large datasets.
CNN-based MethodsProposed in the 1980s, rapidly developed in the early 2010s, and matured and widely applied after 2020Rapidly Growing, High-AdaptabilityHighly automated and adaptable for various types of plants, increasingly driving high-accuracy 3D reconstruction in dynamic environments.
3D Gaussian Splatting (3DGS)First proposed in 2023, widely discussed in academia during 2023-2024, and widely applied across multiple fields since 2024Emerging, Real-Time RenderingEffective for generating high-quality 3D scenes in static environments, with relatively fast training speed and real-time rendering capabilities.
Table 2. Comparison of 3D reconstruction methods.
Table 2. Comparison of 3D reconstruction methods.
Technology TypeImaging TechnologyPrincipleAdvantagesDisadvantages
ActiveStructured Light MethodEncoded structured lightHigh precision, fast speed, resistant to ambient light interferenceDifficult reconstruction for highly reflective or transparent surfaces
Time-of-Flight (TOF) MethodReflection light time differenceCost-effective, simple structure, strong real-time performanceLimited resolution for small targets at close range
Laser Scanning MethodLaser rangingHigh precision, large coverage, non-contact measurementHigh equipment cost, expensive maintenance
PassiveStereo Vision MethodAnalyzing image disparityLow cost, mature technology, suitable for natural light conditionsHigh camera calibration requirements, accuracy affected by image alignment and synchronization
Structure from Motion (SfM)Recovering camera motion and 3D structure from multi-view imagesAllows 3D reconstruction from uncalibrated moving devices, strong adaptabilityRequires multiple perspective images, high image quality and quantity requirements
Deep LearningNeural Radiance Fields (NeRF)Deep learning-based light propagation simulationRealistic rendering effects, capable of handling complex lighting and detailsHigh computational resource demands, long training and rendering time
CNN-based MethodsConvolutional neural network extracts features and predicts 3D structuresHigh automation, strong adaptabilityRequires a large amount of annotated data, performance is scene-dependent
3D Gaussian Splatting (3DGS)Representing the scene as 3D Gaussian primitives, rendered through differentiable rasterizationHigh-quality, realistic scenes; fast, real-time rendering; relatively fast training speedHigh memory and disk usage; poor compatibility with existing rendering pipelines; supports only static scenes
Table 3. Application examples of 3D reconstruction.
Table 3. Application examples of 3D reconstruction.
Application ScenarioMethodAccuracy/ErrorReference
Shape measurement of spherical fruits (apple, citrus, pear)Binocular structured light + multi-view point cloud reconstructionMax diameter RMSE = 1.82 mm, Volume RMSE = 6.01 cm3, Deformation index R2 = 0.97[27]
Maize plant localizationToF camera + RANSAC + ICP registrationPosition RMSE = 3.4 cm, Std. dev. ±1.3 cm[28]
Tree trunk diameter reconstruction in forestsMobile LiDAR + Hough transform + SLAMDBH RMSE = 1.93 cm[79]
Wheat canopy structure extractionStereo vision + Semi-Global Block Matching (SGBM) + SVM segmentationCanopy height RMSE ≈ 1.1 cm, Leaf area RMSE = 0.37 cm2, Angle error < 1.5°[55]
Field soybean plant 3D modelingSfM + Instant-NGP3D model RMSE = 0.15[65]
Seedling vegetable phenotyping (e.g., chili, tomato)NeRF + Point Cloud Library (PCL) analysisPlant height RMSE = 0.86 cm, Stem diameter RMSE = 0.017 cm, Leaf area RMSE = 0.75∼3.22 cm2[69]
Cotton canopy point cloud reconstruction (with occlusion recovery)ISN + GAN + FLPRALeaf completion RMSE ≈ 0.395 cm, Overall reconstruction accuracy = 82.70%[74]
Cotton boll 3D reconstruction and trait analysisMulti-view image + 3D Gaussian Splatting + YOLOv11xBoll count RMSE ≈ 9.23%, Plant height RMSE ≈ 2.38%, Volume RMSE ≈ 8.17%[78]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Cui, Z.; Yang, J.; Wang, B. A Review of Optical-Based Three-Dimensional Reconstruction and Multi-Source Fusion for Plant Phenotyping. Sensors 2025, 25, 3401. https://doi.org/10.3390/s25113401

AMA Style

Li S, Cui Z, Yang J, Wang B. A Review of Optical-Based Three-Dimensional Reconstruction and Multi-Source Fusion for Plant Phenotyping. Sensors. 2025; 25(11):3401. https://doi.org/10.3390/s25113401

Chicago/Turabian Style

Li, Songhang, Zepu Cui, Jiahang Yang, and Bin Wang. 2025. "A Review of Optical-Based Three-Dimensional Reconstruction and Multi-Source Fusion for Plant Phenotyping" Sensors 25, no. 11: 3401. https://doi.org/10.3390/s25113401

APA Style

Li, S., Cui, Z., Yang, J., & Wang, B. (2025). A Review of Optical-Based Three-Dimensional Reconstruction and Multi-Source Fusion for Plant Phenotyping. Sensors, 25(11), 3401. https://doi.org/10.3390/s25113401

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop