Next Article in Journal
Localization of Buried Polyethylene Gas Pipelines Using Compressional Wave Migration Stacking Imaging
Previous Article in Journal
Toward Accurate Cybersickness Prediction in Virtual Reality: A Multimodal Physiological Modeling Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CLCFM3: A 3D Reconstruction Algorithm Based on Photogrammetry for High-Precision Whole Plant Sensing Using All-Around Images

1
National Agriculture and Food Research Organization, Tsukuba 305-8518, Ibaraki, Japan
2
Kazusa DNA Research Institute, Kisarazu 292-0818, Chiba, Japan
3
R&D Initiative, Chuo University, Bunkyo 112-8551, Tokyo, Japan
4
Graduate School of Agricultural and Life Science, The University of Tokyo, Nishitokyo 188-0002, Tokyo, Japan
5
Graduate School of Science and Technology, Chitose Institute of Science and Technology, Chitose 066-8655, Hokkaido, Japan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5829; https://doi.org/10.3390/s25185829
Submission received: 19 May 2025 / Revised: 2 August 2025 / Accepted: 4 September 2025 / Published: 18 September 2025
(This article belongs to the Section Remote Sensors)

Abstract

This research aims to develop a novel technique to acquire a large amount of high-density, high-precision 3D point cloud data for plant phenotyping using photogrammetry technology. The complexity of plant structures, characterized by overlapping thin parts such as leaves and stems, makes it difficult to reconstruct accurate 3D point clouds. One challenge in this regard is occlusion, where points in the 3D point cloud cannot be obtained due to overlapping parts, preventing accurate point capture. Another is the generation of erroneous points in non-existent locations due to image-matching errors along object outlines. To overcome these challenges, we propose a 3D point cloud reconstruction method named closed-loop coarse-to-fine method with multi-masked matching (CLCFM3). This method repeatedly executes a process that generates point clouds locally to suppress occlusion (multi-matching) and a process that removes noise points using a mask image (masked matching). Furthermore, we propose the closed-loop coarse-to-fine method (CLCFM) to improve the accuracy of structure from motion, which is essential for implementing the proposed point cloud reconstruction method. CLCFM solves loop closure by performing coarse-to-fine camera position estimation. By facilitating the acquisition of high-density, high-precision 3D data on a large number of plant bodies, as is necessary for research activities, this approach is expected to enable comparative analysis of visible phenotypes in the growth process of a wide range of plant species based on 3D information.

1. Introduction

The phenotypes of plants and plant organs are important information in plant functional analyses. With the development of molecular biology, it is becoming increasingly necessary to obtain large quantities of highly accurate data. Digital imaging technology is a simple, low-cost method that offers the advantage of nondestructive measurements of samples [1,2]. However, the images captured by cameras are 2D information, and there are limits to how accurately 2D cameras can obtain 3D shapes and structures, including depth. To precisely measure plant phenotypes during the growth process, technology is needed to accurately digitize shapes and structures in 3D space. The development of measurement technology to support this is progressing [3,4,5,6,7].
Two main types of measurement technology are used to acquire 3D information about an object. One is the active method, which uses a device that emits light such as laser light to perform measurements, and the other is the passive method, which reconstructs 3D data from photographed images of the target [8]. An example of the latter is photogrammetry SfM/MVS (Structure from Motion/Multi-View Stereo) [9]. Photogrammetry can be implemented using commercially available digital cameras. It offers the following two advantages. First, the hardware cost is low. Second, the necessary measurements can be taken simply by photographing the target from all directions, so a flexible measurement system can be constructed according to the measurement environment [10]. To obtain phenotypic data for functional analysis, it is necessary to measure different plant bodies of varying sizes and structural complexities, to perform measurements in various locations, and to measure large numbers of individuals. Photogrammetry can solve these issues.
Photogrammetry is a method of reconstructing 3D point cloud data of an object using images taken from various angles. This is realized by combining SfM processing and MVS processing. SfM processing arranges the viewpoint positions (hereafter referred to as camera positions) of images taken in three-dimensional space [11,12]. MVS processing searches for the position in the image that corresponds to the target point and combines this with the camera position information obtained by SfM to ascertain the position of the point in three-dimensional space using stereo matching processing. This process is repeated to reconstruct a dense 3D point cloud of the object in the image [13]. Given its ability to generate dense 3D point clouds from images, this approach has been proposed for use in phenotyping [14].
Plant structures are characterized by the overlapping of thin, narrow parts such as stems and leaves. Thus, image matching using stereo matching is often unsuccessful because of occlusion. Image matching is also likely to fail because leaf and stem surfaces are similar in color and texture. These failures in image matching can lead to problems in estimating the camera position and can result in incomplete sections of the plant body within the sought 3D point cloud [15] or the occurrence of noise points in places other than the plant body [16]. Creating a plant 3D model with low accuracy is relatively easy, but it remains challenging to construct high-precision 3D models at high throughput for measurement purposes. Therefore, when using photogrammetry to measure plants, it is necessary to solve the problem of failures in image matching that occur in SfM and MVS processing.
To address such issues in SfM processing, various proposals have been made to enhance the accuracy of camera position estimation. These methods involve integrating reference points or GPS devices to provide positional information during capture, such as utilizing images taken with 3D reference points placed around the object to be measured [17,18] or employing positional information obtained from GPS [19]. On the other hand, as a solution in image processing, the loop closing method has been proposed. This method minimizes estimation errors by linking the initial image captured with the final image when all surrounding images of an object are taken [20]. In the image matching process during camera position estimation, techniques have been proposed to pre-evaluate the entire image’s similarity and the relation between images. This approach aims to mitigate the buildup of matching errors resulting from similar patterns [21,22,23].
A major issue with MVS processing is the creation of noise points. Zhang et al. reconstructed 3D models of sweet potato and paprika using SfM/MVS. They found that the 3D reconstruction failure was caused by the inaccuracy of background removal and shadows caused by lighting. Additionally, they noted that the lack of effective solutions for matching at the boundaries of the contours between the background and the object contributes to the inaccuracy of 3D reconstruction for plants [4,24]. Some inaccuracies are projected as noise points. One reason for the presence of noise points is the occurrence of “flying pixels.” These isolated points emerge at object boundaries where there is a depth change between objects. They are considered noise points as they appear in locations devoid of actual objects. Flying pixels occur more prominently in active methods such as laser scanners, LiDAR, and ToF cameras than in passive methods. Methods for removing them utilize the characteristics of the scanning data employed in active methods. The problem of flying pixels has been confirmed not only in active methods but also in passive methods that use images, such as stereo cameras [25]. The noise points of flying pixels in the passive method result from errors in the stereo matching process around the edges of the measurement target. In particular, the matching results become unstable at the boundary areas where the object and background are mixed in one pixel of the image, and noise points are generated according to these matching results. However, it is difficult to theoretically distinguish whether these points are background or object and to remove them in post-processing. This problem remains unresolved given the fundamental disparity between passive and active methods. Therefore, many methods have been proposed for plant research, with most of them involving the removal of noise points from the 3D point cloud data after reconstruction for post-processing [26], that is, the noise in the point cloud data is solved by filtering it out in post-processing. For example, one proposed method uses the Euclidean clustering algorithm to segment the plant point cloud data and remove irrelevant background as well as spatially discontinuous points [27].
This research aims to acquire 3D point clouds of individual plants for the phenotypic evaluation of aspects such as plant organ shapes and individual structures. To this end, we have developed a technology that uses photogrammetry to acquire a large amount of high-density, high-precision 3D point cloud data of plants. Non-photogrammetric methods for reconstructing 3D point clouds using images have also been proposed, with deep learning–based methods such as NeRF [28,29] being representative. These methods entail training costs and, when high-resolution images are used as input images, a large amount of flying pixel noise [30]. Moreover, the results have been confirmed to differ from the original shape of the object, which has uneven surfaces on smooth planes and smooth curved surfaces [31]. For this reason, these methods are not suitable for the acquisition of 3D point cloud data, which is our objective. Furthermore, 3D point clouds with a spatial resolution of 1 mm for plants have been reconstructed by utilizing photogrammetry techniques [32], demonstrating the effectiveness of the use of photogrammetry.
This paper proposes a method, closed-loop coarse-to-fine method with multi-masked matching (CLCFM3), for addressing the problems of photogrammetry for plants. CLCFM3 combines two newly proposed methods—multi-masked matching (MMM) and closed-loop coarse-to-fine method (CLCFM). MMM (Multi-Masked Matching) combines the suppression of occlusion by locally generating point clouds (Multi-Matching) using images with adjacent camera positions and Masked Matching, which removes flying pixel noise that tends to occur at the boundary of the target object. Multi-matching generates a 3D point cloud with reduced occlusion defects and then performs a noise point deletion process using images taken from different angles surrounding those utilized in multi-matching. Masked matching effectively reduces flying pixel noise, which has been considered difficult to remove until now. To perform MMM processing, it is necessary to improve the accuracy of the camera position estimation using SfM processing. To achieve this, we proposed CLCFM, which suppresses the accumulation of camera position estimation errors. CLCFM3 can reconstruct high-density 3D point clouds with fewer noise and missing points. We created an analysis pipeline that implements CLCFM3 and applied the point cloud reconstruction process to soybeans. The results showed that it is possible to reconstruct point clouds with fewer missing points by reducing the noise caused by flying pixels. Moreover, the measurement accuracy achieved with this method revealed that the 3D models constructed through it possess sufficient precision for plant measurement. This suggests that the proposed method holds the potential to truly enable the application of 3D models for precise plant measurements.

2. Materials and Methods

2.1. Plant Material and Cultivation

Four soybean varieties were used in this study. Three varieties—GmJMC092 (KUROHIRA) [33], GmJMC112 (FUKUYUTAKA) [34], and HOUJAKU [35]—were derived from the Genebank project, NARO. Meanwhile, the fourth variety, MISUZUDAIZU, was derived from the NBRP Lotus japonicus & Glycine max collection [36]. The seeds were sown in 1/5000a Wagner pots in 2018 using baked soil, and the plants were grown in a greenhouse at the Kazusa DNA Research Institute. Water tubes were installed in the pots, with irrigation occurring twice a day.

2.2. CLCFM3: 3D Reconstruction Algorithm

We propose a 3D reconstruction method called CLCFM3, based on all-around images of a rotating object taken in front of a camera, assuming that each image is accompanied by information on the order in which it was taken. CLCFM3 consists of two processes: MMM and CLCFM, which utilize information on the order in which images were taken. MMM selects adjacent images and performs MVS processing locally to reduce the number of defects caused by occlusion. CLCFM first selects a part of the captured image and performs SfM before iteratively adding portions of the captured image to the SfM results for further processing. An important point of the proposed method is that it selects images using order-of-capture information so as to surround the measurement target, thus minimizing error accumulation common in SfM methods when images of the entire surrounding area are used. In this study, the Python API of Agisoft Metashape 1.8.5 (Agisoft LLC, St. Petersburg, Russia; hereinafter referred to as Metashape API) [37] is used to process CLCFM3. The Metashape API, a Python module released by Agisoft, provides external access to the functions of Agisoft Metashape Professional Edition 1.8.5 (Agisoft LLC, St. Petersburg, Russia; hereinafter referred to as Metashape), a photogrammetry software.

2.2.1. MMM: Multi-Masked Matching

In MMM, our proposed 3D point cloud generation method, a cloud is generated by stereo matching using images captured from closely positioned cameras locally (Multi-matching). This is followed by iterative point removal using mask images derived from the surrounding viewpoint images utilized during point cloud generation (Masked matching).
Figure 1 shows a representative example of the process. In this example, stereo matching is performed using two images from adjacent camera positions (indicated by the pink squares in Figure 1A) to generate a 3D point cloud, referred to as the baseCloud. This process is executed sequentially using the Metashape API functions buildDepthMaps () and buildDenseCloud ().
Next, mask images from other viewpoint images are used to remove noise points of the baseCloud (the 72 pink squares in Figure 1B), taken from viewpoints within a 45-degree range to the left and right, facilitating the point removal process from the baseCloud. The mask image is a binary representation where the area of the plant region to be measured is set to 1, while other areas are designated as 0. The execution of this process involves utilizing the Metashape API functions selectMaskedPoints () and removeSelectedPoints () in that order. For every image captured, this process is carried out for all combinations of adjacent images to obtain the point cloud and remove noise points. The resulting point clouds are amalgamated to construct a 3D point cloud of the target plant.

2.2.2. CLCFM: Closed-Loop Coarse-to-Fine Method for SfM

We propose the CLCFM to improve the stability and accuracy of camera position estimation for images taken from all around the target. Instead of performing SfM simultaneously with all images captured around the target, we utilize a small subset of the images for the initial SfM process. We then iterate the process by incorporating subsets of the remaining images taken around the target. This iterative process employs a small number of all-around images initially, then incorporates all images to perform SfM to estimate the camera positions of all captured images. This method of performing SfM sequentially rather than all at once improves the accuracy of the camera position estimation.
The principle of the method is illustrated through an example where a single camera captures images of an object rotated in 5-degree increments (Figure 2A). First, the images are categorized into groups, with the 72 images around the object being divided into six groups in this instance. Next, the order in which the SfM process is executed (GroupOrder) is determined. The GroupOrder is selected so as to minimize the bias in the estimated camera positions. For example, if Group 1 is the first group subjected to the SfM process, then Group 4, which includes the images halfway between the adjacent images of Group 1, is selected as the second group to be processed, and Group 2, which is between Group 1 and Group 4, is selected as the third.
Next, the procedure shown in Figure 2B is followed to perform SfM sequentially for each group. The images in Group 1 are used to estimate the first camera position. The main process for camera position estimation in Figure 2B, AlignProcess (), is executed for ImgList [] in the order of the Metashape API functions matchPhotos (), alignCameras (), and optimizeCameras (). The results of the first camera position estimation are then used as initial values, and a second SfM is performed by adding the Group 4 images to the Group 1 images. Similarly, images are processed in the order specified by GroupOrder, with SfM being sequentially executed using the camera position estimates obtained from the previous step as initial values. The final camera position estimates are derived by processing all groups. When images are captured using multiple cameras arranged vertically, those taken from the same vertical angle are grouped together for processing.

2.3. Three-Dimensional Point Cloud Reconstruction Pipeline Program

We developed a 3D point cloud reconstruction pipeline that implements our proposed method. This pipeline consists of four subprograms executed sequentially, starting with the captured image as the initial input to reconstruct the 3D point cloud (Figure 3).
  • CreateMask.cpp
This program creates mask images from captured images. Reference images for the object and background regions are manually predefined. Subsequently, each pixel in the captured image is classified by comparing its average color difference with the reference images [38]. The program generates binary mask images where the plant areas vital for reconstructing the 3D point cloud and the relevant object regions are shown in white, while the remaining areas are black. Mask images are produced for all captured images, utilized for mask processing in CLCFM and noise elimination in MMM.
This program was developed in Visual C++ by using Visual Studio Professional 2019 and OpenCV version 3.4.14.
2.
align_camera.py
This script estimates camera position for the CLCFM algorithm using captured and masked images, providing the estimated camera position information as a Metashape project file used in MMM.
This program was developed using Python 3.8 and the Python 3 module of Metashape Professional Edition 1.8.5 (Metashape API).
3.
build_densecloud.py
This script executes the MMM algorithm to construct a 3D point cloud and exports multiple 3D point cloud data files in PLY format. The script was developed using Python 3.8 and the Python 3 module of Metashape Professional Edition 1.8.5 (Metashape API). The results presented in this paper were obtained using the following execution parameters for the processing functions within the Python 3 Module.
  • matchPhotos(): The images were processed at 100% scale with accuracy set to ‘high’.
  • buildDepthMaps(): To evaluate the effectiveness of MMM, the images were processed at 1/4 scale (1500 × 1000 pixels) with quality set to ‘medium’.
4.
merge_subsampling.py
This program integrates the diverse 3D point cloud data output by ‘build_densecloud.py’ and performs subsampling processing at 0.5 mm intervals, resulting in the output of the 3D point cloud of the target object.
This program was developed using Python 3.7 and open3D 0.15.

2.4. Imaging

To obtain the best images for the proposed CLCFM3 method, we built an imaging studio consisting of a rotating stage with plants, a camera, LED lighting, and background panels (Figure 4A). The background panels, equipment, and cords were colored blue to facilitate the creation of mask images.
The imaging stage was devised to allow the rotation of the plant in front of the camera, capturing a comprehensive image of the entire plant (Figure 4B). SGSP-160YAW (Sigma Koki, Tokyo, Japan) and GIP-101 (Sigma Koki) were connected to a personal computer using a serial USB converter (COM-1(USB)H; Contec, Osaka, Japan). The rotations were controlled by a program running on the personal computer.
To hold the cultivation pot, an acrylic disc (35 cm in diameter) was made and attached to the SGSP-160YAW. Small pieces of paper printed with coded targets were attached to the acrylic disc, and additional coded targets were attached to the three pillars fixed outside the perimeter of the disc. The coded targets were used as reference points for integrating the point cloud using MMM and to define the scale that gives the actual dimensions [6]. The coded targets were attached to the following positions: the center of the disc; eight positions at 45-degree intervals around the circumference, 15 cm from the center; and positions separated by 20 cm vertically on the pillars. Each of the three pillars consisted of an aluminum frame of 2.5 cm width Í 1.5 cm thickness Í1 m length, to which was attached a strip of paper printed with a random dot image. These random dots were used to ensure that the feature points obtained in the CLCFM and MMM processing steps were stable [6]. The pillars were attached to the side of the disc at uniform angles (Figure 4B, left).
Images were captured using four digital SLR cameras (D5500; Nikon Corporation, Tokyo, Japan) and lenses (AF-S Micro NIKKOR 60 mm f/2.8G ED, Nikon Corporation). Figure 4C shows the camera arrangement used to capture images of the plants used as targets with heights ranging from 400 to 1200 mm. Images of plants shorter than 400 mm (e.g., lettuce), approximately 1000 mm (e.g., Japanese cedar, soybean, and rice plants), or taller than 2000 mm (e.g., tomato, paprika, and eggplant) can be captured by adjusting the number and arrangement of cameras, as well as the height of the pillars [9]. The cameras were connected to a personal computer via USB cables. The shooting conditions were: shutter speed 1/15 s, ISO sensitivity 100, aperture F/16. The acquired images were saved as 6000 × 4000 pixel JPG files. Shooting and rotation were controlled by the same program. To minimize the vibration effects, the stage rotated 5 degrees in approximately 2 s and paused for 1 s before the next shot.

3. Results

3.1. ‘baseCloud’ Construction Process in MMM

Figure 5 shows the results of an experiment to verify the optimal viewing angle for generating baseCloud point clouds of the target plant. Figure 5a,b display portions of the baseCloud reconstructed using only two adjacent images and four adjacent images, respectively, from different camera positions. These images were captured by rotating the table holding the object in 5-degree increments, resulting in viewing angles of 5 degrees for the baseCloud in Figure 5a and 15 degrees for that in Figure 5b. The number of point clouds is 327,768 in Figure 5a and 364,145 in Figure 5b. For this object, no significant difference was observed between the baseClouds in Figure 5a,b. Nonetheless, while reconstructing a separate sample with Metashape 1.6, we confirmed that configuring the viewing angle to 5 degrees helped reduce the loss of point clouds (Figure S3). The reconstruction time with MMM diminishes with a reduction in the number of images employed. Given these findings, we chose to generate the baseCloud using two adjacent images.

3.2. Viewing Angle of the Mask Image Used in the MMM Point Removal Process

We next investigated the optimal viewing angle for the mask image used in the point removal procedure of the baseCloud performed by MMM. To ascertain the optimal angle, we performed the point removal process of baseCloud using MMM under four different viewing angles of the mask image, similar to the method illustrated in Figure 6. Figure 6 illustrates the outcomes when employing the mask image at viewing angles of (a) 50 degrees, (b) 90 degrees, (c) 180 degrees, and (d) 360 degrees. The corresponding numbers of point clouds are (a) 4,901,663, (b) 4,309,275, (c) 2,793,222, and (d) 1,768,662. The results indicated that as the viewing angle of the surrounding mask increased, more points were eliminated; however, this increase also led to the deletion of points in the plant body that should have been retained. In particular, (c) and (d) exhibit prominent defects in the stems and leaves. This means that noise reduction and defect suppression are in a trade-off relationship. To reduce more noise and minimize these defects, we set a 90-degree viewing angle for the mask image used in point removal.

3.3. Three-Dimensional Point Cloud Reconstruction of Soybean Plants

Using the pipeline developed and the processing conditions established in the previous sections, we successfully reconstructed 3D point clouds for four soybean varieties (Figure 7a and Figure S1). The results demonstrated the proposed method’s ability to reconstruct 3D point clouds of varying shapes, sizes, and structural complexities. For comparison, the identical image sets underwent standard Metashape functions to generate 3D point clouds. The number of point clouds is 4,309,275 in Figure 7a and 1,882,403 in Figure 7b, both subsampled to 0.5 mm to match the conditions. This comparison confirmed that the proposed method effectively mitigated the appearance of blue dots as noise in the outlines of leaves and stems, while also resolving the problem of point clouds failing to form in certain thin-stem regions (Figure 7b, Figures S1 and S2).
Figure 7c displays the outcome when the point removal process was omitted after the baseCloud was generated with MMM, while Figure 7d showcases the result of the point removal process after the baseCloud was generated with MMM. In both cases, point cloud reconstruction and point removal processing were performed after SfM was executed using CLCFM. The number of point clouds is 5,776,168 in Figure 7c and 4,309,275 in Figure 7d. Without point removal processing, noise points emerged in the background and other parts of the captured image (Figure 7c), while point removal processing resulted in the removal of noise points (Figure 7d).

3.4. Effects of the Point Removal Processes in MMM

We compared the differences in the reconstruction of 3D point clouds by MMM with and without point removal processing. The Base clouds were generated using the MMM point cloud reconstruction approach, leading to the production of 3D point clouds after applying point removal processing (Figure 8a) and without such processing (Figure 8b). These 3D point clouds were then overlaid and compared within a virtual space. To facilitate this comparison, only the plant body was extracted, while the support, pot, and rotating stage were manually removed. The distance between corresponding points in the aligned 3D point clouds was calculated. Points with distances of 3 mm or less were deleted, yielding a point cloud that represented the differences between the two 3D reconstructions (Figure 8c). This process was repeated for distance intervals of 0–1 mm, 1–2 mm, 2–3 mm, and 3–22 mm, and the number of points in each point cloud was recorded. As shown in Table 1, in the point cloud post-point removal process, 78,637 points exhibited a distance variance of 3 mm or more, constituting 4.3% of the aggregate points within the point cloud.

3.5. Validation of Position Accuracy in the Reconstructed Point Cloud

We assessed the position estimation errors for the 3D point cloud acquired through the CLCFM3 pipeline using the evaluation model shown in Figure 9. The evaluation model was arranged with 13 pairs of measurement points (Figure 9 (1)–(13)), consisting of coded target pairs. The distance between these targets was measured manually with a caliper and compared with the value measured using the 3D point cloud generated by CLCFM3. The actual values were measured 10 times each, and the average was used. The coded targets attached to the pillars were measured using a ruler to derive the average value (Figure 9 (14)). The reading errors for manual measurement were set at 0.02 mm for the caliper and 1 mm for the ruler. The measurement using CLCFM3 was acquired by obtaining the distances between the targets in the 3D point cloud reconstructed through the pipeline. This procedure was repeated 10 times, and the average value was derived.
Table 2 shows the results. For measurement points (1) to (13), the maximum difference between the manually measured values (actual values) and the values obtained from the 3D point cloud was 0.280 mm, with a maximum standard deviation of 0.0000346 mm for the 3D measurements. For measurement point (14), the difference between the average manually measured value (actual value) and the average value obtained from the 3D point cloud was −0.97 mm, with a standard deviation of 0.00012 mm for the 3D measurements.
A quantitative target of 1 mm was set for the measurement resolution in [9]. As shown in Table 2, both the mean values and standard deviations were less than 1 mm, indicating that the target was successfully achieved.

4. Discussion

4.1. Generating Local 3D Points and Removing Noise with Mask Images

We proposed an approach to reduce the loss due to occlusion by constructing 3D point clouds using images with narrow viewing angles (Multi-matching). However, this process generates many flying pixels due to errors in stereo matching when searching for corresponding points along the outline. Flying pixels appear at object boundaries where a depth discontinuity exists between adjacent objects. They are perceived as noise points because they occur in regions where no object is present. We proposed to solve these issues by eliminating points using a mask image of the peripheral field of view from the images employed in point cloud reconstruction, reducing occlusion-related losses, and achieving point cloud reconstruction with fewer noise points (Masked matching). This section explains the underlying principles of these methods.
Figure 10 shows a schematic of the image-matching process at a plant boundary.
To estimate the 3D position of point P on the object’s contour line, we extract the coordinates of point PA1 and point PA2, representing point P’s location in the images taken by cameras A1 and A2, respectively. Using these coordinates and the camera position information for cameras A1 and A2 acquired through SfM, the depth distance DP between point P and the cameras is calculated. Based on the outcome, the 3D position of point P is ascertained, leading to the creation of a point for the 3D point cloud (Figure 10A).
Here, we focus on point P in the outline of the plant body. Image matching is performed to find position PA2 on the image taken by camera A2 relative to position PA1 on the image taken by camera A1.
The DP value fluctuates according to the matching results. If the position of point P found by camera A1 is correctly matched with point P in the image from camera A2, DP is assigned is accurate value. Consequently, point P on the plant body will be generated in 3D space (Figure 10A). However, if point P is incorrectly matched and the matching result differs from the accurate point P (Figure 10B), the DP value deviates, leading to the acquisition of DPm and DPm’ values (Figure 10A). Based on these values, points are generated at positions Pm and Pm’, which differ from the accurate position of point P. These points, Pm and Pm’, are noise points that do not exist on the plant body. The causes of inaccurate image matching include the resolution of the imaging camera’s image sensor and the matching process with points exhibiting textures akin to leaves and stems close to point P. To address the issue of noise points caused by mismatching, our proposed method does not aim to rectify the mismatches but instead identifies numerous candidate points, incorporating a post-processing step for point removal. This approach reduces the number of noise points.
Figure 10C shows a schematic of the point removal process for MMM. Cameras A1, A2, B1, and B2 represent the camera positions. The A1 and A2 images are used to generate the point cloud (baseCloud), and the mask images generated from the images captured by all four cameras are used for the point removal process (Figure 10C). The mismatch at points on the contour line of the object occurs in the vicinity of the intersection lines PA1Pback and PA2P’back of Cameras A1 and A2. Using the A1 and A2 mask images, the point removal process can eliminate mismatched points beyond the object’s boundaries. However, none of the mismatched points generated on the PA1Pback and PA2P’back boundary lines can be removed. The mismatched points Pm and Pm’ on the boundary line shown in Figure 10C can be eliminated by excluding areas outside the object at the intersection lines PB1Pb1 and PB2Pb2 of Cameras B1 and B2 in the mask image. This approach also enables the removal of mismatched points along the line between PA1Pback and PA2P’back. The same image sets were processed using the standard Metashape procedure (Figure 7b) for comparison with the point cloud generated using MMM (Figure 7a). The number of point clouds is 4,309,275 in Figure 7a and 1,882,403 in Figure 7b, both subsampled to 0.5 mm to match the conditions. This comparison demonstrates that MMM generates point clouds with less noise and fewer defects.
The proposed method uses surrounding mask images to detect flying pixel noise points. Therefore, the accuracy of noise point detection depends on the results of the mask image region extraction, making the image processing of the region extraction an important factor. The developed pipeline employs a method that uses color differences for region extraction [38,39]. Numerous image analysis techniques have been proposed for area extraction, enabling the generation of tailored mask images capable of eliminating noise points adapted to various plants and shooting environments. This versatility allows MMM to be applied across a wide range of plants, facilitating the reconstruction of 3D point clouds with suppressed flying pixel noise. In 3D plant phenotyping using rotating stage-based image capture, background removal is applied to eliminate interference from non-plant elements during reconstruction [10,40,41,42]. However, these studies do not address the issue of noise loss caused by flying pixels and defects.
The proposed method can remove noise points on the outside of the plant body using a point removal process. However, noise points located inside the plant body cannot be detected and removed. Developing a method for removing these internal noise points stands as a future issue.

4.2. Enhancing Camera Position Accuracy for All-Around Images

In this study, we used all-around images taken by rotating the object at 5-degree intervals for 3D point cloud reconstruction. In estimating camera position using SfM processing on omnidirectional images with narrow angles of view, one challenge is that the estimation process is likely to fail due to accumulated errors. Figure 11A shows the result of camera pose estimation using all-around images without pillars and the standard Metashape procedure.
CLCFM estimates camera position while gradually adding images. First, camera position is estimated using only images that surround the object but have an open viewpoint, and the results are used as initial values. Next, we add a few more images and run the estimation process again, updating the results. This process is repeated until the camera positions for all images in the complete surroundings are estimated. This processing method is expected to reduce the failure of position estimation due to error accumulation by suppressing the propagation of estimation errors that occur in the estimation process. Figure 11B shows the result of camera pose estimation using all-around images with pillars and CLCFM.
However, there is a problem with implementing this algorithm. If the cameras are placed far apart in the initial stage of processing, few common feature points would exist between the images, making it difficult to estimate the camera position. We addressed this issue by attaching random patterns to the pillars when taking pictures (Figure 4B). Securing numerous image feature points within the area featuring the random pattern enabled a bundle adjustment with self-calibration [9,43]. This made it possible to stably obtain the camera’s internal and external parameters. Finally, since it is not necessary to perform internal and external calibration in advance, a key advantage is the potential to attain high measurement accuracy without engaging in device setup or maintenance [9].

5. Conclusions

This paper introduced the CLCFM3 algorithm for the 3D reconstruction of all-around plant images. Given that plants possess a distinctive morphology characterized by overlapping thin parts such as leaves and stems, challenges arise when performing 3D point cloud reconstruction via photogrammetry. These challenges include the inability to acquire points in the 3D point cloud due to occlusion in the overlapping parts and the generation of points where the object does not exist due to image matching errors along the object’s outline.
To solve these problems, we proposed the MMM approach. MMM generates a point cloud using images taken from a viewpoint close to the object to avoid occlusion (multi-matching) and then repeatedly executes point cloud removal processing using surrounding mask images to remove the noise points (mask matching) (Figure 1).
To execute MMM, a highly accurate camera position estimation must first be obtained by SfM processing. To meet this requirement, we proposed CLCFM, which estimates camera positions with high accuracy for images encompassing the entire surroundings of the target. However, when SfM is conducted using such images, estimation errors accumulate. CLCFM selects images for SfM processing and iterates the SfM procedure to suppress this accumulation of errors (Figure 2). MMM and CLCFM are based on the assumptions that the target is rotated and photographed from all sides and that the order of the images is known. Using this information, we were able to implement the proposed method by reconstructing local point clouds and selecting surrounding mask images for point removal processing.
We built an imaging studio (Figure 4) and an image analysis pipeline (Figure 3) to implement the proposed method and reconstructed 3D point clouds for soybean plants (Figure 7). We validated the feasibility of reconstructing a high-density 3D point cloud of a plant body that minimizes the loss of thin stems, a characteristic shape of the plant body, as well as reducing noise from 3D points generated in areas devoid of plant structure (Figure 8).
In this study, API of Metashape Professional Edition 1.8.5 [37] was used to execute SfM/MVS. We found that the different versions of Metashape yielded different results in MVS point cloud reconstruction (Figures S3 and S4). However, regardless of the version, the proposed method is effective for reconstructing 3D point clouds of plant bodies characterized by overlapping organ shapes such as leaves and stems. This effectiveness is based on the method’s ability to reduce noise through point cloud reconstruction using local perspectives with MMM and point removal procedures involving mask images.
Although numerous methods for constructing 3D models of plants have been proposed, practical applications for measurement purposes remain limited except for rough assessments of community dynamics in the field. The proposed method enables the robust construction of highly accurate 3D models, a feature anticipated to advance the widespread use of 3D modeling for plant measurements in plant research.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/s25185829/s1, Figure S1 shows the results of reconstruction of 3D point clouds of four soybean varieties using the proposed method. Figure S2 shows the results of reconstruction of 3D point clouds for the image in Figure S1 using Metashape. Figure S3 shows the results of running the analysis in Figure 5 using Metashape 1.6. Figure S4 shows the results of running the analysis in Figure 6 using Metashape 1.6.

Author Contributions

A.H., N.K., S.I., and T.T. contributed to the design of the study. A.H. developed the 3D reconstruction pipeline programs. A.H. and T.T. made imaging hardware. A.H., N.K., and K.K. conducted experiments and analyzed and interpreted data. A.H., N.K., S.I., and T.T. wrote and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a CREST Grant (No. JPMJCR16O1) from the Japan Science and Technology Agency and the Kazusa DNA Research Institute Foundation.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The computer codes and scripts of MMM and CLCFM are deposited in a GitHub repository at https://github.com/tanasoft/MMM-CLCFM (accessed on 3 September 2025). Download links to image and 3D point cloud data that support the findings of this study are also provided at this GitHub repository.

Acknowledgments

We would like to thank Keiko Saegusa, Rumi Fujihira, Satomi Nakayama, Chie Kasashima, and Sayaka Kikuchi for their technical support at the Kazusa DNA Research Institute. We also thank Ryo Akashi, Hidenori Tanaka, Masatsugu Hashiguchi, and Takuyu Hashiguchi for providing soybean materials in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, Z.; Guo, R.; Li, M.; Chen, Y.; Li, G. A review of computer vision technologies for plant phenotyping. Comput. Electron. Agric. 2020, 176, 105672. [Google Scholar] [CrossRef]
  2. Yang, W.; Feng, H.; Zhang, X.; Zhang, J.; Doonan, J.H.; Batchelor, W.D.; Xiong, L.; Yan, J. Crop Phenomics and High-Throughput Phenotyping: Past Decades, Current Challenges, and Future Perspectives. Mol. Plant 2020, 13, 187–214. [Google Scholar] [CrossRef]
  3. Okura, F. 3D modeling and reconstruction of plants and trees: A cross-cutting review across computer graphics, vision, and plant phenotyping. Breed. Sci. 2022, 72, 31–47. [Google Scholar] [CrossRef]
  4. Zhang, Y.; Teng, P.; Aono, M.; Shimizu, Y.; Hosoi, F.; Omasa, K. 3D monitoring for plant growth parameters in field with a single camera by multi-view approach. J. Agric. Meteorol. 2018, 74, 129–139. [Google Scholar] [CrossRef]
  5. Kunita, I.; Morita, M.T.; Toda, M.; Higaki, T. A Three-Dimensional Scanning System for Digital Archiving and Quantitative Evaluation of Arabidopsis Plant Architectures. Plant Cell Physiol. 2021, 62, 1975–1982. [Google Scholar] [CrossRef]
  6. Akhtar, M.S.; Zafar, Z.; Nawaz, R.; Fraz, M.M. Unlocking plant secrets: A systematic review of 3D imaging in plant phenotyping techniques. Comput. Electron. Agric. 2024, 222, 109033. [Google Scholar] [CrossRef]
  7. Li, S.; Cui, Z.; Yang, J.; Wang, B. A Review of Optical-Based Three-Dimensional Reconstruction and Multi-Source Fusion for Plant Phenotyping. Sensors 2025, 25, 3401. [Google Scholar] [CrossRef]
  8. Paulus, S. Measuring crops in 3D: Using geometry for plant phenotyping. Plant Methods 2019, 15, 103. [Google Scholar] [CrossRef] [PubMed]
  9. Kochi, N.; Hayashi, A.; Shinohara, Y.; Tanabata, T.; Kodama, K.; Isobe, S. All-around 3D plant modeling system using multiple images and its composition. Breed. Sci. 2022, 72, 75–84. [Google Scholar] [CrossRef] [PubMed]
  10. Salter, W.T.; Shrestha, A.; Barbour, M.M. Open source 3D phenotyping of chickpea plant architecture across plant development. Plant Methods 2021, 17, 95. [Google Scholar] [CrossRef]
  11. Snavely, N. Bundler: Structure from Motion (sfm) for Unordered Image Collections. Available online: https://www.cs.cornell.edu/~snavely/bundler/ (accessed on 2 April 2025).
  12. Wu, C. Towards Linear-Time Incremental Structure from Motion. In Proceedings of the 2013 International Conference on 3D Vision, Seattle, WA, USA, 29 June–1 July 2013. [Google Scholar] [CrossRef]
  13. Furukawa, Y.; Ponce, J. Accurate, Dense, and Robust Multiview Stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1362–1376. [Google Scholar] [CrossRef]
  14. Townsend, A.J.; Retkute, R.; Chinnathambi, K.; Randall, J.W.P.; Foulkes, J.; Carmo-Silva, E.; Murchie, E.H. Suboptimal Acclimation of Photosynthesis to Light in Wheat Canopies. Plant Physiol. 2018, 176, 1233–1246. [Google Scholar] [CrossRef]
  15. Santos, T.T.; Oliveira, A.A.d. Image-based 3D digitizing for plant architecture analysis and phenotyping. In Proceedings of the Workshop on Industry Applications (WGARI) in SIBGRAPI 2012 (XXV Conference on Graphics, Patterns and Images), Ouro Preto, Barzil, 22–25 August 2012. [Google Scholar] [CrossRef]
  16. Gibbs, J.A.; Pound, M.P.; French, A.P.; Wells, D.M.; Murchie, E.H.; Pridmore, T.P. Active Vision and Surface Reconstruction for 3D Plant Shoot Modelling. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019, 17, 1907–1917. [Google Scholar] [CrossRef] [PubMed]
  17. Barazzetti, L.; Forlani, G.; Remondino, F.; Roncella, R.; Scaioni, M. Experiences and achievements in automated image sequence orientation for close-range photogrammetric projects. In Proceedings of the Videometrics, Range Imaging, and Applications XI, Munich, Germany, 25–26 May 2011; p. 8085. [Google Scholar] [CrossRef]
  18. Anai, T.; Kochi, N.; Yamada, M.; Sasaki, T.; Otani, H.; Sasaki, D.; Nishimura, S.; Kimoto, K.; Yasui, N. Examination about influence for precision of 3D image measurement from the ground control point measurement and surface matching. In Proceedings of the Joint Workshop with ISPRS WG IV/7 and WG V/4, Tokyo, Japan, 21–22 May 2015; pp. 165–170. [Google Scholar] [CrossRef]
  19. Sattler, T.; Torii, A.; Sivic, J.; Pollefeys, M.; Taira, H.; Okutomi, M.; Pajdla, T. Are Large-Scale 3D Models Really Necessary for Accurate Visual Localization? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1637–1646. [Google Scholar]
  20. Scaramuzza, D.; Fraundorfer, F.; Pollefeys, M. Closing the loop in appearance-guided omnidirectional visual odometry by using vocabulary trees. Robot. Auton. Syst. 2010, 58, 820–827. [Google Scholar] [CrossRef]
  21. Havlena, M.; Torii, A.; Pajdla, T. Efficient Structure from Motion by Graph Optimization. In Proceedings of the 11th European Conference on Computer Vision, Crete, Greece, 5–11 September 2010; pp. 110–113. [Google Scholar] [CrossRef]
  22. Sweeney, C.; Sattler, T.; Höllerer, T.; Turk, M.; Pollefeys, M. Optimizing the Viewing Graph for Structure-from-Motion. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 801–809. [Google Scholar]
  23. Schönberger, J.L.; Frahm, J. Structure-from-Motion Revisited. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 4104–4113. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Teng, P.; Shimizu, Y.; Hosoi, F.; Omasa, K. Estimating 3D leaf and stem shape of nursery paprika plants by a novel multi-camera photography system. Sensors 2016, 16, 874. [Google Scholar] [CrossRef] [PubMed]
  25. Sabov, A.; Krüger, J. Identification and correction of flying pixels in range camera data. In Proceedings of the 24th Spring Conference on Computer Graphics, Budmerice Castle, Slovakia, 21–23 April 2008; pp. 135–142. [Google Scholar]
  26. Harandi, N.; Vandenberghe, B.; Vankerschaver, J.; Depuydt, S.; Van Messem, A. How to make sense of 3D representations for plant phenotyping: A compendium of processing and analysis techniques. Plant Methods 2023, 19, 60. [Google Scholar] [CrossRef] [PubMed]
  27. Li, Y.; Liu, J.; Zhang, B.; Wang, Y.; Yao, J.; Zhang, X.; Fan, B.; Li, X.; Hai, Y.; Fan, X. Three-dimensional reconstruction and phenotype measurement of maize seedlings based on multi-view image sequences. Front. Plant Sci. 2022, 13, 974339. [Google Scholar] [CrossRef]
  28. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis. arXiv 2020, arXiv:2003.08934. [Google Scholar] [CrossRef]
  29. Hu, K.; Ying, W.; Pan, Y.; Kang, H.; Chen, C. High-fidelity 3D reconstruction of plants using Neural Radiance Fields. Comput. Electron. Agric. 2024, 220, 108848. [Google Scholar] [CrossRef]
  30. Croce, V.; Billi, D.; Caroti, G.; Piemonte, A.; De Luca, L.; Véron, P. Comparative Assessment of Neural Radiance Fields and Photogrammetry in Digital Heritage: Impact of Varying Image Conditions on 3D Reconstruction. Remote Sens. 2024, 16, 301. [Google Scholar] [CrossRef]
  31. Tsukamoto, Y.; Hayashi, A.; Tokuda, K.; Kochi, N. Comparative Verification of 3D Reconstructed Point Cloud Data: Comparison of AI Estimation and Multi-View Stereo Methods using Nerfstudio and Metashape. In Proceedings of the SICE Festival 2024 with Annual Conference, Kochi, Japan, 27–30 August 2024. FrAT3.2. [Google Scholar]
  32. Kochi, N.; Isobe, S.; Hayashi, A.; Kodama, K.; Tanabata, T. Introduction of All-Around 3D Modeling Methods for Investigation of Plants. Int. J. Autom. Technol. 2021, 15, 301–312. [Google Scholar] [CrossRef]
  33. KUROHIRA. Available online: https://www.gene.affrc.go.jp/databases-plant_search_detail.php?jp=27863 (accessed on 7 July 2025).
  34. FUKUYUTAKA. Available online: https://www.gene.affrc.go.jp/databases-plant_search_detail.php?jp=29668 (accessed on 7 July 2025).
  35. HOUJAKU. Available online: https://www.gene.affrc.go.jp/databases-plant_search_detail.php?jp=28770 (accessed on 7 July 2025).
  36. Misuzudaizu. Available online: https://legumebase.nbrp.jp/glycine/rilParentStrainDetailAction.do?&strainId=0 (accessed on 7 July 2025).
  37. Agisoft Metashape. Available online: https://www.agisoft.com (accessed on 7 July 2025).
  38. Tanabata, T.; Yamada, T.; Shimizu, Y.; Shinozaki, Y.; Kanekatsu, M.; Takano, M. Development of automatic segmentation software for efficient measurement of area on the digital images of plant organs. Hortic. Res. 2010, 9, 501–506. [Google Scholar] [CrossRef][Green Version]
  39. Tanabata, T.; Shibaya, T.; Hori, K.; Ebana, K.; Yano, M. SmartGrain: High-throughput phenotyping software for measuring seed shape through image analysis. Plant Physiol. 2012, 160, 1871–1880. [Google Scholar] [CrossRef] [PubMed]
  40. Yang, D.; Yang, H.; Liu, D.; Wang, X. Research on automatic 3D reconstruction of plant phenotype based on Multi-View images. Comput. Electron. Agric. 2024, 220, 108866. [Google Scholar] [CrossRef]
  41. Zhu, R.; Sun, K.; Yan, Z.; Yan, X.; Yu, J.; Shi, J.; Hu, Z.; Jiang, H.; Xin, D.; Zhang, Z.; et al. Analysing the phenotype development of soybean plants using low-cost 3D reconstruction. Sci. Rep. 2020, 10, 7055. [Google Scholar] [CrossRef] [PubMed]
  42. Leménager, M.; Burkiewicz, J.; Schoen, D.J.; Joly, S. Studying flowers in 3D using photogrammetry. New Phytol. 2022, 237, 1922–1933. [Google Scholar] [CrossRef]
  43. Karara, H.M. Camera-Calibration in Non-Topographic Photogrammetry; American Society for Photogrammetry and Remote Sensing: Bethesda, MD, USA, 1989; Chapter 5. [Google Scholar]
Figure 1. Algorithm of MMM (Multi-Masked Matching): (A) Point clouds are generated locally using images from two adjacent viewpoints. (B) Noise points are removed using mask images derived from the surrounding viewpoint images. Processes 1 and 2 are repeated for each viewpoint, and the resulting point clouds are integrated.
Figure 1. Algorithm of MMM (Multi-Masked Matching): (A) Point clouds are generated locally using images from two adjacent viewpoints. (B) Noise points are removed using mask images derived from the surrounding viewpoint images. Processes 1 and 2 are repeated for each viewpoint, and the resulting point clouds are integrated.
Sensors 25 05829 g001
Figure 2. Algorithm of CLCFM (Closed-Loop Coarse-to-Fine Method): (A) The sequentially captured images are divided into groups according to viewpoint. Six groups consist of 72 images taken at 5-degree intervals. (B) Procedure of CLCFM. First, SfM is performed on group 1 images. Then, group 4 images are appended, and SfM is performed. Similarly, groups 2, 5, 3, and 6 images are appended, and SfM is performed.
Figure 2. Algorithm of CLCFM (Closed-Loop Coarse-to-Fine Method): (A) The sequentially captured images are divided into groups according to viewpoint. Six groups consist of 72 images taken at 5-degree intervals. (B) Procedure of CLCFM. First, SfM is performed on group 1 images. Then, group 4 images are appended, and SfM is performed. Similarly, groups 2, 5, 3, and 6 images are appended, and SfM is performed.
Sensors 25 05829 g002
Figure 3. Three-Dimensional point cloud reconstruction pipeline program. “CreateMaskImage.cpp” makes mask images. “align_camera.py” processes CLCFM. “build_densecloud.py” and “merge_subsampling.py” are MMM processes. These programs are executed sequentially with the captured image as the initial input to reconstruct the 3D point clouds.
Figure 3. Three-Dimensional point cloud reconstruction pipeline program. “CreateMaskImage.cpp” makes mask images. “align_camera.py” processes CLCFM. “build_densecloud.py” and “merge_subsampling.py” are MMM processes. These programs are executed sequentially with the captured image as the initial input to reconstruct the 3D point clouds.
Sensors 25 05829 g003
Figure 4. Imaging studio for 3D point cloud reconstruction by CLCFM3: (A) The imaging studio consists of an imaging stage with a rotating platform on which plants are placed, four cameras positioned in front, LED lights surrounding the plant, and a blue background panel to facilitate mask image generation. (B) The imaging stage has three pillars on which random dot patterns and coded targets for scale are attached. (C) Camera angles and distance from the object are determined for soybeans ranging in height from 400 to 1200 mm.
Figure 4. Imaging studio for 3D point cloud reconstruction by CLCFM3: (A) The imaging studio consists of an imaging stage with a rotating platform on which plants are placed, four cameras positioned in front, LED lights surrounding the plant, and a blue background panel to facilitate mask image generation. (B) The imaging stage has three pillars on which random dot patterns and coded targets for scale are attached. (C) Camera angles and distance from the object are determined for soybeans ranging in height from 400 to 1200 mm.
Sensors 25 05829 g004
Figure 5. Results of ‘baseCloud’ reconstruction with different numbers of images: (a) Two adjacent images. (b) Four adjacent images. Processing time increases with the number of images, while there is no significant difference between (a,b).
Figure 5. Results of ‘baseCloud’ reconstruction with different numbers of images: (a) Two adjacent images. (b) Four adjacent images. Processing time increases with the number of images, while there is no significant difference between (a,b).
Sensors 25 05829 g005
Figure 6. Results for different viewing angles of the mask image used in the point removal process. The upper figures show a magnified view of the area outlined in red: (a) 50 degrees, (b) 90 degrees, (c) 180 degrees, (d) 360 degrees. More points are deleted as the viewing angle increases, while points consisting of leaves and stems are also deleted.
Figure 6. Results for different viewing angles of the mask image used in the point removal process. The upper figures show a magnified view of the area outlined in red: (a) 50 degrees, (b) 90 degrees, (c) 180 degrees, (d) 360 degrees. More points are deleted as the viewing angle increases, while points consisting of leaves and stems are also deleted.
Sensors 25 05829 g006
Figure 7. Differences in results between point removal processes: (a,d) Removed by MMM. (b) Standard Metashape process. (c) Removed by MMM without the point removal process. A distinct difference can be observed between (a,b) at the tip of the red arrow.
Figure 7. Differences in results between point removal processes: (a,d) Removed by MMM. (b) Standard Metashape process. (c) Removed by MMM without the point removal process. A distinct difference can be observed between (a,b) at the tip of the red arrow.
Sensors 25 05829 g007
Figure 8. Results comparing point cloud reconstruction with and without the noise removal process by MMM: (a) Point clouds with noise removal process by MMM. (b) Point clouds without noise removal process by MMM. (c) Point clouds where the distance between (a,b) is 1–3 mm.
Figure 8. Results comparing point cloud reconstruction with and without the noise removal process by MMM: (a) Point clouds with noise removal process by MMM. (b) Point clouds without noise removal process by MMM. (c) Point clouds where the distance between (a,b) is 1–3 mm.
Sensors 25 05829 g008
Figure 9. Model for evaluating accuracy and measurement positions. CLCFM3 is processed 10 times individually, and the distances of measurement positions (1)–(14) are measured.
Figure 9. Model for evaluating accuracy and measurement positions. CLCFM3 is processed 10 times individually, and the distances of measurement positions (1)–(14) are measured.
Sensors 25 05829 g009
Figure 10. Image-matching process at the boundary and the point removal process in MMM: (A) Process of mismatching. (B) Difference between correct point and mismatched point. (C) Point removal process using masked images derived from surrounding viewpoint images (Masked Matching).
Figure 10. Image-matching process at the boundary and the point removal process in MMM: (A) Process of mismatching. (B) Difference between correct point and mismatched point. (C) Point removal process using masked images derived from surrounding viewpoint images (Masked Matching).
Sensors 25 05829 g010
Figure 11. Camera position estimation failure. There is a risk of camera position estimation failure due to the accumulation of camera position estimation errors for all-around images taken at short intervals in SfM: (A) Without CLCFM and pillar. (B) With CLCFM and pillar.
Figure 11. Camera position estimation failure. There is a risk of camera position estimation failure due to the accumulation of camera position estimation errors for all-around images taken at short intervals in SfM: (A) Without CLCFM and pillar. (B) With CLCFM and pillar.
Sensors 25 05829 g011
Table 1. Comparison of the number of point clouds categorized by distance ranges.
Table 1. Comparison of the number of point clouds categorized by distance ranges.
Range (mm)Number of Points%
0–31,732,06695.7
3–78,6374.3
Total1,810,703100.0
Table 2. Evaluation of measurement accuracy by the proposed pipeline. Caliper values are obtained by measuring the length 10 times using a caliper for (1)–(13) and a ruler for (14), and then calculating the average. CLCFM3 values are obtained using the point cloud data obtained using CLCFM3. “Difference” is the difference in distance between the caliper and CLCFM3.
Table 2. Evaluation of measurement accuracy by the proposed pipeline. Caliper values are obtained by measuring the length 10 times using a caliper for (1)–(13) and a ruler for (14), and then calculating the average. CLCFM3 values are obtained using the point cloud data obtained using CLCFM3. “Difference” is the difference in distance between the caliper and CLCFM3.
Measurement PositionCaliper
(mm)
CLCFM3Distance
(mm)
NumberDirectionAvg(mm)SD
(1)width124.24124.3520.0000314−0.112
(2)width84.5184.6510.0000346−0.141
(3)height120.05120.1040.0000271−0.054
(4)height69.1569.1600.0000068−0.010
(5)height63.1163.1510.0000115−0.041
(6)height42.4542.3560.00001100.094
(7)height156.36156.3570.00002600.003
(8)width131.41131.1300.00002250.280
(9)width133.6133.5090.00001760.091
(10)width44.2744.3560.0000099−0.086
(11)height101.05101.1830.0000201−0.133
(12)width47.647.7930.0000067−0.193
(13)height144.38144.5790.0000191−0.199
(14)height593593.9700.0001225−0.970
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hayashi, A.; Kochi, N.; Kodama, K.; Isobe, S.; Tanabata, T. CLCFM3: A 3D Reconstruction Algorithm Based on Photogrammetry for High-Precision Whole Plant Sensing Using All-Around Images. Sensors 2025, 25, 5829. https://doi.org/10.3390/s25185829

AMA Style

Hayashi A, Kochi N, Kodama K, Isobe S, Tanabata T. CLCFM3: A 3D Reconstruction Algorithm Based on Photogrammetry for High-Precision Whole Plant Sensing Using All-Around Images. Sensors. 2025; 25(18):5829. https://doi.org/10.3390/s25185829

Chicago/Turabian Style

Hayashi, Atsushi, Nobuo Kochi, Kunihiro Kodama, Sachiko Isobe, and Takanari Tanabata. 2025. "CLCFM3: A 3D Reconstruction Algorithm Based on Photogrammetry for High-Precision Whole Plant Sensing Using All-Around Images" Sensors 25, no. 18: 5829. https://doi.org/10.3390/s25185829

APA Style

Hayashi, A., Kochi, N., Kodama, K., Isobe, S., & Tanabata, T. (2025). CLCFM3: A 3D Reconstruction Algorithm Based on Photogrammetry for High-Precision Whole Plant Sensing Using All-Around Images. Sensors, 25(18), 5829. https://doi.org/10.3390/s25185829

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop