High-Resolution 3D Crop Reconstruction and Automatic Analysis of Phenotyping Index Using Machine Learning

Beyond the use of 2D images, the analysis of 3D images is also necessary for analyzing the phenomics of crop plants. In this study, we configured a system and implemented an algorithm for the 3D image reconstruction of red pepper plant (Capsicum annuum L.), as well as its automatic analysis. A Kinect v2 with a depth sensor and a high-resolution RGB camera were used to obtain more accurate reconstructed 3D images. The reconstructed 3D images were compared with conventional reconstructed images, and the data of the reconstructed images were analyzed with respect to their directly measured features and accuracy, such as leaf number, width, and plant height. Several algorithms for image extraction and segmentation were applied for automatic analysis. The results showed that the proposed method showed an error of about 5 mm or less when reconstructing and analyzing 3D images, and was suitable for phenotypic analysis. The images and analysis algorithms obtained by the 3D reconstruction method are expected to be applied to various image processing studies.


Introduction
Recently, research on convergence studies based on information and communication technology (ICT) has been actively conducted in order to improve the competitiveness of agriculture and to solve the rapidly developing problem of climate change. With the development of fusion technology for crop analysis, the importance of plant phenomics that observes and analyzes the entire phenotype, including crop shape and biochemical characteristics, is emerging [1]. Plant phenotyping to express complex crop characteristics extends the accuracy of the analysis and the range of applications by using a variety of sensor techniques such as visible light, thermal imaging, hyperspectral, and depth cameras [2].
In addition to sensor technology, a variety of algorithms, especially machine learning, have enabled a variety of analyzes of plant phenotypes. Leaf area and nutrient concentration have been identified through the use of plant images and pattern recognition methods [3,4]. These studies were developed for the purpose of automation. Using deep learning, one of the machine learning algorithms, observations are being made in real fields by analyzing growth status and environmental information [5], and research applied to identification of features and location is in progress [6]. In terms of data collection, several machine learning algorithms are being used for automated analysis such as disease severity, lodging assessment, and growth monitoring for a large area of crops using an Unmanned Aerial Vehicle (UAV) [7][8][9].
One important factor for determining plant phenotype is the shape of crops, referred to as morphology. The shape includes the crop height, leaf area, fruit volume, weight, and direction of growth [10]. These factors reflect both environmental factors such as temperature, humidity, and soil, and the trait elements of the crop itself [11]. Crops may grow toward the light, or environmental stress may cause a lower fruit yield or smaller fruits than the natural size due to genetic factors. Two-dimensional (2D) images have been used to anlyze these various crop shapes. However, observing three-dimensional (3D) objects as 2D cross-section images is difficult, and there are limitations to the analysis due to the portion covered by leaves.
To overcome this problem, crop analysis using 3D images has been attempted, and various studies are underway. Three-dimensional images are created through a process called 3D reconstruction, methods of which can be divided into passive and active methods. Passive methods are slower than active ones, but data acquisition is easier and highresolution analysis is possible. Meanwhile, active methods have a lower resolution than passive ones, but real-time analysis can be performed as a result of the acquisition of accurate depth information.
The passive type is an image-based 3D modeling method using only images acquired by a normal RGB camera without a distance measuring sensor. Typically, the structure from motion (SfM) method is used, and there are various methods to achieve this, such as shape from silhouettes and shape from shading. Snavely et al. [12] encountered many issues when working with photo tourism, restoring images of historic locations that people had photographed under various conditions using the SfM method. Li et al. [13] briefly introduced and summarized an algorithm for multi-image 3D reconstruction. Remondino et al. [14] compared the results of dense image matching algorithms (SURE, Micmac, PMVS, Photoscan) with various data sets captured using a DSLR camera. Some studies have applied such algorithms to crops. Lou et al. [15] estimated camera poses using several featurematching algorithms with images taken with DSLR and performed point cloud generation using the Multi-View Stereo (MVS) method. Similarly, Lou et al. [16] determined that a 3D laser/LiDAR or structured-light scanner is not suitable for plants and compared their algorithms using SfM, stereo matching, depth optimization and merging with Patch-based Multi-View Stereo (PMVS) and CMPMVS [17] algorithms. Ni et al. [18] performed 3D reconstruction using software called visualSfM and tried to use voxels to obtain the volume. Nguyen et al. [19] photographed sunflowers using a stereo camera in an outdoor environment and performed 3D reconstruction using the SfM method. The results were compared with other algorithms in terms of error.
The active type is a depth-based 3D modeling method that uses expensive equipment employing sensors capable of distance measurement. Structured-light, laser scanning, and time-of-flight (TOF) cameras have been used, and many studies have been carried out as a result of the development of inexpensive RGB with depth (RGB-D) sensors. Q. Yang et al. [20] introduced an algorithm that can apply depth image upsampling, mentioning that the passive stereo method fails in the non-textured featureless portion. Zhu et al. [21] and Lee and Ho [22] increased the accuracy of the depth map and created a multi-view video sequence and a depth map using a combination of a TOF range sensor and a stereo sensor. Paulus et al. [23] measured sugar beet using relatively inexpensive equipment such as the DAVID laser scanning system and showed that it can substitute expensive equipment through comparison with reference values. Wasenmüller and Stricker [24] and Yang et al. [25] analyzed the accuracy and precision of a Kinect sensor, a type of RGB-D camera, for temperature effect, color, and depth value. Lachat et al. [26] performed the 3D reconstruction of a small object using a Kinect v2 and compared this result with that of Kinect v1 and the photogrammetry method using a DSLR. Gai et al. [27] and Gai et al. [28] attempted plant detection and the localization of crops such as broccoli and lettuce using a Kinect sensor for weed management. Several clustering and classification algorithms were used to perform plant localization and discrimination for different crop growth stages. In addition, this type of sensor was used for research to acquire various plant growth-related information [29][30][31].
In this study, several steps were performed to obtain more accurate 3D crop reconstruction images in the laboratory, and various methods were applied for automatic analysis. After acquiring depth information using a Kinect v2, which is an RGB-D sensor, we attempted to assist the 3D reconstruction process by performing matching with Agriculture 2021, 11, 1010 3 of 22 a high-resolution RGB image. To be able to use the color image and the depth image together, a color image is matched to a depth image with a relatively small resolution. In this case, a lot of information in the color image is discarded, and it is difficult to perform precise analysis. Accordingly, an algorithm that increases the resolution of the depth image was used to reduce the amount of discarded color information. With respect to obtaining an image, a process to increase the accuracy of 3D reconstruction was also added. The results were compared with those of existing 3D reconstruction algorithms, and model accuracy analysis was performed on the ground-truth data of the directly measured crops for morphology characteristics. Similar to the component that creates the 3D image, the process of automatic analysis is an important aspect. 2D image analysis has been automated in various areas, but in the case of 3D images, there is no sufficient analysis technology. Accordingly, an automated analysis algorithm process for the reconstructed 3D crop image was also developed. Through the classification and segmentation of the image, the desired crop part was identified and divided based on a specific part. Automatic analysis was performed on major crop factors, including crop growth direction, number of leaves, leaf length, width, area, and crop height. Through this, it was possible to reconstruct 3D images using various types of images and to perform accuracy analysis on the model. In addition, for automatic analysis, an extraction algorithm suitable for 3D crop images was selected and various attempts at segmentation were made. This study contributes to broadening the scope of analysis by presenting various plant characteristics and the applicability of sensors and algorithms for 3D information in plant phenotypic analysis.

Crop and Experimental Environment
The crops used in the experiment were the common and Cheongyang red pepper (Capsicum annuum L.), which is one of the spiciest varieties of red pepper cultivated in the Republic of Korea. Seedlings were 17-25 cm. Red pepper was selected as the target crop because it responds quickly to artificial effects and the differentiation of leaves from stems is relatively clear. Images of a specific period were acquired, and 3D images were created and used for analysis. Figure 1 shows the photography environment. A turntable was placed in a space made of a particular color background and the experimental crop was rotated. There are several advantages to rotating crops in this way rather than the cameras. Taking pictures takes up less space, and configuring the experimental environment is cheap and easy. Although the rotating crop may shake due to the direct crop rotation, this method was adopted because it can reduce swing through the stable rotation of the turntable and small shaking does not significantly affect the actual image analysis. It takes about 47 s for a crop to complete one full rotation on the turntable and the Kinect and high-resolution RGB camera took images simultaneously at regular intervals while the crop rotated. Each camera acquired 60 images per crop, and shot from two different angles-a side view and a diagonal top view-to improve accuracy. At the time of photographing, an attempt was made to increase the accuracy of restoration by adding objects with features along with crops. Two types of backgrounds were used to examine the effect of background on 3D crop reconstruction: white and black.

Image Sensor
For 3D reconstruction, a sensor capable of acquiring depth information and a sensor capable of capturing high-resolution images were used. A Kinect v2 sensor (Microsoft Corp., Redmond, WA, USA) was used to acquire depth information. The Kinect v2 is a lowcost RGB-D camera developed by Microsoft that allows users to acquire RGB information and depth/infrared (IR) images. The Kinect v2 has greatly increased color resolution compared to the Kinect v1 and a different depth measurement method. The Kinect v1 uses a structured light method called light coding that acquires depth information through pattern changes in the projected pattern while the Kinect v2 gets the depth value using the TOF method that calculates the distance by measuring the return time for the projected light [25]. In addition to the new measurement method, the Kinect v2 has better depth fidelity (3×) and accuracy than the Kinect v1 in the color images [32]. These Kinect sensors are being used for research in many fields because of their low cost and easy sensor recognition and development through the Kinect SDK provided by Microsoft. Table 1 shows the specifications of the Kinect v2. Although each Kinect includes a color sensor, a smartphone camera (iPhone 6 Plus, Apple Inc., Cupertino, CA, USA) was used to acquire additional high-resolution images for more accurate 3D reconstruction. The camera's image resolution is 3264 × 2448 pixels, and it has features such as ƒ/2.2 image, hybrid IR filter, optical image stabilization, and exposure control.

Image Sensor
For 3D reconstruction, a sensor capable of acquiring depth information and a sensor capable of capturing high-resolution images were used. A Kinect v2 sensor (Microsoft Corp., Redmond, WA, USA) was used to acquire depth information. The Kinect v2 is a low-cost RGB-D camera developed by Microsoft that allows users to acquire RGB information and depth/infrared (IR) images. The Kinect v2 has greatly increased color resolution compared to the Kinect v1 and a different depth measurement method. The Kinect v1 uses a structured light method called light coding that acquires depth information through pattern changes in the projected pattern while the Kinect v2 gets the depth value using the TOF method that calculates the distance by measuring the return time for the projected light [25]. In addition to the new measurement method, the Kinect v2 has better depth fidelity (3×) and accuracy than the Kinect v1 in the color images [32]. These Kinect sensors are being used for research in many fields because of their low cost and easy sensor recognition and development through the Kinect SDK provided by Microsoft. Table 1 shows the specifications of the Kinect v2. Although each Kinect includes a color sensor, a smartphone camera (iPhone 6 Plus, Apple Inc., Cupertino, CA, USA) was used to acquire additional high-resolution images for more accurate 3D reconstruction. The camera's image resolution is 3264 × 2448 pixels, and it has features such as ƒ/2.2 image, hybrid IR filter, optical image stabilization, and exposure control.  [33].

Measurement Method and Application Software
To examine the model accuracy of the reconstructed 3D image, the height of seedlings and the length and width of all leaves were analyzed. Since the height of crops is related to growth and the size of leaves plays an important role in the photosynthetic process, it was used as an index of plant quality. The crop height was the distance from the pot to the highest part of the seedling and the length and width of the leaves were measured ( Figure 2). Each leaf was assigned a number from the bottom and used for analysis. This was called a 'location number'. When measuring the leaf width, the maximum straight-line distance was measured if the leaf was curved. related to growth and the size of leaves plays an important role in the photosynthetic process, it was used as an index of plant quality. The crop height was the distance from the pot to the highest part of the seedling and the length and width of the leaves were measured ( Figure 2). Each leaf was assigned a number from the bottom and used for analysis. This was called a 'location number'. When measuring the leaf width, the maximum straight-line distance was measured if the leaf was curved. The size of each 3D image was analyzed using MeshLab software [34]. The artificially segmented image was measured using MeshLab measurement tool and corrected values were compared with actual measured values. Kinect software development kit, C++, OpenCV library, GMLC++ camera calibration toolbox, and 3D restoration commercial software were used for 3D reconstruction. In addition, Python and MATLAB were used for the segmentation and measurement parts related to automatic analysis of 3D images.

Algorithm for High-Resolution 3D Reconstruction
Agisoft Photoscan (Agisoft LLC, St.Petersburg, Russia) [35] and visualSfM [36] are representative 3D reconstruction software packages. However, when the target crop was rotated in this experiment, modeling was often performed improperly. Several algorithms were investigated to overcome this problem and obtain high-resolution 3D model data. Figure 3 shows the flowchart for image acquisition, 3D reconstruction, and analysis. First, depth images and high-resolution RGB images are acquired using the Kinect and an additional smartphone camera mentioned above. Since the acquired depth image has a low resolution, upsampling was performed to overcome this. The resultant upsampled images were matched by camera calibration and image registration with high-resolution images. Matched depth images and color images were used to acquire 3D point cloud data by 3D reconstruction. The 3D crop image data was segmented for each height and leaf manually and the model's accuracy was analyzed using the actual measured data. The size of each 3D image was analyzed using MeshLab software [34]. The artificially segmented image was measured using MeshLab measurement tool and corrected values were compared with actual measured values. Kinect software development kit, C++, OpenCV library, GMLC++ camera calibration toolbox, and 3D restoration commercial software were used for 3D reconstruction. In addition, Python and MATLAB were used for the segmentation and measurement parts related to automatic analysis of 3D images.

Algorithm for High-Resolution 3D Reconstruction
Agisoft Photoscan (Agisoft LLC, St.Petersburg, Russia) [35] and visualSfM [36] are representative 3D reconstruction software packages. However, when the target crop was rotated in this experiment, modeling was often performed improperly. Several algorithms were investigated to overcome this problem and obtain high-resolution 3D model data. Figure 3 shows the flowchart for image acquisition, 3D reconstruction, and analysis. First, depth images and high-resolution RGB images are acquired using the Kinect and an additional smartphone camera mentioned above. Since the acquired depth image has a low resolution, upsampling was performed to overcome this. The resultant upsampled images were matched by camera calibration and image registration with high-resolution images. Matched depth images and color images were used to acquire 3D point cloud data by 3D reconstruction. The 3D crop image data was segmented for each height and leaf manually and the model's accuracy was analyzed using the actual measured data.
The Kinect's depth image is very low resolution. This low-resolution image can cause problems such as data loss and inaccurate model generation in subsequent analysis. Image upsampling was performed to overcome this problem. The obtained Kinect depth image was subjected to upsampling using Anisotropic Total Generalized Variation [37]. This algorithm formulated a convex optimization problem using higher order regularization for depth image upsampling. In the optimization process, an anisotropic diffusion tensor calculated from a high-resolution intensity image was used. During the upsampling process, it showed better results than other algorithms with respect to data loss and inappropriate expansion. Excessive image enlargement through upsampling shows the disadvantage of producing inaccurate information rather than the advantage of resolution. Therefore, only twice magnification was used for the width and height.  The Kinect's depth image is very low resolution. This low-resolution image c problems such as data loss and inaccurate model generation in subsequent anal age upsampling was performed to overcome this problem. The obtained Kinect d age was subjected to upsampling using Anisotropic Total Generalized Variation [ algorithm formulated a convex optimization problem using higher order regula for depth image upsampling. In the optimization process, an anisotropic diffusio calculated from a high-resolution intensity image was used. During the upsamp cess, it showed better results than other algorithms with respect to data loss and priate expansion. Excessive image enlargement through upsampling shows th vantage of producing inaccurate information rather than the advantage of re Therefore, only twice magnification was used for the width and height.
The RGB, IR sensor of the Kinect and the high-resolution RGB camera canno to photograph from exactly the same position, so the image information does no Even if the picture was taken from the same location, the part that is distorted d characteristics of the sensor should be considered. Solving this problem require calibration and image registration to be performed. Camera calibration is a metho taining camera parameters, and the image registration process helps express diff ordinate systems in one coordinate system [38,39]. Camera parameters consist of parameters indicating the camera's internal structure and extrinsic parameters in camera direction and position. The intrinsic parameters include the focal length o the pixel size of the sensor, and the radial distortion factor. A size-aware check and GML C++ Camera Calibration toolbox were used to obtain these camera par The checkerboard was repeatedly photographed at various angles to extract m points, thereby deriving the camera parameter values. Several unique points are a ically extracted from each camera image, and knowing the distance between th makes it possible to estimate internal and external parameters. In general, when many images, the estimation is improved, but usually about 20 images is sufficien purposes of calibration.

Machine Learning-Based Crop Extraction and Segmentation Algorithm
In order to acquire a 3D image of the crop using the 3D reconstruction algor The RGB, IR sensor of the Kinect and the high-resolution RGB camera cannot be used to photograph from exactly the same position, so the image information does not match. Even if the picture was taken from the same location, the part that is distorted due to the characteristics of the sensor should be considered. Solving this problem requires camera calibration and image registration to be performed. Camera calibration is a method of obtaining camera parameters, and the image registration process helps express different coordinate systems in one coordinate system [38,39]. Camera parameters consist of intrinsic parameters indicating the camera's internal structure and extrinsic parameters indicating camera direction and position. The intrinsic parameters include the focal length of the lens, the pixel size of the sensor, and the radial distortion factor. A size-aware checkerboard and GML C++ Camera Calibration toolbox were used to obtain these camera parameters. The checkerboard was repeatedly photographed at various angles to extract matching points, thereby deriving the camera parameter values. Several unique points are automatically extracted from each camera image, and knowing the distance between the points makes it possible to estimate internal and external parameters. In general, when there are many images, the estimation is improved, but usually about 20 images is sufficient for the purposes of calibration.

3D Crop Extraction
In order to acquire a 3D image of the crop using the 3D reconstruction algorithm, an attempt was made to develop an automated analysis technique to eliminate errors in the artificial part resulting from human intervention and to increase the reproducibility of the analysis. An algorithm suitable for the reconstructed image was applied for image extraction, which is the process of acquiring the desired part of the image, and image segmentation for dividing it into specific parts. Figure 4 shows the overall flow of the 3D reconstruction process and the further analysis of the 3D image. In Figure 3, all processes after '3D image' were performed manually, but in this section, automatic analysis was performed through the application of various algorithms. tation for dividing it into specific parts. Figure 4 shows the overall flow of the 3D recon struction process and the further analysis of the 3D image. In Figure 3, all processes afte '3D image' were performed manually, but in this section, automatic analysis was pe formed through the application of various algorithms. As a result of performing 3D reconstruction of the crop using the above algorithm, point cloud with 3D coordinates could be obtained. Each point contains coordinate value and color information in 3D space. The reconstructed 3D image contains a small amoun of noise information generated during the reconstruction process. It contains backgroun information that is not the target crop, or it might be reconstructed in the wrong location These parts correspond to various sources of noise that interfere with the analysis an should be removed. In addition, although it cannot be seen as noise, it is necessary t separate out parts that are not related to the morphology of the crop, such as potted plan and soil. In addition, when several 3D-reconstructed crop images are drawn in one coo dinate system, they are expressed in various positions and sizes. That is, since the coord nate system and the image scale of the image appear differently depending on the recon struction process, a correction needs to be performed.
Image extraction to remove parts other than crops, which is the part to be analyzed started by identifying the characteristics of the data. The 3D point cloud data contain only coordinate and color information, as previously discussed. With this data, it is no possible to distinguish which points correspond to the crop. That is, since the label, whic is the explicit correct answer for the 3D image, is not known, the unsupervised learnin technique was mainly performed. After applying several machine learning methods, k means clustering was found to be simple but suitable, and additionally, Otsu's threshol and vegetation index were used together.
K-means clustering is an algorithm that divides a given data set into k groups, an this group is called a cluster [40]. Because it groups similar types of data together, it mainly used to discover hidden structures or features in data. The k centroids are assigne to a random location for data on a 2D plane, and a cluster of data is determined accordin to various distance measurement methods. The centroids are updated using the data be longing to each cluster and repeated until there is no change. Through this process, dat can be divided into any number of groups according to the initially specified k. In the 3 crop point cloud, it was used to perform a classification according to color information.
Otsu's thresholding technique is widely used to process image thresholds [41]. Bina rization is the simplest case for separating objects to get information from images. As fo what criteria to separate, Otsu provides criteria for setting thresholds using statistic methods. The vegetation index is an index indicating the vitality and growth status o As a result of performing 3D reconstruction of the crop using the above algorithm, a point cloud with 3D coordinates could be obtained. Each point contains coordinate values and color information in 3D space. The reconstructed 3D image contains a small amount of noise information generated during the reconstruction process. It contains background information that is not the target crop, or it might be reconstructed in the wrong location. These parts correspond to various sources of noise that interfere with the analysis and should be removed. In addition, although it cannot be seen as noise, it is necessary to separate out parts that are not related to the morphology of the crop, such as potted plants and soil. In addition, when several 3D-reconstructed crop images are drawn in one coordinate system, they are expressed in various positions and sizes. That is, since the coordinate system and the image scale of the image appear differently depending on the reconstruction process, a correction needs to be performed.
Image extraction to remove parts other than crops, which is the part to be analyzed, started by identifying the characteristics of the data. The 3D point cloud data contains only coordinate and color information, as previously discussed. With this data, it is not possible to distinguish which points correspond to the crop. That is, since the label, which is the explicit correct answer for the 3D image, is not known, the unsupervised learning technique was mainly performed. After applying several machine learning methods, k-means clustering was found to be simple but suitable, and additionally, Otsu's threshold and vegetation index were used together.
K-means clustering is an algorithm that divides a given data set into k groups, and this group is called a cluster [40]. Because it groups similar types of data together, it is mainly used to discover hidden structures or features in data. The k centroids are assigned to a random location for data on a 2D plane, and a cluster of data is determined according to various distance measurement methods. The centroids are updated using the data belonging to each cluster and repeated until there is no change. Through this process, data can be divided into any number of groups according to the initially specified k. In the 3D crop point cloud, it was used to perform a classification according to color information.
Otsu's thresholding technique is widely used to process image thresholds [41]. Binarization is the simplest case for separating objects to get information from images. As for what criteria to separate, Otsu provides criteria for setting thresholds using statistical methods. The vegetation index is an index indicating the vitality and growth status of plants, and is used to emphasize vegetation characteristic information. By using this, it can be used as an index that can easily distinguish between crop parts and non-crop parts. The data that can use the vegetation index in the 3D crop image contains information in the visible light band as color values. Accordingly, excess green index (ExG), an index mainly used for analysis through RGB images, was used [42]. As in Equation (1), R, G, and B color channels are used, and the normalized value of each channel divided by the maximum value (R m , G m , B m ) is used. At this time, weight is given to green.

3D Crop Segmentation and Automatic Analysis
Various algorithms for correction and automatic analysis were applied to the 3D image of the crop part separated through the image extraction algorithm. First, work was carried out to match the coordinate system of several 3D crop images. Coordinate transformation was performed considering the direction of each crop. In the case of scale, it is assumed that there is a certain error in the entire 3D image, and correction was performed using a reference object, the size of which was known. The main algorithms used for this were Principal Component Analysis (PCA) and Euler Angle Transformation.
PCA is an algorithm that finds components that are judged to explain the given data well [43]. The data to be analyzed have many features. The large number of features means that there are many data dimensions, and an increase in the dimensions slows down the learning process, increasing the likelihood of poor performance. Additionally, features include elements that are important to the analysis, as well as elements that are not. Accordingly, it is necessary to reduce the dimensions, and the representative algorithm for this is PCA. PCA performs an algorithm that maximizes data variance. If the PCA algorithm is performed on the given two-dimensional data, PC1, which is judged to be the most important, and PC2, which is the next, can be obtained. In addition to PCA, there are dimensional reduction algorithms such as Linear Discriminant Analysis (LDA) and Independent Component Analysis (ICA), and algorithms are used according to the target direction.
Euler Angle Transformation through Euler Angles and Rotation Matrix can be performed to perform the rotation of objects [44]. The rotation matrix is obtained by using the direction cosine formed by each axis of the reference coordinate system and the rotation coordinate system, and this is multiplied by the coordinate matrix. The conversion equation is as shown in Equation (2), and can be configured by using the rotation angle for each axis where R is rotation matrix, sθ = sinθ and cθ = cosθ. Next, to analyze the characteristics of the crop, the 3D crop image should be divided into stems and leaves. The partitioning was attempted using a clustering algorithm. Appropriate parameters were set according to the characteristics of the target stem and leaf, and analysis was performed on a satisfactory individual. In addition to automatic measurement of crop height, automatic calculation of leaf length and width, which are phenotyping features related to leaves, was performed on the image, where stems and leaves were segmented. The measurement criteria were implemented in the same manner as in the 3D image reconstruction accuracy.
As the crop coordinate transformation is performed, the height of crop can be measured as the maximum value of the image z-axis coordinate. Compared to heights where simple measurements are possible, values for leaves require additional processing. The crop coordinate transformation algorithm was additionally performed for the divided leaves. The leaves that performed the PCA algorithm were transformed on the basis of an axis similar to the length vector and an axis similar to the width vector. The length of the leaf was measured by the length of two points with maximum Euclidean distance. The width is measured as the Euclidean distance between the minimum and maximum point coordinates of the axis corresponding to the width. The calculation for each part is as shown in Equation (3).
where z is z-axis coordinate value, p i is i-th point, p l and p r are maximum (right) and minimum (left) coordinates on the same axis. Figure 5 shows the results of performing 3D reconstruction. With respect to the method by which images were obtained in this study, the program using the existing SfM method did not show the 3D reconstruction results; therefore, the results could not be included as comparison images. The Kinect Fusion Explorer on the Kinect SDK provided the overall appearance, but did not seem to display the proper 3D objects. Only the large leaves and potted plants were implemented in 3D form, while the rest were either distorted or appeared in 2D. While it was possible to acquire more accurate 3D objects with low-resolution Kinect color images than with the Kinect Fusion Explorer, it is difficult to perform this analysis. Many leaves were not fully shaped, and were heavily influenced by noise. Finally, using high-resolution RGB images showed that the 3D reconstruction demonstrated good performance overall, except for small parts.

High-Resolution 3D Reconstruction Results
The accuracy of the 3D model was analyzed using the high-resolution 3D reconstruction result image. The number of leaves, the crop height, and the size of leaves were compared with the ground-truth data. Table 2 shows a comparison of the true values with the reconstructed 3D image. Four seedlings were photographed twice to obtain eight 3D models. In each image, about two leaves have not been reconstructed whatsoever, and few of the very small leaves are depicted.   Table 3 shows the degree of error between the ground-truth data and the measured values in each plant's 3D reconstruction models. In total, 104 leaves were used for the analysis of the eight 3D models obtained from the four seedlings. The was analyses were divided into whole data and each sample. The error for each seedling was at a similar level, which was attributed to the seedlings being at similar growth stages. Absolute errors were high with respect to leaf width, leaf length, and plant height, depending on the size. The percentage error was the highest for leaf width, which is presumed to have caused several errors due to the method of measuring the straight-line distance of bent leaves.   Table 3 shows the degree of error between the ground-truth data and the measured values in each plant's 3D reconstruction models. In total, 104 leaves were used for the analysis of the eight 3D models obtained from the four seedlings. The was analyses were divided into whole data and each sample. The error for each seedling was at a similar level, which was attributed to the seedlings being at similar growth stages. Absolute errors were high with respect to leaf width, leaf length, and plant height, depending on the size. The percentage error was the highest for leaf width, which is presumed to have caused several errors due to the method of measuring the straight-line distance of bent leaves. Table 3. Absolute error, percentage error, and standard deviation between the ground-truth data and the measured values in 3D reconstruction models.

Crop
Error ( Overall, the absolute error was low but the characteristics that appeared according to the position or size of the object were further analyzed. Therefore, to understand the underlying cause of errors, an analysis was performed with respect to the location of leaves. The total percentage error value was used to compare the length and width of leaves together. Then, percentage errors were compared for leaves with the same location number (Figure 6). Each plant has a different number of leaves. The red pepper used in the experiment had at least 12~15 leaves. If the location number were greater than 13, the number of data points would be different, and the same location analysis would be difficult. Therefore, analysis was only performed until the 12th leaf. On the graph, each location number should contain eight data points corresponding to four red peppers and two backgrounds. However, since the leaves that did not have a complete 3D shape were excluded, fewer actual points were displayed. The percentage errors were high at the top and bottom of the plant, particularly at the top. Overall, the absolute error was low but the characteristics that appeared according to the position or size of the object were further analyzed. Therefore, to understand the underlying cause of errors, an analysis was performed with respect to the location of leaves. The total percentage error value was used to compare the length and width of leaves together. Then, percentage errors were compared for leaves with the same location number (Figure 6). Each plant has a different number of leaves. The red pepper used in the experiment had at least 12~15 leaves. If the location number were greater than 13, the number of data points would be different, and the same location analysis would be difficult. Therefore, analysis was only performed until the 12th leaf. On the graph, each location number should contain eight data points corresponding to four red peppers and two backgrounds. However, since the leaves that did not have a complete 3D shape were excluded, fewer actual points were displayed. The percentage errors were high at the top and bottom of the plant, particularly at the top. On the basis of the results of the analysis of location number, it was deemed necessary to analyze leaf size. The length, width, and relative areas of leaves were used as indices for leaf size. Since the relative area is composed of a simple product, it is different from the actual leaf area, and the calculation formula for this is as shown in Equation (4): Relative area = Leaf length × Leaf width (4) Figures 7 and 8 show the percentage error for the relative area, length, and width of leaves; the smaller the relative area, the greater the percentage error in both length and width. Similar results were obtained when the length and width were lower, respectively. This means that the smaller the leaves, the more likely it is that reconstruction will fail or that the modeling results will be less complete.  On the basis of the results of the analysis of location number, it was deemed necessary to analyze leaf size. The length, width, and relative areas of leaves were used as indices for leaf size. Since the relative area is composed of a simple product, it is different from the actual leaf area, and the calculation formula for this is as shown in Equation (4): Relative area = Leaf length × Leaf width (4) Figures 7 and 8 show the percentage error for the relative area, length, and width of leaves; the smaller the relative area, the greater the percentage error in both length and width. Similar results were obtained when the length and width were lower, respectively. This means that the smaller the leaves, the more likely it is that reconstruction will fail or that the modeling results will be less complete. Overall, the absolute error was low but the characteristics that appeared according to the position or size of the object were further analyzed. Therefore, to understand the underlying cause of errors, an analysis was performed with respect to the location of leaves. The total percentage error value was used to compare the length and width of leaves together. Then, percentage errors were compared for leaves with the same location number (Figure 6). Each plant has a different number of leaves. The red pepper used in the experiment had at least 12~15 leaves. If the location number were greater than 13, the number of data points would be different, and the same location analysis would be difficult. Therefore, analysis was only performed until the 12th leaf. On the graph, each location number should contain eight data points corresponding to four red peppers and two backgrounds. However, since the leaves that did not have a complete 3D shape were excluded, fewer actual points were displayed. The percentage errors were high at the top and bottom of the plant, particularly at the top. On the basis of the results of the analysis of location number, it was deemed necessary to analyze leaf size. The length, width, and relative areas of leaves were used as indices for leaf size. Since the relative area is composed of a simple product, it is different from the actual leaf area, and the calculation formula for this is as shown in Equation (4): Relative area = Leaf length × Leaf width (4) Figures 7 and 8 show the percentage error for the relative area, length, and width of leaves; the smaller the relative area, the greater the percentage error in both length and width. Similar results were obtained when the length and width were lower, respectively. This means that the smaller the leaves, the more likely it is that reconstruction will fail or that the modeling results will be less complete.

3D Crop Extraction and Coordinate Transformation
Image extraction was performed by combining several algorithms for the given 3D crop image. Several clustering algorithms were used as unsupervised learning techniques for image extraction. Mean-shift clustering, density-based spatial clustering of applications with noise (DBSCAN), and k-means clustering with differences in k values were compared. In the case of mean-shift clustering, the size of the radius had to be determined, but there was a difficulty in automatic analysis, because it needed to be changed in consideration of the size of the crop. In the case of DBSCAN, there was an advantage in that there were no parameters to be determined in advance, but the extraction itself was not performed properly in many cases. Finally, for k-means clustering, it was necessary to determine the number of classes, k, and although random elements were included in the extraction, the performance was superior to other algorithms. Because it was extraction rather than image segmentation, it was found that the number of classes could be easily pre-determined for both the target crop part and the other parts, and as a result, fewer random elements appeared. Therefore, in subsequent analysis, k-means clustering was used, as it exhibited the best performance. When performing image extraction by combining several additional algorithms, this was confirmed, as shown in Figure 9. When only the color information of the 3D point cloud and k-means clustering were used, the extraction was not performed properly ( Figure 9A). Various extractions were possible, depending on the number of clusters, but four-classified into crops, pot, sand, and noiseshowed the most optimal results. Nevertheless, various sources of noise remained in the crop part, and various data confirmed that other clusters had not been clearly separated. In addition, because the algorithm contained a random element, we confirmed the instability, with the result being different each time the algorithm was executed. To solve this problem, k-means clustering was performed using the vegetation index value rather than the simple color of the 3D point cloud as shown in Figure 9B. Here, it was performed with three clusters corresponding to the crop, pot, and the rest. Compared to the results of performing only clustering, the crop part was clearly segmented, and other parts were also separated depending on purpose. In the case of configuring the four clusters, the data with respect to the pot and the soil are incomplete, but it can be confirmed that the data were distinguished. Figure 9C shows how to not use clustering. Color information was applied to Otsu's thresholds. The method employed simple color values and vegetation index for color information. Since the part to be extracted was a crop, it was found that the performance was better when the vegetation index was applied. Accordingly, the vegetation index of ExG was applied to Otsu's thresholds and analyzed. Because it was a process of binarization, it was divided into crop and non-crop parts. Using this method, it is also possible to extract the crop by properly excluding noise and other parts. Both

3D Crop Extraction and Coordinate Transformation
Image extraction was performed by combining several algorithms for the given 3D crop image. Several clustering algorithms were used as unsupervised learning techniques for image extraction. Mean-shift clustering, density-based spatial clustering of applications with noise (DBSCAN), and k-means clustering with differences in k values were compared. In the case of mean-shift clustering, the size of the radius had to be determined, but there was a difficulty in automatic analysis, because it needed to be changed in consideration of the size of the crop. In the case of DBSCAN, there was an advantage in that there were no parameters to be determined in advance, but the extraction itself was not performed properly in many cases. Finally, for k-means clustering, it was necessary to determine the number of classes, k, and although random elements were included in the extraction, the performance was superior to other algorithms. Because it was extraction rather than image segmentation, it was found that the number of classes could be easily pre-determined for both the target crop part and the other parts, and as a result, fewer random elements appeared. Therefore, in subsequent analysis, k-means clustering was used, as it exhibited the best performance. When performing image extraction by combining several additional algorithms, this was confirmed, as shown in Figure 9. When only the color information of the 3D point cloud and k-means clustering were used, the extraction was not performed properly ( Figure 9A). Various extractions were possible, depending on the number of clusters, but four-classified into crops, pot, sand, and noise-showed the most optimal results. Nevertheless, various sources of noise remained in the crop part, and various data confirmed that other clusters had not been clearly separated. In addition, because the algorithm contained a random element, we confirmed the instability, with the result being different each time the algorithm was executed. To solve this problem, k-means clustering was performed using the vegetation index value rather than the simple color of the 3D point cloud as shown in Figure 9B. Here, it was performed with three clusters corresponding to the crop, pot, and the rest. Compared to the results of performing only clustering, the crop part was clearly segmented, and other parts were also separated depending on purpose. In the case of configuring the four clusters, the data with respect to the pot and the soil are incomplete, but it can be confirmed that the data were distinguished. Figure 9C shows how to not use clustering. Color information was applied to Otsu's thresholds. The method employed simple color values and vegetation index for color information. Since the part to be extracted was a crop, it was found that the performance was better when the vegetation index was applied. Accordingly, the vegetation index of ExG was applied to Otsu's thresholds and analyzed. Because it was a process of binarization, it was divided into crop and non-crop parts. Using this method, it is also possible to extract the crop by properly excluding noise and other parts. Both algorithms using the vegetation index succeeded in extracting crops. There was no significant difference in crop extraction performance. However, when using clustering, since parts other than crops can be further divided, it is expected that additional analysis will be possible compared to the algorithm using the threshold value. griculture 2021, 11,1010 algorithms using the vegetation index succeeded in extracting crops. There was icant difference in crop extraction performance. However, when using cluste parts other than crops can be further divided, it is expected that additional an be possible compared to the algorithm using the threshold value. A coordinate system matching process was performed through coordinat mation of a 3D image of a crop extracted using a combination of algorithms. shows the results of applying the PCA algorithm to the extracted 3D crop imag nents (pc1, pc2, …, pcn) acquired through the PCA sequentially acquire vertica on the basis of the first component. On the basis of pc1 (red line), it can be se vertical components are displayed as blue and green lines. Since the pepper u experiment is a height-growing crop, pc1 was found to coincide with the stem Accordingly, Euler Angle Transformation was performed by considering pc1 axis of the crop. Coordinate transformation was performed using pc1 acquire PCA as the z-axis of the common coordinate system ( Figure 10C). The conver has the bottom of the crop stem as the origin. Through this, the height of the c easily measured, and the degree and direction of differentiation of leaves from can be measured. It also makes it easier to implement segmentation and autom A coordinate system matching process was performed through coordinate transformation of a 3D image of a crop extracted using a combination of algorithms. Figure 10B shows the results of applying the PCA algorithm to the extracted 3D crop image. Components (pc 1 , pc 2 , . . . , pc n ) acquired through the PCA sequentially acquire vertical elements on the basis of the first component. On the basis of pc1 (red line), it can be seen that the vertical components are displayed as blue and green lines. Since the pepper used in the experiment is a height-growing crop, pc 1 was found to coincide with the stem direction. Accordingly, Euler Angle Transformation was performed by considering pc1 to be the axis of the crop. Coordinate transformation was performed using pc1 acquired through PCA as the z-axis of the common coordinate system ( Figure 10C). The converted image has the bottom of the crop stem as the origin. Through this, the height of the crop can be easily measured, and the degree and direction of differentiation of leaves from the stem can be measured. It also makes it easier to implement segmentation and automatic measurement algorithms. As a result, it was possible to obtain a standardized 3D crop image in a desired format at a constant scale within the same coordinate system.

3D Crop Image Segmentation and Automatic Measurement
The segmentation results for leaves and stems through the clustering algorithm are shown in Figure 11. Regions were divided for the entire 3D crop image, and it was possible to perform independent analysis of each leaf and stem. However, there was no perfect segmentation. While some individuals had properly separated leaves and stems, some parts were not divided, and stems were divided according to specific sizes ( Figure 12). If the leaf was too small, or if several individuals were attached very close together, or contained parts that had not been properly restored, proper division could not be achieved.
The data for point cloud leaves that had been successfully segmented were saved separately. The preprocessing carried out in order to apply the automatic measurement algorithm for length and width is shown in Figure 13. When the transformation was performed on the 3D leaf image obtained through the segmentation algorithm, it was possible to obtain results aligned with each of the z-and y-axes for the length and width of the leaf. Since the shape of the leaf is curved rather than flat, differences arise with respect to the actual length and width of the leaf, but there were no problems with the measurement of length and width for the purpose of analysis.

3D Crop Image Segmentation and Automatic Measurement
The segmentation results for leaves and stems through the clustering algorithm are shown in Figure 11. Regions were divided for the entire 3D crop image, and it was possible to perform independent analysis of each leaf and stem. However, there was no perfect segmentation. While some individuals had properly separated leaves and stems, some parts were not divided, and stems were divided according to specific sizes ( Figure 12). If the leaf was too small, or if several individuals were attached very close together, or contained parts that had not been properly restored, proper division could not be achieved.

3D Crop Image Segmentation and Automatic Measurement
The segmentation results for leaves and stems through the clustering algor shown in Figure 11. Regions were divided for the entire 3D crop image, and it w ble to perform independent analysis of each leaf and stem. However, there was n segmentation. While some individuals had properly separated leaves and stem parts were not divided, and stems were divided according to specific sizes (Figu the leaf was too small, or if several individuals were attached very close together tained parts that had not been properly restored, proper division could not be ac The data for point cloud leaves that had been successfully segmented we separately. The preprocessing carried out in order to apply the automatic meas algorithm for length and width is shown in Figure 13. When the transformation formed on the 3D leaf image obtained through the segmentation algorithm, it was to obtain results aligned with each of the z-and y-axes for the length and width of Since the shape of the leaf is curved rather than flat, differences arise with respe actual length and width of the leaf, but there were no problems with the measure length and width for the purpose of analysis.  The data for point cloud leaves that had been successfully segmented were saved separately. The preprocessing carried out in order to apply the automatic measurement algorithm for length and width is shown in Figure 13. When the transformation was performed on the 3D leaf image obtained through the segmentation algorithm, it was possible to obtain results aligned with each of the z-and y-axes for the length and width of the leaf. Since the shape of the leaf is curved rather than flat, differences arise with respect to the actual length and width of the leaf, but there were no problems with the measurement of length and width for the purpose of analysis.  The divided leaves were automatically measured according to the devised length and width measurement method ( Figure 14). Distance values were measured for the length (red line) and width (blue line) of the leaves, and compared with actual data using the scale correction value. When looking at the results of the automatic measurement itself before comparing it with the ground truth data, it was found that the desired measurement was made in most cases for the length. However, in the case of the width of the leaves, depending on the degree of restoration, the length of the crooked part, rather than the desired part, was measured and returned as the width, so a high degree of error was expected.   The divided leaves were automatically measured according to the devised length and width measurement method ( Figure 14). Distance values were measured for the length (red line) and width (blue line) of the leaves, and compared with actual data using the scale correction value. When looking at the results of the automatic measurement itself before comparing it with the ground truth data, it was found that the desired measurement was made in most cases for the length. However, in the case of the width of the leaves, depending on the degree of restoration, the length of the crooked part, rather than the desired part, was measured and returned as the width, so a high degree of error was expected. The height, leaf length, and leaf width values obtained via automatic measurement were compared with the actual measured ground true values. Since this is performed using the same sample as that used in the 3D reconstruction accuracy measurement, eight heights for eight 3D crop images and about 50 properly segmented leaves were compared.
The accuracy of the crop height estimations as measured by the error was 4.9 mm, and the total percentage error was 2.2%. This was a little lower than the manually measure results, but there was only a minor difference. As for the results of automatic leaf length measurement in comparison with the ground truth value, the error and total percent error were 2.1 mm and 4.65%, respectively. This error was greater than that for crop height and, moreover, it was greater than the manually measured results. In the case of Figure 15A, the graph of the ground truth value and the manually measured value is shown on the basis of the divided leaves only not the whole leaves, and R 2 = 0.9584, thus showing better results than using whole leaves. Figure 15B presents a graph of the ground truth value and the automatically measured value, confirming R 2 = 0.904. The height, leaf length, and leaf width values obtained via automatic measurement were compared with the actual measured ground true values. Since this is performed using the same sample as that used in the 3D reconstruction accuracy measurement, eight heights for eight 3D crop images and about 50 properly segmented leaves were compared.
The accuracy of the crop height estimations as measured by the error was 4.9 mm, and the total percentage error was 2.2%. This was a little lower than the manually measure results, but there was only a minor difference. As for the results of automatic leaf length measurement in comparison with the ground truth value, the error and total percent error were 2.1 mm and 4.65%, respectively. This error was greater than that for crop height and, moreover, it was greater than the manually measured results. In the case of Figure 15A, the graph of the ground truth value and the manually measured value is shown on the basis of the divided leaves only not the whole leaves, and R 2 = 0.9584, thus showing better results than using whole leaves. Figure 15B presents a graph of the ground truth value and the automatically measured value, confirming R 2 = 0.904. Finally, the leaf width was measured with an error of 4.1mm and a total percentage error of 18.6%. It was worse than the length of the leaf, and the percentage error, considering its relative size, was very high. Figure 16 presents a graph comparison in the same way as that provided for the leaf length analysis, and it can be seen that the automatically measured leaf width R 2 was relatively low. Finally, the leaf width was measured with an error of 4.1mm and a total percentage error of 18.6%. It was worse than the length of the leaf, and the percentage error, considering its relative size, was very high. Figure 16 presents a graph comparison in the same way as that provided for the leaf length analysis, and it can be seen that the automatically measured leaf width R 2 was relatively low. Finally, the leaf width was measured with an error of 4.1 mm and a total percentage error of 18.6%. It was worse than the length of the leaf, and the percentage error, considering its relative size, was very high. Figure 16 presents a graph comparison in the same way as that provided for the leaf length analysis, and it can be seen that the automatically measured leaf width R 2 was relatively low.

3D Reconstruction
The background being composed of black and white did not appear to affect sults, because the 3D reconstruction process was not significantly affected by lightin Accordingly, the subsequent results proceeded without considering the backgrou addition, images acquired from two directions with respect to the angle of the ro crop were acquired from all directions with respect to the crop, and this presen difficulties in recreating the 3D image [18].

3D Reconstruction
The background being composed of black and white did not appear to affect the results, because the 3D reconstruction process was not significantly affected by lighting [45]. Accordingly, the subsequent results proceeded without considering the background. In addition, images acquired from two directions with respect to the angle of the rotating crop were acquired from all directions with respect to the crop, and this presented no difficulties in recreating the 3D image [18].
The analysis of the error in the reconstructed 3D image was as follows. In the case of small-sized leaves, there were cases in which reconstruction was unsuccessful. Even when reconstructed, the leaves could not be analyzed, because, for example, only part of the leaf was modeled, or it was modeled overlapping with peripheral leaves. A relatively large error appeared at the top and bottom of the crop, which was attributed to the relatively small shooting angles and leaf sizes, in the case of the bottom of the crop. In the upper part of the plant, the leaves were often overlapping or small. With respect to unmodeled leaves, the error is judged to be higher. In addition to the size and position of the leaves, there are other factors to be considered as error. Since the process of collecting data from the turntable to construct the 3D model and that of measuring the actual length were not carried out simultaneously, changes in the curve of the leaf may have occurred due to time change or movement. Since the length of the straight line, rather than the curved length of the leaf, was measured, this was presumed to exert a lot of influence [46].
The results of the 3D reconstruction were analyzed as having a low degree of error, however, there are limitations with respect to the manual measurement of leaf count, length, and width from reconstructed 3D images. Since people are able to count and measure incomplete 3D reconstruction leaves, the error result is judged as being fairly accurate [18]. In addition, the process of manually removing outliers when performing the 3D reconstruction process may have contributed to a reduction in the error. Automation algorithms need to be developed for these processes, and it was found that the accuracy when applying them was lower than that of human measurements. In addition, it is necessary to consider bending when measuring the width of leaves to obtain accurate leaf areas.

3D Crop Image Automatic Analysis
In the case of leaves, the degree of inclusion of the node was not constant, making analysis difficult. Since the clustering algorithm used for partitioning contained random elements [47], it seems that a slightly more complete segmentation would be possible only with the addition of other mathematical algorithms.
It was found that the automatic measurement of the segmented part showed better performance than previous studies using a single sensor [48,49]. Because the coordinate transformation exhibited good performance, it was found that the coordinates selected to measure the crop height did not significantly affect the measurement error. The leaves showed a relatively high error compared to the height of the crop. This inferiority in accuracy was attributed to the algorithm not being able to accurately segment the leaves, while there were also differences in the locations to be measured. However, because the small leaves, which were found to cause a large error, were not properly segmented, they were excluded from the analysis, and therefore did not result in a higher degree of error. Basically, the widths of the leaves have smaller values than those of the height of the crop or the length of the leaf; therefore, it is regarded as being a larger error, even with similar values of error. As shown with respect to accuracy of reconstruction, the smaller the size, the greater the influence of the unrestored portion. In addition, as seen in the data analysis performed on the basis of the image, if the point cloud on the side of the leaf is not properly displayed, the value in the diagonal direction, rather than the width in the horizontal direction, will be recognized as the width, inevitably resulting in an increase in error.

Conclusions
In this study, the crop image was acquired through 3D reconstruction, and the process of automatically measuring specific information of the crop was performed using the proposed image extraction and segmentation process. The 3D reconstruction system using plant rotation was developed in a laboratory environment. To compensate for the existing reconstruction method, several attempts were made to utilize Kinect depth images and high-resolution color images. As a result, successfully obtaining a high-resolution 3D reconstruction result was possible. However, there is a limitation with respect to the incomplete modeling of small objects and the need to manually process the analysis. To compensate for the manual analysis part, using machine learning algorithms and coordinate transformation, crop images were extracted and segmented, and a reference coordinate system was set. Various algorithms were used to enable the extraction methods to meet the needs of the analysis, and coordinate transformation was performed to ensure consistent analysis. We attempted to segment the extracted image in a simple manner, and the analysis was performed using the designed automatic measurement method, and errors were confirmed. It was found that 3D crop images present large errors for small objects in most cases. The height of the crop and the length of the leaves showed coefficients of determination of 0.9 or higher for the manual measurement, but leaf width was confirmed to be a low value. Accordingly, it is expected that an additional process for elaborate segmentation and automatic measurement is required. Additionally, if a phenotypic analysis experiment is actually conducted using these algorithms, unlike in this paper, it will be necessary to obtain more samples in order to increase the explanatory power. These results confirmed the applicability to various 3D image analyses for high-throughput phenotyping.