Next Article in Journal
Developing Native Fish to Control Spirogyra in Paddy Fields for Improving the Growth, Nutrient Uptake, and Physiological Characteristics of Oryza sativa L.
Previous Article in Journal
Performance of Georgian Grapevine Varieties in a Vineyard Infected by Flavescence Dorée Phytoplasma in Piedmont, Northwestern Italy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Super-Resolution Point Cloud Completion for Large-Scale Missing Data in Cotton Leaves

1
College of Information Engineering, Tarim University, Alaer 843300, China
2
College of Information Science and Engineering, Xinjiang University of Science & Technology, Korla 841000, China
3
Key Laboratory of Tarim Oasis Agriculture, Ministry of Education, Tarim University, Alaer 843300, China
*
Author to whom correspondence should be addressed.
Agriculture 2025, 15(18), 1989; https://doi.org/10.3390/agriculture15181989
Submission received: 9 July 2025 / Revised: 12 September 2025 / Accepted: 19 September 2025 / Published: 22 September 2025
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)

Abstract

Point cloud completion for cotton leaves is critical for accurately reconstructing complete shapes from sparse and significantly incomplete data. Traditional methods typically assume small missing ratios (≤25%), which limits their effectiveness for morphologically complex cotton leaves with severe sparsity (50–75%), large geometric distortions, and extensive point loss. To overcome these challenges, we introduce an end-to-end neural network that combines PF-Net and PointNet++ to effectively reconstruct dense, uniform point clouds from incomplete inputs. The model initially uses a multiresolution encoder to extract multiscale features from locally incomplete point clouds at different resolutions. By capturing both low-level and high-level attributes, these features significantly enhance the network’s ability to represent semantic content and geometric structure. Next, a point pyramid decoder generates missing point clouds hierarchically from layers at different depths, effectively reconstructing the fine details of the original structure. PointNet++ is then used to fuse and reshape the incomplete input point clouds with the generated missing points, yielding a fully reconstructed and uniformly distributed point cloud. To ensure effective task completion at different training stages, a loss function freezing strategy is employed, optimizing the network’s performance throughout the training process. Experimental evaluation on the cotton leaf dataset demonstrated that the proposed model outperformed PF-Net, reducing the Chamfer distance by 80.15% and the Earth Mover distance by 54.35%. These improvements underscore the model’s robustness in reconstructing sparse point clouds for precise agricultural phenotyping.

1. Introduction

Xinjiang cotton is prized for its exceptional fibre purity, strength, softness, and breathability, making the region China’s foremost cotton-producing area [1]. Owing to its extensive cultivation, high yield, and premium quality, cotton has become a pillar of Xinjiang’s economy and one of the nation’s most valuable cash crops [2,3]. However, data from agricultural research institutions revealed a 17.6% decline in China’s cotton yarn exports in 2023, indicating a broader drop in domestic lint production and a weakening of global competitiveness [4,5]. Because growth conditions and phenotypic traits directly determine lint yield and quality, real-time monitoring of leaf size and automated phenotypic analysis via 3D point clouds are essential for improving production outcomes [6,7]. Nevertheless, occlusion, specular reflection, limited sensor resolution, and constrained viewpoints often produce incomplete point clouds that omit critical geometric and semantic information, highlighting the urgent need for robust completion techniques tailored to cotton phenotyping [8].
These challenges have prompted researchers in plant phenotyping to seek effective solutions; consequently, considerable efforts have focused on developing point-cloud completion and super-resolution techniques to overcome incomplete and sparse data. Initial research in this area was dominated by geometric algorithms that utilized hand-crafted features, symmetry assumptions, and surface-smoothness constraints to interpolate missing regions or retrieve similar patches from 3D databases [9,10,11,12]. Jiang et al. [13] restored stem topological continuity through L1-median skeleton extraction. Xie et al. [14] enhanced oilseed rape silique reconstruction via composite feature point registration. Chen et al. [15] filled missing leaf surfaces via Harris corner detection. Although these classical methods, which are primarily based on composite feature registration and corner detection, facilitate 3D reconstruction in certain plant species, they remain inadequate for addressing extensive sparsity or irregular structural variation due to their dependence on manually designed features.
These limitations have catalyzed a shift toward deep learning approaches, which learn permutation-invariant representations directly from raw point sets and can model complex structures end to end. The introduction of neural architectures capable of learning permutation-invariant features directly from unordered point sets [16] laid the groundwork for data-driven point cloud analysis. Building on this foundation, researchers have developed various deep learning-based completion methods tailored to plant morphology. Multiscale pyramidal decoding approaches have been shown to restore complete leaf geometry and enhance the global topological structure, although they often fail to fully recover fine boundary details [17]. Transformer-based semantic encoding methods increase organ-level reconstruction accuracy but exhibit heightened sensitivity when confronted with extremely sparse data [18]. Encoder–decoder frameworks for biomass estimation are effective in reconstructing multiplant scenes; however, they still struggle to achieve high-fidelity restoration of fine local structures [19]. These advances collectively enhanced the reconstruction of both the global topology and local details, overcoming the limitations of traditional hand-crafted methods.
Recent advances in point cloud completion have yielded a series of influential deep learning algorithms, each introducing distinct strategies to address specific limitations of incomplete data. GRNet [20] leverages a gridding structure and 3D convolutional operations to learn global geometric features, which enhances its ability to recover the overall coarse structure of objects with substantial missing regions. However, this global approach often leads to uneven point distributions and limited restoration of fine details in complex morphologies. PF-Net [21] introduces a progressive folding mechanism to iteratively refine point locations, aiming to improve both the accuracy and continuity of completed shapes. Despite better surface refinement, PF-Net may still suffer from oversmoothing and insufficient boundary fidelity, especially when the input data are highly sparse. To address the need for detailed boundary preservation, PMP-Net [22] and its improved version, PMP-Net++ [23], implement point-mixing and progressive point generation strategies. These models excel at capturing local contextual information and reconstructing sharp geometric features, which is particularly beneficial for restoring the intricate contours of plant leaves. Nevertheless, their focus on boundary detail can introduce issues such as internal voids, point clustering, or instability in severely incomplete regions. SnowflakeNet [24] further improves completion performance through the use of a hierarchical snowflake-shaped expansion module, which employs point expansion from coarse to fine to achieve a balance between structural completeness and detail. However, it is prone to layering artefacts and sampling irregularities, which can compromise its morphological fidelity. These methods demonstrate clear progress in point cloud completion, but each retains characteristic limitations in global uniformity, local detail preservation, or robustness to severe sparsity. This underscores the necessity for integrated approaches capable of jointly optimizing geometric completeness, distribution uniformity, and boundary fidelity in complex plant phenotyping scenarios.
In addition to geometric completion, point clouds require effective densification, as severe sparsity can degrade phenotypic measurements even when shapes are fully recovered. Deep learning-based super-resolution addresses this challenge by converting low-resolution point sets into dense, structurally consistent outputs, thereby increasing geometric fidelity, visual realism, and downstream analytic accuracy [25]. Traditional super-resolution techniques have employed interpolation-based strategies [26], statistical filtering methods [27], and geometric or topological algorithms [28] to upsample sparse point sets. While approaches such as Gaussian filtering, kriging interpolation, Laplacian smoothing [29], and IsoMap [30] have achieved varying degrees of success, they often suffer from excessive computational cost or loss of fine-scale structural fidelity in complex, large-scale scenes.
In contrast, recent learning-based super-resolution frameworks leverage hierarchical feature encoding to infer latent high-resolution representations, enabling robust point-density enhancement while preserving local morphology. Inspired by progress in image super-resolution, neural architectures such as the SRCNN [31], VDSR [32], and ESPCN [33] have been extended to 3D point clouds, enabling more accurate recovery of dense structures from sparse data. These models typically employ end-to-end frameworks with hierarchical encoders to capture both global shape information and local geometric features. Current deep learning-based super-resolution methods can be broadly categorized into three technical directions. Feature-enriched encoders, including multibranch MLPs and PointNet++ pyramids, enhance multiscale information extraction and support high-resolution point set reconstruction [34,35]. Adversarial and graph-based frameworks integrate graph convolution, generative adversarial training, and residual connections to accelerate convergence and improve structural realism [36]. Attention-driven and dilated-convolution modules utilize self-attention, dilated kernels, or local CNN blocks to refine the point distribution and detail fidelity, particularly under extreme sparsity [37,38,39]. However, these approaches remain fragmented, and no single framework has yet achieved a comprehensive balance among density enhancement, geometric completeness, and point distribution uniformity, all of which are essential criteria for reliable cotton leaf phenotyping.
Unlike interpolation or voxel-based densification techniques, deep learning models are uniquely capable of modelling complex spatial relationships and generating semantically coherent distributions, even in the presence of extreme sparsity or occlusion. Most existing super-resolution models continue to focus on isolated tasks, often prioritizing apparent density or overall plausibility while neglecting fine-scale morphological accuracy or distribution regularity. The absence of unified frameworks that jointly optimize all critical aspects underscores a persistent gap in current research and highlights the urgent need for more robust and comprehensive solutions for point cloud densification in plant phenotyping applications.
In response to these challenges, this study introduces a super-resolution point cloud completion framework tailored for highly sparse, nonuniform, and geometrically complex cotton leaf data. The proposed method incorporates five key modules within a unified end-to-end architecture. The design begins with a PointNet++ encoder that extracts global geometric features to capture the overall shape information. A decoder then integrates bilinear interpolation with the PF-Net algorithm to reconstruct coarse geometry, followed by an upsampling module that generates additional points in regions of data loss, thereby enhancing local density and detail recovery. The feature-fusion module subsequently merges local and global representations into deconvolutional feature maps, further improving the structural detail of the reconstructed point cloud. A convolutional completion network refines the output, whereas farthest-point sampling ensures that the final point cloud is of high quality and uniformly distributed, meeting the requirements of downstream phenotypic analysis. The entire training pipeline is further optimized via a multiloss freezing scheme, and the inclusion of an attention-augmented PointNet++ module enhances point cloud uniformity throughout the reconstruction process. The main contributions of this paper are as follows:
I. Super-resolution completion for sparse data: A novel completion network reconstructs highly sparse cotton-leaf point clouds. Bilinear-interpolation upsampling improves local feature capture, restores missing regions, and yields accurate leaf-area representations.
II. Uniformity module based on PointNet++: A PointNet++ feature-fusion block merges incomplete inputs with reconstructed points, equalizes the point distribution and improves the geometric consistency.
III. Stagewise loss-freezing strategy: During training, early layers are frozen under CD loss to stabilize global reconstruction; later layers optimize a feature-fusion loss to refine local detail and maintain uniform density, leading to faster convergence and higher fidelity.

2. Materials and Methods

2.1. Cotton Leaf Dataset

Owing to the absence of point cloud data for cotton plant leaves in public datasets, this study collected data from 96 cotton plants over one year in Shier Tuan, Alar city, Xinjiang. The dataset spans five growth stages: sowing and seedling, vegetative, bud, flowering and boll formation, and boll opening. We adopt a stratified split by plant ID and growth stage: 8:1:1 for train: validation: test. Sampling was conducted in adjacent cotton fields with consistent water and fertilizer conditions, providing a robust foundation for future studies on yield and quality estimation. To minimize the effects of uneven lighting and wind-induced motion on image quality and 3D reconstruction, all the experiments were conducted indoors under controlled conditions. The laboratory setup included four soft light lamps to ensure uniform lighting, a black backdrop beneath the plants to eliminate background interference, and curtains to block external light sources. Each plant was photographed sequentially from three angles: approximately 45° downwards, horizontal, and 75° upwards, resulting in approximately 260 images per plant.
The 260 images for each sample are organized by plant ID to maintain image-to-sample correspondence. All plants were photographed with a Canon EOS 200D II camera (Canon Inc., Urumqi, China) at a native resolution of 4000 × 6000 pixels (24 MP). As shown in Figure 1, we employ multiview stereo using RealityCapture (version 1.2.0.11881, Capturing Reality, Bratislava, Slovakia) to align the images, generate a sparse point cloud, and subsequently perform dense reconstruction. This process yields high-fidelity 3D models of entire cotton plants. Although CloudCompare (https://cloudcompare.org/index.html, accessed on 18 September 2025) supports purely manual selection, we adopted a semi-automatic workflow with manual verification following dense plant reconstruction. The plant cloud was first denoised using statistical outlier removal (SOR) and smoothed with moving least squares (MLS). We then computed per-point normals and curvature/roughness scalar fields to guide segmentation. Leaves were coarsely isolated by thresholding these fields and applying connected components; overlapping organs were further separated via region growing (using normal/curvature cues) and interactively refined with the Segmentation tool, consulting RGB textures where boundaries were ambiguous. Verified leaves were exported as individual PLY files.
To construct the dataset required for this study, 3D point cloud data of cotton leaves were extracted and uniformly sampled via farthest point sampling (FPS). Each model was sampled to contain 2048 points. Through leaf segmentation, a total of 6000 cotton leaf point clouds were obtained. Point cloud data were extracted from multiview images, and each leaf was downsampled to 2048 points via FPS. For point clouds containing fewer than 2048 points, upsampling was performed by duplicating neighboring points to achieve the required point count. CloudCompare was used for segmentation and annotation. For training and evaluation, each leaf cloud was denoised, normalized to [−1,1], and converted to a canonical ground truth (GT) of 2048 points using deterministic FPS (fixed seed for reproducibility): iteratively select the point with the largest minimum Euclidean distance to the already chosen set, ensuring uniform coverage. Rare leaves with fewer than 2048 points were first completed to 2048 by nearest-neighbor duplication and then FPS-refined. The resulting 2048-point GT serves as supervision for all losses and as the reference for metrics (CD/EMD/F-score).
To emulate acquisition failures, we derive incomplete inputs from the GT leaves at 50% and 75% missing using two mask families. (i) Spherical removal (local-region erasure): one or several balls are sampled with centers on the leaf point set; radii are adjusted so that the retained-point ratio matches the target. Points whose Euclidean distance to any center is below its radius are removed and overlapping balls form contiguous holes. (ii) Axial cut (one-sided half-space): a cutting plane, whose normal is aligned with the leaf’s principal/midrib axis, removes all points on one side (“one cut”). For reproducibility, we store all mask parameters (centers/radii or plane coefficients) and random seeds. When a method requires a fixed input size, sparse inputs are upsampled to 2048 by nearest-neighbor duplication; at evaluation, both prediction and GT are FPS-downsampled to 2048 before computing CD/EMD/F-score under a unified protocol.

2.2. Super-Resolution Point Cloud Completion Network

Owing to the inherent sparsity and incompleteness of the obtained cotton point cloud data, this model focuses on upsampling and feature representation of cotton leaf point clouds as key technical research. This project is based on the PF-Net and PointNet++ models, focusing on addressing the issue of reduced density accuracy caused by point cloud incompleteness and constructing a network model with both completion and densification capabilities. Figure 2 illustrates the overall framework of the project. To address the insufficient feature extraction from sparse point clouds, bilinear interpolation is used to upsample the original data. The point cloud reconstruction process involves the fusion of complementary and missing point clouds, while also achieving a uniform distribution of the point cloud. Finally, by freezing the loss function, the complementary and plasticity tasks are completed simultaneously during the end-to-end network training.

2.2.1. Point Cloud Completion

To achieve accurate and uniform reconstruction of sparse cotton leaf point clouds, this study proposes a PF-Net-based completion framework that integrates multiresolution feature extraction and hierarchical decoding, as illustrated in Figure 3. The primary objective is to address limitations in traditional feature extraction and enhance the uniformity and geometric fidelity of the reconstructed point clouds.
Bilinear Interpolation and Iterative Upsampling: Let X R B × N × 3 denote a batch of sparse cotton-leaf point clouds (coordinates normalized to [−1,1]). To provide a dense scaffold for completion, we perform bilinear-interpolation upsampling on local tangent patches and obtain a seed set X k R B × K × 3 with K = k N (we fix K = 2048 in all experiments), as shown in Figure 3a. In local parametric coordinates ( x , y ) 0 , 1 2 , bilinear interpolation is given by:
f ( x , y ) = ( 1 x ) ( 1 y ) f 00 + x ( 1 y ) f 10 + ( 1 x ) y f 01 + x y f 11
Here f 00 , f 10 , f 01 , f 11 are the values at the four corners ( 0 , 0 ) , ( 1 , 0 ) , ( 0 , 1 ) , ( 1 , 1 ) ; the weights sum to 1.
Specifically, during the upsampling process, a nearest-neighbor search identifies a set of surrounding points for each reference point. These neighbors are used as both references and supplements, ultimately improving the precision of the upsampling step. Let D i represent the set of distances from point P i to all other points. The nearest neighbors of P i , denoted as N i , are selected from D i based on the minimum distance criterion. For each point P i , the average position F ¯ i of its neighboring points is computed, reflecting the latent structural information of the point cloud. Subsequently, a directional vector v ¯ i is calculated from the reference point P i to the mean position F ¯ i , which describes the trend or direction of the variation in the surrounding point cloud. This vector is then utilized to generate new points P i along the direction of v ¯ i , with one new point being the default, which are sequentially added on the basis of the directional vector. This procedure is iterated for each point, continuously identifying neighboring point sets and generating new points. The entire point cloud is processed in this manner, resulting in a modified set of points. Through the iterative adjustment of each point, analogous modifications and enhancements are applied to the entire point cloud.
Encoding and Multiresolution Feature Extraction: In the encoding stage, the MRE processes an unordered input point cloud of size N , as illustrated in Figure 3b. Bilinear interpolation is applied to upsample the initial data, generating additional scales with ( k 1 ) N and k N points. These multiscale representations enable the encoder to focus on both global and local features. Three distinct MLPs are subsequently utilized to fuse these scales into a comprehensive latent feature vector, facilitating the extraction of rich geometric and semantic information.
Hierarchical Decoding and Point Cloud Refinement: The latent feature vector is then input into the PPD, which hierarchically upsamples and reconstructs the missing data through three levels of linear-convolutional layers, as illustrated in Figure 3c. At each decoding stage, the output from the previous layer is refined and upsampled, progressively generating new points and restoring the detailed structure of the original cloud. This multistage decoding mechanism enhances both the global topology and local geometric details, such as sharp edges and intricate surfaces. The outputs α , β , and γ from each level are fused to produce the final, high-resolution point cloud of size k N .
Multiscale fusion at the output layer: Let the sparse input be X R B × K × 3 (coordinates normalized to [−1,1]). After bilinear-interpolation upsampling, we obtain the seed set X K R B × K × 3 with K = k N (fixed K = 2048 in all experiments). Three parallel multiscale branches operate on the same K seeds and do not generate independent point clouds; instead, each branch predicts an index-aligned residual offset Δ Y ( s ) R B × K × 3 for s { 1 , 2 , 3 } . For every seed point we also learn non-negative weights α , β , γ R B × K × 1 , normalized by softmax so that α + β + γ = 1 . The final reconstruction is obtained by residual fusion:
Y = X k + α Δ Y ( 1 ) + β Δ Y ( 2 ) + γ Δ Y ( 3 ) ( Y R B × K × 3 )
where denotes element-wise (per-point) scaling and broadcasting. Because Δ Y ( s ) are index-aligned with X k , the fusion is computed point-wise rather than by concatenating/unioning three sets. When K 2048 , we FPS-downsample Y to 2048 before computing CD/EMD/F-score; when K = 2048 , no further resizing is required.

2.2.2. Point Cloud Reshaping

The primary objective of this stage is to achieve a complete and uniformly distributed 3D point cloud model suitable for accurate phenotypic analysis. While the preceding completion module reconstructs missing regions within the point cloud, the distribution of points may still be uneven, which can hinder downstream processing and analysis.
To address this, the output of the completion algorithm is first integrated with the original point cloud, forming a unified point cloud set. PointNet++ is employed at this stage to facilitate effective fusion of the generated and original data, leveraging its strong capability to extract both local and global geometric features. Precise alignment strategies are then applied to ensure spatial consistency between the two point clouds, preventing any overlap or misalignment. We term the subsequent post-fusion distribution normalization Reshape. After completion, we denote the prediction by Y and the original input by X (both normalized). Reshape produces a uniformly distributed, boundary-preserving point set used for evaluation and analysis: Y ˜ = R ( Y , X ) . Concretely, the merged cloud is first deduplicated using a radius criterion and then rigidly aligned to ensure fine spatial consistency; local densities are equalized by iteratively dispersing overly clustered points. When necessary, small gaps are interpolated, isolated outliers are lightly denoised, and the final set is brought to a fixed cardinality ( K = 2048 ) via deterministic farthest-point sampling. The resulting cloud achieves uniform spatial sampling while preserving boundaries and details, providing a robust foundation for subsequent phenotypic analysis.
To rigorously evaluate the model’s ability to handle incomplete data, additional experiments are performed on synthetic partially missing point clouds. Specifically, the input data are first aligned to the center of the coordinate system and normalized to fit within the range of [−1, 1]. Each shape is then uniformly sampled with 1024 points to capture its intrinsic geometric features. To simulate missing regions, a random viewpoint is selected as the center, and all points within a specified radius are removed. This controlled generation of incomplete point clouds enables systematic assessment of the completion and Reshape modules under realistic conditions, highlighting their effectiveness in restoring both structure and uniformity. The degree of missing data can be controlled by adjusting the removal radius. PointNet++ is employed to achieve the fusion and uniformization of the missing point cloud P m and the complementary point cloud P c . The fusion process is defined as the union of these two points, described as follows:
P f = P m U P c
PointNet++ processes P f by extracting features from each point P i via an MLP. It then aggregates local and global features to generate a unified representation, denoted as f ( P f ) . The process can be described as follows:
f i = M L P ( P i )   , f ( P f ) = P o i n t N e t + + ( P f )
Here, P o int N e t + + ( P f ) represents the PointNet++ network used for fusing the point cloud f ( P f ) . The resulting feature vector f ( P f ) encapsulates a homogenized representation of the fused point cloud, facilitating subsequent analysis or processing tasks.

2.2.3. Loss Freeze Program

This study employs a loss function freezing strategy to address the critical tasks of point cloud completion and reshaping for cotton leaf data. The loss functions consist of the CD loss ( L C D ) and the reshaping loss ( L R e s ). L C D ensures the uniform distribution of upsampled points across the target surface, while L R e s maintains consistent local structures for points within the same region. To ensure training accuracy, distinct freezing schemes are applied to these loss functions. During the first 15 h of training, L R e s remains constant to enhance learning for point cloud completion and super-resolution. Between the 15th and 80th epochs, the weight of L C D is reduced to 0.5, while L R e s is assigned an equal weight of 0.5. After 80 epochs, L C D is frozen, allowing the network to focus entirely on fusion and morphological reshaping tasks. This adaptive weighting strategy ensures efficient optimization and precise task-specific learning throughout the training process.
The CD loss function quantifies the difference between two points by calculating the average distance from each point in one cloud to its nearest point in the other cloud. This effectively guides the network in learning features, ensuring the precision and uniformity of the generated point clouds. The CD loss function formula used in this study is defined as follows:
L C D ( P , S ) = 1 P p P min s S p s 2 2 + 1 S s S min p P p s 2 2
To improve the control of point cloud fusion and uniform distribution, this study introduces a reshaping loss to constrain the distribution process. The reshaping step in the model builds upon the method used in PointNet++ and incorporates the application of a feature transformation matrix. To ensure that this matrix approximates an orthogonal matrix, a regularization term is added to the SoftMax training loss. The reshaping loss is defined as follows:
L R e s = I A A T F 2
where A represents the feature alignment matrix predicted by the network. The incorporation of orthogonal transformations helps preserve input information, ensuring the integrity of the feature transformation process. The regularization term enhances optimization stability while improving the overall performance of the model.
This strategy enables task-specific loss functions to guide network learning at different training stages, effectively avoiding gradient conflicts while improving convergence speed and training stability. The experimental results demonstrate that the proposed network achieves outstanding performance in both point cloud completion and reshaping tasks for cotton leaves, further validating the effectiveness and applicability of the loss function freezing strategy in multitask deep learning networks.

2.3. Evaluation Measures

2.3.1. Chamfer Distance

This study utilizes the chamfer distance (CD) as the primary evaluation metric for the cotton leaf dataset. CD is widely used in point cloud reconstruction and generation tasks and serves as an effective measure of similarity between two points. A higher CD value between the denoised point cloud and the reference point cloud indicates poor completion performance, while a lower CD value signifies better results. The metric evaluates the difference between the completed and ground-truth point clouds, with values approaching zero indicating greater shape similarity and superior completion quality.
In addition to CD, we report per-point L 1 and L 2 CDs to provide complementary error profiles. The L 1 CD is defined as L 1 = Q i Q ^ i i , where Q i and Q ^ i denote corresponding points in the reference and predicted clouds, respectively. L 1 aggregates absolute coordinate differences, treating all deviations uniformly. The L 2 CD is formulated as L 2 = ( Q i Q ^ i ) 2 i . Because the L 2 norm squares deviations before aggregation, it places greater emphasis on large errors than the L 1 norm does, making it particularly sensitive to substantial discrepancies between point clouds.

2.3.2. Earth Mover’s Distance

The Earth Mover’s Distance (EMD) quantifies the minimum effort required to transport a distribution (conceptualized as “soil”) from one location to another. This effort is calculated as the total amount of “soil” moved multiplied by the distance it travels. As shown in Equation (7), unlike CD, EMD requires that the two point clouds, Q and Q , have the same size. A lower EMD value indicates greater similarity in the resolution and density of the point clouds. This metric captures the spatial alignment between points in the complementary network’s output and the corresponding plant sample point cloud, thereby validating the accuracy and reliability of the point cloud reconstruction.
E M D ( Q , Q ^ ) = min ϕ : Q Q ^ 1 Q q Q K q ϕ ( q ) 2
where π is a bijection. Because EMD is defined only for point sets with the same cardinality Q = Q ^ = K , we fix K = 2048 as described above and compute the assignment using an approximate matching solver. The term q ϕ ( q ) denotes the Euclidean distance between a ground-truth point q and its matched prediction ϕ ( q ) . This formulation enables EMD to capture fine spatial discrepancies between the generated and ground-truth point clouds, often providing a more sensitive similarity assessment than CD.

2.3.3. F-Score

The F-score evaluates model performance by combining precision and recall through the harmonic mean, providing a balanced measure of both metrics. Unlike CD, the F-score mitigates potential inaccuracies arising from point-to-point distance aggregation and serves as a more reliable and comprehensive evaluation metric. By integrating performance across multiple dimensions, the F-score offers a robust assessment of the model’s overall effectiveness.
Let G = { g i } i = 1 K be the ground-truth point set and Y = { y i } i = 1 K the prediction after cardinality normalization. We fix K = 2048 for all evaluations; if a method outputs a different size, both G and Y are FPS-downsampled to 2048 beforehand. Distances are Euclidean and reported in the same units as CD/EMD. Given a global matching radius d , precision and recall are computed on the two point sets via nearest-neighbor distances:
P ( d ) = 1 K y Y 1 min g G y g 2 < d , R ( d ) = 1 K g G 1 min y Y g y 2 < d .
The F-score is the harmonic mean:
F - S c o r e ( d ) = 2 P ( d ) R ( d ) P ( d ) + R ( d )
where K = 2048 (cardinality normalization), and d is a single global matching radius chosen on the validation set and kept constant across all methods.

3. Results

3.1. Training Visualization

At a 50% missing data rate (Figure 4a,b), the training loss curves of state-of-the-art methods, including GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, and PointTR v2 [40], exhibit markedly different convergence behaviors. GRNet is highly unstable, with large oscillations and no clear convergence by 160 epochs. PF-Net and PMP-Net drop rapidly at the start but plateau early, suggesting overfitting or convergence to suboptimal minima. PMP-Net++ and SnowflakeNet are smoother yet stagnate after ~30 epochs. PointTR v2 descends quickly in the first 10–15 epochs and is steadier than GRNet, but it shows an early transient spike and then saturates at a small, non-zero floor above our method. In contrast, our model (Figure 4b) exhibits a fast, near-monotonic decrease within the first ~20 epochs and stabilizes at a substantially lower plateau with minimal oscillation, indicating more reliable optimization and stronger learning at the 50% sparsity level; the visual comparisons at this setting (Figure 5 and Figure 6) further show superior preservation of leaf contours and vein continuity with a more uniform point distribution.
When the missing data rate increases to 75% (Figure 4c,d), these trends become more pronounced. GRNet remains severely unstable and fails to converge. PF-Net, PMP-Net, PMP-Net++, and SnowflakeNet still plateau early and show little additional loss reduction beyond the initial epochs, highlighting the difficulty these methods face under extreme sparsity. PointTR v2 has clear strengths: it descends rapidly at the start and is noticeably steadier than GRNet, reaching a lower loss than PF-Net and PMP-Net at this setting. Nevertheless, it soon saturates at a non-zero floor that remains above our method. By contrast, our model maintains robust convergence: the loss drops to a low level within about 20 epochs and then stays stable with minimal fluctuation (Figure 4d). Consistent with the curves, the visual comparisons at 75% (Figure 5 and Figure 6) show that although PointTR v2 produces reasonable global shapes, it tends to display uneven point density, boundary discontinuities, and vein fragmentation, whereas our reconstructions preserve contours and vein continuity with fewer artifacts and more uniform sampling. These results underline the effectiveness of our architecture, particularly the combination of adaptive density-aware processing and multiscale feature fusion, mitigating the adverse effects of severe data scarcity and sustaining low, stable loss values.
To further investigate the cause of this training stability and efficiency, Figure 4b reveals that the proposed model incorporates a novel parameter-freezing strategy. This technique significantly accelerates early convergence, as evidenced by the rapid decline in loss within the first 15 epochs. During subsequent epochs (15–80), the loss exhibits minor oscillations before stabilizing, a behavior directly attributable to the selective freezing of network parameters that guides optimization toward global minima. As the missing data rate increases from 50% to 75%, the model consistently demonstrates adaptive convergence characteristics, confirming its robustness and flexibility across varying levels of point cloud incompleteness. Taken together, both the comparative and causal analyses establish that the proposed model consistently outperforms state-of-the-art methods in terms of convergence efficiency and reconstruction accuracy, effectively addressing the challenges posed by moderate to severe point cloud data loss.
The completion network was implemented in PyTorch (version 1.10.2, CUDA 11.2) and optimized with Adam. We trained all components for 301 epochs with a global batch size of 16 and an initial learning rate of 0.001. Training was performed on 4 × NVIDIA GTX 2080 Ti with CUDA 11.6, Ubuntu 22.04.1 LTS, and Python 3.8. Evaluation protocol: before scoring, GT and predictions are FPS-normalized to K = 2048; F-score uses a fixed global matching radius d = 0.01.

3.2. Cotton Leaf Completion for Missing Shapes

3.2.1. Completion Results for 75% Missing Data

To rigorously assess point cloud completion performance under high missing rate scenarios, a comparative analysis was conducted using the cotton leaf dataset with a 75% missing-data rate. As summarized in Table 1, the proposed model demonstrates statistically significant improvements over existing advanced methods across all major metrics.
Specifically, in terms of the two mean distance metrics ( L 1 and L 2 ), the proposed model reduces the error by 75.5% and 97.5%, respectively, compared with GRNet, which represents the baseline with the highest error values. Relative to PointTR v2, our model further reduces these two mean distance metrics by 45.7% and 77.0%, respectively. For EMD, our model achieves a 24.6% lower error than the best-performing reference model, SnowflakeNet, and a 60.5% lower error than PointTR v2. Furthermore, the proposed model attains the highest F-score among all compared methods, improving by 2.9% relative to GRNet, by 809.9% relative to SnowflakeNet, and by 129.7% relative to PointTR v2, thereby demonstrating a more effective balance between completion accuracy and structural integrity. Taken together, these statistical results highlight the comprehensive performance advantage of the proposed model for point cloud completion under severe data loss conditions, establishing its effectiveness for reconstruction in complex and highly sparse scenarios.
Following the comparative analysis of various models using quantitative metrics such as EMD, F-score, and CD to evaluate the performance of point cloud reconstruction on the cotton dataset with a 75% missing data rate, a visual assessment was conducted to further substantiate and elucidate these numerical findings. This step is essential, as quantitative metrics alone may not fully capture critical aspects such as the uniformity of the point cloud distribution and the preservation of detailed morphological features. Therefore, to comprehensively evaluate the reconstruction fidelity, Figure 5 presents complementary multiangle renderings of six Gossypium hirsutum leaf samples under the condition of 75% missing data, effectively illustrating the reconstruction capabilities of the various models and providing deeper insight into their practical advantages and shortcomings.
Figure 5 presents a visual comparison of reconstructed cotton-leaf point clouds under 75% missing data. GRNet has difficulty capturing complex structural details, particularly for Leaves (4) and (5), where the results are fragmented, noisy, and far from the ground-truth morphology. PF-Net shows significant distortions and uneven density distributions, most evident in Leaves (3) and (4), with elliptical point patterns that radiate from central regions and cause shape deformation and loss of detail. PMP-Net and PMP-Net++ consistently produce sparse and over-smoothed reconstructions, for example, Leaves (2) and (6), yielding rounded outlines that lack critical geometric nuances. SnowflakeNet provides relatively complete shapes, yet clear layering artifacts and nonuniform densities remain; in Leaves (3) and (4), the detailed contours are not fully recovered. PoinTR v2 improves global shape fidelity compared with GRNet, PF-Net, and the PMP family, but it still underperforms SnowflakeNet: fewer serrations and marginal veins are restored and edges are blurrier with stronger local density fluctuations, especially in Leaves (3) and (4). By contrast, our method reconstructs sharper leaf margins and more continuous venation with fewer artifacts, achieving the closest resemblance to the ground truth.
To further elucidate these microstructural differences, Figure 6 provides a fine-grained visual comparison of leaf edge reconstruction between SnowflakeNet and the proposed method under conditions of 75% missing data. SnowflakeNet generally succeeds in generating globally complete leaf shapes, yet its outputs frequently display uneven point distributions and noticeable layering artifacts. These issues are especially evident in samples (1), (2) and (4), where high levels of input noise result in poorly defined contours and fragmented edge structures. In these cases, the reconstructed point clouds exhibit discontinuities along the boundaries, with local geometric details frequently blurred or absent. Layering effects are visually apparent, manifesting as stratified, nonuniform patches that detract from the morphological realism of the reconstructed leaves.
In contrast, our proposed model clearly outperforms the other methods in terms of reconstruction quality across all six leaf samples, effectively maintaining uniformity in the point cloud distribution and preserving intricate morphological features. Leaf (4) and Leaf (5) highlight the strength of our approach, where reconstructed point clouds closely replicate the original leaf shape, displaying enhanced topological accuracy and detailed structural fidelity. Overall, these visual comparisons highlight the robustness of the proposed model in reconstructing cotton leaf point clouds under conditions of severe data loss. The model consistently maintains structural fidelity and preserves fine morphological features, thereby demonstrating its suitability and effectiveness for high-precision point cloud completion in challenging phenotyping applications.
When the missing data rate is reduced to 50%, our model demonstrates even greater reconstruction fidelity, underscoring its strong robustness and adaptability to varying data conditions. In scenarios with lower missing rates, the reshape module refines the global leaf shape, enhancing geometric accuracy, whereas the upsampling and freezing modules work together to restore local features and improve point cloud uniformity. This progressive enhancement in performance, observed consistently as the amount of missing data decreases, confirms the model’s generalizability and structural integrity, validating its practical applicability across diverse data-loss scenarios.

3.2.2. Completion Results for 50% Missing Data

On the basis of the results presented in Table 2, a statistical analysis of the cotton leaf dataset at a 50% missing data rate reveals notable performance differences among the evaluated architectures. Compared with SnowflakeNet, the proposed model significantly improves the point distribution uniformity, even though SnowflakeNet results in slightly higher local geometric accuracy. This enhanced uniformity contributes to improved morphological integrity and practical applicability in phenotypic analysis, highlighting the overall superiority of the proposed approach for robust point cloud reconstruction under incomplete data conditions. Specifically, the proposed model reduces the EMD by 3.6% relative to SnowflakeNet, highlighting its superior ability to maintain a consistent point cloud distribution. Visual inspection of the reconstructed point clouds further corroborates these findings, as our model produces more uniformly distributed points and preserves greater morphological detail, resulting in reconstructions that closely align with the ground truth.
Further analysis indicates that architectural refinements, such as the strategic reduction in convolutional operations in local feature encoding and the enhancement of global attention weighting, resulted in a predictable trade-off. Although there was a slight increase in localized geometric error, with L 1 increasing by 14.8% and L 2 increasing by 23.5%, the shape completeness as measured by the F-score remained statistically equivalent to SnowflakeNet. This outcome demonstrates that our approach effectively reallocates computational resources from low-level feature extraction to high-level contextual modelling, thereby prioritizing topological fidelity over local metric optimization. Although GRNet achieved the lowest EMD value among all methods, its visual reconstructions presented distribution and completeness deficiencies. PointTR v2, a recent transformer baseline, shows mid-range L 1 / L 2 but the largest EMD and a very low F-score, indicating difficulty in maintaining structural completeness and uniform point density under 50% sparsity. By comparison, the proposed model delivered a strong balance, with competitive EMD and a consistently superior F-score, outperforming PF-Net, PMP-Net, and PMP-Net++ across all the key metrics. These results establish the model’s effectiveness in achieving both uniform point distribution and detailed morphological preservation in challenging, incomplete data scenarios.
To rigorously assess the performance of various models in the task of cotton leaf point cloud reconstruction, this study designed a controlled experiment involving 50% data omission. As illustrated in Figure 7, a comparative analysis reveals substantial disparities among the models, particularly in terms of reconstruction quality and fidelity to fine details.
The experimental results indicate that the conventional GRNet, despite demonstrating fundamental reconstruction capabilities, has significant limitations in preserving complex local details and maintaining uniform point cloud distributions. These shortcomings are particularly evident in Leaf (4), where GRNet cannot reconstruct critical high-curvature features, including apical indentations, and produces substantial noise artifacts. PF-Net, while exhibiting an improvement over GRNet in terms of global structural accuracy, still suffers from uneven point density distributions. Notably, Leaf (4) exhibited an elliptical scattering pattern originating from the center, resulting in pronounced distortions in overall leaf shape and a considerable loss of local detail. Similarly, PMP-Net and PMP-Net++ yield sparse point clouds, with reconstructed contours tending toward simplistic arc-shaped forms that inadequately represent intricate leaf structures. PointTR v2, though capable of recovering a plausible global outline, frequently shows density clustering and over-smoothed details; in Leaf (2) and Leaf (3) points concentrate along the midrib leaving lamina voids, and in Leaf (5) and Leaf (6) the margins remain jagged with broken veins, indicating incomplete structural recovery under sparsity. As illustrated in Figure 8, although SnowflakeNet produces relatively complete reconstructions, it frequently results in uneven point distributions and pronounced layering artifacts. These shortcomings become especially apparent in challenging cases such as Leaf (5) and Leaf (6), where substantial input noise leads to significant contour ambiguity and a degraded edge definition in the reconstructed results.
In contrast, the proposed model demonstrates consistently superior reconstruction performance across all six leaf samples, maintaining both global morphological integrity and high-fidelity local detail. The reconstructed point clouds exhibit substantially improved uniformity and edge continuity, closely match the ground truth and effectively preserve fine structural features. Qualitative analysis further reveals that the model not only restores the overall leaf shape with high accuracy but also excels in capturing intricate microstructural details. Notably, the following advantages are observed: (1) the uniformity and continuity of the point cloud distribution surpass those achieved by the control models; (2) the accuracy of leaf edge contour reconstruction is significantly enhanced, showing excellent agreement with the morphological features of the ground truth; and (3) the preservation of surface detail fidelity is superior to that of all other methods, highlighting the model’s strong capability in feature extraction and reconstruction. These experimental results unequivocally demonstrate the model’s enhanced ability to address both large-scale completeness and fine-grained accuracy in complex cotton leaf point cloud reconstruction, thereby underscoring its potential as a robust solution for high-precision phenotypic analysis and advancing the state-of-the-art in point cloud reconstruction research.

3.3. Cotton Leaf Completion for Sparse Point Cloud

To evaluate the performance of cotton leaf point clouds with 75% sparsity systematically, quantitative and qualitative comparisons were conducted against state-of-the-art completion methods and the proposed algorithm. Performance was assessed via four widely adopted metrics: L 1 distance, L 2 distance, EMD, and F-score.
On the basis of the results in Table 3, a systematic evaluation was conducted to compare state-of-the-art completion methods on sparse cotton-leaf point clouds. Compared with the next-best method, PMP-Net++, the proposed model achieves markedly better reconstruction quality: L 2 (CD) decreases from 5.51 to 2.11 (a 61.7% reduction), L 1 decreases from 44.63 to 28.76 (a 35.6% reduction), and F-score increases from 0.48% to 17.8%; EMD remains comparable. Notably, compared with SnowflakeNet, which reports the lowest L 1 / L 2 /EMD, our method attains a substantially higher F-score (an 8.8 percentage-point improvement), indicating more complete and uniform reconstructions under severe sparsity. These results highlight the model’s robust balance between reconstruction fidelity and structural completeness in the presence of large missing regions and challenging occlusions.
Figure 9 provides a qualitative comparison of point cloud completion results for sparse cotton leaves using different methods. As illustrated, the input point clouds are highly sparse, presenting significant challenges for accurate reconstruction. GRNet produced outputs with substantial noise and incomplete structural recovery, while PF-Net captured only the coarse outline of leaves, resulting in interior voids and irregular point distributions. PMP-Net and PMP-Net++ maintained coherent global morphology, but persistent internal gaps and surface discontinuities were still evident due to sparsity, compromising local topological integrity. PointTR v2 recovers a reasonable outline and cleaner interiors than GRNet and PF-Net, but under high sparsity, it shows density clustering and over-smoothed lamina; margins remain jagged and veins partly broken, indicating incomplete fine-detail recovery.
In contrast, both SnowflakeNet and the proposed method generated more complete and structurally coherent point clouds. However, the proposed algorithm more effectively preserved fine morphological features, such as leaf serrations and vein patterns, especially in Leaves 3 and 4, resulting in reconstructions closely aligned with the ground truth. This qualitative evidence is consistent with the quantitative results presented in Table 3 and further confirms the enhanced capability of the proposed approach for detailed phenotypic analysis in agricultural applications, particularly for accurate cotton leaf structural recovery and subsequent phenotypic parameter extraction.

3.4. Ablation Study

To comprehensively evaluate the effectiveness of each architectural component, an ablation study was conducted on the point cloud completion model using a self-constructed cotton dataset with a 75% missing data rate. Three key modules (upsample, reshape, and frozen) were systematically enabled or disabled, resulting in five model configurations, as summarized in Table 4. Each configuration was assigned a unique identifier (ID 1–5) for systematic comparison. Through these targeted comparisons, the influence of each individual network component and the combined contributions to the final reconstruction quality were clearly identified.
ID 1 included only the upsampling module while disabling both reshaping and freezing.
ID 2 enables the reshaping module without upsampling and freezing.
ID 3 activated both upsampling and reshaping but excluded freezing.
ID 4 integrates the upsampling and reshaping modules, disabling the freezing module.
ID 5 represents the complete model, which simultaneously incorporates upsampling, reshaping, and freezing modules.
On the basis of the results in Table 4, the full model (ID 5), which integrates upsampling, reshaping, and freezing, achieves the best overall performance. Relative to the baseline configuration (ID 1), ID 5 reduces EMD by 39.6% and increases the F1 score by 104.2%, indicating substantial gains in geometric fidelity and structural completeness. Models incorporating the reshaping module (IDs 2–5) consistently outperform those without it, yielding on average 67.5% lower L 2 error and 25.7% higher F1 than the non-reshaping baseline. Adding the upsampling module further enhances performance: within the reshaping family, configurations that include both upsampling and reshaping (IDs 4 and 5) achieve up to 16.4% lower L 1 error and 17.2% lower EMD compared with their reshape-only counterparts (IDs 2 and 3).
Interestingly, the freezing parameters had varying impacts on performance. While it enhanced the results when it was combined with both upsampling and reshaping (ID 5), it slightly diminished performance when it was used with reshaping alone (as seen when comparing IDs 2 and 3). These findings validate the effectiveness of our architectural design, demonstrating that the full integration of all three components yields the most robust and accurate point cloud completion results.

3.5. Robustness Test

We conduct all robustness tests on the cotton-leaf dataset. In the first test, we vary the extent of incompletion to 50% and 75% while keeping the output cardinality fixed (K = 2048) and evaluating under the same protocol (normalization, FPS, CD/L1, CD/L2, EMD, F-score). The missing pattern is Axial-Cut, where a half-space cut removes a large sheet-like portion of the lamina. As shown in Figure 10, our network reconstructs coherent contours, restores serrations and vein continuity, and maintains uniform sampling even when three quarters of the points are removed. In the second test, we switch to Sphere-Removal missing, where spherical windows carve local voids across the leaf surface at the same 50%/75% ratios. Figure 11 illustrates that the model reliably infills these punctate holes, suppresses artifacts, and preserves global shape and boundary sharpness with balanced point density. Together, the two tests demonstrate that the proposed method is robust to both sheet-like and localized occlusions under severe sparsity. Across Axial-Cut and Sphere-Removal tests at 50%/75% missing ratios, the model sustains low, stable losses and produces visually faithful, uniformly sampled completions, verifying that the framework generalizes well to distinct occlusion geometries and severe sparsity and is applicable to diverse data-loss conditions.

4. Discussion

4.1. Quantitative and Quality Analysis

The quantitative results across multiple benchmarks consistently showed that the proposed method either outperformed or closely matched state-of-the-art models, particularly under challenging conditions with 75% missing data. While PMP-Net attains the lowest L 1 and L 2 errors and SnowflakeNet the lowest EMD among the reference methods, the proposed model achieves the highest F-score, exceeding SnowflakeNet by 809.9%, reflecting superior structural completeness and point-distribution uniformity. In addition, the proposed model reduces the L 2 distance by 76.4% relative to PMP-Net and the EMD by 24.6% relative to SnowflakeNet.
Importantly, quantitative metrics alone may not provide a comprehensive assessment of reconstruction quality. Previous studies have demonstrated that models that achieve high scores on global metrics such as EMD or CD can still produce suboptimal outcomes with respect to spatial uniformity and perceptual realism [34,41]. Although SnowflakeNet achieves superior results in certain numerical indicators, qualitative analysis reveals significant limitations in its practical application. In particular, qualitative evaluations reveal that SnowflakeNet often generates nonuniform point distributions and pronounced layering artifacts, corroborating the challenges previously identified by Sun et al. [24] and Weng et al. [42]. These deficiencies are especially evident in the reconstruction of fine-scale structures such as leaf edges and venation, thereby compromising both the morphological integrity and the perceptual fidelity of the reconstructed point clouds. This limitation has also been highlighted in prior evaluations of generative point cloud models [43]. In contrast, the proposed model consistently achieves visually superior reconstructions, preserving intricate structural details and maintaining a uniform distribution, both of which are critical for practical applications and downstream phenotypic analysis. This observation underscores a key limitation of relying exclusively on global distance-based metrics, because these metrics may fail to capture local structural fidelity and visual realism. The practical utility of point cloud completion methods is therefore more accurately reflected through qualitative visualization and a comprehensive, multimetric evaluation strategy rather than through single-metric rankings alone.
Building upon this insight, further analysis was conducted through ablation studies to elucidate the specific contributions of individual architectural modules. Both the reshape and upsampling modules were shown to be indispensable for achieving optimal geometric fidelity and improving the F-score, while the parameter-freezing module played a crucial role in stabilizing the training process and accelerating convergence. The integration of these modules enables the network to robustly reconstruct complex geometric patterns and maintain a uniform point distribution, even in the presence of substantial data loss. Together, these findings underscore the necessity of combining strong quantitative performance with qualitative robustness to ensure reliable application in real-world phenotypic analysis tasks.

4.2. Loss Function Freezing Strategy

The loss function freezing strategy demonstrates strong applicability in multitask deep learning networks. By selectively freezing the loss function parameters of specific tasks at designated training stages, this approach reduces task interference and enables each task to learn independently and efficiently. In multitask learning scenarios, the optimization of multiple objectives often introduces gradient conflicts, which can impede training stability and hinder convergence. Numerous studies have shown that loss freezing and adaptive loss scheduling are effective in mitigating these conflicts, thereby facilitating stable and efficient optimization [44,45,46]. This is especially important in complex networks where shared parameters are influenced by competing gradients from different tasks. Loss function freezing has been successfully applied in computer vision and natural language processing to improve training stability, accelerate convergence, and enhance the robustness of multitask architectures [47]. Techniques such as GradNorm [48], adaptive loss balancing [49], and curriculum-based or phased training [50] have improved in both convergence performance and generalization performance by dynamically adjusting the contributions of different losses during training. In the context of point cloud completion, the loss function freezing strategy implemented in this study effectively stabilized the training process, mitigated gradient conflicts, and improved convergence rates. Consequently, the model exhibited greater robustness when processing complex datasets, particularly cotton leaf point clouds with extensive missing regions. These results are consistent with broader findings in multitask learning literature, where loss freezing strategies are recognized for enhancing optimization efficiency and overall predictive quality.
Despite these strengths, several limitations should be acknowledged. The proposed approach was validated exclusively on a self-constructed cotton leaf dataset, and its generalizability to other plant species and diverse point cloud modalities remains to be established. The computational complexity introduced by multistage feature extraction and upsampling could also limit its feasibility for real-time deployment in large-scale or high-throughput scenarios. The current framework is designed primarily for geometric completion and does not yet incorporate color or spectral information, which could provide additional improvements in the precision and utility of phenotypic analysis. Future research will focus on extending the model’s applicability to a wider variety of plant phenotypes and real-world field datasets. The integrating of multimodal data, such as hyperspectral or RGB point clouds, has the potential to further increase the effectiveness of plant health monitoring. Continued algorithmic advancements aimed at lowering computational cost, improving scalability, and incorporating uncertainty estimation for more reliable completion in highly ambiguous cases will also be pursued.

5. Conclusions

To address the challenges posed by large-scale missing data (missing ratio ≥ 50%) and severe sparsity (points ≤ 128) in cotton leaf point clouds, an end-to-end super-resolution completion network is proposed. This network is designed to reconstruct uniform, high-resolution, and fully complete 3D point cloud models. It integrates an MRE to extract hierarchical features from locally incomplete point clouds, thereby enhancing the representation of both semantic and geometric attributes. A PPD is employed to iteratively generate missing point clouds layer by layer, enabling precise reconstruction of both missing regions and intricate details. To ensure consistency and uniformity, PointNet++ is leveraged to fuse the incomplete input point cloud with the generated completed point cloud, producing a fully reconstructed and evenly distributed point cloud model. Additionally, a loss function freezing strategy is implemented, allowing the network to efficiently execute task-specific objectives across different training stages, ultimately enhancing the overall reconstruction accuracy and quality.
Extensive experiments demonstrated that the proposed model delivers robust and accurate completion even under extreme scenarios characterized by large-scale data loss and severe sparsity. The reconstructed point sets are dense, uniform, and boundary-preserving, reducing fragmentation and sampling bias and supporting consistent geometric analysis across growth stages and occlusion patterns. This strong performance under the most challenging conditions implies that the model is also well-suited for small-scale or mild sparsity cases, where completion quality can be expected to be even higher. These results indicate that the proposed framework provides a reliable and generalizable solution for 3D point cloud completion in diverse agricultural phenotyping applications.

Author Contributions

H.G.: Investigation, Writing—original draft, Methodology, Data curation, Writing—review and editing. C.S.: Writing—review and editing, Formal analysis, Resources, Conceptualization. M.S.: Funding acquisition, Supervision, Project administration. J.P.: Writing—review and editing. Z.Y.: Visualization, Methodology, Validation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the South Xinjiang Key Industry Innovation Development Support Plan [grant number 2023AB063].

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, J.; Li, K.; Gao, Y.; Feng, D.; Zheng, C.; Cao, C.; Sun, J.; Dang, H.; Hamani, A.K. Evaluation of saline water irrigation on cotton growth and yield using the AquaCrop crop simulation model. Agric. Water Manag. 2022, 261, 107355. [Google Scholar] [CrossRef]
  2. Shuchun, M.; Yabing, L.; Zhanbiao, W.; Xiaoyu, Z.; Pengcheng, L.; Huiyun, X. Cotton Production in China; Wiley: Hoboken, NJ, USA, 2019; pp. 175–191. [Google Scholar] [CrossRef]
  3. Arekhi, E.; Ghasemi Bezdi, K.; Ajam Norozei, H.; Faghani, E. The effect of growth regulators on biochemical properties, yield, and fiber quality of different cultivars of cotton (Gossypium hirsutum) under different irrigation intervals. J. Plant Growth Regul. 2023, 42, 5574–5586. [Google Scholar] [CrossRef]
  4. Li, P.S.; Lu, Y.D.; Yu, G.; Zhang, Q.P.; Liu, T.F.; Wang, T.H.; Yang, M.F.; Xiang, D.; Tian, J.S.; Zhang, W.F. The Regional Distribution of Raw Cotton Quality in xinjiang Based on Notarized Inspection Data for Cotton. Sci. Agric. Sin. 2025, 58, 58–74. [Google Scholar] [CrossRef]
  5. Tan, S.; Wang, Q.; Zhang, J.; Chen, Y.; Shan, Y.; Xu, D. Performance of AquaCrop model for cotton growth simulation under film-mulched drip irrigation in southern Xinjiang, China. Agric. Water Manag. 2018, 196, 99–113. [Google Scholar] [CrossRef]
  6. Pabuayon, I.L.; Sun, Y.; Guo, W.; Ritchie, G.L. High-throughput phenotyping in cotton: A review. J. Cotton Res. 2019, 2, 18. [Google Scholar] [CrossRef]
  7. Xu, R.; Li, C.; Paterson, A.H. Multispectral imaging and unmanned aerial systems for cotton plant phenotyping. PLoS ONE 2019, 14, e0205083. [Google Scholar] [CrossRef]
  8. Yuniarti, A.; Suciati, N. A review of deep learning techniques for 3D reconstruction of 2D images. In Proceedings of the 2019 12th International Conference on Information & Communication Technology and System (ICTS), Surabaya, Indonesia, 18 July 2019; IEEE: New York, NY, USA, 2019; pp. 327–331. [Google Scholar] [CrossRef]
  9. Munir, M.; Chattha, M.A.; Dengel, A.; Ahmed, S. A comparative analysis of traditional and deep learning-based anomaly detection methods for streaming data. In Proceedings of the 2019 18th IEEE International Conference on Machine Learning and Applications (ICMLA), Boca Raton, FL, USA, 16–19 December 2019; IEEE: New York, NY, USA, 2019; pp. 561–566. [Google Scholar] [CrossRef]
  10. Wang, P.; Fan, E.; Wang, P. Comparative analysis of image classification algorithms based on traditional machine learning and deep learning. Pattern Recognit. Lett. 2021, 141, 61–67. [Google Scholar] [CrossRef]
  11. Shah, D.; Gopan, K.G.; Sinha, N. An investigation of the multi-dimensional (1D vs. 2D vs. 3D) analyses of EEG signals using traditional methods and deep learning-based methods. Front. Signal Process. 2022, 2, 936790. [Google Scholar] [CrossRef]
  12. Engel, N.; Belagiannis, V.; Dietmayer, K. Point transformer. IEEE Access 2021, 9, 134826–134840. [Google Scholar] [CrossRef]
  13. Jiang, J.; Li, L.; Wang, M. Research on plant stem complement based on L1-medial skeleton extraction. J. Nanjing For. Univ. 2022, 46, 40–50. [Google Scholar] [CrossRef]
  14. Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3dmatch: Learning local geometric descriptors from RGB-D reconstructions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1802–1811. [Google Scholar] [CrossRef]
  15. Qiu, Q.; Sun, N.; Bai, H.; Wang, N.; Fan, Z.; Wang, Y.; Meng, Z.; Li, B.; Cong, Y. Field-based high-throughput phenotyping for maize plant using 3D LiDAR point cloud generated with a “Phenomobile”. Front. Plant Sci. 2019, 10, 554. [Google Scholar] [CrossRef]
  16. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar] [CrossRef]
  17. Xiao, H.; Xu, H.; Li, Y.; Kang, W. Multi-dimensional graph interactional network for progressive point cloud completion. IEEE Trans. Instrum. Meas. 2022, 72, 1–2. [Google Scholar] [CrossRef]
  18. Wang, J.; Chakraborty, R.; Stella, X.Y. Transformer for 3D point clouds. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4419–4431. [Google Scholar] [CrossRef]
  19. Pan, L.; Liu, L.; Condon, A.G.; Estavillo, G.M.; Coe, R.A.; Bull, G.; Stone, E.A.; Petersson, L.; Rolland, V. Biomass prediction with 3D point clouds from LiDAR. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2022; pp. 1330–1340. [Google Scholar]
  20. Xie, H.; Yao, H.; Zhou, S.; Mao, J.; Zhang, S.; Sun, W. Grnet: Gridding residual network for dense point cloud completion. In European Conference on Computer Vision; Springer International Publishing: Cham, Germany, 2020; pp. 365–381. [Google Scholar] [CrossRef]
  21. Huang, Z.; Yu, Y.; Xu, J.; Ni, F.; Le, X. Pf-net: Point fractal network for 3d point cloud completion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7662–7670. [Google Scholar] [CrossRef]
  22. Wen, X.; Xiang, P.; Han, Z.; Cao, Y.P.; Wan, P.; Zheng, W.; Liu, Y.S. Pmp-net: Point cloud completion by learning multi-step point moving paths. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; IEEE: New York, NY, USA, 2021; pp. 7443–7452. [Google Scholar] [CrossRef]
  23. Wen, X.; Xiang, P.; Han, Z.; Cao, Y.P.; Wan, P.; Zheng, W.; Liu, Y.S. PMP-Net++: Point cloud completion by transformer-enhanced multi-step point moving paths. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 852–867. [Google Scholar] [CrossRef]
  24. Xiang, P.; Wen, X.; Liu, Y.S.; Cao, Y.P.; Wan, P.; Zheng, W.; Han, Z. Snowflakenet: Point cloud completion by snowflake point deconvolution with skip-transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; IEEE: New York, NY, USA, 2021; pp. 5499–5509. [Google Scholar] [CrossRef]
  25. Dinesh, C.; Cheung, G.; Bajić, I.V. 3D point cloud super-resolution via graph total variation on surface normals. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 4390–4394. [Google Scholar] [CrossRef]
  26. Levin, D. The approximation power of moving least-squares. Math. Comput. 1998, 67, 1517–1531. [Google Scholar] [CrossRef]
  27. Wang, X.; Hu, J.; Guo, L.; Zhang, D.; Qin, H.; Hao, A. Feature-preserving, mesh-free empirical mode decomposition for point clouds and its applications. Comput. Aided Geom. Des. 2018, 59, 1–16. [Google Scholar] [CrossRef]
  28. Kumbar, A.; Anvekar, T.; Vikrama, T.A.; Tabib, R.A.; Mudenagudi, U. Tp-node: Topology-aware progressive noising and denoising of point clouds towards upsampling. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 2272–2282. [Google Scholar] [CrossRef]
  29. Field, D.A. Laplacian Smoothing Delaunay triangulations. Commun. Appl. Numer. Methods 1988, 4, 709–712. [Google Scholar] [CrossRef]
  30. Balasubramanian, M.; Schwartz, E.L. The isomap algorithm and topological stability. Science 2002, 295, 7. [Google Scholar] [CrossRef]
  31. Ward, C.M.; Harguess, J.; Crabb, B.; Parameswaran, S. Image quality assessment for determining efficacy and limitations of Super-Resolution Convolutional Neural Network (SRCNN). Appl. Digit. Image Process. XL 2017, 10396, 19–30. [Google Scholar] [CrossRef]
  32. Vint, D.; Di Caterina, G.; Soraghan, J.J.; Lamb, R.A.; Humphreys, D. Evaluation of performance of VDSR super resolution on real and synthetic images. In Proceedings of the 2019 Sensor Signal Processing for Defence Conference (SSPD), Brighton, UK, 9–10 May 2019; pp. 1–5. [Google Scholar] [CrossRef]
  33. Talab, M.A.; Awang, S.; Najim, S.A. Super-low resolution face recognition using integrated efficient sub-pixel convolutional neural network (ESPCN) and convolutional neural network (CNN). In Proceedings of the 2019 IEEE International Conference on Automatic Control and Intelligent Systems (I2CACIS), Selangor, Malaysia, 29 June 2019; pp. 331–335. [Google Scholar] [CrossRef]
  34. Yu, L.; Li, X.; Fu, C.W.; Cohen-Or, D.; Heng, P.A. Pu-net: Point cloud upsampling network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2790–2799. [Google Scholar] [CrossRef]
  35. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep hierarchical feature learning on point sets in a metric space. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar] [CrossRef]
  36. Wu, H.; Zhang, J.; Huang, K. Point Cloud Super Resolution with Adversarial Residual Graph Networks. arxiv 2019, arXiv:1908.02111. [Google Scholar] [CrossRef]
  37. Li, R.; Li, X.; Fu, C.W.; Cohen-Or, D.; Heng, P.A. Pu-gan: A point cloud upsampling adversarial network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 November–2 December 2019; pp. 7203–7212. [Google Scholar] [CrossRef]
  38. Yu, F.; Koltun, V.; Funkhouser, T. Dilated Residual Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 636–644. [Google Scholar] [CrossRef]
  39. Kathariya, B.; Akthar, A.; Li, Z.; Van Der Auwera, G. PointCU: Multiscale Sparse Convolutional Learning for Point Cloud Color Upsampling. In Proceedings of the 2024 IEEE International Conference on Visual Communications and Image Processing (VCIP), Tokyo, Japan, 8–11 December 2024; pp. 1–5. [Google Scholar] [CrossRef]
  40. Wu, X.; Lao, Y.; Jiang, L.; Liu, X.; Zhao, H. Point Transformer V2: Grouped Vector Attention and Partition-based Pooling. arXiv 2022, arXiv:2210.05666. [Google Scholar] [CrossRef]
  41. Wen, X.; Li, T.; Han, Z.; Liu, Y.S. Point Cloud Completion by Skip-attention Network with Hierarchical Folding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1936–1945. [Google Scholar] [CrossRef]
  42. Wang, X.; Ang, M.H.; Lee, G.H. Point cloud completion by learning shape priors. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 10719–10726. [Google Scholar] [CrossRef]
  43. Yuan, W.; Khot, T.; Held, D.; Mertz, C.; Hebert, M. PCN: Point Completion Network. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 728–737. [Google Scholar] [CrossRef]
  44. Sener, O.; Koltun, V. Multi-Task Learning as Multi-Objective Optimization. in: Advances in neural information processing systems. arXiv 2018, arXiv:1810.04650. [Google Scholar] [CrossRef]
  45. Chen, Z.; Badrinarayanan, V.; Lee, C.Y.; Rabinovich, A. Gradnorm: Gradient normalization for adaptive loss balancing in deep multitask networks. Int. Conf. Mach. Learn. 2018, 80, 794–803. [Google Scholar] [CrossRef]
  46. Liu, S.; Johns, E.; Davison, A.J. End-to-end multi-task learning with attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1871–1880. [Google Scholar] [CrossRef]
  47. Zhang, Y.; Yang, Q. An overview of multi-task learning. Natl. Sci. Rev. 2018, 5, 30–43. [Google Scholar] [CrossRef]
  48. Gong, T.; Lee, T.; Stephenson, C.; Renduchintala, V.; Padhy, S.; Ndirango, A.; Keskin, G.; Elibol, O.H. A comparison of loss weighting strategies for multi task learning in deep neural networks. IEEE Access 2019, n7, 141627–141632. [Google Scholar] [CrossRef]
  49. Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar] [CrossRef]
  50. Xiao, Y.; Tian, Z.; Yu, J.; Zhang, Y.; Liu, S.; Du, S.; Lan, X. A review of object detection based on deep learning. Multimed. Tools Appl. 2020, 79, 23729–23791. [Google Scholar] [CrossRef]
Figure 1. Workflow for cotton leaf dataset acquisition and point-cloud generation. (a) Multiview image capture setup with a rotating turntable and arrayed cameras. (b) Weekly image collection from the top, side, and bottom viewpoints. (c) Sparse point-cloud reconstruction via structure-from-motion. (d) Dense point-cloud reconstruction through multiview stereo. (e) Manual labelling of point-cloud data into plant organs (red: cotton leaves, blue: stems and bolls). (f) Segregation of cotton leaf points (red) from other organs (blue). (g) Normalization of segmented leaf point clouds (green) into a canonical coordinate frame.
Figure 1. Workflow for cotton leaf dataset acquisition and point-cloud generation. (a) Multiview image capture setup with a rotating turntable and arrayed cameras. (b) Weekly image collection from the top, side, and bottom viewpoints. (c) Sparse point-cloud reconstruction via structure-from-motion. (d) Dense point-cloud reconstruction through multiview stereo. (e) Manual labelling of point-cloud data into plant organs (red: cotton leaves, blue: stems and bolls). (f) Segregation of cotton leaf points (red) from other organs (blue). (g) Normalization of segmented leaf point clouds (green) into a canonical coordinate frame.
Agriculture 15 01989 g001
Figure 2. Architecture of the super-resolution point-cloud completion network. (a) The point-cloud completion and super-resolution module integrates multiresolution extraction (MRE) and point-pyramid decomposition (PPD) for hierarchical upsampling and fine-detail refinement. (b) Point-cloud reshape module adapted from PointNet++ for geometry adjustment, featuring loss-freeze connections to stabilize training.
Figure 2. Architecture of the super-resolution point-cloud completion network. (a) The point-cloud completion and super-resolution module integrates multiresolution extraction (MRE) and point-pyramid decomposition (PPD) for hierarchical upsampling and fine-detail refinement. (b) Point-cloud reshape module adapted from PointNet++ for geometry adjustment, featuring loss-freeze connections to stabilize training.
Agriculture 15 01989 g002
Figure 3. Improved PF-Net architecture. (a) The bilinear interpolation module upsamples the input point cloud to generate multiscale representations at N , ( k 1 ) N , and k N . (b) The MRE extracts and fuses hierarchical features from the multiscale inputs to form a latent embedding. (c) The PPD hierarchically upsamples and reconstructs the latent embedding through three decoding layers, producing and fusing outputs ( α , β , and γ ) to generate the final k N high-resolution point cloud.
Figure 3. Improved PF-Net architecture. (a) The bilinear interpolation module upsamples the input point cloud to generate multiscale representations at N , ( k 1 ) N , and k N . (b) The MRE extracts and fuses hierarchical features from the multiscale inputs to form a latent embedding. (c) The PPD hierarchically upsamples and reconstructs the latent embedding through three decoding layers, producing and fusing outputs ( α , β , and γ ) to generate the final k N high-resolution point cloud.
Agriculture 15 01989 g003
Figure 4. Training loss comparisons of the point-cloud completion methods on the cotton leaf dataset under 50% and 75% missing-data scenarios. (a) Loss curves of GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, and PointTR v2 at a 50% missing rate; (b) Loss curve of the proposed model at a 50% missing rate; (c) Loss curves of GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, and PointTR v2 at a 75% missing rate; (d) Loss curve of the proposed model at a 75% missing rate.
Figure 4. Training loss comparisons of the point-cloud completion methods on the cotton leaf dataset under 50% and 75% missing-data scenarios. (a) Loss curves of GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, and PointTR v2 at a 50% missing rate; (b) Loss curve of the proposed model at a 50% missing rate; (c) Loss curves of GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, and PointTR v2 at a 75% missing rate; (d) Loss curve of the proposed model at a 75% missing rate.
Agriculture 15 01989 g004
Figure 5. Qualitative comparison of cotton leaf point-cloud reconstructions under a 75% missing-data rate. Each row corresponds to one representative leaf sample; columns from left to right show the input sparse cloud, baseline reconstructions (GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, PoinTR v2), the proposed method, and the ground-truth dense point cloud. Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Figure 5. Qualitative comparison of cotton leaf point-cloud reconstructions under a 75% missing-data rate. Each row corresponds to one representative leaf sample; columns from left to right show the input sparse cloud, baseline reconstructions (GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, PoinTR v2), the proposed method, and the ground-truth dense point cloud. Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Agriculture 15 01989 g005
Figure 6. Fine-grained comparison of leaf edge reconstruction with SnowflakeNet and the proposed method under 75% missing data. Columns from left to right: Input, SnowflakeNet, Ours, and Ground Truth (G.T.). Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Figure 6. Fine-grained comparison of leaf edge reconstruction with SnowflakeNet and the proposed method under 75% missing data. Columns from left to right: Input, SnowflakeNet, Ours, and Ground Truth (G.T.). Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Agriculture 15 01989 g006
Figure 7. Qualitative comparison of cotton leaf point-cloud reconstructions under a 50% missing-data rate. Each row represents a distinct leaf sample; columns from left to right represent the input sparse cloud, baseline reconstructions (GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, PointTR v2), the proposed method, and the ground-truth dense point cloud. Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Figure 7. Qualitative comparison of cotton leaf point-cloud reconstructions under a 50% missing-data rate. Each row represents a distinct leaf sample; columns from left to right represent the input sparse cloud, baseline reconstructions (GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, PointTR v2), the proposed method, and the ground-truth dense point cloud. Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Agriculture 15 01989 g007
Figure 8. Fine-grained visual comparison of leaf edge reconstruction under 50% missing data between SnowflakeNet and the proposed method. Columns from left to right: Input, SnowflakeNet, Ours, and Ground Truth (G.T.). Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Figure 8. Fine-grained visual comparison of leaf edge reconstruction under 50% missing data between SnowflakeNet and the proposed method. Columns from left to right: Input, SnowflakeNet, Ours, and Ground Truth (G.T.). Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Agriculture 15 01989 g008
Figure 9. Qualitative comparison of point cloud completion results for sparse cotton leaves using different methods. The figure shows, from left to right: input point clouds, results generated by GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, PointTR v2, the proposed method, and ground truth. Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Figure 9. Qualitative comparison of point cloud completion results for sparse cotton leaves using different methods. The figure shows, from left to right: input point clouds, results generated by GRNet, PF-Net, PMP-Net, PMP-Net++, SnowflakeNet, PointTR v2, the proposed method, and ground truth. Numbers (1)–(6) correspond to six representative incomplete cotton leaf samples from the dataset.
Agriculture 15 01989 g009
Figure 10. Examples of repairing results on the Sphere-Removal Missing dataset. Columns show Input, 50%, 75%, and G.T.; green denotes the prediction, brown denotes the undamaged input, and purple denotes the ground truth.
Figure 10. Examples of repairing results on the Sphere-Removal Missing dataset. Columns show Input, 50%, 75%, and G.T.; green denotes the prediction, brown denotes the undamaged input, and purple denotes the ground truth.
Agriculture 15 01989 g010
Figure 11. Examples of repairing results on the Axial-Cut Missing dataset. Columns show Input, 50%, 75%, and G.T.; green denotes the prediction, brown denotes the undamaged input, and purple denotes the ground truth.
Figure 11. Examples of repairing results on the Axial-Cut Missing dataset. Columns show Input, 50%, 75%, and G.T.; green denotes the prediction, brown denotes the undamaged input, and purple denotes the ground truth.
Agriculture 15 01989 g011
Table 1. Quantitative evaluation of point-cloud completion methods on the cotton leaf dataset with a 75% missing-data rate.
Table 1. Quantitative evaluation of point-cloud completion methods on the cotton leaf dataset with a 75% missing-data rate.
Means L 1 L 2 EMDF-Score
GRNet [20]99.981260.62020.06290.1158
PF-Net [21]49.10547.75510.08850.0765
PMP-Net [22]47.29246.52680.07310.0285
PMP-Net++ [23]57.20519.30050.09350.01
SnowflakeNet [24]56.15539.85520.05360.0131
PointTR v2 [40]45.19876.69140.10220.0519
Ours24.52461.53960.04040.1192
Table 2. Quantitative evaluation of point-cloud completion methods on the cotton leaf dataset with a 50% missing-data rate.
Table 2. Quantitative evaluation of point-cloud completion methods on the cotton leaf dataset with a 50% missing-data rate.
Means L 1 L 2 EMDF-Score
GRNet [20]57.959626.43450.04660.1794
PF-Net [21]53.94369.46550.08020.0857
PMP-Net [22]41.75315.20570.08080.0764
PMP-Net++ [23]47.98356.60250.08200.0230
SnowflakeNet [24]32.24163.73450.05530.1858
PointTR v2 [40]42.20605.84530. 14070. 0119
Ours37.02534.61370.05330.1329
Table 3. Quantitative evaluation of point-cloud completion methods on sparse cotton leaf data.
Table 3. Quantitative evaluation of point-cloud completion methods on sparse cotton leaf data.
Means L 1 L 2 EMDF-Score
GRNet [20]108.1170 83.90740.09750.0291
PF-Net [21]34.69403.41820.05130.1004
PMP-Net [22]39.5188 4.49920.07640.0139
PMP-Net++ [23]44.6338 5.51380.07950.0048
SnowflakeNet [24]26.6929 1.79990.02060.0900
PointTR v2 [40]25.29292.32160.34930.0787
Ours28.75652.11110.17800.0869
Table 4. Ablation study of module contributions to point-cloud completion performance under a 75% missing-data rate on the cotton leaf dataset. Model variants include upsampling, reshaping, and freezing modules.
Table 4. Ablation study of module contributions to point-cloud completion performance under a 75% missing-data rate on the cotton leaf dataset. Model variants include upsampling, reshaping, and freezing modules.
ID 1UpsampleReshapeFrozen L 1 L 2 F1EMD
1××45.65286.7480 0.08380.0669
2××31.5832.61880.07630.0458
3×30.77752.51970.08480.0488
4×27.00971.81670.08930.0410
526.38921.81160.17110.0404
1 Each combination is assigned a unique ID ranging from 1 to 5. Bold values indicate the best performance across the compared settings.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Geng, H.; Yin, Z.; Shi, M.; Pan, J.; Si, C. Super-Resolution Point Cloud Completion for Large-Scale Missing Data in Cotton Leaves. Agriculture 2025, 15, 1989. https://doi.org/10.3390/agriculture15181989

AMA Style

Geng H, Yin Z, Shi M, Pan J, Si C. Super-Resolution Point Cloud Completion for Large-Scale Missing Data in Cotton Leaves. Agriculture. 2025; 15(18):1989. https://doi.org/10.3390/agriculture15181989

Chicago/Turabian Style

Geng, Hui, Zhiben Yin, Mingdeng Shi, Junzhang Pan, and Chunjing Si. 2025. "Super-Resolution Point Cloud Completion for Large-Scale Missing Data in Cotton Leaves" Agriculture 15, no. 18: 1989. https://doi.org/10.3390/agriculture15181989

APA Style

Geng, H., Yin, Z., Shi, M., Pan, J., & Si, C. (2025). Super-Resolution Point Cloud Completion for Large-Scale Missing Data in Cotton Leaves. Agriculture, 15(18), 1989. https://doi.org/10.3390/agriculture15181989

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop