Next Article in Journal
Efficient Modeling of Deterministic Decision Trees for Recognition of Realizable Decision Rules: Bounds on Weighted Depth
Previous Article in Journal
Derivation of a Closed-Form Asymptotic Variance for the Coefficient of Variation Under the Reparameterized Birnbaum–Saunders Distribution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion

1
Key Laboratory of Digital Economy and Social Computing Science of Gansu, Lanzhou 730020, China
2
School of Statistics and Data Science, Lanzhou University of Finance and Economics, Lanzhou 730020, China
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(11), 793; https://doi.org/10.3390/axioms14110793 (registering DOI)
Submission received: 2 October 2025 / Revised: 24 October 2025 / Accepted: 25 October 2025 / Published: 28 October 2025

Abstract

The integrity of real-time monitoring data is paramount to the accuracy of scientific research and the reliability of decision-making. However, data incompleteness arising from transmission interruptions or extreme weather disrupting equipment operations severely compromises the validity of statistical analyses and the stability of modelling. From a mathematical view, real-time monitoring data may be regarded as continuous functions, exhibiting intricate correlations and mutual influences between different indicators. Leveraging their inherent smoothness and interdependencies enables high-precision data imputation. Within the functional data analysis framework, this paper proposes a Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion (DCAGMFMC) method. Integrating multi-view learning with an adaptive graph strategy, this approach comprehensively accounts for complex correlations between data from different views while extracting differential information across views, thereby enhancing information utilization and imputation accuracy. Random simulation experiments demonstrate that the DCAGMFMC method exhibits significant imputation advantages over classical methods such as KNN, HFI, SFI, MVNFMC, and GRMFMC. Furthermore, practical applications on meteorological datasets reveal that, compared to these imputation methods, the root mean square error (RMSE), mean absolute error (MAE), and normalized root mean square error (NRMSE) of the DCAGMVNFMC method decreased by an average of 39.11% to 59.15%, 54.50% to 71.97%, and 43.96% to 63.70%, respectively. It also demonstrated stable imputation performance across various meteorological indicators and missing data rates, exhibiting good adaptability and practical value.

1. Introduction

With the iterative advancement of data acquisition technologies, functional data exhibiting curve characteristics have proliferated across diverse domains. Examples include air pollutant concentration data and meteorological records. However, due to factors such as data transmission interruptions and extreme weather disrupting equipment operation, discrete sampling data often remains incomplete. This manifests as extensive random, strip-like, or block-shaped gaps, rendering data missingness a prevalent quality issue in functional datasets. Given that mainstream data analysis tools rely on complete datasets, imputation of missing values becomes a crucial preliminary step to ensure the accuracy of analytical outcomes.
In recent years, scholars have conducted in-depth research into functional data imputation methods. Existing imputation approaches [1] primarily encompass statistical analysis methods, machine learning methods, and deep learning methods [2]. Traditional statistical imputation techniques, such as mean-based imputation, perform poorly with functional data due to their failure to adequately account for the data’s curvilinear characteristics and the uncertainty inherent in missing patterns [3]. When non-linear relationships exist between variables, imputation based on linear regression yields suboptimal results [4]. For sparse functional data, Yao et al. [5] proposed Conditional Expectation Principal Component Estimation (PACE), leveraging principal component information. As a single imputation method, its performance is severely constrained by the data distribution and struggles to effectively capture non-linear relationships. The Multiple Imputation by Chain Equations (MICE) method proposed by Royston et al. [6] imputes missing values by generating multiple plausible yet distinct datasets. However, it remains sensitive to specific non-repetitive patterns. He et al. [7] introduced a functional multiple imputation approach under functional mixed-effects models, but this method imposes linearity constraints on the model assumptions and requires that the cause of data missingness be independent of the unobserved data itself. Ciarleglio et al. [8] combined MICE with functional regression to develop the fregMICE algorithm, although this approach is computationally intensive, necessitating advanced processing capabilities. When the growth rates of curve observations and sample size diverge, model parameter estimation lacks consistency, prompting the development of modern multiple imputation methods for functional data [9]. While multiple imputation is suitable for small-scale missingness in traditional datasets, it struggles to address large-scale missingness in high-dimensional matrices.
Furthermore, distance-based spatial imputation methods possess the ability to capture local characteristics in functional data analysis. Among these, inverse distance weighting (IDW) imputation stands as a classical approach, estimating missing values by calculating distance weights with respect to the adjacent sampling points. Its core assumption is that spatially proximate objects exhibit greater similarity; however, this method suffers from the flaw that weights tend towards infinity during close-range imputation [10]. Liu Yan et al. [11] further noted that such methods inadequately account for global data correlations. Spline imputation demonstrates significant advantages in handling missing values due to its smoothness and local adaptability. Its core principle involves constructing a globally smooth curve using piecewise low-order polynomials [12]. Compared to traditional linear imputation, this method better captures non-linear trends. However, spline imputation is relatively sensitive to data noise [11], meaning reliance solely on spline imputation may amplify the impact of high-frequency noise [13].
Traditional functional data imputation methods, such as soft functional matrix completion (SFI) [14] and hard functional matrix completion (HFI) [14], demonstrate certain efficacy in functional data imputation. However, they exhibit insufficient information extraction from the view of multivariate functional data [15]. KNN imputation [16], based on similarity metrics, performs imputation by identifying the K nearest samples to the missing data points, demonstrating excellent performance in preserving local data characteristics [17]. However, KNN imputation is sensitive to missing patterns. Research indicates that the regularized Expectation–Maximization (EM) method performs better when input variables exhibit little to no interdependence [18]. Matrix Factorization (MF) aims to discover a set of latent features (i.e., factor matrices), generating complete matrix data through linear combinations of these latent features to achieve missing value imputation [19]. To enhance MF’s ability to handle time series data, Yu et al. [20] introduced temporal regularization constraints, subsequently proposing a TRMF framework capable of data-driven temporal learning and forecasting. This method automatically learns dependencies between time series. Concurrently, deep learning approaches such as the time series imputation method proposed by He et al. [21] leverage joint tensor completion and recurrent neural networks to account for both temporal dependencies within individual time series and correlations between multiple time series. Furthermore, while the time series missing value estimation method based on recurrent neural networks [22] does not require assumptions about the underlying generative process of the data, it struggles to capture long-term dependencies. The Neural Context-Aware Anomaly Detection (NCAD) method proposed by Carmona et al. [23] excels at capturing long-term dependencies and complex patterns, but it suffers from high computational complexity and computational cost. Yoon et al. [24] proposed an iterative optimization of missing value imputation using a generator-discriminator game, though this approach demands substantial computational resources and requires large datasets.
In summary, when addressing the imputation of complex non-linear functional data, the aforementioned methods fail to adequately account for the functional characteristics of the data, the complexity of large-scale missing data, and the inter-variable correlations inherent in functional data. Functional Data Analysis (FDA) [25] represents discrete observations of data in functional form. By extracting multivariate data information and modelling from a functional data view, it can effectively reduce data noise when interpolating missing values [26]. Matrix completion, as an imputation method for handling large-scale data missingness, is widely adopted in practical applications [27]. Furthermore, research has demonstrated that the functional data recovery problem is equivalent to rank-constrained matrix completion [28]. Considering that functional data are typically multivariate, and that multi-view learning (MVL) outperforms single-view learning [29], utilizing information from multiple views can extract more accurate data features [30]. Methods for handling missing data within a multi-view learning framework can be categorized into two types: firstly, integrating consistency information across multiple views into matrix completion; secondly, treating the neighboring information of one view as complementary information or another view, thereby mutually providing neighboring information between different views and ultimately integrating complementary information from multiple views into matrix completion [31]. Therefore, within the functional data analysis framework, incorporating multi-view learning into matrix completion enables more effective imputation of missing data by extracting inter-view correlation information. For instance, Xue Jiao et al. [15] developed the Multi-View Non-Negative Functional Matrix Completion (MVNFMC) algorithm, which demonstrates superior data restoration performance compared to existing single-view functional data completion methods. Gao Haiyan and Ma Wenjuan [32] proposed a novel multi-view functional matrix completion method (GRMFMC), considering the higher-order domain relationships and complementary information between functional data views. This approach demonstrated superior imputation performance relative to other typical methods.
The aforementioned method achieved certain data imputation effects, but it only utilized the shared and complementary information within multi-view data, while neglecting the differential information between these views [33]. Therefore, fully leveraging the synergistic effects of multiple views can significantly enhance the precision and reliability of data feature extraction. Based on this, this study explores a more efficient functional data imputation method within the functional data analysis framework. This approach fully considers the multivariate and large-scale missing characteristics of functional data, as well as the differential information between views. Leveraging the principle that stronger vector orthogonality reduces information overlap, multi-view learning not only considers shared information across views but also enhances imputation accuracy by introducing diversity constraints between views to utilize their differential information [34,35]. This ensures the preservation of unique attributes from different views during multi-view data fusion. Capturing geometric properties through the spatial structure and local relationships between data samples can significantly enhance representation performance [36]. Specifically, this involves constructing a nearest neighbor graph to model sample relationships, utilizing the regularized decomposition process of the graph Laplacian matrix. However, data graphs constructed using traditional KNN methods may misclassify neighbors, thereby degrading clustering performance [33,37]. Furthermore, while higher similarity matrices yield more stable models, stricter similarity matrices diminish model generalization capability. Consequently, adaptively learning similarity matrices by maximizing information entropy enables more uniform information distribution within the similarity matrix, balancing model stability and generalization capability [38].
Given the intricate interdependencies and mutual influences among multi-view data, this paper integrates diversity constraints and adaptive graph learning into a functional matrix completion framework within the functional data analytics paradigm, incorporating multi-view learning. We propose a Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion method (DCAGMFMC). This approach effectively addresses the limitations of existing functional data imputation methods in modelling complex nonlinear structures and integrating information across views by constructing a multi-view functional data representation framework and incorporating diversity constraints alongside adaptive graph learning strategies. It provides novel theoretical methods and technical pathways for missing data imputation in complex functional datasets. The principal contributions of this study are summarized as follows:
(1) Within the functional data analysis framework, integrating multi-view learning principles, we propose a novel multi-view functional matrix completion method. This approach pioneers the fusion of diversity constraints with adaptive graph learning within the functional matrix completion paradigm. By leveraging the diversity characteristics inherent in multi-view data, diversity constraints extract discriminative information across different views. By introducing information entropy, it adaptively learns similarity matrices, overcoming the shortcomings of traditional K-Nearest Neighbors methods in misclassifying neighbors during graph construction. Based on an adaptive graph strategy, it effectively captures complex correlations among data across views, deeply mining differential information between views. This significantly enhances data utilization efficiency, thereby substantially improving the imputation accuracy of functional data.
(2) Verification through random simulation experiments demonstrates that the DCAGMFMC method exhibits significant imputation advantages over classical methods such as KNN, HFI, SFI, MVNFMC, and GRMFMC, robustly proving its effectiveness and advancement in the field of functional data imputation.
(3) Practical application on meteorological datasets indicates that the DCAGMFMC method achieves substantial reductions in root mean square error (RMSE), mean absolute error (MAE), and normalized root mean square error (NRMSE). It also delivers stable imputation results across diverse meteorological indicators and varying levels of data missingness. This fully demonstrates the method’s excellent adaptability, practical value, and robust performance in repairing complex functional data gaps.
The structure of this article is as follows: Section 2 introduces prior work on curve fitting and functional matrix completion methods. Section 3 primarily details the formulation and optimization process of the DCAGMFMC method, including the convergence and time complexity of the submatrix update iteration. Section 4 presents simulation experiments and results analysis for the proposed method. Section 5 demonstrates practical applications of the proposed method. Section 6 summarizes the findings of the proposed method.

2. Related Work

2.1. Curve Fitting of Multi-View Functional Data

Dataset X ( t ) = { X 1 ( t ) , X 2 ( t ) , , X v ( t ) , , X n v ( t ) } comprises n v views, with X v ( t ) representing the data for the v -th viewpoint. Each viewpoint X v ( t ) contains n samples, whose observational dimensions are denoted as m v . Within X v ( t ) = { x 1 v ( t ) , x 2 v ( t ) , , x n v ( t ) } , there are n curves in total. The curve for the v -th viewpoint is x i v ( t ) ( i = 1 , 2 , , n ) . The j -th discrete observation value y ˜ i j v of the i -th curve under the v -th viewpoint is generated by the general regression model Equation (1):
y ˜ i j v = x i v ( t i j ) + ε i j v
where v denotes the random error term, and j = 1 , 2 , , m v represents the observation time points over the time interval T . Consequently, x i v ( t ) can be approximated by the following finite-dimensional Gaussian function expansion:
x i v ( t ) l = 1 r α i l v φ i l v ( t ) = φ i v ( t ) T α i v
Given a set of basis vectors φ i v ( t ) = ( φ i 1 v ( t ) , φ i 2 v ( t ) , , φ i r v ( t ) ) T and an estimated coefficient vector α i v = ( α i 1 v , α i 2 v , , α i r v ) T , Equations (1) and (2) can be expressed in matrix form respectively as follows:
X ˜ v + Φ v A v
Y ˜ v + = Φ v A v + E v
where Φ v = ( φ 1 v ( t ) , φ 2 v ( t ) , , φ m v v ( t ) ) T denotes the basis functions computed at the observation time point, and represents the basis function matrix shared across all viewpoints, A v = ( α 1 v , α 2 v , , α n v ) T , E v = ( ε 1 v , ε 2 v , , ε n v ) , ε i v = ( ε i 1 v , ε i 2 v , , ε i m v v ) .

2.2. Multi-View Functional Matrix Completion

Given Y = { Y ˜ 1 + , Y ˜ 2 + , , Y ˜ n v + } as a multi-view data matrix, the symbol ( · ) + denotes taking the positive part of the matrix, ensuring the non-negativity of the observed data, the fundamental objective function based on multi-view functional matrix completion [15] is:
min U v , V v 0 v = 1 n v θ v O ( Y ˜ v + Φ U v V v T ) F 2
Here, θ v denotes the weighting parameter for different views, represents the Hadamard matrix product, and Φ m × r and O m × n are projection matrices of the same type as Y ˜ v + . Specifically, if the entries in Y ˜ v + are observable, then O i j = 1 ; otherwise, O i j = 0 , U v r × d , and V v n × d respectively denote the basis matrix and coefficient matrix for the v -th view.

3. Methodology

3.1. DCAGMFMC

Multi-view data possess richer content than single-view data, exhibiting multiple features. However, existing methods overly rely on shared information across different views, neglecting the distinctiveness between each view. This leads to information loss and underutilization. By learning constrained coefficient matrices, we obtain diversified coefficient matrices, thereby extracting complementary information between different views of the same sample. Let the i -th coefficient vector corresponding to the v -th view be denoted as c, where v = 1 , 2 , , n v , i = 1 , 2 , n , and the i -th coefficient vector corresponding to the w -th view be denoted as V w i d , where w = 1 , 2 , , n v , i = 1 , 2 , n . The stronger the orthogonality between two vectors, the more diverse and non-redundant the information or features they carry [34]. Therefore, by minimizing the inner product between two coefficient vectors from different views, we extract the unique information specific to each view. For n vectors, we have:
min i = 1 n v v i T v w i = min tr ( V v V w T )
To understand the differences in information between the coefficient matrices of the v -th and other views across different samples, the diversity constraint can be expressed as:
min v = 1 n v w = 1 n v tr ( V v V w T )
Concurrently, high-dimensional spaces exhibit data redundancy. Sampling data from low-dimensional manifolds embedded within these high-dimensional spaces [36] enables the extraction of key information while reducing computational complexity. Owing to the multivariate nature of functional data, samples sharing identical views exhibit strong interconnections, revealing inherent patterns within the data. Consequently, preserving the intrinsic geometric structure of functional data is paramount. As a dimension reduction and feature extraction technique, NMF focuses on numerical relationships within data while neglecting its geometric structure. Graph-based NMF captures this geometric structure, with the graph regularization expressed as:
min 1 2 i = 1 n j = 1 n v v i v v j 2 2 s v i j = min tr ( V v T L v V v )
Here, S v denotes the similarity matrix, representing the degree of similarity between data points. L v = D v W v denotes the Laplacian matrix. D v is the degree matrix, W v is the adjacency matrix, D v i i = j = 1 n w v i j , and since S v is an asymmetric matrix, let W v = ( S v + S v T ) / 2 .
For learning similarity matrices, the KNN method is commonly employed. However, if the majority of data points cluster near the intersection of multiple subspaces, KNN fails to partition these subspaces correctly. Furthermore, the KNN approach necessitates frequent adjustments to the number of neighbors [39]. In information theory, information entropy serves as a physical quantity describing the unpredictability of information content. Let information q = ( q 1 , q 2 , , q n ) consist of n sources, with q i denoting the probability of the i -th source. The information entropy q can be expressed as:
y = i = 1 n q i log 1 q i
The higher the similarity, the more stable the model; however, a stricter similarity matrix reduces the model’s generalization capability. In Equation (9), greater entropy indicates richer information content within the matrix. By maximizing the entropy of the similarity matrix with temperature parameters, one can effectively learn the similarity matrix, promoting more uniform information distribution within the matrix and balancing model stability with generalization capability [38]. Consequently, the information-entropy-based adaptive graph method can overcome the drawback of misclassifying neighbors. Its mathematical expression is:
max i = 1 n j = 1 n s v i j log 1 / s v i j τ s.t. j = 1 n s v i j = 1 , s v i j 0
Here, the temperature parameter τ controls the strength of entropy.
Combining diversity constraints and adaptive graphs [40], we propose a multi-view functional matrix completion model based on diversity constraints and adaptive graphs.
min v = 1 n v θ v O ( Y ˜ v + Φ U v V v T ) F 2 + 2 θ v s v = 1 n v w = 1 n v tr ( V v V w T ) + μ v = 1 n v tr ( V v T L v V v ) + λ i = 1 n j = 1 n s v i j log s v i j τ s.t. U v 0 , V v 0 , j = 1 n s v i j = 1 , s v i j 0
Here, v = 1 , 2 , , n v , θ v s , μ , λ 0 are parameters adjusting the weighting relationships between components. The first term constitutes the fundamental objective function for multi-view functional matrix completion; the second term represents a diversity constraint designed to integrate disparate information across different sites; the third and fourth terms form an adaptive graph constraint. The fourth component adapts to learn similarity matrices by maximizing their entropy, while the third component incorporates the geometric structure of the data through the graph Laplacian matrix. Figure 1 presents the schematic diagram of the proposed DCAGMFMC method.

3.2. Optimization Algorithm

The non-convexity of objective function (11) with respect to U v and V v poses challenges for global optimization. To address this issue, this paper employs an alternating minimization strategy to solve Equation (11). Each variable is updated while holding the others constant. The augmented Lagrangian function for (11) is given by:
L = v = 1 n v θ v O ( Y ˜ v + Φ U v V v T ) F 2 + 2 θ v s v = 1 n v w = 1 n v tr ( V v V w T ) + μ v = 1 n v tr ( V v T L v V v ) + λ i = 1 n j = 1 n s v i j log s v i j τ tr ( Ψ v T U v ) tr ( Γ v T V v ) i = 1 n ξ i ( j = 1 n s v i j 1 ) i = 1 n j = 1 n π i j s v i j
(1) Fixing V v and S v , update U v . Given V v and S v , find a U v that better reconstructs the observed data matrix Y ˜ v + . The part of Equation (12) concerning U v simplifies to:
L 1 = θ v tr O ( Y ˜ v + Φ U v V v T ) O ( Y ˜ v + Φ U v V v T ) T tr ( Ψ v T U v )
Taking the partial derivative of Equation (13) with respect to U v and setting it equal to zero yields:
2 θ v Φ T ( O Y ˜ v + ) V v + 2 θ v Φ T ( O Φ U v V v T ) V v = Ψ v
According to the KKT conditions, satisfying Ψ v U v = 0 , we have:
2 θ v Φ T ( O Y ˜ v + ) V v + 2 θ v Φ T ( O Φ U v V v T ) V v i j U v i j = 0
Thus, the updated formula for U v is obtained as follows:
U v i j U v i j Φ T ( O Y ˜ v + ) V v i j Φ T ( O Φ U v V v T ) V v i j
(2) Fixing U v and S v , update V v . The part of Equation (12) concerning V v simplifies to:
L 2 = θ v tr O ( Y ˜ v + Φ U v V v T ) O ( Y ˜ v + Φ U v V v T ) T + 2 θ v s w = 1 n v tr ( V v V w T ) + μ tr ( V v T L v V v ) tr ( Γ v T V v )
Taking the partial derivative of Equation (17) with respect to V v and setting it to zero and combining with the KKT conditions satisfying Γ v V v = 0 yields:
( 2 θ v ( O T Y ˜ v + T ) Φ U v + 2 θ v ( O T V v U v T Φ T ) Φ U v + 2 θ v s w = 1 , w v n v V w + 4 θ v s V v + 2 μ ( D v W v ) V v ) i j V v i j = 0
The updated formula for V v is:
V v i j V v i j θ v ( O T Y ˜ v + T ) Φ U v + μ W v V v i j θ v ( O T V v U v T Φ T ) Φ U v + θ v s w = 1 , w v n v V w + 2 θ v s V v + μ D v V v i j
where Equation (19) updates the latent representation V v , shared across all viewpoints, integrating multi-view information. The diversity constraint is embodied by the θ v s w = 1 , w v n v V w + 2 θ v s V v term, which ensures the latent representation V v of the current viewpoint remains orthogonal to representations from all other viewpoints. This captures complementary information and prevents learning redundant features. Image regularization is embodied by the μ W v V v and μ D v V v terms.
(3) Fixing U v and V v , update S v . The augmented Lagrangian function with respect to S v is:
L 3 = 2 tr ( V v T L v V v ) + 2 λ i = 1 n j = 1 n s v i j log s v i j τ i = 1 n ξ i ( j = 1 n s v i j 1 ) i = 1 n j = 1 n π i j s v i j
Taking the partial derivative with respect to S v in Equation (20) and setting it equal to zero yields:
v v i v v j 2 2 + 2 λ τ ( log s v i j + 1 ) ξ i π i j = 0
Applying the KKT condition π i j s v i j = 0 , when s v i j > 0 , then π i j = 0 , it follows that:
v v i v v j 2 2 + 2 λ τ ( log s v i j + 1 ) ξ i = 0
Thereby:
s v i j = exp ( ξ i 2 λ τ ) / 2 λ τ exp v v i v v j 2 2 / 2 λ τ = ρ i exp v v i v v j 2 2 / 2 λ τ
where ρ i = exp ( ξ i 2 λ τ ) / 2 λ τ . Given that j = 1 n s v i j = 1 , the calculation of ρ i yields:
ρ i = 1 / j = 1 n exp v v i v v j 2 2 / 2 λ τ
Combining Equation (24), the updated formula for S v is obtained as:
s v i j exp v v i v v j 2 2 / 2 λ τ / j = 1 n exp v v i v v j 2 2 / 2 λ τ

3.3. Convergence Analysis

Below, we conduct a convergence analysis of the update formulas for U v , V v , and S v in Equations (16), (19), and (25), that is, prove Theorem 1. Our optimization algorithm can be incorporated into the Block Coordinate Descent (BCD) framework [41]. Within this framework, since each subproblem is solved precisely and ensures sufficient descent of the objective function, the sequence generated by the algorithm converges to a stationary point of the objective function according to its convergence theory.
Theorem 1.
Given initial values, applying the update rules derived in this study (Equations (16), (19), and (25)) for optimization results in the objective function (11) decreasing monotonically at each iteration.
Proof. 
According to the convergence theory established by Tseng [41], if solving each subproblem leads to a sufficient decrease in the objective function value, then any limit point of the sequence is a stationary point of the objective function. The convergence of updates to expressions (16), (19) and (25) is demonstrated by con structing auxiliary functions. In each update, terms related to U v are retained from the objective function expression (11), while unrelated terms are discarded.
L ( U v ) = tr ( U v T Φ T ( O Φ U v V v T ) V v 2 U v T Φ T ( O Y ˜ v + ) V v )
The auxiliary function G ( U v , U v t ) is constructed based on the control inequality commonly used in the MM algorithm [42]. For terms of the form x 2 , we employ the inequality x 2 / y 2 x y , where y > 0 . For terms of the form log x , we employ the inequality log x 1 x / y + log y . Replacing each term L ( U v ) with its upper bound at U v = U v t yields the expression for the auxiliary function G ( U v , U v t ) as follows:
G ( U v , U v t ) = i , j Φ T ( O Φ U v t V v T ) V v i j U v i j 2 U v i j t 2 i , j Φ T ( O Y ˜ v + ) V v i j U v i j t ( 1 + log U v i j U v i j t )
This construction ensures that when U v = U v t , all inequalities hold with equality, hence G ( U v , U v ) = L ( U v ) . For other values of U v , since each term is an upper bound, we have G ( U v , U v t ) L ( U v ) if U v t + 1 is chosen such that:
U v t + 1 = arg min t G ( U v , U v t )
Established, it is evident that L ( U v ) is monotonically decreasing, with:
G ( U v t + 1 , U v t + 1 ) G ( U v t + 1 , U v t ) G ( U v t , U v t )
To minimize the auxiliary function Equation (27), solve for U v t + 1 subject to the constraint in Equation (28). Taking the partial derivative of Equation (27) with respect to U v i j and setting it to zero, while substituting U v i j with U v i j t + 1 , yields:
Φ T ( O Φ U v t V v T ) V v i j U v i j t + 1 U v i j t = Φ T ( O Y ˜ v + ) V v i j U v i j t U v i j t + 1
Furthermore:
U v i j t + 1 = U v i j t Φ T ( O Y ˜ v + ) V v i j Φ T ( O Φ U v t V v T ) V v i j
Hence, Equation (31) is derived as the update rule Equation (16) for U v .
Below we prove the convergence of the update rule Equation (19) for V v . In Equation (11), only the terms relating to V v are retained:
L ( V v ) = tr ( 2 θ v V v T ( O T Y ˜ v + T ) Φ U v 2 θ v ( O T V v U v T Φ T ) Φ U v ) + 2 θ v s w = 1 n v tr ( V v V w T ) + μ tr ( V v T L v V v )
Similarly, construct auxiliary function G ( V v , V v t ) of L ( V v ) , defined as:
G ( V v , V v t ) = i , j θ v ( O T V v U v T Φ T ) Φ U v + θ v s w = 1 , w v n v V w + 2 θ v s V v + μ D v V v i j V v i j 2 V v i j t 2 θ v i , j ( O T Y ˜ v + T ) Φ U v + μ W v V v i j V v i j t ( 1 + log V v i j V v i j t )
Then, conditions G ( V v , V v ) = L ( V v ) and G ( V v , V v t ) L ( V v ) are satisfied. If V v t + 1 is chosen such that:
V v t + 1 = arg min G ( V v , V v t )
This established, it is evident that L ( V v ) is monotonically decreasing, with:
G ( V v t + 1 , V v t + 1 ) G ( V v t + 1 , V v t ) G ( V v t , V v t )
To minimize the auxiliary function Equation (33), solve for V v t + 1 subject to the constraint in Equation (34). Taking the partial derivative of Equation (34) with respect to V v i j and setting it to zero yields:
θ v ( O T Y ˜ v + T ) Φ U v + μ W v V v i j V v i j t V v i j t + 1 = θ v ( O T V v U v T Φ T ) Φ U v + θ v s w = 1 , w v n v V w + 2 θ v s V v + μ D v V v i j V v i j t V v i j t + 1
Thereby:
V v i j t + 1 V v i j t θ v ( O T Y ˜ v + T ) Φ U v + μ W v V v i j θ v ( O T V v U v T Φ T ) Φ U v + θ v s w = 1 , w v n v V w + 2 θ v s V v + μ D v V v i j
Hence, Equation (37) is derived as the update rule (19) for U v . Similarly, the update rule Equation (25) for S v can be demonstrated. □
In summary, the specific implementation steps of the DCAGMFMC method are summarized in Algorithm 1 as follows:
Algorithm 1. DCAGMFMC
Input: Multi-view data matrix Y , parameters θ v , θ v s , μ , λ , ε 0 = 10 7 , iteration count t
Output: { U 1 t + 1 , U 2 t + 1 , , U n v t + 1 } and { V 1 t + 1 , V 2 t + 1 , , V n v t + 1 } .
1. Generate random values for U v 0 and V v 0 , then initialize S v 0 according to Equation (25);
2. Fixing V v and S v , update U v t + 1 according to Equation (16);
3. Fixing U v and S v , update V v t + 1 according to Equation (19);
4. Fixing U v and V v , update S v t + 1 according to Equation (25);
5. Repeat the aforementioned optimization steps (steps 2–4) until U v t + 1 V v t + 1 T U v t V v t T 2 / U v t V v t T 2 ε 0 is achieved.
Given the global non-convexity of the objective function (11), the convergence of the algorithm to a fixed point depends on the initialization choice, and multiple local minima may exist. To mitigate the potential impact of this issue on algorithm performance and the reliability of conclusions, and to obtain stable, robust solutions, we will implement the following measures in the numerical experimental design:
(1) Multiple Random Initializations: In each independent experiment run, the weight matrices U v and V v are initialized using random values generated from a chi-squared distribution. We conducted ntest = 50 independent experiments, each starting from distinct random initial points to thoroughly explore the solution space and simulate the algorithm’s convergence behavior from different regions.
(2) Stability Assessment of Results: All performance metrics reported in our final analysis (e.g., RMSE, MAE) represent the arithmetic mean and standard deviation of results from 50 independent experiments. As shown in Section 4 of the experimental results, the standard deviation of each performance metric remains at a low level. This indicates that although the convergence path may differ each time, the algorithm consistently converges to locally optimal points with similar performance. This demonstrates the method’s strong robustness to initialization.
(3) Under such extensive random initialization sampling, the superior and stable average performance demonstrated by the proposed method strongly indicates that the local optima found by the algorithm are of high quality and can effectively accomplish the matrix completion task.

3.4. Time Complexity Analysis

The primary computational cost of the proposed DCAGMFMC algorithm lies in the iterative updates of variables U v , V v and S v . The cost of each iteration is analyzed below:
Update U v : Its computation is primarily dominated by matrix multiplication Φ T ( O Φ U v V v T ) V v . Considering Φ m × r , U v r × d , and V v n × d , the complexity of this step is O ( m r n ) .
Update V v : This step involves matrix operations similar to Update U v , with a complexity of O ( m r n ) . It also includes the calculation of the Laplacian term, where constructing the sample similarity matrix requires computing the Euclidean distance between all sample pairs, with a complexity of O ( n 2 m ) . Therefore, the overall complexity of this step is O ( m r n + n 2 m ) .
Update S v : The computational complexity of calculating the Euclidean distance between all sample pairs is O ( n 2 m ) . Subsequently, the computational complexity of computing the similarity matrix S v is O ( n 2 ) .
In summary, the overall computational complexity per iteration is O ( n v ( m r n + n 2 m ) ) , where n v represents the number of viewpoints. For large datasets (i.e., when both m , n are substantial), the dominant cost of the algorithm is O ( n v · n 2 · m ) . This indicates that the algorithm’s runtime scales approximately linearly with the sample size m and the number of samples n , demonstrating excellent scalability.

4. Experiments

To evaluate the imputation performance and effectiveness of the DCAGMFMC method, this study compares it with five typical missing value handling methods—KNN, SFI, HFI, MVNFMC, and GRMFMC—through simulation experiments and practical applications. Specifically, the imputation strategy of KNN [43] involves first calculating the distance between samples based on observed values to quantify their similarity, subsequently estimating missing values using information from similar samples. SFI [14] employs matrix completion techniques to reconstruct complete functional data from partial observations. HFI [14] replaces the original nuclear norm penalty term in SFI with the l 0 norm. MVNFMC [15] builds upon SFI and HFI by employing multi-view learning to estimate missing data. GRMFMC [32] extends multi-view matrix completion methods by utilizing the HSIC criterion and graph regularization to extract inter-view information for missing data estimation. The imputation performance was evaluated using root mean square error (RMSE), mean absolute error (MAE), and normalized root mean square error (NRMSE). Experiments were conducted using R4.3.0 on the following computer environment: Intel(R) Core (TM) i5-8250U CPU @ 1.60GHz 1.80 GHz, 8GB RAM, Windows 10 64-bit operating system.

4.1. Simulation Data Generation

Case a: Generating functional data through linear combinations of trigonometric and polynomial functions [44]:
Y ˜ + = Φ A + E
Among these, the number of sample curves is n = 20 , the time interval is t [ 1 , 92 ] , and each curve comprises m = 92 discrete sampling points, namely t = 1 , 2 , , 92 and j = 1 , 2 , , m . According to Equation (2):
View 1: take
φ j ( t ) = ( 1 , cos 2 ( t / 10 ) , sin 2 ( t / 10 ) ) , α i = ( 21 / 2 + t , α i 1 , α i 2 ) , r = 3
View 2: take
φ j ( t ) = ( 1 , sin 2 ( t / 10 ) , cos 2 ( t / 10 ) ) , α i = ( 21 / 2 + t , α i 1 , α i 2 ) , r = 3
View 3: take
φ j ( t ) = ( 1 , cos 2 ( t / 10 ) , sin 2 ( t / 10 ) , ( t / 10 ) 2 + t / 10 + 1 ) , α i = ( 21 / 2 + t , α i 1 , α i 2 , α i 3 ) , r = 4
Here, Φ = ( φ 1 ( t ) , φ 2 ( t ) , , φ m ( t ) ) T m × r , A = ( α 1 , α 2 , , α n ) r × n , and E = ( ε 1 , ε 2 , , ε n ) m × n , yield the respective fitted curves x i ( t ) ( i = 1 , 2 , , n ) and α i l ~ N ( 1 , 2 ) ( l = 1 , 2 , 3 ) . As shown in Figure 2, the multi-view data visualization results for different noise generation scenarios under Case a are presented. Subfigures (a–c) display multi-view data generated with added normal noise, while subfigures (d–f) show multi-view data generated with added exponential noise.
Case b: Generating functional data through the combination of B-spline basis matrices and non-negative matrices:
Y ˜ + = Φ U V T + E
Among these, the basis matrix is formed by taking Φ m × r as the B-spline basis functions at { 10 , 20 , 30 } equidistant nodes [45], with m = 92 , the number of basis functions r = { 10 , 20 , 30 } , and rank p = 4 . U r × p , V n × p , n = 20 and U , V ~ χ 2 ( 2 ) is initialized. As shown in Figure 3, the multi-view data visualization results for different noise generation scenarios under Case b are presented. Subfigures (a–c) display multi-view data generated with added normal noise, while subfigures (d–f) show multi-view data generated with added exponential noise.
Based on the observational data generated from Equations (38) and (39), plots were produced for the four types of noise:
Mode 1 is ε i j ~ N ( 0 , 1 ) , Mode 2 is ε i j ~ E ( 1 ) , Mode 3 is ε i j ~ U ( 0 , 1 ) , and Mode 4 is ε i j ~ A L D ( 0 , 1 , τ ) ; with a noise level of 0.1, a non-negative discrete data matrix Y ˜ + m × n is obtained.

4.2. Experimental Results and Analysis

In the SFI, HFI, MVNFMC, GRMFMC and DCAGMFMC methods, equidistant nodal B-spline basis functions are employed, with r = { 10 , 20 , 30 } .
Simulated data were artificially generated with missing entries according to a hybrid PM/IM missing data pattern [46]. Here, PM denotes pointwise generation of missing entries; IM denotes generation of distinct missing intervals, with interval lengths following a uniform distribution; PM/IM denotes the combined generation of missing entries using both PM and IM methods. Within the hybrid missing pattern, the PM and IM missing patterns each account for 50%. For the DCAGMFMC method presented in this study, parameters μ and λ were selected from {10−3, 10−2, 10−1, 1, 101, 102, 103} respectively. The RMSE results corresponding to different parameter combinations of μ and λ under normal noise and exponential noise for situations a and b are plotted in Figure 4, Figure 5, Figure 6 and Figure 7.
Figure 4, Figure 5, Figure 6 and Figure 7 demonstrate that the DCAGMFMC method exhibits low sensitivity to parameters, indicating strong model robustness. Through 5-fold cross-validation, the optimal parameter combination μ = 100 , λ = 1000 was selected from { 1 0 3 , 1 0 2 , 1 0 1 , 1 , 1 0 1 , 1 0 2 , 1 0 3 } . The rank of the coefficient matrix V was set to 4, and the temperature parameter τ was optimized via grid search within the range 0.1 , 0.5 , 1.0 , 2.0 , 5.0 . Experimental results show that the interpolation RMSE is relatively small when τ = 0.5 , indicating that moderate entropy intensity effectively balances the model’s discriminative capability and generalization performance. The simulated data were repaired using the DCAGMFMC, GRMFMC, MVNFMC, KNN, SFI, and HFI methods. Below, only the RMSE results for interpolation of simulated data generated under situations a and b with normal and exponential noise are presented. As shown in Table 1.
As shown in Table 1, the proposed DCAGMFMC method demonstrates distinct advantages over conventional approaches. It yields lower imputation RMSE across three views under varying situations and noise levels, with minimal divergence in the imputation results.

5. Example Application

5.1. Meteorological Experiment Data

Meteorological data constitute a long-term objective record of the state and evolution of the Earth’s atmospheric environment. Meteorological indicators encompass numerous types. To validate the effectiveness of the proposed DCAGMFMC method for interpolating missing meteorological data, this study selected only four representative meteorological indicators: air temperature (TEMP), visibility (VISIB), wind speed (WDSP), and maximum wind speed (MXSPD). Data sourced from the National Climatic Data Center (NCDC), with meteorological data for the period 1 January 2024 to 2 September 2024 obtained from the website https://www.ncei.noaa.gov/data/global-summary-of-the-day/archive (accessed on 1 October 2024) [47]. For instance, selecting meteorological indicator data—specifically TEMP, VISIB, WDSP, and MXSPD—from 30 stations across western China, we employed a sample to understand the pattern of missing meteorological data. Observing Figure 8 reveals that the four meteorological indicators exhibit varying degrees of random point missingness, strip-shaped, or block-shaped missingness.
Complete meteorological data for temperature (TEMP), visibility (VISIB), wind speed (WDSP), and maximum wind speed (MXSPD) were selected from 20 meteorological stations across Sichuan Province, Guizhou Province, Chongqing Municipality, Hunan Province, Hubei Province, and Jiangxi Province, covering the period from 1 June 2024 to 31 August 2024. The 20 stations are denoted by the abbreviations A, B, C, D, E, F, G, H, I, J, K, L, M, N, O, P, Q, R, S, and T respectively. Their specific locations are illustrated in Figure 9. This map is based on the standard map with review number GS(2024)0650, sourced from the National Geographic Information Public Service Platform (Tianditu), with no modifications made to the base map.
Temperature (TEMP), visibility (VISIB), wind speed (WDSP), and maximum wind speed (MXSPD) may be regarded as four views describing meteorological conditions. Figure 10 illustrates the daily variation in these four meteorological indicators. Among them, the red curves represent the time series of each meteorological indicator, and the black curve denotes their average value. As evident from Figure 10, these indicators exhibit distinct functional characteristics, with their variations following certain patterns.
To analyze the correlation between various meteorological indicators, their Pearson correlation coefficients were calculated. The closer the absolute value of this coefficient approaches 1, the stronger the correlation between the variables. Observation of Figure 11 reveals that the correlation coefficient between WDSP and MXSPD is 0.95; that between TEMP and VISIB is 0.85; and that all four meteorological indicators exhibit positive correlations. Given the correlations among these four indicators, it is reasonable to employ their complementary characteristics for interpolating missing data.
Using complete datasets as empirical data allows for the deliberate control of the proportion, type, and location of missing data, facilitating comparative performance assessments of different imputation methods under varying conditions. Furthermore, employing complete datasets with known true values enables the precise calculation of imputation accuracy, providing a more comprehensive evaluation that facilitates objective comparisons of different methods’ performance. Table 2 presents the statistical description of the selected meteorological data.

5.2. Ablation Experiment

We artificially generate missing entries in meteorological data using a hybrid PM/IM missing data model, with specified missing rates of 20%, 30%, 40%, and 50%.
The role of the diversity constraint and adaptive graph in the objective function Equation (11) is validated through ablation experiments. Two special cases of the DCAGMFMC method are considered: eliminating the effect of the adaptive graph yields algorithm DCAGMFMC1; eliminating the effect of the diversity constraint yields algorithm DCAGMFMC2.
When the missing rate is 20%, 30%, 40%, and 50%, the RMSE and MAE of the DCAGMFMC method, DCAGMFMC1 method, and DCAGMFMC2 method are shown in Table 3, with bold text indicating superior comparative results. Observation of Table 3 reveals that, across varying missing data rates, the DCAGMFMC method consistently exhibits lower RMSE values than both the DCAGMFMC1 and DCAGMFMC2 methods. with RMSE values consistently lower than those of DCAGMFMC1 and DCAGMFMC2. Across the four meteorological indicators, RMSE reductions ranged from 10.36% to 11.10% and 9.55% to 11.12%, respectively. This demonstrates that both diversity constraints and adaptive graphs contribute to enhancing imputation accuracy.

5.3. Application Effect and Analysis

An empirical application was conducted using complete meteorological indicator data to validate the practical effectiveness of the DCAGMFMC method. The DCAGMFMC method was employed to interpolate missing data, with root mean square error, mean absolute error, and normalized root mean square error selected as metrics to assess deviations between predicted values and actual observations. Comparisons were made against KNN, HFI, SFI, MVNFMC, and GRMFMC methods. In Equation (11), θ v = 0.8 , θ v s = 0.8 , μ = 100 , λ = 1000 . The RMSE, MAE, and NRMSE results for the six imputation methods at 20%, 30%, 40%, and 50% missing data rates are presented in Table 4, Table 5 and Table 6, respectively, with bold text indicating superior comparative performance.
During the experiments, it was observed that the proposed method becomes unsuitable when the missing rate reaches 60%. As demonstrated in Table 4, Table 5 and Table 6, under various missing rate conditions within the experimental setup, the DCAGMFMC method consistently yields lower imputation metrics including RMSE, MAE, and NRMSE compared to KNN, HFI, SFI, MVNFMC, and GRMFMC.
To further demonstrate the validity of the DCAGMFMC method’s interpolated values, scatter plots were plotted for the four viewpoint interpolated values against the true values at 20% and 50% missing data rates, respectively. The results are shown in Figure 12, Figure 13, Figure 14 and Figure 15. Among them, the black diamonds represent the data points of true versus imputed values, and the red dashed line denotes the 1:1 ideal fit line ( y = x ).
Figure 12 and Figure 13 demonstrate that the interpolated values exhibit good agreement with the true values at 20% and 50% missing data rates. This indicates that the DCAGMFMC method delivers satisfactory imputation results, with performance remaining stable even as the missing data rate increases. Consequently, the DCAGMFMC method is suitable for interpolating meteorological data with extensive missing values. Figure 14 and Figure 15 demonstrate that, at 20% and 50% missing data rates, the interpolated values are uniformly distributed along the 45-degree line relative to the true values. This indicates minimal error between interpolated and true values, confirming the high predictive accuracy of the DCAGMFMC method.

6. Conclusions

This study proposes a Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion method (DCAGMFMC) within the functional data analysis framework, integrating multi-view learning with diversity constraints and adaptive graph regularization strategies. During missing value handling, the diversity constraint term accounts for inter-data variability, enhancing information utilization. The adaptive graph regularization term preserves the inherent geometric structure of the data while adaptively learning similarity matrices by maximizing their entropy. This approach enhances imputation accuracy whilst preventing model overfitting. Random simulation experiments demonstrate that the DCAGMFMC method exhibits significant imputation advantages over classical methods such as KNN, HFI, SFI, MVNFMC, and GRMFMC. Moreover, practical applications on real meteorological datasets demonstrate that the proposed method can adaptively capture dynamic correlations among multi-view observational data while preserving the intrinsic manifold structure of functional data, thereby substantially enhancing the accuracy and robustness of missing data imputation. Specifically, compared to the KNN, HFI, SFI, MVNFMC, and GRMFMC imputation methods, at different missing rates, the DCAGMFMC method achieved average RMSE reductions of 52.18–74.96%, 45.49–65.84%, 45.63–66.09%, 30.27–32.43%, and 21.97–56.43% when processing meteorological data. MAE was reduced by an average of 73.01% to 91.16%, 71.36% to 88.61%, 71.36% to 88.65%, 34.49% to 35.33%, and 22.27% to 56.10%, respectively; The NRMSE was reduced by an average of 91.25–95.78%, 44.56–65.05%, 44.56–65.51%, 25.54–33.90%, and 13.89–58.26%, respectively. Compared to these baseline methods, the DCAGMVNFMC approach achieved average reductions in RMSE, MAE, and NRMSE of 39.11–59.15%, 54.50–71.97%, and 43.96–63.70%, respectively. It also demonstrated stable interpolation performance across varying meteorological indicators and data missing rates, exhibiting strong adaptability.
The experiments in this study primarily focus on the mixed pointwise missing (PM) and interval missing (IM) patterns. However, we acknowledge that real-world missing data, particularly non-randomly absent (MNAR) data, exhibit more complex mechanisms—for instance, the probability of missingness may be correlated with the unobserved values themselves. Exploring model performance under MNAR mechanisms represents a key direction for our future work. Nevertheless, successfully handling complex interval missing (IM) data already demands robust inference and reconstruction capabilities from models, providing positive evidence for the applicability of this approach in more practical scenarios. In subsequent research, we will continue exploring novel multi-perspective functional matrix completion methods, with a focus on enhancing their interpolation performance. While fully considering the complementary information across perspectives, we strive to reduce redundancy and improve information utilization. Therefore, future research will prioritize integrating the HSIC criterion with dual-information graph strategies into the multi-perspective functional matrix completion framework. Additionally, we will explore applying the proposed DCAGMFMC method to broader practical scenarios, such as medical data interpolation and financial data interpolation.

Author Contributions

Conceptualization, H.G. and Y.B.; methodology, H.G. and Y.B.; software, Y.B.; validation, H.G. and Y.B.; formal analysis, H.G. and Y.B.; investigation, H.G. and Y.B.; resources, H.G. and Y.B.; data curation, Y.B.; writing—original draft preparation, Y.B.; writing—review and editing, H.G. and Y.B.; visualization, Y.B.; supervision, H.G.; project administration, H.G. and Y.B.; funding acquisition, H.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by National Social Science Fund of China (NO. 21BTJ042), National Key Statistical Science Research Project (NO. 2025LZ007), Gansu Provincial Natural Science Foundation (NO. 23JRRA1186), and Gansu Provincial Universities’ Young Doctor Support Program (NO. 2025QB-058).

Data Availability Statement

Publicly available datasets were analyzed in this study.

Acknowledgments

The authors sincerely appreciate the editors and reviewers for their valuable comments and professional suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

PACEPrincipal Components Analysis through Conditional Expectation
MICEMultiple Imputation by Chained Equations
fregMICEFunctional Regression MICE
IDWInverse Distance Weighting
SFISoft Functional Impute
HFIHard Functional Impute
KNNk-Nearest Neighbors
EMExpectation Maximization
MFMatrix Factorization
TRMFTemporal Regularized Matrix Factorization
NCADNeural Contextual Anomaly Detection
FDAFunctional Data Analysis
MVLMulti-view Learning
MVNFMCMulti-view Non-negative Functional Matrix Completion
GRMFMCGraph-Regularized Multi-view Functional Matrix Completion
DCAGMFMCDiversity Constraint and Adaptive Graph Multi-view Functional Matrix Completion
PMPointwise Missing
IMInterval Missing

References

  1. Miao, X.; Wu, Y.; Chen, L.; Gao, Y.; Yin, J. An experimental survey of missing data Imputation algorithms. IEEE Trans. Knowl. Data Eng. 2022, 35, 6630–6650. [Google Scholar] [CrossRef]
  2. Kong, X.; Zhou, W.; Shen, G.; Zhang, W.; Liu, N.; Yang, Y. Dynamic graph convolutional recurrent imputation network for spatiotemporal traffic missing data. Knowl.-Based Syst. 2023, 261, 110188. [Google Scholar] [CrossRef]
  3. Gao, H.Y.; Li, W.X. Function-based multiple imputation method based on cross-sectional and longitudinal information. Tongji Yu Juece/Stat. Decis. 2025, 41, 37–42. [Google Scholar] [CrossRef]
  4. Bertsimas, D.; Pawlowski, C.; Zhuo, Y.D. From predictive methods to missing data imputation: An optimization approach. J. Mach. Learn. Res. 2018, 18, 1–39. Available online: https://jmlr.org/papers/v18/17-073.html (accessed on 15 October 2024).
  5. Yao, F.; Müller, H.G.; Wang, J.L. Functional data analysis for sparse longitudinal data. J. Am. Stat. Assoc. 2005, 100, 577–590. [Google Scholar] [CrossRef]
  6. Royston, P.; White, I.R. Multiple Imputation by Chained Equations (MICE): Implementation in Stata. J. Stat. Softw. 2011, 45, 1–20. [Google Scholar] [CrossRef]
  7. He, Y.; Yucel, R.; Raghunathan, T.E. A functional multiple imputation approach to incomplete longitudinal Data. Stat. Med. 2011, 30, 1137–1156. [Google Scholar] [CrossRef]
  8. Ciarleglio, A.; Petkova, E.; Harel, O. Elucidating age and sex-dependent association between frontal EEG asymmetry and depression: An application of multiple imputation in functional regression. J. Am. Stat. Assoc. 2022, 117, 12–26. [Google Scholar] [CrossRef]
  9. Rao, A.R.; Reimherr, M. Modern multiple imputation with functional data. Stat 2021, 10, e331. [Google Scholar] [CrossRef]
  10. Li, J.W.; Sun, Y.H.; Wang, S.; Zhang, Z.W.; Wang, Y. Ultra-short-term load forecasting considering missing data in abnormal situations. Autom. Electr. Power Syst. 2025, 49, 133–143. [Google Scholar] [CrossRef]
  11. Liu, Y.; Zhao, Y.P.; Liu, F.Q.; Kong, X.F.; Ma, R. Research progress on abnormal and missing data processing methods in marine environmental monitoring. J. Appl. Oceanogr. 2025, 44, 388–401. [Google Scholar] [CrossRef]
  12. Liang, Y.P.; Li, X.Y.; Li, Q.G.; Mao, S.R.; Zhen, M.H.; Li, J.B. Research on intelligent prediction of gas concentration in working face based on CS-LSTM. J. Mine Saf. Environ. Prot. 2022, 49, 80–86. [Google Scholar] [CrossRef]
  13. Wu, Y.H.; Wang, Y.S.; Xu, H.; Chen, Z.; Zhang, Z.; Guan, S.J. Review on wind power output prediction technology. Comput. Sci. Explor. 2022, 16, 2653–2677. [Google Scholar] [CrossRef]
  14. Kidzinski, Ł.; Hastie, T. Longitudinal data analysis using matrix completion. Stat 2018, 1050, 24. [Google Scholar] [CrossRef]
  15. Xue, J.; Fu, D.Y.; Han, H.B.; Gao, H.Y. A non-negative functional matrix completion algorithm based on multi-view learning. Stat. Decis. 2022, 38, 5–11. [Google Scholar] [CrossRef]
  16. Kramer, O. K-nearest neighbors. In Dimensionality Reduction with Unsupervised Nearest Neighbors; Springer: Berlin/Heidelberg, Germany, 2013; pp. 13–23. [Google Scholar] [CrossRef]
  17. Gao, H.Y.; Liu, C.; Ma, W.J. Comparison and application of time series data imputation methods for surface water quality monitoring. Hydrology 2024, 44, 63–69. [Google Scholar] [CrossRef]
  18. Nelwamondo, F.V.; Mohamed, S.; Marwala, T. Missing data: A comparison of neural network and expectation maximization techniques. Curr. Sci. 2007, 93, 1514–1521. Available online: https://www.jstor.org/stable/24099079 (accessed on 15 October 2024).
  19. Koren, Y.; Bell, R.; Volinsky, C. Matrix factorization techniques for recommender systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  20. Yu, H.F.; Rao, N.; Dhillon, I.S. Temporal regularized matrix factorization for high-dimensional time series prediction. Adv. Neural Inf. Process. Syst. 2016, 29. Available online: https://proceedings.neurips.cc/paper_files/paper/2016/hash/85422afb467e9456013a2a51d4dff702-Abstract.html (accessed on 24 October 2025).
  21. He, J.; Lai, Z.; Shi, K. Time series imputation method based on joint tensor completion and recurrent neural network. Data Acquis. Process. 2024, 39, 598–608. [Google Scholar] [CrossRef]
  22. Cao, W.; Wang, D.; Li, J.; Zhou, H.; Li, L.; Li, Y. BRITS: Bidirectional recurrent imputation for time series. Adv. Neural Inf. Process. Syst. 2018, 31. Available online: https://proceedings.neurips.cc/paper_files/paper/2018/hash/734e6bfcd358e25ac1db0a4241b95651-Abstract.html (accessed on 24 October 2025).
  23. Carmona, C.U.; Aubet, F.X.; Flunkert, V.; Gasthaus, J. Neural contextual anomaly detection for time series. arXiv 2021, arXiv:2107.07702. [Google Scholar] [CrossRef]
  24. Yoon, J.S.; Jordon, J.; Schaar, M. GAIN: Missing data imputation using generative adversarial nets. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar] [CrossRef]
  25. Wang, J.L.; Chiou, J.M.; Müller, H.G. Functional data analysis. Annu. Rev. Stat. Appl. 2016, 3, 257–295. [Google Scholar] [CrossRef]
  26. Ullah, S.; Finch, C.F. Applications of functional data analysis: A systematic review. BMC Med. Res. Methodol. 2013, 13, 43. [Google Scholar] [CrossRef]
  27. Yao, X.H. Research on Several Multivariate Functional Clustering Methods Based on Multi-View Learning. Ph.D. Thesis, Lanzhou University of Finance and Economics, Lanzhou, China, 2022. Available online: https://library.lzufe.edu.cn/asset/detail/0/20471970033 (accessed on 24 October 2025).
  28. Descary, M.H.; Panaretos, V.M. Functional data analysis by matrix completion. Ann. Stat. 2019, 47, 1–38. [Google Scholar] [CrossRef]
  29. Kong, S.; Wang, X.; Wang, D.; Wu, F. Multiple feature fusion for face recognition. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–7. [Google Scholar] [CrossRef]
  30. Xu, H.; Zhang, X.; Xia, W.; Gao, Q.; Gao, X. Low-rank tensor constrained co-regularized multi-view spectral clustering. Neural Netw. 2020, 132, 245–252. [Google Scholar] [CrossRef] [PubMed]
  31. Wu, K.; Xiao, Y.S.; Liu, B. Multi-view semi-supervised marked distribution learning. Comput. Appl. Res. 2025, 42, 1–8. Available online: https://www.arocmag.cn/abs/2025.04.0114 (accessed on 24 October 2025).
  32. Gao, H.Y.; Ma, W.J. Air quality data restoration based on graph regularization and multi-view function matrix completion. China Environ. Sci. 2024, 44, 5357–5370. [Google Scholar] [CrossRef]
  33. Li, C.; Che, H.; Leung, M.F.; Liu, C.; Yan, Z. Robust multi-view non-negative matrix factorization with adaptive graph and diversity constraints. Inf. Sci. 2023, 634, 587–607. [Google Scholar] [CrossRef]
  34. Wang, J.; Tian, F.; Yu, H.; Liu, C.H.; Zhan, K.; Wang, X. Diverse non-negative matrix factorization for multiview data representation. IEEE Trans. Cybern. 2017, 48, 2620–2632. [Google Scholar] [CrossRef] [PubMed]
  35. Liang, N.; Yang, Z.; Li, Z.; Sun, W.; Xie, S. Multi-view clustering by non-negative matrix factorization with co-orthogonal constraints. Knowl.-Based Syst. 2020, 194, 105582. [Google Scholar] [CrossRef]
  36. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar] [CrossRef] [PubMed]
  37. Li, J.; Zhou, G.; Qiu, Y.; Wang, Y.; Zhang, Y.; Xie, S. Deep graph regularized non-negative matrix factorization for multi-view clustering. Neurocomputing 2020, 390, 108–116. [Google Scholar] [CrossRef]
  38. Yang, X.; Che, H.; Leung, M.F.; Liu, C. Adaptive graph nonnegative matrix factorization with the self-paced regularization. Appl. Intell. 2023, 53, 15818–15835. [Google Scholar] [CrossRef]
  39. Luo, P.; Peng, J.; Guan, Z.; Fan, J. Dual regularized multi-view non-negative matrix factorization for clustering. Neurocomputing 2018, 294, 1–11. [Google Scholar] [CrossRef]
  40. Li, X.; Zhang, H.; Zhang, R.; Liu, Y.; Nie, F. Generalized uncorrelated regression with adaptive graph for unsupervised feature selection. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1587–1595. [Google Scholar] [CrossRef]
  41. Hunter, D.R.; Lange, K. A tutorial on MM algorithms. Am. Stat. 2004, 58, 30–37. [Google Scholar] [CrossRef]
  42. Tseng, P. Convergence of a block coordinate descent method for nondifferentiable minimization. J. Optim. Theory Appl. 2001, 109, 475–494. [Google Scholar] [CrossRef]
  43. Mucherino, A.; Papajorgji, P.J.; Pardalos, P.M. K-nearest neighbor classification. In Data Mining in Agriculture; Springer: New York, NY, USA, 2009; pp. 83–106. [Google Scholar] [CrossRef]
  44. Jacques, J.; Preda, C. Model-based clustering for multivariate functional data. Comput. Stat. Data Anal. 2014, 71, 92–106. [Google Scholar] [CrossRef]
  45. Boor, C.D. A Practical Guide to Splines; Springer: New York, NY, USA, 1978. [Google Scholar] [CrossRef]
  46. Chiou, J.M.; Zhang, Y.C.; Chen, W.H.; Chang, C.W. A functional data approach to missing value imputation and outlier detection for traffic flow data. Transp. B Transp. Dyn. 2014, 2, 106–129. [Google Scholar] [CrossRef]
  47. NOAA National Centers for Environmental Information. Global Summary of the Day—Weather Data. 2024. Available online: https://www.ncei.noaa.gov/data/global-summary-of-the-day/archive (accessed on 1 October 2024).
Figure 1. Schematic diagram of the DCAGMFMC algorithm.
Figure 1. Schematic diagram of the DCAGMFMC algorithm.
Axioms 14 00793 g001
Figure 2. Multi-View data generated by normal noise and exponential noise in Case a.
Figure 2. Multi-View data generated by normal noise and exponential noise in Case a.
Axioms 14 00793 g002
Figure 3. Multi-View data generated by normal noise and exponential noise in Case b.
Figure 3. Multi-View data generated by normal noise and exponential noise in Case b.
Axioms 14 00793 g003
Figure 4. RMSE results for different parameters μ and λ under normal noise in case a at different views.
Figure 4. RMSE results for different parameters μ and λ under normal noise in case a at different views.
Axioms 14 00793 g004
Figure 5. RMSE results for different parameters μ and λ under exponential noise in case a at different views.
Figure 5. RMSE results for different parameters μ and λ under exponential noise in case a at different views.
Axioms 14 00793 g005
Figure 6. RMSE results for different parameters μ and λ under normal noise in case b at different views.
Figure 6. RMSE results for different parameters μ and λ under normal noise in case b at different views.
Axioms 14 00793 g006
Figure 7. RMSE results for different parameters μ and λ under exponential noise in case b at different views.
Figure 7. RMSE results for different parameters μ and λ under exponential noise in case b at different views.
Axioms 14 00793 g007
Figure 8. Visualization of data missing situations for four meteorological indicators.
Figure 8. Visualization of data missing situations for four meteorological indicators.
Axioms 14 00793 g008
Figure 9. Distribution map of 20 meteorological stations. (Note: The map coordinate system is WGS 1984 (EPSG:4326)).
Figure 9. Distribution map of 20 meteorological stations. (Note: The map coordinate system is WGS 1984 (EPSG:4326)).
Axioms 14 00793 g009
Figure 10. Trends in the numerical changes of four meteorological indicators.
Figure 10. Trends in the numerical changes of four meteorological indicators.
Axioms 14 00793 g010
Figure 11. Heatmap of correlation coefficients for four meteorological indicators.
Figure 11. Heatmap of correlation coefficients for four meteorological indicators.
Axioms 14 00793 g011
Figure 12. Scatter plot of true values vs. imputed values under 20% missing rate.
Figure 12. Scatter plot of true values vs. imputed values under 20% missing rate.
Axioms 14 00793 g012
Figure 13. Scatter plot of true values vs. imputed values under 50% missing rate.
Figure 13. Scatter plot of true values vs. imputed values under 50% missing rate.
Axioms 14 00793 g013
Figure 14. Graph of the relationship between true values and imputation values at a 20% missing rate.
Figure 14. Graph of the relationship between true values and imputation values at a 20% missing rate.
Axioms 14 00793 g014
Figure 15. Graph of the relationship between true values and imputation values at a 50% missing rate.
Figure 15. Graph of the relationship between true values and imputation values at a 50% missing rate.
Axioms 14 00793 g015
Table 1. Imputation RMSE of simulated data with normal noise and exponential noise under different cases (Mean ± Standard Deviation of 50 Results).
Table 1. Imputation RMSE of simulated data with normal noise and exponential noise under different cases (Mean ± Standard Deviation of 50 Results).
Imputation MethodsNormal NoiseExponential Noise
View 1View 2View 3View1 View 2View 3
Case aKNN6.32 ± 0.3725.56 ± 1.46159.17 ± 15.536.60 ± 0.4127.56 ± 1.91174.85 ± 17.66
SFI2.65 ± 0.0018.40 ± 0.00121.78 ± 0.002.81 ± 0.0020.56 ± 0.00136.37 ± 0.00
HFI2.64 ± 0.0018.36 ± 0.00122.48 ± 0.002.80 ± 0.0020.62 ± 0.01136.88 ± 0.00
MVNFMC8.41 ± 0.0414.25 ± 0.0964.95 ± 0.598.46 ± 0.0415.01 ± 0.0971.45 ± 0.55
GRMFMC16.60 ± 0.8223.99 ± 1.0288.90 ± 3.0215.83 ± 1.3025.06 ± 1.9395.59 ± 1.64
DCAGMFMC3.18 ± 0.109.82 ± 0.0958.39 ± 0.573.27 ± 0.1110.92 ± 0.0966.91 ± 0.75
Case bKNN0.61 ± 0.051.01 ± 0.081.60 ± 0.110.98 ± 0.091.51 ± 0.112.63 ± 0.18
SFI0.95 ± 0.001.69 ± 0.002.76 ± 0.001.79 ± 0.002.92 ± 0.004.24 ± 0.00
HFI0.94 ± 0.001.67 ± 0.002.75 ± 0.001.78 ± 0.002.91 ± 0.004.21 ± 0.00
MVNFMC0.52 ± 0.000.87 ± 0.001.46 ± 0.001.01 ± 0.001.57 ± 0.002.39 ± 0.01
GRMFMC0.82 ± 0.051.23 ± 0.071.95 ± 0.131.33 ± 0.051.99 ± 0.143.13 ± 0.27
DCAGMFMC0.47 ± 0.000.78 ± 0.001.22 ± 0.000.88 ± 0.001.32 ± 0.002.01 ± 0.01
Note: Bold text indicates superior results.
Table 2. Statistical description of four meteorological index data.
Table 2. Statistical description of four meteorological index data.
VariableMeanSDMaxMin25th %PercentileMedian75th %PercentileNumber
TEMP (°F)55.122.4760.1149.1053.5555.3657.151840
VISIB (mile)16.722.1618.608.6015.8017.3018.601840
WDSP (knots)4.500.556.063.504.094.434.761840
MXSPD (knots)8.591.1111.996.347.988.499.051840
Table 3. RMSE and MAE results of four meteorological indicators under different missing rates (Mean ± Standard Deviation of 50 Results).
Table 3. RMSE and MAE results of four meteorological indicators under different missing rates (Mean ± Standard Deviation of 50 Results).
Missing Rate (%)Imputation MethodsEvaluation MetricTEMP (°C)VISIB (Mile)WDSP (Knots)MXSPD (Knots)
20DCAGMFMCRMSE1.55 ± 0.02831.47 ± 0.02240.70 ± 0.02141.16 ± 0.0232
MAE0.63 ± 0.01040.56 ± 0.01050.25 ± 0.00610.44 ± 0.0058
DCAGMFMC1RMSE2.06 ± 0.02541.58 ± 0.02330.73 ± 0.01611.24 ± 0.0232
MAE0.89 ± 0.00800.61 ± 0.01120.26 ± 0.00480.47 ± 0.0070
DCAGMFMC2RMSE2.08 ± 0.02621.57 ± 0.02610.74 ± 0.01471.25 ± 0.0208
MAE0.90 ± 0.00780.61 ± 0.01210.26 ± 0.00510.47 ± 0.0053
30DCAGMFMCRMSE1.28 ± 0.02871.21 ± 0.02290.57 ± 0.01700.97 ± 0.0187
MAE0.43 ± 0.00920.38 ± 0.00850.16 ± 0.00420.30 ± 0.0041
DCAGMFMC1RMSE1.70 ± 0.02341.27 ± 0.01860.62 ± 0.01751.01 ± 0.0151
MAE0.60 ± 0.00570.40 ± 0.00710.18 ± 0.00390.31 ± 0.0038
DCAGMFMC2RMSE1.71 ± 0.02181.29 ± 0.02020.60 ± 0.01091.01 ± 0.0156
MAE0.61 ± 0.00580.41 ± 0.00690.17 ± 0.00310.31 ± 0.0038
40DCAGMFMCRMSE1.12 ± 0.02611.05 ± 0.01870.52 ± 0.00930.86 ± 0.0200
MAE0.32 ± 0.00680.28 ± 0.00550.13 ± 0.00220.22 ± 0.0038
DCAGMFMC1RMSE1.47 ± 0.01901.14 ± 0.02000.57 ± 0.01480.89 ± 0.0186
MAE0.45 ± 0.00450.31 ± 0.00650.13 ± 0.00250.24 ± 0.0032
DCAGMFMC2RMSE1.48 ± 0.02211.14 ± 0.01570.54 ± 0.01280.88 ± 0.0145
MAE0.45 ± 0.00520.31 ± 0.00500.13 ± 0.00300.23 ± 0.0032
50DCAGMFMCRMSE0.99 ± 0.02560.94 ± 0.01620.46 ± 0.01040.77 ± 0.0185
MAE0.26 ± 0.00590.23 ± 0.00430.10 ± 0.00210.18 ± 0.0028
DCAGMFMC1RMSE1.35 ± 0.01971.02 ± 0.01960.49 ± 0.01360.80 ± 0.0132
MAE0.37 ± 0.00440.25 ± 0.00420.11 ± 0.00210.19 ± 0.0024
DCAGMFMC2RMSE1.31 ± 0.02091.03 ± 0.01830.49 ± 0.01410.80 ± 0.0146
MAE0.36 ± 0.00400.25 ± 0.00510.11 ± 0.00220.19 ± 0.0023
Note: Bold text indicates superior results.
Table 4. RMSE between the real data and predicted values of meteorological indicators (mean ± standard deviation of 50 results).
Table 4. RMSE between the real data and predicted values of meteorological indicators (mean ± standard deviation of 50 results).
Missing Rate (%)Imputation MethodsTEMP (°C)VISIB (Mile)WDSP (Knots)MXSPD (Knots)
RMSERMSERMSERMSE
20KNN2.09 ± 0.10723.22 ± 0.17092.04 ± 0.25393.12 ± 0.2267
HFI2.13 ± 0.00152.93 ± 0.00191.49 ± 0.00262.41 ± 0.0012
SFI2.14 ± 0.00112.92 ± 0.00081.49 ± 0.00192.43 ± 0.0022
MVNFMC3.63 ± 0.01551.99 ± 0.02370.90 ± 0.01341.47 ± 0.0245
GRMFMC5.89 ± 0.30162.73 ± 0.20301.55 ± 0.07042.37 ± 0.2344
DCAGMFMC1.55 ± 0.02831.47 ± 0.02240.70 ± 0.02141.16 ± 0.0232
30KNN2.37 ± 0.13493.42 ± 0.14742.10 ± 0.25383.26 ± 0.1969
HFI2.14 ± 0.00422.94 ± 0.00251.52 ± 0.00452.42 ± 0.0046
SFI2.14 ± 0.00262.94 ± 0.00091.51 ± 0.00302.43 ± 0.0038
MVNFMC2.98 ± 0.01451.64 ± 0.01780.77 ± 0.01381.22 ± 0.0205
GRMFMC4.31 ± 0.33722.34 ± 0.25331.29 ± 0.06601.81 ± 0.0898
DCAGMFMC1.28 ± 0.02871.21 ± 0.02290.57 ± 0.01700.97 ± 0.0187
40KNN2.68 ± 0.13733.63 ± 0.11622.24 ± 0.18373.40 ± 0.1933
HFI2.17 ± 0.00383.01 ± 0.03051.51 ± 0.00402.46 ± 0.0083
SFI2.18 ± 0.00602.97 ± 0.00791.52 ± 0.00302.43 ± 0.0040
MVNFMC2.56 ± 0.01711.39 ± 0.01580.65 ± 0.00721.08 ± 0.0156
GRMFMC3.15 ± 1.77031.51 ± 0.84500.83 ± 0.46791.22 ± 0.6879
DCAGMFMC1.12 ± 0.02611.05 ± 0.01870.52 ± 0.00930.86 ± 0.0200
50KNN3.02 ± 0.12893.84 ± 0.14472.27 ± 0.14323.40 ± 0.129
HFI2.23 ± 0.01923.03 ± 0.02721.53 ± 0.00892.47 ± 0.0100
SFI2.26 ± 0.02103.12 ± 0.03361.52 ± 0.00262.45 ± 0.0077
MVNFMC2.31 ± 0.01941.25 ± 0.01340.60 ± 0.00810.97 ± 0.0180
GRMFMC2.07 ± 1.89181.01 ± 0.92790.59 ± 0.53770.83 ± 0.7606
DCAGMFMC0.99 ± 0.02560.94 ± 0.01620.46 ± 0.01040.77 ± 0.0185
Note: Bold text indicates superior results.
Table 5. Mean Absolute Error (MAE) between the real data and predicted values of meteorological indicators (mean ± standard deviation of 50 results).
Table 5. Mean Absolute Error (MAE) between the real data and predicted values of meteorological indicators (mean ± standard deviation of 50 results).
Missing Rate (%)Imputation
Methods
TEMP (°C)VISIB (Mile)WDSP (Knots)MXSPD (Knots)
MAEMAEMAEMAE
20KNN1.55 ± 0.07422.28 ± 0.11881.18 ± 0.09162.04 ± 0.1235
HFI1.71 ± 0.00152.19 ± 0.00410.96 ± 0.00191.71 ± 0.0006
SFI1.71 ± 0.00192.17 ± 0.00290.95 ± 0.00241.72 ± 0.0008
MVNFMC1.70 ± 0.00670.80 ± 0.00870.33 ± 0.00350.58 ± 0.0064
GRMFMC2.53 ± 0.13871.09 ± 0.08650.54 ± 0.02580.83 ± 0.0833
DCAGMFMC0.63 ± 0.01040.56 ± 0.01050.25 ± 0.00610.44 ± 0.0058
30KNN1.76 ± 0.09102.45 ± 0.10851.23 ± 0.08122.12 ± 0.0963
HFI1.71 ± 0.00142.18 ± 0.00350.96 ± 0.00231.72 ± 0.0013
SFI1.71 ± 0.00172.20 ± 0.00370.96 ± 0.00131.73 ± 0.0020
MVNFMC1.14 ± 0.00480.54 ± 0.00470.22 ± 0.00280.38 ± 0.0043
GRMFMC1.47 ± 0.12130.76 ± 0.08230.36 ± 0.02010.53 ± 0.0329
DCAGMFMC0.43 ± 0.00920.38 ± 0.00850.16 ± 0.00420.30 ± 0.0041
40KNN1.99 ± 0.09302.62 ± 0.08201.30 ± 0.06012.21 ± 0.0881
HFI1.75 ± 0.00132.23 ± 0.00930.97 ± 0.00301.76 ± 0.0069
SFI1.73 ± 0.00132.18 ± 0.00310.97 ± 0.00291.73 ± 0.0011
MVNFMC0.84 ± 0.00500.40 ± 0.00370.17 ± 0.00160.29 ± 0.0028
GRMFMC0.93 ± 0.52800.42 ± 0.23680.20 ± 0.11440.32 ± 0.1770
DCAGMFMC0.32 ± 0.00680.28 ± 0.00550.13 ± 0.00220.22 ± 0.0038
50KNN2.24 ± 0.09842.78 ± 0.10891.34 ± 0.04892.24 ± 0.0642
HFI1.75 ± 0.00832.24 ± 0.00160.98 ± 0.00381.76 ± 0.0047
SFI1.78 ± 0.00672.22 ± 0.00680.98 ± 0.00231.76 ± 0.0031
MVNFMC0.68 ± 0.00540.32 ± 0.00310.13 ± 0.00160.24 ± 0.0029
GRMFMC0.55 ± 0.50170.25 ± 0.23300.13 ± 0.12040.19 ± 0.1737
DCAGMFMC0.26 ± 0.00590.23 ± 0.00430.10 ± 0.00210.18 ± 0.0028
Note: Bold text indicates superior results.
Table 6. NRMSE between observed and predicted values of meteorological indicators (mean ± standard deviation of 50 results).
Table 6. NRMSE between observed and predicted values of meteorological indicators (mean ± standard deviation of 50 results).
Missing Rate (%)Imputation
Methods
TEMP (°C)VISIB (Mile)WDSP (Knots)MXSPD (Knots)
NRMSENRMSENRMSENRMSE
20KNN0.50 ± 0.02580.67 ± 0.03550.68 ± 0.08450.73 ± 0.0530
HFI0.09 ± 0.00010.16 ± 0.00010.05 ± 0.00010.07 ± 0.0000
SFI0.09 ± 0.00000.16 ± 0.00000.05 ± 0.00010.07 ± 0.0001
MVNFMC0.16 ± 0.00070.11 ± 0.00130.03 ± 0.00050.04 ± 0.0007
GRMFMC0.26 ± 0.01310.15 ± 0.01090.05 ± 0.00240.07 ± 0.0067
DCAGMFMC0.08 ± 0.00120.08 ± 0.00120.02 ± 0.00070.03 ± 0.0007
30KNN0.57 ± 0.03250.71 ± 0.03060.70 ± 0.08440.76 ± 0.0460
HFI0.09 ± 0.00020.16 ± 0.00010.05 ± 0.00020.07 ± 0.0001
SFI0.09 ± 0.00010.16 ± 0.00000.05 ± 0.00010.07 ± 0.0001
MVNFMC0.13 ± 0.00060.09 ± 0.00100.03 ± 0.00050.03 ± 0.0006
GRMFMC0.19 ± 0.01470.13 ± 0.01360.04 ± 0.00230.05 ± 0.0026
DCAGMFMC0.06 ± 0.00120.07 ± 0.00120.02 ± 0.00060.03 ± 0.0005
40KNN0.65 ± 0.03310.75 ± 0.02410.75 ± 0.06110.79 ± 0.0452
HFI0.09 ± 0.00020.16 ± 0.00160.05 ± 0.00010.07 ± 0.0002
SFI0.09 ± 0.00030.16 ± 0.00040.05 ± 0.00010.07 ± 0.0001
MVNFMC0.11 ± 0.00070.07 ± 0.00090.02 ± 0.00020.03 ± 0.0004
GRMFMC0.14 ± 0.07700.08 ± 0.04540.03 ± 0.01620.03 ± 0.0197
DCAGMFMC0.05 ± 0.00110.06 ± 0.00100.02 ± 0.00030.02 ± 0.0006
50KNN0.73 ± 0.03110.80 ± 0.03000.76 ± 0.04760.79 ± 0.0301
HFI0.10 ± 0.00140.16 ± 0.00070.05 ± 0.00010.07 ± 0.0002
SFI0.10 ± 0.00090.17 ± 0.00180.05 ± 0.00010.07 ± 0.0002
MVNFMC0.10 ± 0.00080.07 ± 0.00070.02 ± 0.00030.03 ± 0.0005
GRMFMC0.09 ± 0.08230.05 ± 0.04990.02 ± 0.01860.02 ± 0.0217
DCAGMFMC0.04 ± 0.00110.05 ± 0.00090.02 ± 0.00040.02 ± 0.0005
Note: Bold text indicates superior results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gao, H.; Bian, Y. Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion. Axioms 2025, 14, 793. https://doi.org/10.3390/axioms14110793

AMA Style

Gao H, Bian Y. Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion. Axioms. 2025; 14(11):793. https://doi.org/10.3390/axioms14110793

Chicago/Turabian Style

Gao, Haiyan, and Youdi Bian. 2025. "Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion" Axioms 14, no. 11: 793. https://doi.org/10.3390/axioms14110793

APA Style

Gao, H., & Bian, Y. (2025). Diversity Constraint and Adaptive Graph Multi-View Functional Matrix Completion. Axioms, 14(11), 793. https://doi.org/10.3390/axioms14110793

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop