Next Article in Journal
Beyond Traditional Forecasting Methods: Evaluating LSTM Performance on Diverse Time Series
Previous Article in Journal
ADAT: Adaptive Dynamic Anonymity and Traceability via Privacy-Aware Random Forest and Truncated Local Differential Privacy in a Trusted Execution Environment (TEE)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Topology Learning and Latent Input Identification Using Spatio-Temporally Linear Structured SEM

College of Information Science and Engineering, Hohai University, Changzhou 213200, China
*
Author to whom correspondence should be addressed.
Mathematics 2026, 14(5), 837; https://doi.org/10.3390/math14050837
Submission received: 26 January 2026 / Revised: 24 February 2026 / Accepted: 26 February 2026 / Published: 1 March 2026
(This article belongs to the Special Issue Advanced Computational and Intelligent Methods in Signal Processing)

Abstract

Topology identification and signal inference are cornerstone tasks in graph signal processing (GSP). Structural Equation Modeling (SEM) is particularly effective for network inference as it explicitly captures causal dependencies. However, a major bottleneck in existing SEM-based approaches is the reliance on fully observable exogenous inputs. In many practical applications, systems are driven by latent stimuli, rendering traditional estimation methods ineffective. To overcome this, we propose a novel SEM framework for the joint inference of graph topology and unknown exogenous inputs. The core innovation lies in the spatio-temporal modeling of these latent inputs: each stimulus is decomposed into a rank-one component characterized by nodal sparsity (spatial localization) and temporal piecewise smoothness (temporal persistence). This structured formulation transforms an otherwise ill-posed blind identification problem into a tractable regularized optimization task. We develop an efficient algorithm based on the Alternating Direction Method of Multipliers (ADMM) to solve the resulting convex problem. Numerical experiments on synthetic and real-world datasets demonstrate that the proposed method effectively disentangles endogenous network interactions from latent exogenous influences, outperforming baseline approaches in both topology and signal recovery.

1. Introduction

Graph signal processing (GSP) [1,2,3,4,5] offers a principled framework for analyzing data on non-Euclidean domains by representing signals on graph nodes and dependencies through edges. This structured formulation enables the study of complex networks—such as social systems—where GSP has proven effective in uncovering patterns in information diffusion, opinion dynamics, and community structures.
Among the core tasks in graph-based learning are topology identification and signal inference. Structural Equation Modeling (SEM) [6,7,8] has become a foundational tool in this setting, particularly for network analysis. Its strength lies in modeling directed relationships between nodes, thereby distinguishing direct interactions from indirect dependencies—addressing a critical limitation of correlation-based methods. In this work, we focus on linear SEMs, which characterize nodal relationships through a set of linear equations. This linear assumption is justified in many large-scale networked systems where signals fluctuate near a steady-state equilibrium or where the coupling between agents is relatively weak. Consequently, linear SEM has been widely applied to network topology identification across various domains including effective connectivity mapping in brain networks [9], monitoring of distributed sensor systems [10], and causal analysis of financial time-series [11,12,13,14,15,16,17].
The evolution of SEM estimation techniques reflects a continuous pursuit of computational efficiency and scalability. Early foundational works by Jöreskog [8] and Bentler [7] primarily relied on Maximum Likelihood (ML) estimation. While statistically rigorous, these methods often entailed high computational costs and strict distributional assumptions. To address these issues, Generalized Least Squares (GLS) and Two-Stage Least Squares (2SLS) estimators were subsequently developed by Browne [18] and Bollen [19], offering more computationally tractable alternatives. Building on these LS-based foundations, recent advances by Cai et al. [13] and Baingana et al. [12] integrated sparsity-inducing regularization (e.g., 1 -norm) into the SEM framework. This crucial innovation enabled the effective identification of high-dimensional topologies, such as gene regulatory networks and social cascades, from limited data.
However, a critical limitation persists even in these state-of-the-art approaches. While robust variations (e.g., TLS-SEM) have been developed to mitigate measurement noise [20,21], they fundamentally rely on the availability of observed input regressors. Consequently, existing frameworks typically operate under the restrictive assumption that exogenous inputs are either fully observable (albeit noisy) or modeled merely as white noise. In many practical scenarios, however, systems are driven by latent or unmeasured stimuli. For instance, in brain connectivity analysis, neural activity is often triggered by sensory inputs not explicitly captured by scanning modalities [22]; similarly, in social networks, information cascades may be initiated by unobserved external events [23]. When applied to such data, existing methods like [12,13] fail to distinguish between internal causal edges and external driving forces, leading to biased or erroneous topology estimates. Effectively addressing this limitation requires solving a blind identification problem—a challenging task recently explored in the context of graph filters [24], but which remains under-addressed in causal SEM frameworks.
To tackle this challenge, we propose a novel SEM-based model that jointly infers network topology and the structure of unknown exogenous inputs within a unified optimization framework. Drawing inspiration from sparse component analysis [25] and dictionary learning for graph signals [26], we model the latent inputs as structured components rather than random noise. Specifically, our model explicitly accounts for the active nodes and temporal duration of external influences by enforcing spatio-temporal sparsity constraints, such as the fused lasso penalty for piecewise temporal smoothness [27]. This structured formulation enables the model to capture complex system dynamics while remaining robust under realistic assumptions. To solve the resulting convex optimization problem efficiently, we develop an algorithm based on the Alternating Direction Method of Multipliers (ADMM) [28,29,30,31].
The remainder of this paper is structured as follows. Section 2 reviews the related work in graph signal processing and Structural Equation Modeling. Section 3 introduces the proposed algorithm, including the spatio-temporal SEM model, the MAP-based problem formulation, and a theoretical discussion on identifiability and well-posedness. It also details the ADMM-based numerical method developed to solve the resulting convex optimization task. Section 4 presents comprehensive numerical experiments on both synthetic and real-world datasets to validate the performance of our approach. Finally, Section 5 concludes the paper and discusses potential future research directions.

2. Related Work

Current approaches to graph topology inference can be broadly categorized by their modeling assumptions, ranging from classical estimation with known inputs to robust methods handling measurement noise, and ultimately to blind identification strategies.

2.1. Sparse Structural Equation Modeling

The classical SEM framework, rooted in psychometrics and econometrics, originally relied on ML estimation [7,8]. While asymptotically efficient, ML estimators involve convex optimization and scale poorly with network size, rendering them impractical for high-dimensional signal processing.
To mitigate the curse of dimensionality, recent advances have integrated compressive sensing principles into the SEM framework. By assuming a sparse underlying topology, researchers have formulated inference as a convex optimization problem regularized by the 1 -norm. For instance, discrete-time approaches [12] and their vector autoregressive (VAR) extensions [13] successfully recover directed edges by minimizing least-squares (LS) prediction errors.
A fundamental weakness of these sparse LS methods is their reliance on the perfect information assumption. They treat the exogenous input matrix X as ground truth. In practice, inputs are often corrupted or only partially observed. When the regressor X contains noise, standard LS estimators minimize a biased objective, leading to the identification of spurious edges and asymptotic inconsistency.

2.2. Robust Inference Under Measurement Uncertainty

Acknowledging that real-world data is rarely pristine, subsequent research shifted toward robust topology identification via an “errors-in-variables” perspective. This approach acknowledges that both the dependent variables (network outputs Y ) and the independent variables (external inputs X ) are subject to perturbations.
The TLS-SEM framework [20,21] was introduced to jointly correct the noisy input regressors and estimate the adjacency matrix. By minimizing the Frobenius norm of the perturbation matrix alongside the fitting error, TLS-SEM provides a statistically consistent estimator under Gaussian noise assumptions. Further extensions, such as structured TLS (sTLS) [20], incorporate sparsity constraints to simultaneously handle input uncertainty and topological sparsity.
While TLS methods mitigate measurement noise, they are correctional rather than generative. They operate on the premise that a baseline observation of the input exists. Consequently, these methods falter in “blind” scenarios where driving stimuli are latent (fully unobserved). Treating a missing input as “zero plus Gaussian noise” is mathematically ill-posed and fails to capture the inherent spatio-temporal structure (e.g., block sparsity) of real-world latent drivers.

2.3. Blind Identification and Joint Recovery

The challenge of identifying system dynamics driven by unobserved inputs has been extensively explored in the domain of graph filter identification. Approaches in this category [24,26] leverage the assumption that graph signals are smooth or sparse with respect to underlying topology, typically solving a joint matrix factorization problem to recover the Graph Shift Operator (GSO) and sparse inputs simultaneously.
However, graph filter-based methods frequently assume that the GSO is symmetric (undirected) or that the filter coefficients commute with the adjacency matrix. These spectral assumptions are too restrictive for causal SEMs, which must model directed, asymmetric information flows (e.g., i j does not imply j i ).
The existing literature presents a clear dichotomy: SEM-based methods [12,20] offer causal interpretability but demand observed inputs, while blind graph filter methods [24] handle latent inputs but often sacrifice the directionality inherent to SEMs. Our work bridges this gap by proposing a blind SEM framework that enforces spatio-temporal sparsity on latent inputs, thereby preserving causal directionality without requiring prior input measurements.

3. Proposed Algorithm

3.1. SEM Model

Consider a directed network of N nodes represented by an adjacency matrix A R N × N , where an entry A i , j 0 denotes a directed edge from node j to node i. We assume the graph contains no self-loops, such that A i , i = 0 for all i. Let y i ( t ) denote the observation at node i at time t, which is modeled as a linear combination of measurements from its incoming neighbors and an exogenous input x i ( t )
y i ( t ) = j i A i , j y j ( t ) + x i ( t ) , t = 1 , , T .
By defining the observation vector y ( t ) = y 1 ( t ) , , y N ( t ) and the input vector x ( t ) = x 1 ( t ) , , x N ( t ) , the system can be expressed in compact form as y ( t ) = A y ( t ) + x ( t ) . Over T time steps, the relationship is captured by the SEM
Y = AY + X ,
where Y = y ( 1 ) , , y ( T ) R N × T and X = x ( 1 ) , , x ( T ) R N × T are the observation and exogenous input matrices, respectively.
It should be noted that the relationship in (2) describes the system’s response to latent stimuli across the entire observation window. While a steady-state analysis could simplify the estimation of A , our framework is specifically designed to capture the transient dynamics and temporal evolution of X . Regarding the initial conditions, we assume that the system is either observed from a state of rest or that the effect of the initial state y ( 0 ) is encapsulated within the identified latent component X during the initial time steps, ensuring the tractability of the blind identification problem.
The structural equation in (2) implies a mapping Y = ( I A ) 1 X . For the inferred network to be physically plausible and stable, the matrix ( I A ) must be invertible, which is generally satisfied when the spectral radius of the adjacency matrix follows ρ ( A ) < 1 . In the context of real-world networks (e.g., neural or sensor networks), this condition is naturally maintained by the inherent energy dissipation and negative feedback mechanisms within the system. While our optimization framework in (5a) does not explicitly enforce a spectral constraint to preserve convexity, the row-sparsity and no-self-loop ( A i , i = 0 ) priors effectively regularize the eigenvalues of A , ensuring that the identified topology represents a stable dynamical system.
Our formulation is guided by two primary assumptions regarding the latent exogenous inputs:
  • A single exogenous stimulus may affect a subset of nodes, and a single node may be influenced by multiple distinct stimuli.
  • Each stimulus typically persists over a specific duration, with onset and offset times that vary across different inputs.
Based on these assumptions, the exogenous input matrix is modeled as a sum of Q rank-one components plus stochastic noise:
X = q = 1 Q d q c q + E ,
where Q N represents the total number of latent inputs and E N ( 0 , σ 2 I ) denotes the i.i.d. Gaussian noise. The vector d q R N identifies the spatial profile (excited nodes), while c q R T represents the temporal evolution of the q-th input. We emphasize that while E represents unpredictable noise, the structured component d q c q captures the deterministic driving forces external to the network’s internal dynamics.
Because exogenous inputs generally persist over continuous intervals, c q is expected to exhibit piecewise constant behavior. This is characterized by the first-order temporal difference operator D T R T × ( T 1 ) , defined as the bidiagonal matrix
D T = 1 1 1 1 1 1 ,
where the sparsity of the vector c q D T reflects the temporal persistence of the q-th input.
Given that X is the sum of Q rank-one matrices, its rank is at most Q. To ensure that the identification of X is tractable, we employ nuclear norm regularization X to promote a low-rank structure. Furthermore, since each spatial profile d q is sparse, X inherently exhibits row-sparsity, which motivates the use of an 2 , 1 -norm penalty to encourage joint sparsity across the rows of the input matrix.
It is important to clarify that while the SEM in (2) assumes instantaneous dependencies, it is closely related to dynamic models such as vector autoregressive (VAR) processes. The instantaneous formulation captures the contemporaneous causal effects occurring within a single sampling interval, which is particularly relevant in systems where information propagation is faster than the measurement frequency. By enforcing temporal consistency on X , we bridge the gap between a static network structure and the dynamic nature of graph signals.

3.2. Problem Formulation

In practice, observations in the SEM may be corrupted by additive noise. To provide a rigorous statistical foundation, we formulate the joint identification task within a maximum a posteriori (MAP) estimation framework. Assuming the observation noise follows a Gaussian distribution and the structural priors (sparsity and smoothness) follow corresponding Laplacian or fused-lasso distributions, the MAP estimator is equivalent to solving the following regularized optimization problem:
min A , X Y A Y X F 2 + α A 1 + β X + γ g δ X
A i , i = 0 , i = 1 , . . . , N ,
where the constraint A i , i = 0 excludes self-loops in the adjacency matrix A . Here, α , β , and γ are regularization coefficients. The penalty function g δ ( X ) is defined as a convex combination of two structural priors   
g δ X = 1 δ X 2 , 1 + δ X D T 1
with 0 δ 1 . In this formulation, X 2 , 1 = i = 1 N x i 2 represents the 2 , 1 -norm, where x i denotes the i-th row of X . The first term in g δ ( X ) is introduced to promote row-sparsity, while the second term is used to promote temporal consistency.
The inclusion of these specific regularization terms is justified by the following structural priors:
  • A 1 promotes a sparse network topology, consistent with the belief that most real-world networks are characterized by a relatively small number of direct causal links to highlight the major dependences.
  • X encourages a low-rank structure in X , reflecting the assumption that the network dynamics are driven by a limited number of latent stimuli.
  • X 2 , 1 facilitates row-sparsity to identify specific nodes that serve as the primary entry points for exogenous influences.
  • X D T 1 enforces piecewise temporal smoothness, ensuring that the identified inputs persist over meaningful durations rather than manifesting as transient noise.
It is important to distinguish the estimation principle from numerical optimization. While the MAP framework motivates the objective function in (5a), the Alternating Direction Method of Multipliers (ADMM) is employed as the computational tool to efficiently solve this convex optimization problem. By decomposing the joint identification task into smaller, tractable sub-problems, the ADMM-based solver ensures numerical stability and provides reliable estimates of the network topology A and latent stimuli X that are consistent with the prescribed structural priors.
While expectation maximization (EM) is a common strategy for models with latent variables, we adopt a direct optimization approach via ADMM. This choice is motivated by the presence of multiple non-smooth regularizers (e.g., 1 and nuclear norms), for which ADMM provides efficient proximal updates. Furthermore, given the convexity of the formulated MAP problem, the ADMM-based solver ensures global convergence and higher computational efficiency in high-dimensional settings compared to the iterative E-steps required in EM.

Discussion on Identifiability and Well-Posedness

The joint recovery of the adjacency matrix A and the latent input matrix X from observations Y = ( I A ) 1 X is inherently a blind identification problem, which immediately raises the question of identifiability: can different pairs ( A , X ) produce exactly the same observations Y ? Such observational equivalence would render the decomposition non-unique and the estimation ill-posed [32]. In our framework, however, the potential for ambiguity is significantly mitigated by the non-overlapping structural constraints imposed on the two variables.
To understand this, consider any hypothetical linear transformation M that attempts to reparameterize the system as
A = A + M , X = X M Y ,
while preserving the observation equation Y = A Y + X . For such a transformation to be observationally equivalent, it must satisfy A Y + X = A Y + X . The key observation is that any non-trivial M inevitably violates at least one of the prescribed structural priors:
  • Violation of topological constraints: The adjacency matrix A is constrained to have zero diagonal entries ( A i , i = 0 ) and is regularized by the 1 norm to encourage sparsity. If a transformation attempts to absorb part of X into A (i.e., M 0 ), the resulting A would generally acquire non-zero diagonal elements or become densely populated, directly conflicting with the no-self-loop constraint and incurring a large 1 penalty.
  • Violation of input structure: Conversely, transferring information from A into X would disrupt the low-rank and piecewise smooth temporal structure of X . The nuclear norm penalty X would increase because the rank of X would likely exceed the true number of latent sources Q. Moreover, the temporal consistency term X D T 1 would penalize any newly introduced abrupt changes that are not characteristic of the actual exogenous stimuli.
Thus, the joint spatio-temporal and low-rank regularizers act as a mathematical filter that enforces a unique decomposition by ensuring that A and X reside in disjoint low-complexity manifolds [25]. Incoherence between these manifolds—one promoting sparse, diagonal-free matrices, the other promoting low-rank and piecewise constant row structures—guarantees that any mixing of the two components would be heavily penalized by the objective function in (5a).
Beyond this qualitative argument, the well-posedness of the problem can also be examined through a degrees-of-freedom perspective. The number of unknown parameters in A (non-zero edges s) and in the structured part of X (the low-rank factorization Q ( N + T ) ) must be small relative to the total number of observations N × T . Concretely, for the inverse problem to be well-conditioned, the total effective degrees of freedom s + Q ( N + T ) should be less than N T . The regularization parameters α , β , γ in (5a) serve as “complexity controllers” that automatically balance this trade-off during optimization.
A rigorous theoretical proof of identifiability under general conditions remains an open challenge, as is common in blind source separation, dictionary learning, and robust principal component analysis (RPCA) [25]. Nevertheless, the structural disparity between the two components, combined with the convex regularization framework, provides strong empirical guarantees. The stable recovery results reported in Section 4.1.2 and Section 4.1.3 (e.g., the monotonic improvement with increasing sample size T and the high F1-scores in topology recovery) confirm that the proposed decomposition is identifiable under realistic spatio-temporal assumptions.

3.3. Numerical Method

While expectation maximization (EM) is a conventional choice for latent variable estimation, it can be computationally intensive when the model incorporates multiple non-smooth regularizers. In this work, we adopt the Alternating Direction Method of Multipliers (ADMM), as it allows for the seamless integration of proximal mappings for the 1 , nuclear, and 2 , 1 norms, providing a more direct and efficient computational path for high-dimensional topology inference.
The problem formulated above is convex and can be tackled by well-established solvers. To improve computational efficiency and scalability for large-scale networks, we develop a customized framework based on the Alternating Direction Method of Multipliers (ADMM). We first introduce an auxiliary variable Z = X to decouple the low-rank and row-sparse regularizers. The augmented Lagrangian is given by
L ρ A , X , Z , Θ = Y A Y X F 2 + α A 1 + β Z + γ g δ X + Tr Θ X Z + ρ 2 X Z F 2 ,
where Θ is the matrix of Lagrange multipliers corresponding to the constraint X = Z and ρ > 0 . The ADMM alternately updates A , X , Z , and Θ , while fixing the other variables. The major steps are summarized below.

3.3.1. Update of A

The adjacency matrix A is updated by
A ( k + 1 ) = arg min A A Y R ( k ) F 2 + α A 1 , s . t . A i , i = 0
where R ( k ) = Y X ( k ) . Each row a i of A (excluding the diagonal element A i , i ) can be updated in parallel:
a i ( k + 1 ) = arg min a R N 1 r i ( k ) Y i a 2 2 + α a 1 ,
where Y i is Y with its i-th row removed and r i ( k ) denotes the i-th row of R ( k ) . This is the standard Lasso model [33]. We employ the fast iterative shrinkage thresholding algorithm (FISTA) to solve it [34]. The main steps are summarized as follows:
b i ( t ) = a i ( t 1 ) + η ( t 1 ) 1 η ( t ) a i ( t 1 ) a i ( t 2 ) ,
a i ( t ) = S α / L b i ( t ) 2 L Y i Y i b i ( t ) r i ( k ) ,
η ( t + 1 ) = 1 + 1 + 4 η ( t ) 2 2 .
In these updates, t denotes the inner iteration index with initial values a i ( 1 ) = a i ( 0 ) = a i ( k ) and η ( 0 ) = 1 . The constant L 2 λ max ( Y i Y i ) represents the Lipschitz constant of the gradient of the smooth least-squares loss, where λ max ( · ) denotes the maximum eigenvalue operator. The soft-thresholding operator S τ ( x ) = sgn ( x ) · max ( | x | τ , 0 ) is the proximal operator of the 1 -norm, which induces the topological sparsity in A .

3.3.2. Update of X

The input matrix X is updated by
X ( k + 1 ) = arg min X B ( k ) X F 2 + γ g δ X + ρ 2 X Δ ( k ) F 2 }
where B ( k ) = Y A ( k + 1 ) Y and Δ ( k ) = Z ( k ) 1 ρ Θ ( k ) . By decomposing this problem row-wise, for each row x i we solve the following optimization problem
x i ( k + 1 ) = arg min x R T x c i ( k ) 2 2 + γ x 2 + δ D T x 1
where c i ( k ) = 2 b i ( k ) + ρ d i ( k ) 2 + ρ , and the scaled regularization parameters are γ = 2 γ 1 δ 2 + ρ and δ = 2 γ δ 2 + ρ . Here, b i ( k ) and d i ( k ) denote the i-th row of B ( k ) and Δ ( k ) , respectively.
We employ the Primal-Dual Hybrid Gradient (PDHG) algorithm [35] to solve this subproblem. By introducing the dual variable g R T 1 , the 1 norm term is reformulated via its convex conjugate as max g δ g D T x . The subproblem is thus equivalent to finding the saddle point of the following Lagrangian:
min x max g δ x c i ( k ) 2 2 + γ x 2 + g D T x .
The saddle-point problem is solved iteratively using the Primal-Dual Hybrid Gradient (PDHG) algorithm with an inner iteration index m. At each iteration, the variables are updated using the dual and primal step sizes, denoted by σ > 0 and τ > 0 , respectively. To guarantee the convergence of the algorithm, these parameters are chosen to satisfy the stability condition σ τ L 2 < 1 , where L =   D T 2 denotes the operator norm of the difference matrix D T . Furthermore, the dual update involves an orthogonal projection onto the -ball with radius δ , which is denoted by proj · δ ( · ) .
  • Dual Update (Projection Step):
    g ( m + 1 ) = proj · δ g ( m ) + σ D T x ¯ ( m ) .
    This is performed via element-wise clamping: g j = sgn q j · min q j , δ , where q = g ( m ) + σ D T x ¯ ( m ) . Here, x ¯ ( m ) is the extrapolated primal variable used to ensure convergence stability.
  • Primal Update (Proximal Step): The update for x combines the quadratic fidelity term and the 2 penalty. First, an intermediate vector v ^ is computed by solving the optimality condition of the quadratic part of the Lagrangian:
    v ^ = 2 τ c i ( k ) + x ( m ) τ D T g ( m + 1 ) 1 + 2 τ .
    Subsequently, the primal variable x ( m + 1 ) is obtained by applying the 2 proximal operator (block soft-thresholding) to v ^ :
    x ( m + 1 ) = v ^ v ^ 2 max v ^ 2 τ γ 1 + 2 τ , 0 .
  • Extrapolation Step: To stabilize the interaction between primal and dual variables, the extrapolated variable x ¯ is updated for the next dual step:
    x ¯ ( m + 1 ) = x ( m + 1 ) + θ x ( m + 1 ) x ( m )
    where θ [ 0 , 1 ] is the relaxation parameter used in the extrapolation step to stabilize the interaction between primal and dual variables, ensuring the algorithm converges to the saddle point of the subproblem.

3.3.3. Update of Z

The auxiliary variable Z is updated by
Z ( k + 1 ) = arg min Z 1 2 Z X ˜ ( k ) F 2 + β ρ Z = D β ρ X ˜ ( k )
where X ˜ ( k ) = X ( k + 1 ) + 1 ρ Θ ( k ) , and D λ X is the singular-value thresholding operator, given by
D λ X = U D λ Σ V ,
with X = U Σ V as the singular value decomposition of X , Σ = diag σ i i = 1 r containing singular values σ i , and D λ Σ = diag σ i λ + with t + = max 0 , t . The operator ( · ) + is defined as the positive part of a real number, such that ( t ) + = max ( 0 , t ) . Note that the auxiliary variable Z is introduced as a proximal surrogate to decouple the low-rank and row-sparse constraints within the ADMM framework.

3.3.4. Update of Θ

The Lagrangian multipliers Θ are updated by
Θ ( k + 1 ) = Θ ( k ) + ρ X ( k + 1 ) Z ( k + 1 ) ,
where ρ > 0 is the penalty parameter of the augmented Lagrangian. It controls the step size of the dual update and determines the weight of the quadratic penalty term added to ensure strong convexity during the variable updates.
The iterative procedure continues until the relative updates of A and X between consecutive iterations fall below a predefined threshold. Given that the original optimization problem is convex, the convergence of this ADMM-based framework to a global optimum is theoretically guaranteed. The complete execution of the proposed method is summarized in Algorithm 1, and the algorithm is henceforth referred to as XLS-SEM.    
Algorithm 1: Major steps of proposed XLS-SEM
Mathematics 14 00837 i001

4. Numerical Experiments

In this section, we conduct numerical experiments on both synthetic and real-world data to evaluate the performance of the proposed algorithm. To demonstrate its efficacy in handling latent exogenous stimuli, we benchmark our framework against several state-of-the-art baselines that conventionally require direct input measurements:
  • LS-SEM [12,13]: The conventional sparse Structural Equation Model based on least-squares estimation, which utilizes 1 regularization to recover network topology.
  • TLS-SEM and sTLS-SEM [20]: Robust variants designed to account for errors in variables. While TLS-SEM addresses measurement noise, sTLS-SEM further incorporates sparsity regularization on the input matrix E .
Crucially, a defining advantage of the XLS-SEM framework is its blind identification capability. To simulate a realistic scenario where exogenous inputs exist but cannot be perfectly captured, the baseline methods (LS-SEM, TLS-SEM, and sTLS-SEM) are provided with noisy observations of the inputs. In contrast, the proposed XLS-SEM receives no prior information regarding exogenous inputs and must infer the network topology solely from observed output signals. All model regularization parameters are determined via exhaustive grid search.

4.1. Synthetic Data

In this suite of experiments, we evaluate the proposed framework using a synthetic network of N = 20 nodes generated via an Erdős-Rényi (ER) random graph with an edge connection probability of 0.2 . The system dynamics are driven by Q latent exogenous input sources. To faithfully reflect the spatio-temporal characteristics of real-world stimuli, the inputs are synthesized as follows: for each source q, a spatial signature d q is created by randomly selecting P active nodes. The corresponding temporal profile c q is modeled as a piecewise constant signal containing up to three non-overlapping active segments with randomized onset times and durations (constrained by T / 3 ). This construction naturally ensures the sparsity of the temporal finite difference D T c q . Non-zero amplitudes of these latent inputs are independently sampled from a uniform distribution over [ 1 , 1 ] .
Given the topology A and the exogenous matrix X , the observed signals are generated as Z = Y + E , where Y = I A 1 X represents the latent clean graph signals and E denotes i.i.d. Gaussian measurement noise is sampled from N 0 , σ E 2 . The noise variance σ E 2 is calibrated to a prescribed output signal-to-noise ratio ( SNR Y ), defined as SNR Y = 10 log 10 y ¯ 2 2 N σ E 2 , where y ¯ = 1 T Y 1 is the temporal mean of the signals.
To assess the robustness of blind identification under measurement uncertainty, we simulate a scenario where baseline methods (LS-SEM, TLS-SEM, and sTLS-SEM) are provided with corrupted input measurements X ˜ = X + R , where R is adjusted to achieve a target input SNR X . Crucially, the proposed XLS-SEM algorithm performs inference in a completely blind manner, ignoring X ˜ and relying solely on the observed output Z .
The performance is averaged over 100 independent Monte Carlo trials for each configuration of SNR Y , SNR X , Q, P, and T. We employ two primary metrics for performance evaluation and comparison:
  • Mean squared error (MSE) of the adjacency matrix: MSE A = 1 N 2 A A ^ F 2 where A ^ denotes the estimated adjacency matrix, evaluating topology recovery.
  • MSE of the signal reconstruction (denoising): MSE Y = 1 N T Y Y ^ F 2 where Y ^ denotes the estimated output matrix.

4.1.1. Visual Illustration and Convergence

Before analyzing the statistical performance metrics, we first provide a qualitative assessment of the proposed XLS-SEM using a representative trial. The synthetic data was generated based on an ER graph with N = 20 nodes and connection probability p = 0.2 . The observation period was set to T = 100 . To simulate the spatio-temporal sparsity of the latent drivers, we generated X with K = 3 temporal segments and P = 3 active nodes per segment. Both the input and output signal-to-noise ratios were set to SNR Y = SNR X = 15 dB.
Figure 1 visually compares the ground truth and estimated matrices. As observed in Figure 1a, the latent input X exhibits a distinct block-sparse structure, reflecting the spatio-temporal activation patterns of the exogenous sources. The recovered X ^ (Figure 1b) successfully captures these sparsity patterns and active segments with high fidelity. For visual clarity, a small hard threshold (<0.1) was applied to the estimated sparse matrix to suppress negligible numerical residuals inherent to the iterative solver. Consequently, the recovery error map (Figure 1c) shows minimal discrepancies, validating the effectiveness of the joint 1 and nuclear norm regularization. Similarly, the second row (Figure 1d–f) demonstrates the signal reconstruction capability; despite the observed signals being corrupted by noise, the estimated Y ^ in Figure 1e closely matches the clean ground truth Y (Figure 1d), corroborating the denoising properties of our framework.
The convergence behavior of the ADMM-based solver is illustrated in Figure 2. The plot displays the relative change of the updated variables A and X versus the iteration count on a logarithmic scale. Both curves exhibit a rapid and monotonic decay, dropping below the tolerance threshold ( 10 4 ) within a few hundred iterations. This empirical evidence confirms the computational efficiency and numerical stability of the proposed algorithm.

4.1.2. Impact of Input and Output SNR

We first examine the sensitivity of the algorithms to noise levels by varying SNR Y and SNR X independently, while fixing the non-varying parameter at 15 dB. The remaining parameters are chosen as Q = 2 , P = 3 , and T = 100 . Figure 3a,b illustrate topology recovery and reconstruction accuracy under varying observation noise ( SNR Y ), while Figure 3c,d present the results for varying input noise ( SNR X ).
As anticipated, the performance of all methods improves monotonically with increasing SNR. Notably, XLS-SEM consistently outperforms the baseline methods across the entire SNR spectrum. To provide a more granular assessment of structural recovery capability—as requested by the statistical motivation of our MAP framework—we summarize the precision, recall, and F1-scores in Table 1 for the case of SNR Y = 15 dB.
While LS-SEM and TLS-variants are limited by the quality of the provided input measurements X ˜ , XLS-SEM effectively “denoises” the system by leveraging the low-rank and spatio-temporal structural priors of the latent inputs. As shown in Table 1, although the restricted sample size ( T = 100 ) makes joint identification highly challenging, XLS-SEM achieves the highest precision and F1-score. This advantage remains consistent despite the fact that XLS-SEM performs the task in a completely blind manner, whereas the baselines rely on noisy prior observations of the inputs. This confirms that the joint structural priors effectively compensate for the absence of direct measurements, allowing XLS-SEM to exceed the performance ceiling typically imposed by corrupted input data in conventional SEM frameworks.

4.1.3. Impact of Sample Size T

The influence of the temporal observation length T is evaluated with SNR Y and SNR X fixed at 15 dB. As shown in Figure 4, an increase in T provides more data points for the ADMM framework to resolve the underlying causal links and latent profiles, leading to a visible reduction in MSE for all methods. XLS-SEM exhibits superior sample efficiency, achieving lower estimation errors even at smaller values of T. For the output reconstruction ( MSE Y ), the gap between XLS-SEM and the baselines widens as T increases, suggesting that our model more effectively exploits temporal consistency ( X D T 2 ) as more temporal information becomes available.
To further investigate the asymptotic identifiability of the proposed framework, we evaluate structural recovery performance under an extended observation window. As the sample size T increases from 100 to 1000, the F1-scores of XLS-SEM exhibit a steady and significant upward trend, rising from 0.340 to 0.482 . While the joint identification of A and X remains inherently challenging in a completely blind setting, this performance gain confirms the statistical consistency of our approach. It demonstrates that the incorporated joint spatio-temporal and low-rank priors allow the algorithm to effectively resolve structural ambiguities and suppress observation noise as more temporal information becomes integrated.

4.1.4. Impact of Latent Input Complexity (Q and P)

Beyond signal quality and length, we investigate the impact of the structural complexity of X by varying the number of input channels Q and the spatial influence per channel P, with SNR Y = 15 dB and T = 100 .
When varying Q with P = 2 (Figure 5a,b), most baselines maintain a relatively stable yet inferior performance. XLS-SEM consistently maintains the highest accuracy, although its performance shows a slight, gradual degradation as Q increases. This phenomenon is likely due to the fact that a larger number of input channels increases the density of the exogenous matrix X , which marginally weakens the effectiveness of the row-sparsity ( 2 , 1 ) and low-rank (nuclear norm) priors. A similar trend is observed when varying the number of influenced nodes P (Figure 5c,d). Despite the increased complexity of the exogenous stimuli, XLS-SEM remains robust, demonstrating that the joint regularization framework can effectively separate dense or multi-channel latent inputs from the underlying graph topology.

4.2. Real-World Dataset

4.2.1. Diabetes Clinical Records

In this subsection, we evaluate the proposed framework using a real-world diabetes dataset from the UCI Machine Learning Repository [36], which comprises longitudinal clinical records from 70 patients. Each record provides comprehensive daily profiles of over ten clinical indicators, with four measurements per day capturing blood glucose levels, insulin dosages, and physical activity schedules.
The model variables are constructed through rigorous preprocessing. The output matrix Z R 70 × 120 represents blood glucose measurements across N = 70 patients over T = 120 time points. To handle data irregularities, missing values are forward-filled along the temporal dimension, and the entries of Z are subsequently mapped to the [ 0 , 1 ] range via min–max normalization.
To accurately reflect the physiological impact of interventions, we incorporate clinically established pharmacokinetic duration windows: regular insulin (5 h), NPH insulin (12 h), ultra-long-acting insulin (27 h), and meals/exercise (≈2 h). These effects are modeled as discrete-time diffusion processes by convolving each input at time t with a rectangular window w ( τ ) of width duration / sampling interval . For instance, the NPH insulin signal is diffused as X ˜ NPH ( t ) = τ = 1 1 w ( τ ) · X NPH ( t τ ) , while ultra-long-acting insulin spans five time steps. Exercise inputs are assigned a 2-h window, influencing both the current and the subsequent time step.
In this context, Z tracks blood glucose trajectories, while exogenous factors—including exercise and insulin administration—serve as latent input signals. To account for patient heterogeneity, the graph topology A is designed to encode similarities among patients within the same diabetes subtype. This allows the model to capture individualized glucose dynamics while identifying shared regulatory mechanisms across the cohort. By applying the same structural priors validated in our synthetic experiments, XLS-SEM identifies a sparse yet physically interpretable interaction network, which reflects the shared physiological responses among patients under similar insulin–meal protocols.
Figure 6 illustrates the normalized fitting loss as a function of the time horizon T, defined as Z A Z X F 2 / Z F 2 , where Z denotes the observation matrix, and A , X represent the estimated parameters. As expected, increasing the observation window T provides more data for the model to resolve underlying dependencies, thereby improving reconstruction accuracy. Notably, XLS-SEM consistently achieves the lowest reconstruction loss. Consistent with the MAP-based formulation, the superior alignment with clinical data demonstrates that the joint incorporation of low-rank and spatio-temporal smoothness priors effectively regularizes the identification process, allowing the model to distinguish true physiological dependencies from the measurement noise and sensor missingness inherent in real-world longitudinal records.
Beyond numerical fitting, the graph topology A identified by XLS-SEM reveals the latent physiological interdependencies within the patient cohort. Specifically, the sparse connections in A tend to cluster patients who exhibit similar glucose–insulin sensitivity profiles. This suggests that the model effectively captures shared regulatory mechanisms by identifying correlations in glucose fluctuations. By disentangling the individual-specific transient noise through X , the recovered A provides a robust representation of common metabolic dynamics, which is consistent with the physiological variations expected in type-1 diabetes heterogeneity.

4.2.2. Beijing Air Quality Data

We further evaluate the framework by inferring the topology of an air quality monitoring network in Beijing, focusing on the influence of extreme meteorological events. This study utilizes hourly data from 12 national monitoring stations (UCI Repository). To characterize the overall pollution level, we construct a composite air quality signal Z by summing the min–max normalized concentrations of PM 2.5 and O 3 .
The exogenous input X represents the aggregate intensity of extreme meteorological conditions. It is computed by summing the normalized values of temperature, dew point, atmospheric pressure, and wind speed, specifically when these variables exceed severity thresholds ( 0.8 for temperature, dew point, and pressure; 0.7 for wind speed). The spatial relationships inferred by the model are validated against a ground-truth adjacency matrix A constructed based on the geographical proximity of the stations.
Figure 7 depicts the variation of the normalized fitting loss with respect to the time span T. The results indicate that all algorithms exhibit improved fitting performance as the sample size increases. However, XLS-SEM maintains a consistent performance lead over LS-SEM, TLS-SEM, and sTLS-SEM. The identified spatial relationships inferred by XLS-SEM reveal a topology that closely aligns with the geographical diffusion patterns of pollutants between monitoring stations. This superior alignment with observational data demonstrates that the comprehensive temporal features captured by XLS-SEM allow the model to better account for the delayed and persistent effects typically associated with meteorological impacts on air quality, further validating the structural consistency of our framework in uncovering physical network dependencies without prior knowledge of exact input trajectories.
The structural validity of the inferred topology A is further confirmed by its alignment with the geographical distribution of the monitoring stations. The dominant edges in A correspond to stations located along potential pollutant transport paths, such as the North–South axis of the Beijing metropolitan area. This indicates that XLS-SEM successfully identifies the directional diffusion patterns of PM 2.5 and O 3 pollutants. Unlike baseline methods that are easily biased by unmeasured local emissions, our framework utilizes the low-rank prior on X to isolate these local disturbances, thereby uncovering intrinsic spatial dependencies that are consistent with regional atmospheric transport patterns.

5. Conclusions

This paper addresses the challenge of blind network topology identification under latent exogenous inputs. Unlike conventional SEM approaches that often confound external stimuli with endogenous interactions, we proposed a spatio-temporally structured SEM framework that jointly infers graph topology and unobserved driving signals.
The core innovation is the structure-aware modeling of latent inputs, incorporating spatial sparsity and temporal piecewise smoothness. Through the development of an MAP-based joint identification framework, this prior transforms the ill-posed blind identification task into a tractable regularized optimization problem.
Experimental results on synthetic and real-world datasets demonstrate that our method significantly outperforms state-of-the-art sparse SEM baselines. Specifically, the proposed framework achieves superior performance not only in numerical reconstruction accuracy but also in structural recovery metrics, as evidenced by the high F1-scores and the observed asymptotic consistency. By effectively disentangling internal dynamics from latent drivers, the proposed framework provides a robust tool for analyzing complex systems with unmeasured inputs. Future work will extend this model to time-varying topologies and develop automated criteria for determining the number of latent sources.

Author Contributions

Conceptualization, J.Z.; methodology, J.Z.; software, J.Z.; validation, J.Z., R.Y. and X.S.; formal analysis, J.Z.; investigation, J.Z.; resources, J.Z.; data curation, J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, J.Z., R.Y., X.S. and S.F.; visualization, J.Z.; supervision, J.Z.; project administration, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors would like to thank their supervisors and colleagues for their invaluable guidance, technical support, and constructive comments throughout the preparation of this manuscript. The authors also appreciate the anonymous reviewers for their insightful suggestions that helped improve the quality of this work.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

SEMStructural Equation Modeling
ADMMAlternating Direction Method of Multipliers
MSEMean Squared Error
SNRSignal-to-Noise Ratio
PDHGPrimal-Dual Hybrid Gradient

References

  1. Shuman, D.I.; Narang, S.K.; Frossard, P.; Ortega, A.; Vandergheynst, P. The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains. IEEE Signal Process. Mag. 2013, 30, 83–98. [Google Scholar] [CrossRef]
  2. Leus, G.; Marques, A.G.; Moura, J.M.F.; Ortega, A.; Shuman, D.I. Graph signal processing: History, development, impact, and outlook. IEEE Signal Process. Mag. 2023, 40, 49–60. [Google Scholar] [CrossRef]
  3. Yan, Y.; Hou, J.; Song, Z.; Kuruoglu, E.E. Signal processing over time-varying graphs: A systematic review. arXiv 2024, arXiv:2412.00462. [Google Scholar] [CrossRef]
  4. Sandryhaila, A.; Moura, J.M.F. Discrete signal processing on graphs. IEEE Trans. Signal Process. 2013, 61, 1644–1656. [Google Scholar] [CrossRef]
  5. Ortega, A.; Frossard, P.; Kovacevic, J.; Moura, J.M.F.; Vandergheynst, P. Graph Signal Processing: Overview, Challenges, and Applications. Proc. IEEE 2018, 106, 808–828. [Google Scholar] [CrossRef]
  6. Kaplan, D. Structural Equation Modeling: Foundations and Extensions, 2nd ed.; Sage Publications: Thousand Oaks, CA, USA, 2009. [Google Scholar]
  7. Bentler, P.M.; Weeks, D.G. Linear structural equations with latent variables. Psychometrika 1980, 45, 289–308. [Google Scholar] [CrossRef]
  8. Jöreskog, K.G. A general method for estimating a linear structural equation system. In Structural Equation Modeling; Academic Press: New York, NY, USA, 1973; pp. 85–112. [Google Scholar]
  9. Friston, K.J. Functional and effective connectivity: A review. Brain Connect. 2011, 1, 13–36. [Google Scholar] [CrossRef]
  10. Giannakis, G.B.; Shen, Y.; Karanikolas, G.V. Monitoring and optimization of cyber-physical networks: A graph signal processing approach. IEEE Signal Process. Mag. 2018, 35, 34–46. [Google Scholar]
  11. Mei, J.; Moura, J.M.F. Signal processing on graphs: Estimating the structure of a graph. IEEE Trans. Signal Process. 2017, 65, 2045–2058. [Google Scholar]
  12. Baingana, B.; Mateos, G.; Giannakis, G.B. Proximal-gradient algorithms for tracking cascades over social networks. IEEE J. Sel. Top. Signal Process. 2014, 8, 563–575. [Google Scholar] [CrossRef]
  13. Cai, X.; Bazerque, J.A.; Giannakis, G.B. Inference of gene regulatory networks with sparse structural equation models exploiting genetic perturbations. PLoS Comput. Biol. 2013, 9, e1003068. [Google Scholar] [CrossRef]
  14. Giannakis, G.B.; Shen, Y.; Karanikolas, G.V. Topology identification and learning over graphs: Accounting for nonlinearities and dynamics. Proc. IEEE 2018, 106, 787–807. [Google Scholar] [CrossRef]
  15. Goldberger, A.S. Structural equation methods in the social sciences. Econometrica 1972, 40, 979–1001. [Google Scholar] [CrossRef]
  16. Muthén, B. A general structural equation model with dichotomous, ordered categorical, and continuous latent variable indicators. Psychometrika 1984, 49, 115–132. [Google Scholar] [CrossRef]
  17. Ning, X.; Karypis, G. SLIM: Sparse linear methods for top-N recommender systems. In Proceedings of the 2011 IEEE 11th International Conference on Data Mining (ICDM), Vancouver, BC, Canada, 11–14 December 2011; pp. 497–506. [Google Scholar]
  18. Browne, M.W. Generalized least squares estimators in the analysis of covariance structures. S. Afr. Stat. J. 1974, 8, 1–24. [Google Scholar] [CrossRef]
  19. Bollen, K.A. An alternative two stage least squares (2SLS) estimator for latent variable equations. Psychometrika 1996, 61, 109–121. [Google Scholar] [CrossRef]
  20. Ceci, E.; Shen, Y.; Giannakis, G.B.; Barbarossa, S. Graph-based learning under perturbations via total least-squares. IEEE Trans. Signal Process. 2020, 68, 2870–2882. [Google Scholar] [CrossRef]
  21. Shames, I.; Teixeira, A.M.H.; Sandberg, H.; Johansson, K.H. Distributed identification of network topology with noisy node data. In Proceedings of the 51st IEEE Conference on Decision and Control (CDC), Maui, HI, USA, 10–13 December 2012; pp. 5205–5210. [Google Scholar]
  22. Preti, M.G.; Bolton, T.A.W.; Griffa, A.; Van De Ville, D. Graph signal processing for neuroimaging to reveal dynamics of brain structure-function coupling. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  23. Mazurek, K.; Hohol, M. Revisiting information cascades in online social networks. Mathematics 2024, 13, 77. [Google Scholar] [CrossRef]
  24. Segarra, S.; Marques, A.G.; Mateos, G.; Ribeiro, A. Blind identification of graph filters. IEEE Trans. Signal Process. 2017, 65, 1146–1159. [Google Scholar] [CrossRef]
  25. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
  26. Thanou, D.; Shuman, D.I.; Frossard, P. Learning parametric dictionaries for graph signals. IEEE Trans. Signal Process. 2017, 65, 2517–2530. [Google Scholar]
  27. Tibshirani, R.; Saunders, M.; Rosset, S.; Zhu, J. Sparsity and smoothness via the fused lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 91–108. [Google Scholar] [CrossRef]
  28. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
  29. Wang, Y.; Liu, J.; Zhang, X. Low-rank matrix recovery via nonconvex optimization and ADMM algorithm. Mathematics 2023, 11, 652. [Google Scholar]
  30. Hong, M.; Luo, Z.Q.; Razaviyayn, M. Convergence analysis of alternating direction method of multipliers for a family of nonconvex problems. SIAM J. Optim. 2016, 26, 337–364. [Google Scholar] [CrossRef]
  31. Liu, Q.; Shen, Z.; Gu, Y. Linearized ADMM for nonconvex nonsmooth optimization with convergence analysis. IEEE Access 2019, 7, 76131–76144. [Google Scholar] [CrossRef]
  32. Bollen, K.A. Structural Equations with Latent Variables; John Wiley & Sons: New York, NY, USA, 1989. [Google Scholar]
  33. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  34. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  35. Chambolle, A.; Pock, T. A first-order primal-dual algorithm for convex problems with applications to imaging. J. Math. Imaging Vis. 2011, 40, 120–145. [Google Scholar] [CrossRef]
  36. Kahn, M. Diabetes Data Set. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/34/diabetes (accessed on 15 February 2026).
Figure 1. Visual illustration of topology inference and signal recovery on synthetic data ( N = 20 , T = 100 ). (ac) Input recovery performance: (a) ground truth latent input X ; (b) estimated input X ^ (proposed); (c) recovery error | X X ^ | . (df) Signal reconstruction performance: (d) clean graph signal Y ; (e) reconstructed signal Y ^ (proposed); (f) reconstruction error | Y Y ^ | .
Figure 1. Visual illustration of topology inference and signal recovery on synthetic data ( N = 20 , T = 100 ). (ac) Input recovery performance: (a) ground truth latent input X ; (b) estimated input X ^ (proposed); (c) recovery error | X X ^ | . (df) Signal reconstruction performance: (d) clean graph signal Y ; (e) reconstructed signal Y ^ (proposed); (f) reconstruction error | Y Y ^ | .
Mathematics 14 00837 g001
Figure 2. Convergence analysis of the proposed ADMM solver. The curves show the relative change of variables A and X over iterations.
Figure 2. Convergence analysis of the proposed ADMM solver. The curves show the relative change of variables A and X over iterations.
Mathematics 14 00837 g002
Figure 3. Performance comparison under varying signal-to-noise Ratio (SNR). (a,b) Metrics for varying SNR Y ; (c,d) metrics for varying SNR X .
Figure 3. Performance comparison under varying signal-to-noise Ratio (SNR). (a,b) Metrics for varying SNR Y ; (c,d) metrics for varying SNR X .
Mathematics 14 00837 g003
Figure 4. Performance comparison under varying sample size T. (a) Topology estimation error; (b) signal reconstruction error.
Figure 4. Performance comparison under varying sample size T. (a) Topology estimation error; (b) signal reconstruction error.
Mathematics 14 00837 g004
Figure 5. Performance comparison under varying latent input complexity Q or P. (a,b) Performance versus number of channels Q; (c,d) performance versus spatial influence P.
Figure 5. Performance comparison under varying latent input complexity Q or P. (a,b) Performance versus number of channels Q; (c,d) performance versus spatial influence P.
Mathematics 14 00837 g005
Figure 6. Normalized fitting loss versus observation length T for the Diabetes dataset.
Figure 6. Normalized fitting loss versus observation length T for the Diabetes dataset.
Mathematics 14 00837 g006
Figure 7. Normalized fitting loss versus observation length T for the Beijing air quality dataset.
Figure 7. Normalized fitting loss versus observation length T for the Beijing air quality dataset.
Mathematics 14 00837 g007
Table 1. Topology recovery performance comparison under T = 100 and SNR Y = 15 dB.
Table 1. Topology recovery performance comparison under T = 100 and SNR Y = 15 dB.
MethodPrecisionRecallF1-ScoreMSEA
LS-SEM0.2330.3550.2810.228
TLS-SEM0.3200.3160.3180.218
sTLS-SEM0.2610.3030.2800.193
XLS-SEM (Ours)0.3520.3290.3400.184
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, J.; Yang, R.; Shi, X.; Feng, S. Joint Topology Learning and Latent Input Identification Using Spatio-Temporally Linear Structured SEM. Mathematics 2026, 14, 837. https://doi.org/10.3390/math14050837

AMA Style

Zhou J, Yang R, Shi X, Feng S. Joint Topology Learning and Latent Input Identification Using Spatio-Temporally Linear Structured SEM. Mathematics. 2026; 14(5):837. https://doi.org/10.3390/math14050837

Chicago/Turabian Style

Zhou, Jie, Rui Yang, Xintong Shi, and Shuyang Feng. 2026. "Joint Topology Learning and Latent Input Identification Using Spatio-Temporally Linear Structured SEM" Mathematics 14, no. 5: 837. https://doi.org/10.3390/math14050837

APA Style

Zhou, J., Yang, R., Shi, X., & Feng, S. (2026). Joint Topology Learning and Latent Input Identification Using Spatio-Temporally Linear Structured SEM. Mathematics, 14(5), 837. https://doi.org/10.3390/math14050837

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop