Next Article in Journal
2D/3D Pattern Formation Comparison Using Spectral Methods to Solve Nonlinear Partial Differential Equations of Condensed and Soft Matter
Previous Article in Journal
Algorithmic Optimal Control of Screw Compressors for Energy-Efficient Operation in Smart Power Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PISI: Physical Information Based Solver-Interactive Network Structure Reconstruction

1
School of Technology, Beijing Forestry University, Beijing 100083, China
2
Technologies of Vision (TeV), Fondazione Bruno Kessler, 38123 Trento, Italy
*
Author to whom correspondence should be addressed.
Algorithms 2025, 18(9), 584; https://doi.org/10.3390/a18090584
Submission received: 6 August 2025 / Revised: 3 September 2025 / Accepted: 12 September 2025 / Published: 16 September 2025
(This article belongs to the Section Evolutionary Algorithms and Machine Learning)

Abstract

Inference of the interactive network structure of the physical world that is captured by nonlinear dynamic systems is a long-standing goal for machine learning. Existing inference methods have shown limited incorporation of physical system information and solver interaction capabilities. We present a comprehensive Physical Information based Solver-Interactive (PISI) network structure identification framework that incorporates network topology, physical constraints, and bidirectional solver interaction in nonlinear dynamical systems. To this end, we first develop a physical information-based graphical neural network (PIGNN). The PIGNN cells are embedded as the basic integration units to iterative interact with dynamical solver. The dynamical systems’s physical information can be flexibly added to the iterative interaction for PIGNN training. The above stages are trained end-to-end using a Runge–Kutta solver. The network structure inferring capability of the proposed framework is demonstrated through two kuramoto systems. Our PISI methodology, integrating graph topology, physical constraints, and solver interactivity shows advantages in trajectory prediction and structure reconstruction compared to state-of-the-art methods.

1. Introduction

A wide range of dynamical systems in ecology [1], power systems [2], social networks [3], epidemics [4], and other areas can be seen as groups of interaction components. The interaction patterns among individual components give rise to complex dynamic behaviors at the component level, while also inducing emergent complexity at the holistic system level. The classical approach to modeling interactions between individuals is through complex networks [5]. In reality, we can only access the component observations through sensors, with the underlying interaction as a black box. However, the unknown interactions, i.e., the network structure, are of great importance in dynamic evolution. For example, identifying synaptic adjacency relationships between neurons can reveal abnormal discharge propagation pathways in epileptic seizures [6]. Accurately identifying physical adjacency relationships among generator nodes is critical for predicting cascading failures. Also, adjacency relationship identification based on social contact networks enables precise prediction of virus transmission hot spots [7]. Thus arises the problem of network structure inference in nonlinear dynamical systems.
Classic structure identification methods primarily rely on observations to reconstruct network topology. This approach fails to fully exploit the inherent physical properties of the underlying dynamical systems. In fact, the dynamical evolution process embodies substantial physical information, including but not limited to initial conditions, boundary conditions, and the governing equations themselves. Moreover, the numerical solvers of dynamical evolution are intrinsically coupled with network structure, where each iteration can be viewed as an interaction among initial conditions, network topology, and individual dynamics. Certain numerical solvers inherently generate physical information during iterative computations, such as automatic differentiation of parameters. However, such critical information remains underutilized in traditional identification paradigms.
The advancement of artificial intelligence algorithms has given rise to diverse methodologies capable of exploiting physical system information from multiple perspectives [8]. First, regarding observational data, whereas classic approaches require extensive sampling to predict short-term system states, contemporary neural network-based methods demand significantly fewer samples while aiming for extended temporal prediction horizons [9]. Secondly, from the dynamic solver interaction perspective, parameter estimation methods with solver interactivity have been developed for individual physical oscillators [10]. However, for network structure identification in systems composed of multiple oscillators, the parameter dimensionality escalates from a fixed number to a quadratic problem, rendering solver-interactive network structure identification exceptionally challenging. Thirdly, from the perspective of physical information, Physics-Informed Neural Networks (PINNs) have emerged in recent years [11]. These networks aim to integrate physical knowledge by embedding physical constraints into the loss function, thereby combining the learning capacity of neural networks with physical principles to enhance learning efficiency and identification accuracy. However, in the context of network structure identification, PINNs face challenges similar to those encountered in solver-interactive methods specifically. How to extend their advantages in single-oscillator parameter identification to network topology inference still remains a challenge.
In response to these challenges, we propose an integrated physical information-based solver-interactive network structure reconstruction framework (PISI). Specifically, our framework employs a physical information based graph neural network (PIGNN) as the foundational component and embeds PIGNN into each solver iteration. By leveraging automatic differentiation (AD) information generated during the solving process, the proposed framework achieves end-to-end network structure learning. We validate the effectiveness of the PISI method in state prediction and structure reconstruction using an Runge–Kutta (RK) solver for two Kuramoto systems. The primary innovations and contributions of this research are outlined as follows.
  • The proposed PISI is solver-interactive, enabling simultaneous estimation of network topology and system states.
  • By integrating multiple information sources, PISI dramatically reduces the required sample size for a high network structure identification accuracy.
  • The proposed PISI method realizes extending from single-node parameter identification to multi-agent system structure reconstruction by leveraging the interaction between graph neural networks and solvers.
The rest of this study is organized as follows. Section 2 reviews related works in three aspects. Section 3 describes the methodology module of this study. Section 4 presents the results of this study and discusses the state of the art of the methodology. Section 5 concludes this study.

2. Related Work

In this section, we present research related to our work in three aspects: (1) graph neural network (GNN)-based structure reconstruction, (2) PINN-based parameter estimation of nonlinear dynamics, and (3) solver-interactivity parameter estimation of single oscillators.

2.1. GNN Based Network Structure Reconstruction

GNN was initially designed to process data with known graph structures by employing message-passing mechanisms to predict or estimate node states [12]. Subsequent research has gradually expanded to infer or reconstruct unknown network structures from observed data (such as node states or interaction behaviors), forming several important branches in this direction [13]. Thomas [14] et al. took the form of a variational auto-encoder, in which the latent code represents the underlying interaction graph and the reconstruction is based on graph neural networks. Zang [15] et al. proposed combining Ordinary Diferential Equation Systems (ODEs) and GNNs to learn continuous-time dynamics on complex networks in a data-driven manner. Rakshit [16] et al. presented DyRep, a novel modeling framework for dynamic graphs that posits representation learning as a latent mediation process bridging two observed processes, namely the dynamics of the network (realized as topological evolution) and dynamics on the network (realized as activities between nodes). Zhang [17] et al. proposed an LLM-based robust graph structure inference framework, LLM4RGNN, which distills the inference capabilities of GPT-4 into a local LLM for identifying malicious edges and an LM-based edge predictor for finding missing important edges, so as to recover a robust graph structure. Although GNNs are capable of handling structure estimation and state prediction tasks, their lack of physics-informed mechanisms compared to our PISI results in a high demand for training data. For example, in [14], to reconstruct a system consisting of five oscillators, more than 50,000 samples are needed.

2.2. PINN Based Parameter Estimation of Nonlinear Dynamics

PINN integrates neural networks with physical equations by using automatic differentiation to compute PDE residuals, which are incorporated as part of the loss function. The usage of physical information requires no large training datasets, due to the fact that the neural network can be trained solely based on PDE constraints [18]. Raissi et al. devised a new family of data efficient spatio-temporal function approximators for continuous systems; for discrete systems, the model allows the use of arbitrarily accurate implicit Runge–-Kutta time stepping schemes with an unlimited number of stages. Ji [19] et al. investigated the performance of the PINN in solving stiff chemical kinetic problems with governing equations of stiff ordinary differential equations. The results elucidate the challenges of utilizing the PINN in stiff ODE systems. However, the loss function of PINN faces difficulties when solving network structure reconstruction compared to our PISI. The aforementioned applications of PINNs have primarily focused on individual systems, with insufficient research on network structure identification.

2.3. Solver-Interactivity Parameter Estimation of Nonlinear Dynamics

In both dynamical evolution and inverse problem parameter estimation, differential equation solvers for dynamical systems invariably play a crucial and inescapable role [20]. In all of the literature reviewed in Section 2.1 and Section 2.2 above, iterative solvers have been consistently employed. Therefore, the integration of parameter estimation and structure reconstruction with solver iterations is of critical importance. Kiwon [21] et al. integrated the solver into the training loop and thereby allowed the model to interact with the PDE during training. Philipp [22] et al. presented a novel hierarchical predictor-corrector scheme enabling neural networks to learn to understand and control complex nonlinear physical systems over long time frames, in which the predictor network plans optimal trajectories and the control network infers the corresponding control parameters. Drawing inspiration from classical numerical integration solution rules for differential equations, Zhai [10] et al. proposed a new recurrent neural network architecture, taking PINN cells as the basic integration units of the architecture to introduce prior physics-based biases. Although existing research has explored the integration of solver-in-the-loop approaches with PINNs, the insufficient understanding of graph properties in network structure identification tasks compared to our PISI hinders their ability to effectively address such problems.

3. Methodology

In this study, we propose a unified neural network framework that integrates neural graph, physical, and solver information, aiming to reconstruct the network structure of nonlinear dynamics systems and at the same time predict the system state. The framework comprises two modules: a physical information-based GNN, and an iterative solver. We propose an end-to-end network structure reconstruction solution. The method employs the fourth-order Runge–Kutta (RK4) method for differential equation solving as its backbone. We employ RK4 as an illustrative example, while other differential equation solvers are equally applicable. The PIGNN takes the observation data as input and outputs, with edge type posterior p ϕ ( A | u ) . With a sampling module, we obtain the estimated network structure A ˜ . A ˜ will compensate for the unknown terms in the dynamical system, thereby enabling the operation of the RK4 solver. The PIGNN and RK4 module interact as follows. At each stage of RK4, the PIGNN provide structure estimation value and the receive the stage node state to obtain the physical loss. At the end of each RK4 iteration, state loss is provided to PIGNN, thereby achieving end-to-end training across all four involved PIGNNs.

3.1. Modeling of the Network Structure Reconstruction Problem

Consider a system composed of n coupled oscillators, captured by parametrized and nonlinear differential equations of the general form
u ˙ + F [ u ; A ] = 0 ,
where u ( · , t ) consists of solutions u i , i = 1 , N defined in Banach space X , and u ˙ = ( u ˙ 1 , , u ˙ N ) denotes the time derivative term of each oscillators. F ( u ; A ) X N X N is a parametrized nonlinear operator, and A R N × N represents system coupling strength parameters, where A i j [ 0 , 1 ] is a discrete categorical variable representing the edge type between oscillators u i and u j .
Without loss of generality, for each oscillator i, the nonlinear operator N [ · , · ] consists of three parts
u ˙ i + L u i + G ( u ; A ) + D u i = 0
with L u i describing the intrinsic oscillatory linear properties, G ( u i ; A ) describing the phase interaction between oscillators, and D u i describing the spatial diffusion/dissipation of the phase field. This decomposition represents many scenarios in mathematical physics, including the extended Kuramoto model describing power grid frequency synchronization [23], the XY model [24], and the theta-gamma coupled oscillation model [25] for brain rhythm synchronization [26]. As an important model that we will focus on in the following discussion, the Kuramoto model corresponds to the case when L u i = D i u ˙ i , G ( u i ; A ) = K A i j j i sin ( u j u i ) , D u i = M i u ¨ i , with D i , M i , and K as the time-independent constant. The first problem we focus on is: how to incorporate the solver with the network inferring procedure thus to improve the state predict accuracy? The second problem we focus on is: how can we infer the coupling strength, i.e., the network structure A from as less as observed data?

3.2. Runge–Kutta Methods

The inputs consist of trajectories of the N oscillators. We denote the observed trajectories of the oscillator u i at time t as u i t . All observed data for N oscillators from time t = 1 : T are denoted as u = { u 1 , , u T } , with the u t = { u 1 t , , u N t } as all trajectories of N oscillator at time t. We denote the trajectories of node u i as u i = { u i 1 , , u i T } .
To solve the considered two problems, we first introduce one of the most widely used differential equation solvers. The Runge–Kutta method can advance temporal solutions for non-stiff ordinary differential equations and semi-discretion partial differential equations. The flexible and extensible framework accommodates both explicit and implicit formulations for temporal discretizations. The RK4 is self-starting and eliminates the dependency on historical data required by multi-step methods. The Runge–-Kutta methods are linear extrapolations based on Taylor series expansion. Since A is unknown during the iteration, we explicitly represent A in F . Next, we employ the fourth-order Runge–Kutta (RK4) method for temporal discretization as an illustrative example. For time step Δ t , from time t n to t n + 1 = t n + Δ t , the solution can be written as
u i n + 1 = u i n + h 6 ( k 1 + 2 k 2 + 2 k 3 + k 4 )
where
k 1 = F ( u i n , t n , A ) ,
k 2 = F ( u i n + k 1 2 , t n + h 2 , A ) ,
k 3 = F ( u i n + k 2 2 , t n + h 2 , A ) ,
k 4 = F ( u i n + k 3 , t n + h , A ) .

3.3. Design of PISI

It should be noticed that among the RK4 procedure, the discretization of phase interaction G ( u , A ) encodes the underlying reconstructed A, thus leading to the underlying inferred F ( u i , t , A ) . During the RK4 procedure, the unknown vector field F ( u i , t , A ) is evaluated four times at distinct intermediate stages. Higher-order Runge–Kutta methods require progressively more evaluations of the vector field. Thus, the key of structure reconstruction and states prediction lies in the refer of F ( u i , t , A ) . Our basic idea is to approximate F with a graph based neural network during the RK4 procedure. The designed neural network undergoes four invocations during a single RK4 iteration while maintaining parameter sharing across all calls. In designing the neural network, we prioritize three key considerations. First, the architecture should effectively utilize intermediate RK iteration data and physical information. Second, the PISI should effectively incorporate graph attributes of the identifiable matrix A. Finally, the network’s input and output data should maintain compatibility with the RK4 iterative procedure.

3.3.1. The Physic Information Based Graph Neural Network

Overall, we adopt the graph neural network structure from [14], but redesign the loss function to satisfy the aforementioned three requirements for our neural network. The PIGNN supports the RK4 procedure as follows by predicting the interaction graph given the trajectories.
First, we state the graph neural network part of PIGNN, then we present how the traditional GNN incorporates the physical information. The physical information-based graph neural network can be denoted as edge type posterior p ϕ ( A | u ) = P I G N N ( u , ϕ ) with ϕ the shared parameter of the four PIGNNs. The PIGNN returns a factorized distribution of A.
We state a brief introduction of GNN, which operates directly on graph-structured data by passing local messages [27]. We define a single node-to-node message passing operation in a GNN as follows:
Node Pair Embedding h i j = f ϕ 1 ( h i , h j , u , A )
Node Embedding h i = f ϕ 2 ( j i h j , u , A ) .
in which the h i and h i j are the embedding of node i and edge between nodes i and j. The function f ϕ 1 denotes a neural network like MLP [28], CNN [29] with under trainable parameter ϕ 1 . Equation (8) enables iterative bidirectional mapping between edge and node representations through successive message-passing operations.
Adopting the GNN structure from [14], we design a message passing procedure consists of two messages passing from nodes to edges as in Figure 1, with the first passing encoding the information of local node pairs and the second passing encoding global information corresponding to the considered nodes pair. Specifically, starting from a fully-connected graph without self-loop, the two nodes to edges message passing for A i j of the graph neural network is as follows
Node Embedding: h j 1 = M L P 1 ( u j )
Node Pair Embedding: h ( i j ) 1 = M L P 2 ( [ h i 1 , h i 1 ] )
Global Node Embedding: h j 2 = M L P 3 ( j i h ( i j ) 1 )
Global Pair Embedding: h ( i j ) 2 = M L P 4 ( [ h i 2 , h j 2 ] ) .
Then, from the global feature h ( i j ) 2 , we can calculate the edge type posterior p ϕ ( A i j | u ) = softmax ( h ( i j ) 2 ) .
To bridge the gap between the PIGNN and the RK4 procedure, we sample from the edge type posterion and obtain A ˜ i j (into Appendix A for sampling method). The resulting edge value should be substituted into the partial unknown vector field F feeding to Equations (4)–(7).
Above all, the interaction between the PIGNN and RK4 is as follows. In the n-th time period of RK4, from trajectories input u n , run the first PIGNN ( u , ϕ ) and obtain A ˜ 1 . Substitute A ˜ 1 into the RK4 procedure, and oobtain the evolution function F ( u n , t n , A ˜ 1 ) . Calculate the second intermediate variable k 2 . The process is then iterated cyclically until reaching the fourth stage of RK4. From Equation (3), we can then obtain the estimated trajectories u ˜ n + 1 . Within a time step of RK4, we employ adjacency matrices A ˜ n , n = 1 , 2 , 3 , 4 at four stages to enhance the model’s learning capacity. To ensure the numerical stability of the RK4 algorithm, we introduce a consistency loss term for the adjacency matrices within each time step, guaranteeing evolutionary continuity between stages. The specific design of the penalty term is detailed in next subsection.

3.3.2. Loss Function

The four PIGNNs share parameter ϕ , and the training loss consists of three components: (1) physical information: automatic differentiation of the structure, (2) the trajectory prediction loss term, and (3) the structure reconstruction loss. The incorporation of tri-dimensional information endows our method with sampling efficiency.
Automatic differentiation is a numerical method for computing derivatives of functions with respect to the input, serving as the core operation in PINNs. Its fundamental principle involves decomposing complex functions into elementary operations and systematically applying the chain rule, making it inherently compatible with most modern neural network training frameworks. The underlying structure can be calculated from the k-th PIGNN as
A phy , k n = A D ( u n , P I G N N k ( u , ϕ ) ) , k = 1 , 2 , 3 , 4 .
with A D ( · , · ) the automatic differentiation calculation in the PIGNN training procedure. As the four PIGNN shares parameter and the ground truth of the structure A is time-invariant among the four stages in RK4, we take coefficient of variation as physic information loss
Loss p h y = c v 2 ( A phy )
= 1 4 T n = 1 T k = 1 4 ( A phy , k n A ¯ phy n ) 2 / ( A ¯ phy n ) 2
where c v 2 is the coefficient of variation and A ¯ phy the mean of A ¯ phy , k , k = 1 , 2 , 3 , 4 .
For the trajectory prediction, we focus on the end-to-end loss. In one iteration of RK4, the end-to-end loss uses the sample u n + 1 and the output u ˜ n + 1 . The PIGNN parameter is trained via the MSE of the trajectories as follows
Loss t r a = 1 T n = 1 T 1 ( u n + 1 u ˜ n + 1 ) 2
For the structure reconstruction loss, we delve into the four stages of RK4 while maintaining an end-to-end learning process. For each stage, the Kullback–Leibler (KL) divergence is employed. Across different stages, we constrain the overall distribution matching by averaging the KL divergences, thereby compelling all stages to jointly optimize the alignment between latent variables and priors, rather than focusing on any single network. The resulting structure reconstruction loss of the k-th step at the n-th time period A k n is as follows:
Loss s t r = 1 4 T n = 1 T k = 1 4 KL ( p ( A k n | u ) | | p ( A ) )
The prior P ( A ) is a factorized uniform distribution over edge types. The core assumption of the factorized uniform prior is that edge types are independent and unbiased.
The total loss function of the total training loss combines the above three losses, using the weighting parameters α 1 , α 2 and α 3 to control the relative importance of each part of the optimization process. We define the total training L o s s total as follows:
Loss t o t a l = α 1 Loss p h y + α 2 Loss t r a + α 3 Loss s t r
By employing the combined loss function Loss t o t a l , the model can optimize state prediction and structure estimation. Now that we have described all the elements in the framework in Figure 2, the training goes as follows.

4. Experiment

We used PyTorch (version 2.7.0) to implement our network. All models in the experiment were trained for 300 epochs. We train the framework with Adam [30], using an initial learning rate of 3 × 10 5 that is halved every 10 epochs. The entire network was trained and evaluated using an NVIDIA GeForce RTX 4090 GPU with a batch size of 64. The grid search method was utilized to find the proper values of weights of the three loss terms.

4.1. Kuramoto Model

The Kuramoto model was independently introduced in the context of power grids in [31] and has a long history of study there. We use the following notation:
M i d 2 θ i d t 2 + D i d θ i d t = P i + K j = 1 N A i j sin ( θ j θ i )
where θ i denotes the phase angle of oscillator i, and the first-order and second derivative of θ i denote voltage angle ω i and its time derivatives, respectively. The injected power P i is selected in { 1 , 1 } . K = 9 is the uniform coupling coefficient [32]. The classical Kuramoto model can be transformed into an equivalent state-space representation for trajectories sampling:
d θ i d t = ω i d ω i d t = 1 M i ( P i D i ω i + K j = 1 N A i j sin ( θ j θ i ) ) .
The initial voltage angle is a random vector from [ 5 , 5 ] . The initial phase angle is sampled from a Logistic map to generate highly dispersed initial phase angles. We introduce the outlier value to prevent the system from reaching synchronization prematurely, as synchronization can significantly interfere with structural identification [33,34]. Specifically, we multiply an outlier parameter with the logistic map during the generation of initial values. We present the order parameter in sampling time period to quantify the level of synchronization in Figure 3. The outlier value is selected from [ 3.5 , 4 ] .

4.2. System Structure

Given the initial conditions, we can simulate the trajectories by solving Equation (21) by RK4 method. The time step is set to 0.01 in the RK4. By default, training and validation trajectories takes length 49 while test trajectories continue for another 20 time steps. We explicitly hard-coded the first edge type as a “non-edge”, i.e., no messages are passed along edges of this type. The MLPs have hidden and output dimensions of 256. The overall input/output dimension of our model is 2 for the Kuramoto model experiments. During training, we use teacher forcing in every 10th time step (i.e., every 10th time step, the model receives a ground truth input; otherwise, it receives its previous prediction as input). The detailed PIGNN structure is in Figure 1. To characterize the small-world [35] properties of smart grids and demonstrate the applicability of our method to diverse network topologies, we employ two distinct small-world networks. Small-world networks are characterized by high clustering coefficients and short average path lengths. We simulate two networks with N { 10 , 20 } nodes, representing the small world network structure in Figure 4. The generator nodes are connected by a global connection strength of 0.01. We implement the classical Watts–Strogatz model using the NetworkX library [36]. The network parameters are as in Table 1.

4.3. Physic Information

According to Equation (21), we will derive the automatic differentiation information for the Kuramoto system. We dnote
g ( ω ˙ i , ω i ) = 1 K ( M i ω ˙ i P i + D i ω i ) .
Then, we have
A D ( ( θ i , ω i ) , P G I N N ( ( θ i , ω i ) , ϕ ) ) = g ( ω ˙ i , ω i ) sin ( θ j θ i ) .
In PIGNN training, g ( · , · ) does not directly depend on the variable sin ( θ j θ i ) in PyTorch’s automatic differentiation. Therefore, we introduce an intermediate variable θ j θ i , and obtain the AD information as follows:
A D ( ( θ i , ω i ) , P G I N N ( ( θ i , ω i ) , ϕ ) ) = g ( ω ˙ i , ω i ) ( θ j θ i ) · ( θ j θ i ) sin ( θ j θ i )
Here, ω ˙ i is from the teacher forcing ω ω Δ t (i.e., the ground truth of next time step).

4.4. Experimental Results

We compare our model with the state-of-the-art structure reconstruction method and evaluate it on the two Kuramoto systems with node size 10 and 20 both in the trajectory prediction and structure reconstruction. To demonstrate our method’s sample efficiency as stated in Section 3.3.2, we conduct comparative Mean Squared Error (MSE) tests across different models with varying sample quantities. Table 2 compares the state prediction MSE of single LSTM, joint LSTM [37], the global PISI, the local PISI, and static under different sample sizes for systems with 20 nodes. The best result under each sample size is highlighted in bold. We run all the methods with T = 50 time steps, and predict the following 20 time steps. The LSTM baseline labeled “single” employs separate but weight-shared LSTMs for each object. The alternative “joint” approach concatenates all state vectors as input to a single LSTM trained for simultaneous multi-object state prediction. The global PISI corresponds to the two message passing GNN in Equation (10). The local PISI corresponds to the classic GNN in Equation (8). The static baseline uses u t + 1 = u t . In RK4PINN [10], a total of 6K samples are required for estimating three parameters in the Lorenz system. We set the minimum sampling size to 50 K, as the neural relational inference (NRI) [14] model reconstructs 10-node graphs with only graph information at this sample size. The results demonstrate that our method exhibits advantages when the sample size is limited. When the sample size is over 60 K, our global PISI reaches 0.01 level MSE.
Table 3 studies the MSE of state prediction with node size 20 and sample size 100 K when the prediction steps are increasing. For one-step prediction, the joint LSTM, local PISI, and the global PISI reaches a 1 × 10 4 level. Our global PISI shows advantages for prediction steps 1 and 20.
For the structure reconstruction accuracy, we compare our PISI with correlation LSTM, the correlation path method, and the NRI in Table 4 with node sizes of 10 and 20 and sample size increasing. The final accuracy is the average over all batches, and the accuracy for each batch is calculated as 1 N × N i , j = 1 N δ ( A ˜ i j , A i j ) with δ ( · , · ) the Kronecker delta function. The correlation LSTM method trains an LSTM network with shared parameters [38] to individually model each trajectory, computes correlations between the final hidden states, and generates an interaction matrix through thresholding. The correlation path method estimates the interaction graph by applying thresholding to the correlation matrix between trajectory feature vectors. Table 4 demonstrates a slight improvement in reconstruction accuracy as the sample size increases. Our global PISI demonstrates superior performance with 20 nodes, outperforming state-of-the-art methods by 0.5% with sample size 100 K.
Table 5 investigates the impact of outlier values and the dynamic evolution time step on the recognition results. As stated in Section 4.1, the system synchronization has significant drawbacks for network structure identification. We set the outlier values 0.5, 3, and 3.8 with an evolution time step of 50. When the outlier value is 0.5, the network achieves synchronization at time step 10. The considered four methods all fail to reconstruct the structure. When the outlier reaches 3.5, the system fails to achieve synchronization. Compared to baseline methods, our approach shows improved reconstruction accuracy. The evolution time scales linearly with the sampling size, where the scaling factor is the sampling interval. We set it to 10, 50, and 100 to maintain consistency with the state prediction experiment.

4.5. Ablation Studies

Finally, we set α 1 = 100 , α 2 = 1 , and α 3 = 100 to balance the loss terms for different tasks. In this subsection, we will first perform an ablation study on weight parameters α 1 , α 2 , and α 3 to verify the effect of each loss term in Equation (19). Among the three weight parameters, α 1 and α 3 capture the structure importance. α 2 captures the trajectory importance. Increasing the weight parameter α 2 accelerates the convergence of loss term of trajectory. Figure 5 shows the trajectory loss convergence with α 2 = 1 , α 1 = 100 , and α 3 = 100 .
When α 1 = 0 , PISI degenerates into a structure identification algorithm that does not rely on physical information but depends on the graph neural network and solver iteration. In Table 6, we investigate the impact of physical information on structural identification for networks with N { 10 , 20 } . It can be observed that when α 1 = 0 , the structural identification accuracy decreases. This is due to insufficient utilization of physical information from the adjacency matrix. When α 1 is too large (relative to α 2 ), the structural identification accuracy also declines, as excessive physical information loss leads to slower convergence in fitting node states.

5. Conclusions

In this study, we propose a physical information-based solver-interactive network structure reconstruction method. The PISI effectively predicts the system trajectory. Furthermore, it is able to output the system structure adjacent matrix. The core concept of PISI lies in employing neural networks to approximate the unknown dynamic functions during the solution process of dynamic systems, thereby integrating physical information, system structure, and trajectory data. Specifically, PISI comprises two main modules: the structure reconstruction PIGNNs and the dynamic solver RK4. For the input observation, the RK4 modular evolves the state, thus providing the trajectory loss for PIGNN. At the end of each RK4 iteration, we obtain a trajectory loss. PIGNN provides the estimated structure for each RK4 stage. We employ four parameter-shared PIGNNs to approximate the unknown dynamic function in RK4. The four PIGNNs are trained end-to-end in a RK4 iteration. Each PIGNN internally consists of a global GNN. The global GNN consists of node-to-edge and edge-to-node message passing operations to enable more efficient exploitation of global structural information. Each information channel is processed by an MLP. Finally, the PISI loss function comprises three components: trajectory loss from RK4 output, physics-informed loss derived from physical constraints at each RK4 stage, and estimated structural loss for stage-wise topology reconstruction. We evaluate our method in two Kuramoto systems. Compared to the baseline, the global PISI shows advantage in the trajectory prediction and the structure reconstruction. Ablation studies demonstrate that the balance among the three loss terms significantly impacts recognition performance.

Author Contributions

Conceptualization, J.L. and G.M.; methodology, J.L. and G.M.; software, J.L. and G.M.; validation, J.L. and G.M.; formal analysis, J.L. and G.M.; investigation, J.L. and G.M.; resources, J.L. and G.M.; data curation, J.L. and G.M.; writing—original draft preparation, J.L.; writing—review and editing, J.L. and G.M.; visualization, J.L. and G.M.; supervision, J.L. and G.M.; project administration, J.L. and G.M.; funding acquisition, J.L. and G.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Fundamental Research Funds for the Central Universities under Grants BLX202338. This work was supported by National Natural Science Foundation of China under Grants 62303052. This work was supported by the PNRR project FAIR-Future AI Research (PE00000013), under the NRRP MUR program funded by the NextGenerationEU.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to privacy.

Acknowledgments

We thank the editor and the anonymous reviewers for their valuable suggestions to improve the quality of this study.

Conflicts of Interest

The authors declare that they have no known conflicting financial interests or personal relationships that could have appeared to influence the work reported in this study.

Appendix A

To enable backpropagation through sampling from our discrete latent variables, we sample from the continuous approximation of p ϕ ( A i j | u ) . Specifically, we employ the Concrete distribution [39] with the sampling procedure defined as follows:
A ˜ i j = softmax ( ( h i j 2 + g ) / τ )
in which g is an i.i.d. sample from the Gubel(0,1) distribution, and τ is a a parameter that controls the “smoothness” of the samples.

References

  1. Hu, J.; Jeff, G. Emergent phases of ecological diversity and dynamics mapped in microcosms. Science 2022, 378, 85–89. [Google Scholar] [CrossRef]
  2. Ehsan, A.; Milad, S. A Novel Modified Fuzzy-predictive Control of Permanent Magnet Synchronous Generator Based Wind Energy Conversion System. Chin. J. Electr. Eng. 2023, 9, 107–121. [Google Scholar] [CrossRef]
  3. Nurisso, M.; Raviola, M.; Tosin, A. Network-based kinetic models: Emergence of a statistical description of the graph topology. Eur. J. Appl. Math. 2024. [Google Scholar] [CrossRef]
  4. Zhang, X.; Barzel, B. Epidemic spreading under mutually independent intra- and inter-host pathogen evolution. Nat. Commun. 2022, 13, 6218. [Google Scholar] [CrossRef]
  5. Battiston, F.; Amico, E. Barrat, A. The Physics of Higher-Order Interactions in Complex Systems. Nat. Phys. 2021, 17, 1093–1098. [Google Scholar] [CrossRef]
  6. Jiang, S.; Pei, H.; Chen, J.; Li, H.; Liu, Z.; Wang, Y.; Gong, J.; Wang, S.; Li, Q.; Duan, M.; et al. Striatum- and Cerebellum-Modulated Epileptic Networks Varying Across States with and without Interictal Epileptic Discharges. Int. J. Neural Syst. 2024, 34, 2450017. [Google Scholar] [CrossRef]
  7. Kuchler, T.; Russel, D.; Stroebel, J. The Geographic Spread of COVID-19 Correlates with the Structure of Social Networks as Measured by Facebook. Available online: https://www.ifo.de/en/cesifo/publications/2020/working-paper/geographic-spread-covid-19-correlates-structure-social-networks (accessed on 6 August 2025).
  8. Zhang, L.; Xu, Z.; Wang, S.; He, G. Clustering Dimensionless Learning for Multiple-Physical-Regime Systems. Comput. Methods Appl. Mech. Eng. 2024, 415, 116728. [Google Scholar] [CrossRef]
  9. Chen, C.; Ma, H.; Aihara, K. Predicting future dynamics from short-term time series by anticipated learning machine. Natl. Sci. Rev. 2020, 7, 1070–1091. [Google Scholar] [CrossRef]
  10. Zhai, W.; Tao, D.; Bao, W. Parameter estimation and modeling of nonlinear dynamical systems based on Runge–-Kutta physics-informed neural network. Nonlinear Dyn. 2023, 111, 21117–21130. [Google Scholar] [CrossRef]
  11. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  12. Scarselli, F.; Gori, M.; Tsoi, A. The Graph Neural Network Model. IEEE Trans. Neural Netw. 2009, 20, 61–80. [Google Scholar] [CrossRef]
  13. Zhou, J.; Cui, G.; Hu, S. Graph Neural Networks for Network Reconstruction: From Dynamics to Topology. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 3124–3140. [Google Scholar]
  14. Kipf, T.; Fetaya, E.; Wang, K.; Welling, M.; Zemel, M. Neural Relational Inference for Interacting Systems. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Volume 80, pp. 2688–2697. [Google Scholar]
  15. Zang, C.; Wang, F. Neural Dynamics on Complex Networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (KDD ’20), Virtual Event, 6–10 July 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 892–902. [Google Scholar]
  16. Trivedi, D.; Farajtabar, M.; Biswal, P.; Zha, H. Learning Representations Over Dynamic Graphs. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  17. Zhang, Z.; Wang, X.; Zhou, H.; Yu, Y.; Zhang, M.; Yang, C.; Shi, C. Can Large Language Models Improve the Adversarial Robustness of Graph Neural Networks? In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD ’25), Toronto, ON, Canada, 3–7 August 2025; ACM: New York, NY, USA, 2025.
  18. Khoo, Y.; Lu, J.; Ying, L. Solving Parametric PDE Problems with Artificial Neural Networks. J. Comput. Phys. 2021, 430, 110092. [Google Scholar] [CrossRef]
  19. Ji, W.; Qiu, W.; Shi, Z.; Pan, S.; Deng, S. Stiff-PINN: Physics-Informed Neural Network for Stiff Chemical Kinetics. J. Phys. Chem. A 2021, 125, 8098–8106. [Google Scholar] [CrossRef]
  20. Morrison, M.; Kutz, J.N. Solving Nonlinear Ordinary Differential Equations Using the Invariant Manifolds and Koopman Eigenfunctions. SIAM J. Appl. Dyn. Syst. 2024, 23, 924–960. [Google Scholar] [CrossRef]
  21. Um, K.; Brand, R.; Fei, Y.; Holl, P.; Thuerey, N. Solver-in-the-Loop: Learning from Differentiable Physics to Interact with Iterative PDE-Solvers. Adv. Neural Inf. Process. Syst. 2020, 33, 6111–6122. [Google Scholar]
  22. Holl, P.; Koltun, V.; Thuerey, N.; Um, K. Learning to Control PDEs with Differentiable Physics. In Proceedings of the 38th International Conference on Machine Learning (ICML), Virtual, 18–24 July 2021; Volume 139, pp. 1234–1245. [Google Scholar]
  23. Skardal, P.; Arenas, A. Higher-order interactions in complex networks of phase oscillators promote abrupt synchronization switching. Commun. Phys. 2020, 218, 3. [Google Scholar] [CrossRef]
  24. Wang, Y.; Han, R.; Liu, X.; Horiguchi, T. Phase transitions of XY model in ultra-thin magnetic film with a triangular latt ice. Acta Phys. Sin. 2023, 52, 1776–1782. [Google Scholar] [CrossRef]
  25. Pirazzini, G.; Ursino, M. Modeling the contribution of theta-gamma coupling for sequential memory Pirazzini Ursino. Front. Neural Circuits 2024, 18, 1326609. [Google Scholar] [CrossRef]
  26. Lin, A.; Liu, K.; Bartsch, R. Dynamic network interactions among distinct brain rhythms as a hallmark of physiologic state and function. Commun. Biol. 2020, 197, 3. [Google Scholar]
  27. Gori, M.; Monfardini, G.; Scarselli, F. A new model for learning in graph domains. In Proceedings of the 2005 IEEE International Joint Conference on Neural Networks, 2005, Montreal, QC, Canada, 31 July–4 August 2005; Volume 2, pp. 729–734. [Google Scholar]
  28. Wu, Z.; Pan, S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef]
  29. Chen, G.; Wang, G.; Wang, K.; Hu, M.; Deng, J. KCNN: A Neural Network Lightweight Method and Hardware Implementation Architecture. J. Comput. Res. Dev. 2025, 62, 532–541. [Google Scholar]
  30. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  31. Bergen, A.; Hill, D. A Structure Preserving Model for Power System Stability Analysis. IEEE Trans. Power Appar. Syst. 1981, PAS-100, 25–35. [Google Scholar] [CrossRef]
  32. Nitzbon, J.; Schultz, P.; Heitzig, J.; Kurths, J.; Hellmann, F. Deciphering the imprint of topology on nonlinear dynamical network stability. New J. Phys. 2017, 19, 033029. [Google Scholar] [CrossRef]
  33. Shanmugasundaram, S.; Kashkynbayev, A.; Udhayakumar, K.; Rakkiyappan, R. Centralized and decentralized controller design for synchronization of coupled delayed inertial neural networks via reduced and non-reduced orders. Neurocomputing 2022, 469, 91–104. [Google Scholar] [CrossRef]
  34. Udhayakumar, K.; Fathalla, A.R.; Li, R.R. Quasi-bipartite synchronisation of multiple inertial signed delayed neural networks under distributed event-triggered impulsive control strategy. IET Control Theory Appl. 2021, 15, 1615–1627. [Google Scholar]
  35. Watts, D.J.; Strogatz, S.H. Collective dynamics of ’small-world’ networks. Nature 1998, 393, 440–442. [Google Scholar] [CrossRef]
  36. Schultz, P.; Heitzig, J.; Kurths, J. A random growth model for power grids and other spatially embedded infrastructure networks. Eur. Phys. J. Spec. Top. 2014, 223, 2593–2610. [Google Scholar] [CrossRef]
  37. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  38. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social LSTM: Human Trajectory Prediction in Crowded Spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
  39. Jang, E.; Gu, S.; Poole, B. Categorical Reparameterization with Gumbel-Softmax. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
Figure 1. Node-to-edge, edge-to-node, and node-to-edge operations for moving between node and edge representations in Equations (10)–(13). The black line denotes the MLP network, and the black arrow denotes the nodes or edges embedding aggregation.
Figure 1. Node-to-edge, edge-to-node, and node-to-edge operations for moving between node and edge representations in Equations (10)–(13). The black line denotes the MLP network, and the black arrow denotes the nodes or edges embedding aggregation.
Algorithms 18 00584 g001
Figure 2. Schematic of the overall framework. Our framework comprises two main modules: the structure reconstruction PIGNNs and the dynamic solver RK4. For the input observation u n , the RK4 module evolves the state u ˜ n + 1 . PIGNN provides the estimated structure A ˜ for each RK4 stage. PIGNN internally consists of node-to-edge and edge-to-node message passing operations. The PIGNN loss function comprises three components: trajectory loss from RK4 output, physics-informed loss derived from automatic differentiation, and estimated structural loss for stage-wise topology reconstruction.
Figure 2. Schematic of the overall framework. Our framework comprises two main modules: the structure reconstruction PIGNNs and the dynamic solver RK4. For the input observation u n , the RK4 module evolves the state u ˜ n + 1 . PIGNN provides the estimated structure A ˜ for each RK4 stage. PIGNN internally consists of node-to-edge and edge-to-node message passing operations. The PIGNN loss function comprises three components: trajectory loss from RK4 output, physics-informed loss derived from automatic differentiation, and estimated structural loss for stage-wise topology reconstruction.
Algorithms 18 00584 g002
Figure 3. The phase dynamics and the order parameter for the Kuramoto model. Subfigures (a,b) present the phase dynamics and order parameters for outlier values of 3.5 and 4, respectively.
Figure 3. The phase dynamics and the order parameter for the Kuramoto model. Subfigures (a,b) present the phase dynamics and order parameters for outlier values of 3.5 and 4, respectively.
Algorithms 18 00584 g003
Figure 4. Two network structures, with each blue node representing a generator nodes in the Kuramoto model. Subfigures (a,b) present the network structures with 10 and 20 nodes respectively.
Figure 4. Two network structures, with each blue node representing a generator nodes in the Kuramoto model. Subfigures (a,b) present the network structures with 10 and 20 nodes respectively.
Algorithms 18 00584 g004
Figure 5. The trajectory training and validation loss over 300 epochs for a 20 nodes Kuramoto system with α 2 = 1 , α 1 = 100 , and α 3 = 100 .
Figure 5. The trajectory training and validation loss over 300 epochs for a 20 nodes Kuramoto system with α 2 = 1 , α 1 = 100 , and α 3 = 100 .
Algorithms 18 00584 g005
Table 1. The small world parameters of synthetic networks.
Table 1. The small world parameters of synthetic networks.
Node NumberNeighbors NumberRewire Rate
1040.3
2060.4
Table 2. The state prediction MSE for 20 prediction steps with node size 20.
Table 2. The state prediction MSE for 20 prediction steps with node size 20.
Sample SizeStaticLSTM (Single)LSTM (Joint)PISI (Local)PISI (Global)
Node Size 20
50 K2.261.352.24 × 10 1 7.34 × 10 1 1.56 × 10 1
60 K1.878.94 × 10 1 5.11 × 10 1 6.25 × 10 2 4.31 × 10 2
70 K1.131.57 × 10 1 2.49 × 10 2 7.13 × 10 2 2.57 × 10 2
90 K9.88 × 10 1 9.86 × 10 2 2.46 × 10 2 4.83 × 10 2 2.41 × 10 2
100 K4.01 × 10 1 9.20 × 10 2 2.21 × 10 2 4.31 × 10 2 2.29 × 10 2
Table 3. The state prediction MSE for 100 K sample size with 20 nodes.
Table 3. The state prediction MSE for 100 K sample size with 20 nodes.
Prediction StepsStaticLSTM (Single)LSTM (Joint)PISI (Local)PISI (Global)
Node Size 20
15.75 × 10 3 1.41 × 10 3 7.23 × 10 4 9.96 × 10 4 4.80 × 10 4
109.33 × 10 1 9.27 × 10 2 4.41 × 10 3 6.90 × 10 2 8.71 × 10 3
204.25 × 10 1 8.56 × 10 2 2.58 × 10 2 5.59 × 10 2 2.26 × 10 2
Table 4. The structure reconstruction accuracy (in %).
Table 4. The structure reconstruction accuracy (in %).
Correlation LSTMCorrelation PathNRIPISI (Global)
Node Size1020102010201020
50 K52.450.657.454.096.195.796.396.2
60 K52.751.257.154.296.395.796.296.3
70 K52.551.557.954.996.395.896.496.4
80 K52.051.457.954.896.495.696.396.3
100 K53.151.857.854.896.495.896.596.3
Table 5. The structure reconstruction accuracy with three outlier values and evolution time steps for the 20 nodes network.
Table 5. The structure reconstruction accuracy with three outlier values and evolution time steps for the 20 nodes network.
Outlier ValueEvolution Time Step
Method0.533.51050100
Correlation LSTM32.432.862.729.462.762.5
Correlation Path31.631.062.328.662.363.0
NRI50.754.796.949.396.997.9
Ours50.553.997.850.797.697.7
Table 6. The structure reconstruction accuracy with increasing α 1 for node size 10 and 20.
Table 6. The structure reconstruction accuracy with increasing α 1 for node size 10 and 20.
α 1 Accuracy
Node Size 10Node Size 20
052.952.6
10096.196.3
50057.757.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Mei, G. PISI: Physical Information Based Solver-Interactive Network Structure Reconstruction. Algorithms 2025, 18, 584. https://doi.org/10.3390/a18090584

AMA Style

Liu J, Mei G. PISI: Physical Information Based Solver-Interactive Network Structure Reconstruction. Algorithms. 2025; 18(9):584. https://doi.org/10.3390/a18090584

Chicago/Turabian Style

Liu, Juan, and Guofeng Mei. 2025. "PISI: Physical Information Based Solver-Interactive Network Structure Reconstruction" Algorithms 18, no. 9: 584. https://doi.org/10.3390/a18090584

APA Style

Liu, J., & Mei, G. (2025). PISI: Physical Information Based Solver-Interactive Network Structure Reconstruction. Algorithms, 18(9), 584. https://doi.org/10.3390/a18090584

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop