Next Article in Journal
ABAQUS Subroutine-Based Implementation of a Fractional Consolidation Model for Saturated Soft Soils
Previous Article in Journal
Existence, Uniqueness, and Stability of Weighted Fuzzy Fractional Volterra–Fredholm Integro-Differential Equation
Previous Article in Special Issue
A Mixed Finite Volume Element Method for Nonlinear Time Fractional Fourth-Order Reaction–Diffusion Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Networks-Based Analytical Solver for Exact Solutions of Fractional Partial Differential Equations

1
School of Mathematics and Big Data, Dezhou University, Dezhou 253023, China
2
School of Mathematics and Statistics, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
3
School of Automation and Software Engineering, Shanxi University, Taiyuan 030013, China
4
Hubei Key Laboratory of Applied Mathematics, Hubei University, Wuhan 430062, China
*
Authors to whom correspondence should be addressed.
Fractal Fract. 2025, 9(8), 541; https://doi.org/10.3390/fractalfract9080541
Submission received: 16 July 2025 / Revised: 5 August 2025 / Accepted: 12 August 2025 / Published: 16 August 2025

Abstract

This paper introduces an innovative artificial neural networks-based analytical solver for fractional partial differential equations (fPDEs), combining neural networks (NNs) with symbolic computation. Leveraging the powerful function approximation ability of NNs and the exactness of symbolic methods, our approach achieves notable improvements in both computational speed and solution precision. The efficacy of the proposed method is validated through four numerical examples, with results visualized using three-dimensional surface plots, contour mappings, and density distributions. Numerical experiments demonstrate that the proposed framework successfully derives exact solutions for fPDEs without relying on data samples. This research provides a novel methodological framework for solving fPDEs, with broad applicability across scientific and engineering fields.

1. Introduction

In recent years, fractional operators have emerged as a powerful mathematical tool for modeling complex systems [1,2,3,4], offering superior capabilities over integer-order calculus in capturing memory effects, hereditary properties, and non-local dynamics with greater accuracy. By incorporating non-local memory effects and capturing anomalous dynamics [5,6], fractional-order models provide a more accurate and flexible framework for describing phenomena such as viscoelasticity, diffusion in heterogeneous media, and biological processes with long-range dependencies. Their ability to characterize multi-scale behaviors and fit real-world data more precisely makes them invaluable in physics, engineering, and biology. Nevertheless, solving fractional partial differential equations (fPDEs) presents significantly greater complexity. The development of effective numerical methods to approximate fPDEs has been the goal of some researchers. The traditional solving methods include finite-difference method [7,8], finite element method [9,10], spectral method [11,12] and virtual-element method [13,14], etc. However, these algorithms require discretization, which incurs significant time costs due to the vast amount of data processing and introduces approximate errors.
The rapid development of artificial intelligence has driven widespread applications of deep learning in scientific and technological domains [15,16,17,18]. This trend is primarily propelled by the strong function approximation capability [19,20,21] of neural networks (NNs), which has established them as powerful computational tools for solving differential and integral equations [22,23]. Particularly in computational mathematics, physics-informed neural networks (PINNs) have emerged as a significant methodology for solving partial differential equations (PDEs), attracting considerable scholarly attention [24,25,26,27]. PINNs are machine learning models that combine deep learning with physical knowledge. By serving as a surrogate model, NNs establish a mapping between spatio-temporal points and the corresponding PDE solutions. The residual information of PDEs is embedded into the loss function of NNs, thereby optimizing the training process through minimization with respect to the parameters of NNs, ultimately leading to the derivation of an optimal model. Due to the inapplicability of the standard chain rule in fractional calculus, automatic differentiation cannot be directly used with fractional operators. To solve this problem, fractional physics-informed neural networks (fPINNs) [28,29,30] discretize the fractional operators numerically. By employing automatic differentiation for integer-order operators and numerical approximation for fractional operators, the residual information from fPDEs can be integrated into the NN’s loss function during training. However, fPINNs require numerical discretization to solve fPDEs, which inherently introduces approximation errors. The fPINNs are data-driven models [31,32,33] that require a large number of training points for training and have high time costs. During network training, the NNs often become stuck in the local optimal value and cannot find the global optimal value. Consequently, enhancing NN training efficiency and developing advanced optimization algorithms remain active research priorities.
Nevertheless, existing approaches universally employ NNs to compute numerical approximations for diverse classes of fPDEs. While numerical solutions of fPDEs have broad applicability in practical applications, analytical solutions still possess irreplaceable value in pursuing accuracy and theoretical depth. Generally, it becomes difficult to find exact analytical solutions of fPDEs. In prior work [34,35,36], we developed an NN framework that derives exact PDE solutions through equation-specific trial functions constructed via NN architectures. Currently, NNs have not yet been employed to obtain exact analytical solutions of fPDEs.
Motivated by these advances, a novel artificial neural networks-based analytical solver is newly introduced herein, addressing fPDEs through an unprecedented computational paradigm. This novel approach integrates NNs with a symbolic computation strategy, enabling the rapid acquisition of precise analytical solutions to fPDEs. The proposed method solely employs the feedforward computation of NNs to obtain exact solutions for fPDEs without involving any training mechanisms, thereby significantly improving computational efficiency and accuracy. Moreover, this analytical method exhibits high customizability and flexibility, allowing the construction of diverse trial functions by adjusting the parameter configuration of NNs, making it widely applicable to various types of fPDEs. The main focus of our contributions lies in the following aspects:
  • We introduce an innovative framework for deriving exact analytical solutions to fPDEs, ensuring mathematically rigorous results free from computational errors.
  • Potential analytical solutions of fPDEs are constructed via NN architectures. Transformed inputs undergo feedforward propagation to yield network outputs, which subsequently serve as trial functions in the fPDE solutions framework.
  • Our approach simplifies the fPDE into computationally feasible algebraic systems through trial function application. The synaptic weights and biases of NNs are then resolved through undetermined coefficient optimization.
  • Exact analytical solutions for fPDEs are obtained in a data-independent manner through this computational framework. NNs are employed to impose structural constraints on trial function formulation, enhancing mathematical tractability.
The paper adopts the following structure: Theoretical principles of the proposed solver are formalized within the Section 2. In Section 3, the accuracy and feasibility of the artificial neural networks-based analytical solver for fPDEs are verified through fractional wave equation [37], fractional telegraph equation [38], fractional Sharma–Tasso–Olever equation [39], and fractional biological population model [40]. In Section 4, the proposed method is discussed. Finally, the conclusions of this paper are given in Section 5.

2. Methodology

In this section, we introduce the idea of an artificial neural network-based analytical solver for finding solutions of fPDEs. Analytical solutions of fPDEs are obtained by embedding NNs as explicit functional representations. The novelty of the technique stems from its ability to impose explicit mathematical constraints on the solution space representation. Consider the following general fractional partial differential equation
P ( u , D t α u , D x β u , D t 2 α u , D x 2 β u , ) = 0 ,
where the function u ( x , t ) satisfies the governing equation, the fractional orders 0 < α 1 and 0 < β 1 , D t α u and D x β u are the conformable fractional derivatives of u ( x , t ) with respect to t and x, respectively.
We employ a precisely formulated explicit model during the forward propagation phase of the NNs to compute the solution u ( x , t ) for Equation (1). The neuron model is composed of input, output, and computational functions. The input can be analogized as dendrites of neurons, while the output can be analogized as axons of neurons, and the computing can be analogized as the nucleus. Figure 1 shows a neuron model with three inputs, one output, and two computational functions. The arrow lines in Figure 1 are called connections, and each connection has a weight on it. In this networks, the weighted sum of the three inputs c 1 , c 2 , and c 3 is first computed as s u m = w 1 c 1 + w 2 c 2 + w 3 c 3 . Subsequently, they go through a nonlinear transformation to obtain
z = f ( s u m ) = f ( w 1 c 1 + w 2 c 2 + w 3 c 3 + b ) ,
here, f denotes the activation function, enhancing the neuron’s ability to model nonlinear relationships, while b represents the bias term. Consequently, the solution to the equation can be explicitly represented by a single neuron’s output. Given the high expressive power of neural networks in approximating functions, solutions to fPDEs can be computed using NNs with limited neurons or hidden layers. In this work, we employ an NN-based model to approximate the solution u ( x , t ) of the fPDEs, as illustrated in Figure 2. Where X and T are inputs of the NNs, and
X = x β β , T = t α α ,
After applying the aforementioned independent variable transformation, all fractional derivatives in Equation (1) are converted into integer-order derivatives
P ( u , u T , u X , u T T , u X X , ) = 0 ,
thus simplifying the subsequent calculation. Consider an NN architecture comprising n hidden layers, with each layer containing m neurons. For the first hidden layer ( L 1 ), the output of its n-th neuron ( z 1 n ) can be expressed as
z 1 n = f ( s u m ) = f ( w x n X + w t n T ) ,
where f is the activation function, while w x n and w t n denote the weights linking the n-th neuron to inputs X and T, respectively. The output z n n of the n-th neuron in layer ( L n ) can be expressed as
z n n = f ( s u m ) = f ( w 1 n n 1 z ( n 1 ) 1 + + w m n n 1 z ( n 1 ) m ) ,
let z ( n 1 ) 1 denote the activation value of the first neuron in the ( n 1 ) -th hidden layer, and z ( n 1 ) m represent the output of the m-th neuron in the same layer. The connection weight from neuron z ( n 1 ) 1 to z n n is denoted as w 1 n n 1 , whereas w m n n 1 represents the weight between z ( n 1 ) m and z n n . The system’s final output is computed as the weighted linear combination of all activation outputs from the final layer’s neurons. Consequently, the potential closed-form solution u ( x , t ) of the fPDEs, derived via the NNs framework, takes the following form:
u ( x , t , w ) = w 1 u z n 1 + w 2 u z n 2 + + w m u z n m .
The trial function takes the expanded representation u ( x , t , w , b ) when neuronal outputs contain bias components.
In the realm of mathematical modeling for physical phenomena, utilizing the trial function method to derive analytical solutions for field partial differential equations related to these domains is a prevalent technique. Nevertheless, the cornerstone of this methodology lies in the artful construction of the trial function itself, which necessitates a meticulous and strategic construction process. This paper presents an analytical solver based on an NN architecture, employing its explicit mathematical formulation as the trial solution for fPDEs. The architecture of the NNs is designed to be regular, and the overall flow of the proposed method is shown in Figure 3.
The method uses the trial function method to construct potential analytical solutions of fPDEs. The potential analytical solutions to fPDEs may incorporate various elementary functions, including rational term ( 1 / ( · ) ), trigonometric components ( sin ( · ) , cos ( · ) ), and exponential term ( e ( · ) ), etc. This is directly related to the activation function of the selected NNs. Fortunately, the selection of the activation function can be facilitated by leveraging the initial or boundary conditions as a guiding principle, significantly expediting the process of designing the architecture of NNs. The NNs have a structured architecture with a standardized pattern, mainly requiring specification of hidden layer neuron quantities and individual neuron activation functions, as detailed in Table 1. Consequently, the proposed methodology significantly reduces the difficulties involved in formulating trial functions.
Suppose there is an architecture of NNs with one hidden layer and two neurons, and sin ( · ) is chosen as the activation function for each neuron. If X = x β / β and T = t α / α are taken as inputs to the NNs, the explicit model u ( X , T , w , b ) of the NNs can be expressed as
u ( X , T , w , b ) = w 1 u sin ( w x 1 X + w t 1 T + b 1 ) + w 2 u sin ( w x 2 X + w t 2 T + b 2 ) + b 3 ,
the connection weights between neurons are denoted by w, while b corresponds to the bias term in each neuron’s output. By substituting the Equation (8) into the Equation (4), a nonlinear equation about weights and biases is obtained:
Eqs ( X , T , w , b ) = 0 ,
by combining the similar terms of Equation (9), extracting the coefficients, and setting all coefficients to zero, we derive a system of nonlinear algebraic equations. Finally, the weight coefficients, biases, and Equation (3) obtained by the solution are substituted into the NN model Equation (8), and the analytical solution u ( x , t ) of the Equation (1) can be obtained.
To facilitate the practical application of our analytical methodology, we have implemented a dedicated Maple-based computational tool that automates all solution procedures.

3. Applications

In this section, we demonstrate the accuracy and feasibility of an artificial neural network-based analytical solver to solve fPDEs by fractional wave equation [37], fractional telegraph equation [38], fractional Sharma–Tasso–Olever equation [39], and fractional biological population model [40].

3.1. Fractional Wave Equation

The fractional wave equation under consideration takes the form
D t 2 α u ( x , t ) = k 2 D x 2 β u ( x , t ) ,
with the following initial condition
u ( x , 0 ) = 2 x 2 + x ,
where k is a positive coefficient, x R , t R + , 1 < 2 α 2 , 1 < 2 β 2 , the function u ( x , t ) satisfies the governing equation.
Via the variable transformation Equation (3), Equation (10) is rewritten as its integer-order counterpart:
u T T = k 2 u X X ,
Case I: To obtain analytical solutions for the fractional wave equation, the NNs architecture is utilized. Design considerations incorporate the initial condition of the equation. We implement a network with two hidden layers, each consisting of two neurons. The corresponding architectural schematic appears in Figure 4a. Using neural networks, the trial function is parameterized by weights and biases. Therefore, the mathematical representation of the explicit model can be expressed as follows:
N 1 = T w t 1 + X w x 1 + b 1 , N 2 = T w t 2 + X w x 2 + b 2 , N 3 = w 23 f 2 ( N 2 ) + w 13 f 1 ( N 1 ) + b 3 , N 4 = w 24 f 2 ( N 2 ) + w 14 f 1 ( N 1 ) + b 4 , u ( X , T ) = w 3 u f 3 ( N 3 ) + w 4 u f 4 ( N 4 ) + b 5 ,
where X = x β / β , T = t α / α , u ( X , T ) denote the exact solution of the Equation (10). In the networks architecture, the initial hidden layer generates outputs N 1 and N 2 from its first and second neurons, respectively, followed by outputs N 3 and N 4 from the subsequent layer’s corresponding neurons, w t 1 , w t 2 , w x 1 , w x 2 , w 13 , w 23 , w 14 , w 24 , w 3 u , w 4 u are the weights between the each neurons, f 1 , f 2 , f 3 , f 4 are the activation functions. The bias terms are denoted as b 1 , b 2 , b 3 , b 4 , and b 5 , corresponding to each neuron’s offset parameter. Within the NNs framework, the trial solution u ( x , t ) for Equation (13) is parameterized by both connection weights and these bias components.
The first hidden layer employs exponential functions e ( · ) for all neuronal activations, while in the second hidden layer, the third and fourth neurons utilize ( · ) and ( · ) 2 activation mappings, respectively. Consequently, the explicit model u ( X , T ) admits the following mathematical representation:
u ( X , T ) = w 14 e T w t 1 + X w x 1 + b 1 + w 24 e T w t 2 + X w x 2 + b 2 + b 4 2 w 4 u + w 13 e T w t 1 + X w x 1 + b 1 + w 23 e T w t 2 + X w x 2 + b 2 + b 3 w 3 u + b 5 .
By substituting the NNs-based Equation (14) into the Equation (12), we obtain
2 w 4 u w 14 w t 1 e T w t 1 + X w x 1 + b 1 + w 24 w t 2 e T w t 2 + X w x 2 + b 2 2 + 2 w 4 u w 14 e T w t , 1 + X w x 1 + b 1 + w 24 e T w t 2 + X w x 2 + b 2 + b 4 × w 14 w t 1 2 e T w t 1 + X w x 1 + b 1 + w 24 w t 2 2 e T w t 2 + X w x 2 + b 2 + w 3 , u w 13 w t 1 2 e T w t 1 + X w x 1 + b 1 + w 23 w t 2 2 e T w t 2 + X w x 2 + b 2 k 2 2 w 4 u w 14 w x 1 e T w t 1 + X w x 1 + b 1 + w 24 w x 2 e T w t 2 + X w x 2 + b 2 2 + 2 w 4 u w 14 e T w t 1 + X w x 1 + b 1 + w 24 e T w t 2 + X w x 2 + b 2 + b 4 × w 14 w x 1 2 e T w t 1 + X w x 1 + b 1 + w 24 w x 2 2 e T w t 2 + X w x 2 + b 2 + w 3 u w 13 w x 1 2 e T w t 1 + X w x 1 + b 1 + w 23 w x 2 2 e T w t 2 + X w x 2 + b 2 = 0 .
By collecting terms involving the basis set { X , T , e T w t 1 + X w x 1 + b 1 , e T w t 2 + X w x 2 + b 2 } and enforcing zero coefficients for each independent component, we derive the following equation system.
2 k 2 b 4 w 14 w 4 u w x 1 2 k 2 w 13 w 3 u w x 1 2 + 2 b 4 w 14 w 4 u w t 1 2 + w 13 w 3 u w t 1 2 = 0 , 2 k 2 b 4 w 24 w 4 u w x 2 2 k 2 w 23 w 3 u w x 2 2 + 2 b 4 w 24 w 4 u w t 2 2 + w 23 w 3 u w t 2 2 = 0 , 2 k 2 w 14 w 24 w 4 u w x 1 2 4 k 2 w 14 w 24 w 4 u w x 1 w x 2 2 k 2 w 14 w 24 w 4 u w x 2 2 + 2 w 14 w 24 w 4 u w t 1 2 + 4 w 14 w 24 w 4 u w t 1 w t 2 + 2 w 14 w 24 w 4 u w t 2 2 = 0 , 4 k 2 w 14 2 w 4 u w x 1 2 + 4 w 14 2 w 4 u w t 1 2 = 0 , 4 k 2 w 24 2 w 4 u w x 2 2 + 4 w 24 2 w 4 u w t 2 2 = 0 .
Through the systematic solving of the equation system, we derive 20 constraint conditions, where each condition corresponds to a general solution in the solution space. Due to limited space, a few representative examples are cited below:
Solution 1:
k = k , b 4 = b 4 , w 13 = w 13 , w 14 = 0 , w 23 = w 23 , w 24 = w 24 , w 3 u = 0 , w 4 u = w 4 u , w t 1 = w t 1 , w t 2 = k w x 2 , w x 1 = w x 1 , w x 2 = w x 2 .
The general solution for this system of constraints can be expressed as follows:
u ( X , T ) = w 24 e T k w x 2 + X w x 2 + b 2 + b 4 2 w 4 u + b 5 .
Substituting X = x β / β , T = t α / α into the Equation (18) yields the exact solution
u 1 ( x , t ) = w 24 e t α k w x 2 α + x β w x 2 β + b 2 + b 4 2 w 4 u + b 5 .
Solution 2:
k = k , b 4 = b 4 , w 13 = 0 , w 14 = 0 , w 23 = w 23 , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = w t 1 , w t 2 = k w x 2 , w x 1 = w x 1 , w x 2 = w x 2 ,
and for the given constraints, the general solution u ( x , t ) takes the following form:
u 2 ( x , t ) = w 24 e t α k w x 2 α + x β w x 2 β + b 2 + b 4 2 w 4 u + w 23 e t α k w x 2 α + x β w x 2 β + b 2 + b 3 w 3 u + b 5 .
Solution 3:
A fully connected NN structure can better approximate the analytical solution of an equation. However, the weight coefficients in the constraint conditions corresponding to the three examples introduced above are not all non-zero, which means the corresponding NNs are not fully connected network structures. For example, the network structure corresponding to the set of constraints Equation (17) of the equation only utilizes the second and fourth neurons. Due to limited space, we provide the constraint conditions corresponding to a fully connected NN structure as follows
k = k , b 4 = b 4 , w 13 = w 13 , w 14 = w 14 , w 23 = w 23 , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = k w x 1 , w t 2 = k w x 2 , w x 2 = w x 2 ,
and the general solution for this system of constraints can be expressed as follows:
u 3 ( x , t ) = w 14 e t α k w x 1 α + x β w x 1 β + b 1 + w 24 e t α k w x 2 α + x β w x 2 β + b 2 + b 4 2 w 4 u + w 13 e t α k w x 1 α + x β w x 1 β + b 1 + w 23 e t α k w x 2 α + x β w x 2 β + b 2 + b 3 w 3 u + b 5 .
In order to further simplify the general solution of the Equation (23), we substitute the initial condition into Equation (23) to obtain
u ( x , t ) = w 14 e T k w x 1 + X w x 1 + b 1 + w 24 e T k w x 2 + X w x 2 + b 2 + b 4 2 w 4 u + w 13 e T k w x 1 + X w x 1 + b 1 + w 23 e T k w x 2 + X w x 2 + b 2 + b 3 w 3 u e 2 ( X w x 1 + b 1 ) w 14 2 + 2 e X w x 1 + b 1 e X w x 2 + b 2 w 14 w 24 + e 2 ( X w x 2 + b 2 ) w 24 2 w 4 u 2 e X w x 1 + b 1 b 4 w 14 + e X w x 2 + b 2 b 4 w 24 w 4 u b 4 2 w 4 u + 2 X 2 b 3 w 3 u + X e X w x 1 + b 1 w 13 + e X w x 2 + b 2 w 23 w 3 u ,
where X = x β / β and T = t α / α .
If one chooses coefficients { k = 1 , b 1 = 0.1 , b 2 = 0.1 , b 3 = 0.1 , b 4 = 0.1 , w 13 = 0.5 , w 14 = 1 , w 23 = 1 , w 24 = 1 , w 3 u = 1 , w 4 u = 0.1 , w x 1 = 0.1 , w x 2 = 0.5 , α = 0.6 , β = 0.6 } in the Equation (24), the results are shown in Figure 5. In this way, the solution’s performance becomes more apparent. Figure 5a presents a 3D plot illustrating the solution over the spatio-temporal domain [ 0 , 5 ] × [ 0 , 5 ] . To observe the local behavior of the solution, we plot the x-curves in Figure 5b at interval t [ 0 , 5 ] . In Figure 5c,d, we illustrate the contour and density representations of the fractional wave equation. As can be seen from the figure, the initial condition takes on a parabolic shape. Over time, the shape and size of the solution evolve, forming distinct peaks and troughs while exhibiting characteristics of propagation and diffusion. The variation in color intuitively reflects the dynamic changes in the solution, providing a crucial visualization tool for understanding the solutions to the fractional wave equation.
Case II: Therefore, the artificial neural networks-based analytical solver can effectively obtain the exact solutions of the Equation (10). Further demonstrating the adjustable nature of our approach, we employed a different NN parameter configuration for Equation (10), visible in Figure 4b. In this case, we take the activation functions of neurons in the second hidden layer as the sin ( · ) and the cos ( · ) , respectively, while other settings in the NN structure remain unchanged. According to the analytical formulation Equation (13), the following trial function can be constructed:
u ( X , T ) = w 14 sin ( T w t 1 + X w x 1 + b 1 ) + w 24 cos ( T w t 2 + X w x 2 + b 2 ) + b 4 2 w 4 u + w 13 sin ( T w t 1 + X w x 1 + b 1 ) + w 23 cos ( T w t 2 + X w x 2 + b 2 ) + b 3 w 3 u + b 5 ,
Substituting Equation (25) into Equation (12) and combining similar terms involving { X , T , cos ( T w t 1 + X w x 1 + b 1 ) , cos ( T w t 2 + X w x 2 + b 2 ) , sin ( T w t 1 + X w x 1 + b 1 ) , sin ( T w t 2 + X w x 2 + b 2 ) } . Setting every coefficient equal to zero generates 20 distinct constraint groups, each associated with a generalized solution of the equation that includes weights and biases. We give an example of the constraint conditions corresponding to a fully connected NN structure (where the weights are all non-zero), as shown below:
k = k , b 4 = b 4 , w 13 = w 13 , w 14 = w 14 , w 23 = w 23 , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = k w x 1 , w t 2 = k w x 2 , w x 1 = w x 1 , w x 2 = w x 2 ,
and the general solution u ( x , t ) takes the form
u ( x , t ) = w 14 sin T k + X w x 1 + b 1 + w 24 cos T k + X w x 2 + b 2 + b 4 2 w 4 u + w 13 sin T k + X w x 1 + b 1 + w 23 cos T k + X w x 2 + b 2 + b 3 w 3 u + b 5 ,
where X = x β / β and T = t α / α .
Substituting the Equation (11) into the Equation (27), we obtain
u ( x , t ) = w 14 sin T k + X w x 1 + b 1 + w 24 cos T k + X w x 2 + b 2 + b 4 2 w 4 u + w 13 sin T k + X w x 1 + b 1 + w 23 cos T k + X w x 2 + b 2 + b 3 w 3 u sin 2 X w x 1 + b 1 w 14 2 w 4 u 2 sin X w x 1 + b 1 cos X w x 2 + b 2 w 14 × w 24 w 4 u cos 2 X w x 2 + b 2 w 24 2 w 4 u 2 sin X w x 1 + b 1 b 4 w 14 w 4 u 2 cos X w x 2 + b 2 b 4 w 24 w 4 u sin X w x 1 + b 1 w 13 w 3 u cos X w x 2 + b 2 w 23 w 3 u b 4 2 w 4 u + 2 X 2 b 3 w 3 u + X ,
where X = x β / β and T = t α / α .
With the coefficients configuration { k = 2 , b 1 = 2 , b 2 = 2 , b 3 = 2 , b 4 = 2 , w 13 = 1 , w 14 = 1 , w 23 = 1 , w 24 = 1 , w 3 u = 1 , w 4 u = 1 , w x 1 = 2 , w x 2 = 2 , α = 0.6 , β = 0.6 } specified in Equation (28), the 3D plot, x-curves plot, contour lines, and density patterns of the Equation (10) are collectively illustrated in Figure 6. The Figure 6 exhibits a distinct wave-like pattern, indicating that the solution possesses wave characteristics in both space and time. The amplitudes of the wave peaks and troughs gradually change over time, revealing the dynamic evolution of the wave properties. In the spatial direction, the waveform displays a certain periodicity, suggesting that the solution exhibits periodic wave characteristics in space. In the Figure 6b, it is also evident that the phase of the curves shifts at different time points, indicating the presence of phase variation as the solution evolves dynamically over time.

3.2. Fractional Telegraph Equation

The following fractional telegraph equation is investigated:
D t α u ( x , t ) + D t 2 α u ( x , t ) + u ( x , t ) = D x β u ( x , t ) ,
with the following initial condition
u ( x , 0 ) = e x ,
where 0 < α 1 , 0 < β 1 , x R , t R + , the function u ( x , t ) satisfies the equation.
Through the independent variable transformation Equation (3), Equation (29) is rewritten as
u T + u T T + u ( x , t ) = u X X ,
Case I: Following the same approach, the NNs framework is employed to derive exact solutions for the fractional telegraph equation, where the initial conditions serve as the basis for designing the network structure. NN architecture 3, as shown in Figure 7a, is used to solve the equation. The network configuration consists of two hidden layers, both containing two neuronal units. Neurons in the initial hidden layer utilize exponential activation e ( · ) , whereas the subsequent layer adopts reciprocal activation 1 / ( · ) . According to the analytical formulation Equation (13), we can determine the trial function
u ( X , T ) = w 3 u w 13 e T w t 1 + X w x 1 + b 1 + w 23 e T w t 2 + X w x 2 + b 2 + b 3 + w 4 u w 14 e T w t 1 + X w x 1 + b 1 + w 24 e T w t 2 + X w x 2 + b 2 + b 4 + b 5 ,
where X = x β / β and T = t α / α .
Substituting Equation (32) into Equation (31) and grouping similar terms involving { X , T , e T w t 1 + X w x 1 + b 1 , e T w t 2 + X w x 2 + b 2 } . Setting every coefficient equal to zero generates 18 distinct constraint groups, each associated with a generalized solution of the equation that includes weights and biases. We give an example of the constraint conditions corresponding to a fully connected NN structure, as shown below:
b 3 = 0 , b 4 = 0 , b 5 = 0 , w 13 = w 13 , w 14 = w 14 , w 23 = w 23 , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = w t 2 , w t 2 = w t 2 , w x 1 = w t 2 2 + w t 2 1 , w x 2 = w t 2 2 + w t 2 1 ,
and the analytical solution u ( x , t ) takes the form
u ( x , t ) = w 3 u w 13 e T w t 2 + X w t 2 2 + w t 2 1 + b 1 + w 23 e T w t 2 + X w t 2 2 + w t 2 1 + b 2 + w 4 u w 14 e T w t 2 + X w t 2 2 + w t 2 1 + b 1 + w 24 e T w t 2 + X w t 2 2 + w t 2 1 + b 2 ,
where X = x β / β and T = t α / α , and by substituting the Equation (30) into the above equation, we obtain the analytic solution in the following form
u ( x , t ) = w 14 2 w 1 , 3 e A + 3 b 1 b 2 + 2 w 24 w 13 w 14 e A + 2 b 1 + w 23 w 14 2 e A + 2 b 1 + w 24 2 w 13 e A + b 1 + b 2 + 2 w 24 w 14 w 23 e A + b 1 + b 2 + w 24 2 w 23 e A + 2 b 2 e b 1 b 2 w 14 + w 24 w 13 e B + b 1 + w 23 e B + b 2 w 14 e B + b 1 + w 24 e B + b 2 ,
where A = 2 X w t 2 2 + ( T + 2 X ) w t 2 3 X , B = X w t 2 2 + ( T + X ) w t 2 X , X = x β / β , and T = t α / α .
Case II: Furthermore, the NNs architecture 4 as shown in Figure 7b is used to solve the fractional telegraph equation. The network structure includes three hidden layers, and the activation functions of neurons in the first hidden layer, second hidden layer, and the third hidden layer are set to identity mapping ( · ) , e ( · ) , and reciprocal 1 / ( · ) , respectively. Following the architecture of the established NNs, we formulate the novel trial function as
u ( X , T ) = w 5 u w 13 e A 1 w 13 + A 2 w 23 + b 3 + w 23 e A 1 w 14 + A 2 w 2 , 4 + b 4 + b 5 + w 6 u w 14 e A 1 w 13 + A 2 w 23 + b 3 + w 24 e A 1 w 14 + A 2 w 24 + b 4 + b 6 + b 7 ,
where A 1 = T w t 1 + X w x 1 + b 1 , A 2 = T w t 2 + X w x 2 + b 2 , X = x β Γ ( 1 + β ) and T = t α Γ ( 1 + α ) .
Compared with previously discussed trial functions, this formulation exhibits significantly greater complexity. Substituting Equation (36) into Equation (31) and grouping similar terms involving { X , T , e A 1 w 13 + A 2 w 23 + b 3 , e A 1 w 14 + A 2 w 24 + b 4 } . When all coefficients are set to zero, the system yields 25 constraint groups, mapping to 18 weighted solutions of the equation with bias terms. We give an example of the constraint conditions corresponding to a fully connected NN structure, as shown below:
b 5 = 0 , b 6 = 0 , b 7 = 0 , w 13 = w 14 , w 14 = w 14 , w 23 = w 24 , w 24 = w 24 , w 5 u = w 5 u , w x 1 = w 14 2 w t 1 2 + 2 w 14 w 24 w t 1 w t 2 + w 24 2 w t 2 2 w 14 w t 1 w 24 w t 2 + w 24 w x 2 + 1 / w 14 , w 6 u = w 6 u , w t 1 = w t 1 , w t 2 = w t 2 , w x 2 = w x 2 .
Consequently, the exact solution satisfying these constraint conditions can be expressed as
u ( x , t ) = w 5 u + w 6 u w 14 e A 3 + A 4 + b 3 + w 24 e A 3 + A 4 + b 4 ,
where A 3 = X w 24 2 w t 2 2 + ( ( 2 X w 14 w t 1 + T + X ) w t 2 + b 2 ) w 24 , A 4 = X w 14 2 w t 1 2 + ( ( T + X ) w t 1 + b 1 ) w 14 X , X = x β / β , and T = t α / α .
By substituting the initial condition into Equation (38), we obtain the solution in the following form:
u ( x , t ) = w 14 e A 5 + b 3 + w 24 e A 5 + b 4 w 14 e A 6 + b 3 + w 24 e A 6 + b 4 ,
where A 5 = ( 2 w 14 2 w t 1 2 + w t 1 ( 2 w 24 w t 2 + 1 ) w 14 w 24 2 w t 2 2 + w 24 w t 2 ) X + b 1 w 14 + b 2 w 24 , A 6 = X w 24 2 w t 2 2 + ( ( 2 X w 14 w t 1 + T + X ) w t 2 + b 2 ) w 24 X w 14 2 w t 1 2 + ( ( T + X ) w t 1 + b 1 ) w 14 X , X = x β / β , and T = t α / α .
With the coefficients configuration { b 1 = 1 , b 2 = 1 , b 3 = 1 , b 4 = 1 , w 14 = 1 , w 24 = 1 , w t 1 = 0.5 , w t 2 = 0.5 , α = 0.4 , β = 0.4 } specified in Equation (39), the various figures of the Equation (29) are collectively illustrated in Figure 8. The exact solution of the fractional telegraph equation exhibits a smooth and monotonically decreasing behavior as the absolute values of x and t increase. The 3D plot, x-curves, contour plot, and density plot all consistently show this trend, with the solution starting higher at x = 0 and t = 0 and gradually declining as x and t change. The visualizations provide a clear understanding of the spatio-temporal dynamics of the solution.

3.3. Fractional Sharma–Tasso–Olever Equation

Let us examine the fractional-order Sharma–Tasso–Olever equation expressed as
D t α u + 3 a ( D x β u ) 2 + 3 a u 2 D x β u + 3 a u D x 2 β u + a D x 3 β u = 0 ,
where a is an arbitrary constant, 0 < α 1 , 0 < β 1 .
Through the variable transformation Equation (3), Equation (40) is rewritten as
u T + 3 a u X 2 + 3 a u 2 u X + 3 a u u X X + a u X X X = 0 ,
Case I: The NNs architecture depicted in Figure 9a, featuring two hidden layers with two neurons each, is employed to solve the governing equation. The first hidden layer employs distinct activation functions, with tanh ( · ) for its first neuron and coth ( · ) for the second, while all neurons in the second hidden layer utilize identity mapping ( · ) . According to the analytical formulation Equation (13), we can derive the trial function
u ( X , T ) = w 13 tanh T w t 1 + X w x 1 + b 1 + w 23 coth T w t 2 + X w x 2 + b 2 + b 3 w 3 u + w 14 tanh T w t 1 + X w x 1 + b 1 + w 24 coth T w t 2 + X w x 2 + b 2 + b 4 w 4 u + b 5 ,
where X = x β / β and T = t α / α .
Substituting Equation (42) into Equation (41) and grouping similar terms involving { X , T , tanh ( T w t 1 + X w x 1 + b 1 ) , coth ( T w t 2 + X w x 2 + b 2 ) } . When all coefficients are set to zero, the system yields 19 constraint groups, mapping to 18 weighted solutions of the equation with bias terms. We give an example of the constraint conditions corresponding to a fully connected NN structure, as shown below:
a = a , b 3 = b 3 , b 4 = b 4 , b 5 = b 3 w 3 u b 4 w 4 u , w 13 = w 13 , w 14 = w 14 , w 23 = w 23 , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = a ( w 13 w 3 u + w 14 w 4 u ) ( ( w 13 2 + 3 w 23 2 ) w 3 u 2 + 2 w 4 u ( w 13 w 14 + 3 w 23 w 24 ) w 3 u + w 4 u 2 ( w 14 2 + 3 w 24 2 ) ) , w t 2 = 3 a ( w 23 w 3 u + w 24 w 4 u ) × ( ( w 13 2 + w 23 2 / 3 ) w 3 u 2 + 2 ( w 13 w 14 + w 23 w 24 / 3 ) w 4 u w 3 u + w 4 u 2 ( w 14 2 + w 24 2 / 3 ) ) , w x 1 = w 13 w 3 u + w 14 w 4 u , w x 2 = w 23 w 3 u + w 24 w 4 u ,
and the exact solution satisfying this constraint condition can be expressed as
u ( x , t ) = tanh T w t 1 + X w x 1 + b 1 w x 1 + coth T w t 2 + X w x 2 + b 2 w x 2 ,
where w x 1 , w x 2 , w t 1 , and w t 2 are denoted by Equation (43), X = x β / β and T = t α / α .
If one chooses coefficients { a = 0.5 , b 1 = 3 , b 2 = 3 , w 13 = 2 , w 14 = 0.5 , w 23 = 0.5 , w 24 = 0.5 , w 3 u = 0.5 , w 4 u = 1 , α = 0.5 , β = 0.5 } in the Equation (44), the results are shown in Figure 10. In this way, the solution’s performance becomes more apparent. Figure 10a presents a 3D plot illustrating the solution over the spatio-temporal domain [ 0 , 100 ] × [ 0 , 20 ] . To observe the local behavior of the solution, we plot the x-curves in Figure 10b at interval t [ 0 , 20 ] . And the contour and density plots of the Equation (40) are shown in Figure 10c and Figure 10d, respectively. From the figure, it can be observed that the solution exhibits a series of fine oscillations, which emerge abruptly on a smooth plane, demonstrating distinct soliton characteristics. The oscillations are unevenly distributed across the plane, with larger amplitudes in certain regions and smaller amplitudes in others. The frequency and amplitude of the oscillations vary both spatially and temporally. The density plot clearly illustrates the vibrational patterns and peaks of the waveform.
Case II: Furthermore, the NNs architecture 6 as shown in Figure 9b is employed to analytically solve the Equation (40). We configure tan(·) activation for all first hidden layer neurons, while implementing identity mapping ( · ) across the second hidden layer. According to the corresponding analytical formulation Equation (13), we derive
u ( X , T ) = w 13 tan T w t 1 + X w x 1 + b 1 + w 23 tan T w t 2 + X w x 2 + b 2 + b 3 w 3 u + w 14 tan T w t 1 + X w x 1 + b 1 + w 24 tan T w t 2 + X w x 2 + b 2 + b 4 w 4 u + b 5 ,
where X = x β / β and T = t α / α .
Substituting Equation (45) into Equation (41) and grouping similar terms involving { X , T , tan ( T w t 1 + X w x 1 + b 1 ) , tan ( T w t 2 + X w x 2 + b 2 ) } . Setting every coefficient equal to zero generates 19 distinct constraint groups, each associated with a generalized solution of the equation that includes weights and biases. We give an example of the constraint conditions corresponding to a fully connected NN structure, as shown below:
a = a , b 3 = b 3 , b 4 = b 4 , b 5 = b 5 , w 13 = w 14 w 4 u / w 3 u , w 14 = w 14 , w 23 = w 23 , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = w t 1 , w x 2 = w 23 w 3 u w 24 w 4 u , w t 2 = 3 a ( w 23 w 3 u + w 24 w 4 u ) ( ( b 3 2 w 23 2 / 3 ) w 3 u 2 + ( ( 2 b 3 b 4 ( 2 w 23 w 24 ) / 3 ) w 4 u + 2 b 3 b 5 ) w 3 u + ( b 4 2 w 24 2 / 3 ) w 4 u 2 + 2 w 4 u b 4 b 5 + b 5 2 ) , w x 1 = w x 1 .
Consequently, the analytical solution associated with this constraint condition can be expressed as
u ( x , t ) = tan T w t 2 + X w x 2 + b 2 w x 2 + b 3 w 3 , u + b 4 w 4 , u + b 5 ,
where w x 2 and w t 2 are denoted by Equation (46), X = x β / β and T = t α / α .
If one chooses coefficients { a = 1 , b 2 = 1 , b 3 = 1 , b 4 = 1 , b 5 = 1 , w 14 = 1 , w 23 = 1 , w 24 = 1 , w 3 u = 1 , w 4 u = 1 , α = 0.5 , β = 0.5 } in the Equation (47), the results are shown in Figure 11. In this way, the solution’s performance becomes more apparent. The surface of the image exhibits significant sharp peaks and depressions in certain regions, indicating abrupt changes in the solution within these areas. The distribution of sharp peaks and depressions reveals the nonlinear behavior of the solution in both space and time. The distribution of colors in the density plot exhibits a complex wave-like pattern. This wave-like structure indicates that the solution undergoes periodic or quasi-periodic variations in both space and time, while the changes in color intensity reveal the amplitude and frequency characteristics of the solution.

3.4. Fractional Biological Population Model

We investigate a fractional biological population model in (2 + 1) dimensions with the following formulation:
D t α u = D x 2 β ( u 2 ) + D y 2 γ ( u 2 ) + h ( u 2 r ) ,
where h and r are arbitrary constants, 0 < α 1 , 0 < β 1 , 0 < γ 1 .
Applying the independent variable transformation
X = x β Γ ( 1 + β ) , Y = y γ Γ ( 1 + γ ) , T = t α Γ ( 1 + α ) ,
we can rewrite Equation (48) as
u T ( u 2 ) X X ( u 2 ) Y Y h ( u 2 r ) = 0 .
Case I: The NNs architecture 7 as shown in Figure 12a is used to solve the equation. The network architecture consists of three inputs, two hidden layers with two neurons each. For the first hidden layer, we adopt the exponential activation function e ( · ) , whereas ( · ) is implemented throughout the second hidden layer. According to the corresponding analytical formulation, we derive
u ( X , Y , T ) = w 13 e T w t 1 + X w x 1 + Y w y 1 + b 1 + w 23 e T w t 2 + X w x 2 + Y w y 2 + b 2 + b 3 w 3 u + w 14 e T w t 1 + X w x 1 + Y w y 1 + b 1 + w 24 e T w t 2 + X w x 2 + Y w y 2 + b 2 + b 4 w 4 u + b 5 ,
where X = x β / β , Y = y γ / γ , and T = t α / α .
Substituting Equation (51) into Equation (50) and grouping similar terms involving { X , Y , T , e T w t 1 + X w x 1 + Y w y 1 + b 1 , e T w t 2 + X w x 2 + Y w y 2 + b 2 } . Setting every coefficient equal to zero generates four distinct constraint groups, each associated with a generalized solution of the equation that includes weights and biases. We give an example of the constraint conditions, as shown below:
h = 4 w x 1 2 4 w y 1 2 , r = b 3 w 3 u + b 4 w 4 u + b 5 2 , b 3 = b 3 , b 4 = b 4 , b 5 = b 5 , w 13 = w 13 , w 14 = w 14 , w 23 = w 4 u w 24 / w 3 u , w 24 = w 24 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = 6 w x 1 2 + w y 1 2 b 3 w 3 u + b 4 w 4 u + b 5 , w t 2 = w t 2 , w x 1 = w x 1 , w x 2 = w x 2 , w y 1 = w y 1 , w y 2 = w y 2 .
Based on the aforementioned constraints, we arrive at the following general solution
u ( x , y , t ) = w 13 w 3 u + w 14 w 4 u e 6 w x 1 2 + w y 1 2 b 3 w 3 u + b 4 w 4 u + b 5 T + X w x 1 + Y w y 1 + b 1 + b 3 w 3 u + b 4 w 4 u + b 5 ,
where X = x β / β , Y = y γ / γ , and T = t α / α .
By selecting the appropriate values { b 1 = 1 , b 3 = 1 , b 4 = 1 , b 5 = 1 , w 14 = 1 , w 13 = 1 , w 3 u = 1 , w 4 u = 1 , w y 1 = 0.3 , w t 1 = 1 , w x 1 = 0.6 , w x 2 = 1 , α = 0.6 , β = 0.6 , γ = 0.6 , t = 0.5 } in Equation (53), the results are shown in Figure 13, which displays the various plots of the Equation (48) over the spatial domain [ 0 , 20 ] × [ 0 , 20 ] . This set of images presents the characteristics of the fractional biological population model from various perspectives, aiding in a more comprehensive understanding of the model’s behavior and evolutionary trends. The figure shows that u increases with the growth of x and y, presenting a smooth, rainbow-like gradient, indicating that the value of u varies continuously in space. From the curve plots, it can be observed that as x and y increase, the curve shifts upward overall and the rate of growth accelerates. From the contour plot, it can be observed that the population density exhibits a gradient trend in both the x and y directions, with the density of the contour lines reflecting the rate of change in population density.
Case II: The NNs architecture 8 as shown in Figure 12b is utilized to solve the equation. For the first hidden layer, we adopt exponential activation functions tanh ( · ) and e ( · ) , whereas identity mapping ( · ) and reciprocal 1 / ( · ) are implemented throughout the second hidden layer. According to the analytical formulation, the new trial function is
u ( X , Y , T ) = w 3 u w 13 tanh T w t 1 + X w x 1 + Y w y 1 + b 1 + w 23 e T w t 2 + X w x 2 + Y w y 2 + b 2 + b 3 + w 4 u w 14 tanh T w t 1 + X w x 1 + Y w y 1 + b 1 + w 24 e T w t 2 + X w x 2 + Y w y 2 + b 2 + b 4 + b 5 ,
where X = x β / β , Y = y γ / γ , and T = t α / α .
Substituting Equation (54) into Equation (50) and combining similar terms involving { X , Y , T , tanh ( T w t 1 + X w x 1 + Y w y 1 + b 1 ) , e T w t 2 + X w x 2 + Y w y 2 + b 2 } . Setting every coefficient equal to zero generates 10 distinct constraint groups, each associated with a generalized solution of the equation that includes weights and biases. We give an example of the constraint conditions, as shown below:
h = 4 w x 2 2 4 w y 2 2 , r = b 3 w 3 u + b 5 w 14 w 4 u / 2 2 / w 14 2 , b 3 = b 3 , b 4 = w 14 , b 5 = b 5 , w 13 = 0 , w 14 = w 14 , w 23 = w 23 , w 24 = 0 , w 3 u = w 3 u , w 4 u = w 4 u , w t 1 = 3 w x 2 2 + w y 2 2 b 3 w 3 u + b 5 w 14 w 4 u / 2 / w 14 , w t 2 = 2 w t 1 , w x 1 = w x 2 / 2 , w x 2 = w x 2 , w y 1 = w y 2 / 2 , w y 2 = w y 2 .
Based on the aforementioned constraints, we arrive at the following general solution
u ( x , y , t ) = w 23 w 3 u e T w t 1 + X w x 2 + Y w y 2 + b 2 + b 2 + b 3 w 23 w 3 u + b 5 w 4 u w 14 tanh ( 2 T w t 1 + X w x 2 / 2 + Y w y 2 / 2 + b 1 ) w 14 ,
where w t 1 is denoted by Equation (55), X = x β / β , Y = y γ / γ , and T = t α / α .
By selecting the appropriate values { b 1 = 1 , b 2 = 1 , b 3 = 1 , b 5 = 1 , w 14 = 1 , w 23 = 1 , w 3 u = 1 , w 4 u = 1 , w y 2 = 1 , w x 2 = 1 , α = 1 , β = 1 , γ = 1 , t = 2 } in Equation (56), the results are shown in Figure 14, which displays the various plots of the Equation (48) over the spatial domain [ 5 , 5 ] × [ 10 , 10 ] . From the figures, it can be observed that the image of the solution abruptly transitions from a smooth plane to a surface characterized by a series of sharp peaks and depressions. The diagonal structures in the Figure 14d,e suggest that the waveform evolves steadily with spatial progression.
Remark 1. 
To validate the correctness of the derived solutions, they are substituted back into the original equations, yielding exact agreement on both sides. Furthermore, all equation solutions computed by the proposed analytical method are verified using Maple.
Remark 2. 
Several tailored NNs configurations are presented for obtaining analytical solutions to fPDEs, including the fractional wave equation, fractional telegraph equation, fractional Sharma–Tasso–Olever equation, and fractional biological population model. Owing to its adaptable framework, the approach permits alternative network designs to solve these and extended classes of fPDEs.

4. Discussions

In recent years, deep learning approaches have emerged as promising and popular methods for solving various numerical problems of partial differential equations. An NN serves as a computational surrogate to establish the mapping from input variables ( x , t ) to the output solution u ( x , t ) . The fundamental innovation of PINNs resides in embedding the residual terms of governing equations within the NN’s loss function during optimization. However, PINNs can only provide numerical solutions for solving the equation. Although PINNs have achieved great success in solving equations, there are still many deficiencies. PINNs are data-driven models, thus requiring a large amount of data for training. Optimizing NN parameters necessitates employing iterative algorithms for progressive refinement. Therefore, solving equations using PINNs entails significant computational costs and involves approximation errors.
Owing to their exceptional function approximation properties, NNs enable the derivation of analytical solutions for fPDEs. This work specifically leverages mathematical architectures of NNs to formulate closed-form expressions that solve these equations. Since shallow NNs also possess strong function approximation capabilities, this significantly reduces the complexity of the model. Through judicious configuration of activation functions, layer depth, and neuron density, customized trial solutions can be formulated for diverse fPDE classes. Moreover, the present methodology selectively adopts only the structural framework of neural networks, explicitly excluding their parameter optimization processes. The proposed method utilizes trial functions constructed using NNs to convert the calculation of fPDEs into the computation of a set of nonlinear algebraic equations containing weights and biases. Modifications to the governing equation’s boundary or initial conditions necessitate that PINNs repeat their training procedure. The process requires reformulating the objective function to include the updated physical relationships and then performing full network retraining. Thus, the computational load is heavy and the efficiency is low. Once the general solution to the equation is obtained using the analytical solver, only the corresponding different conditions need to be substituted into the general solution for computation, without the need for repeated calculations.
To validate the computational efficiency of the analytical method in this research, a comparison of numerical examples will be conducted between PINNs and our proposed method.

4.1. Neural Networks-Based Analytical Solver

The proposed methodology is employed to obtain exact solutions for the fractional wave equation under varying initial and boundary conditions, yielding multiple analytical solutions to the Equation (10). The proposed method is implemented using Maple 2024.0 for computation.
Case I: Consider Equation (10) with initial condition Equation (11). Since the general solution to the equation has already been obtained in Section 3.1, the general solution requires substitution of the specified initial and boundary conditions to obtain particular solutions. Substitute Equation (11) into a general solution Equation (19) of Equation (10) to obtain a particular solution.
u ( x , t ) = w 4 u w 24 e T k w x 2 + X w x 2 + b 2 + b 4 2 w 24 2 w 4 u e X w x 2 + b 2 2 2 b 4 w 24 w 4 u e X w x 2 + b 2 b 4 2 w 4 u + 2 X 2 + X ,
where X = x β / β and T = t α / α .
For enhanced solution characterization, we adopt the parameter configuration { k = 1 , b 2 = 1 , b 4 = 1 , w 24 = 0.5 , w 4 u = 0.4 , w x 2 = 0.4 , α = 1 , β = 1 } specified in Equation (57). The solution’s spatio-temporal behavior is visualized through multiple representations in Figure 15, including three-dimensional surface visualization, spatial profile analysis, contour mapping, and density distribution.
Case II: The Equation (10) is investigated under the following initial and boundary conditions:
u ( x , 0 ) = x ( x 1 ) , u ( 0 , t ) = u ( 1 , t ) = 0 .
Similarly, substitute the Equation (58) into Equation (19) to obtain the solution
u ( x , t ) = X ( X 1 ) Ω , Ω = e 2 T k + X w x 2 + 2 b 2 + e 2 T k + 1 w x 2 + 2 b 2 + e 2 T k + 2 X w x 2 + 2 b 2 e 2 T k + X + 1 w x 2 + 2 b 2 e 2 b 2 + w x 2 e X w x 2 + 2 b 2 + w x 2 + e 2 X w x , 2 + 2 b 2 e X w x 2 + 2 b 2 ,
where X = x β / β and T = t α / α .
For enhanced solution characterization, we adopt the parameter configuration { k = 1 , b 2 = 1 , w 4 u = 0.45 , w x 2 = 0.5 , α = 1 , β = 1 } specified in Equation (59). The solution’s spatio-temporal behavior is visualized through multiple representations in Figure 16, including three-dimensional surface visualization, spatial profile analysis, contour mapping, and density distribution.

4.2. Physics-Informed Neural Networks

We employ PINNs to numerically solve both Case I and Case II presented in Section 4.1, subsequently conducting a systematic comparison between the computational performance and solution accuracy of PINNs and our proposed methodology. The spatio-temporal domain is [ 0 , 1 ] × [ 0 , 1 ] and k = α = β = 1 in Case I and Case II. The physics-informed loss function in PINNs depends on the governing equation’s initial or boundary conditions. Consequently, varying conditions between Case I and Case II require distinct loss formulations, mandating separate network training procedures for each scenario. We use the following form of L 2 relative error
i [ u ( x t e s t , i , t t e s t , i ) u ˜ ( x t e s t , i , t t e s t , i ) ] 2 1 2 i [ u ( x t e s t , i , t t e s t , i ) ] 2 1 2 ,
mean absolute error, and maximum absolute error to measure the performance of the NNs, where the neural approximation u ˜ and analytical solution u are evaluated at discrete test locations ( x t e s t , i , t t e s t , i ) for each sample point index i.
The computational framework is implemented in Python 3.9, leveraging TensorFlow’s built-in automatic differentiation functionality. For optimization, we employ the Adam stochastic gradient descent algorithm to minimize the loss function. We initialize the NN parameters using normalized Glorot initialization. For these computational tasks, we implement an NN architecture consisting of 4 hidden layers with a uniform width of 50 neurons per layer. Some other hyperparameters of the NNs, such as the learning rate, the activation function, and the training epoch, are set to 1 × 10 3 , tanh ( x ) , and 20,000, respectively. A total of 1000 points are sampled within the spatio-temporal domain, with 500 points sampled on both the initial and boundary conditions.
The performance comparison results between PINNs and our analytical solution method are summarized in Table 2. Since PINNs require an iterative training mechanism to optimize the loss function, the training time is relatively long, and the values 0.0210 and 0.0240 in Table 2 represent the inference time of the NNs. PINNs can obtain approximate solutions to the equations with a certain error, as illustrated in Figure 17 and Figure 18.
Therefore, compared with PINNs, this symbolic computation method can obtain exact solutions for equations without data samples, and it reduces computational costs. To facilitate comparison, Table 3 summarizes the performance metrics of both methodologies. The proposed framework presents an innovative technique for deriving exact solutions to fPDEs, offering significant potential for advancing computational mathematics through its elimination of extensive data requirements and iterative optimization procedures.

5. Conclusions

In this paper, an artificial neural networks-based analytical solver that combines NNs with a symbolic computation method is proposed to analytically solve fPDEs. The computational capability of this analytical approach is verified through successful solutions to the fractional wave equation, fractional telegraph equation, fractional Sharma–Tasso—Olever equation, and fractional biological population model. Employing NN architecture, the solver establishes analytical solutions for fPDEs. New trial formulations emerge from the flexible implementation of different activation functions across multiple network models. This study establishes that 3D surface plots, contour diagrams, and density heatmaps collectively provide a robust framework for analyzing solution dynamics. This research introduces an innovative solution strategy for fPDEs with significant potential for cross-disciplinary implementation in both fundamental and applied research domains.
Compared with traditional numerical methods, the method completely avoids approximation errors and significantly improves the computational efficiency of fPDEs. In contrast to the existing bilinear neural network method [41], our methodology removes the requirement for bilinear transformation, significantly improving accessibility for research workers. Furthermore, not all equations possess a bilinear form. The neural network architecture in our proposed framework exhibits significant adaptability, enabling application to diverse fractional PDEs through customizable configurations of network depth, node density, and nonlinear activation selection. For the first time, NNs are employed to derive exact solutions for fPDEs. However, as an exact solution methodology, the proposed technique is specifically designed for obtaining analytical solutions to fPDEs, rather than numerical approximations. The method employs some basic functions to construct potential analytical solutions for fPDEs, and the activation function provides the basic function for this method. For equations without initial and boundary conditions, the selection of activation functions becomes blind. As the complexity of differential equations and NN models increases, it may lead to higher computational complexity. In the future, we will further apply this method to more complex equations, including high-dimensional fPDEs and nonlinear coupled systems, and attempt to construct new activation functions to reduce computational parameters.

Author Contributions

Conceptualization, Y.L. and L.Y.; Methodology, Y.L. and L.Y.; Software, R.Z.; Validation, S.W.; Visualization, S.Y.; Writing-original draft preparation, S.Y.; Writing-review and editing, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of Shandong Province (No. ZR2023MA062), the Belt and Road Special Foundation of The National Key Laboratory of Water Disaster Prevention (No. 2023491911), Tianyuan Fund for Mathematics of the National Natural Science Foundation of China (No. 12426105), and the Scientific and Technological Innovation Programs (STIP) of Higher Education Institutions in Shanxi (No. 2024L022).

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cristofaro, L.; Garra, R.; Scalas, E.; Spassiani, I. A fractional approach to study the pure-temporal Epidemic Type Aftershock Sequence (ETAS) process for earthquakes modeling. Fract. Calc. Appl. Anal. 2023, 26, 461–479. [Google Scholar] [CrossRef]
  2. Zhang, Y.; Sun, H.G.; Stowell, H.H.; Zayernouri, M.; Hansen, S.E. A review of applications of fractional calculus in Earth system dynamics. Chaos Solitons Fractals 2017, 102, 29–46. [Google Scholar] [CrossRef]
  3. Molina, M.I. Fractional electrical impurity. New J. Phys. 2024, 26, 013020. [Google Scholar] [CrossRef]
  4. Yang, Y.Q.; Qi, Q.W.; Hu, J.Y.; Dai, J.S.; Yang, C.D. Adaptive fault-tolerant control for consensus of nonlinear fractional-order multi-agent systems with diffusion. Fractal Fract. 2023, 7, 760. [Google Scholar] [CrossRef]
  5. Baliarsingh, P.; Nayak, L. Fractional derivatives with variable memory. Iran. J. Sci. Technol. Trans. A Sci. 2022, 46, 849–857. [Google Scholar] [CrossRef]
  6. Hu, J.B. Studying the memory property and event-triggered control of fractional systems. Inf. Sci. 2024, 662, 120218. [Google Scholar] [CrossRef]
  7. Guo, J.; Xu, D.; Qiu, W.L. A finite difference scheme for the nonlinear time-fractional partial integro-differential equation. Math. Methods Appl. Sci. 2020, 43, 3392–3412. [Google Scholar] [CrossRef]
  8. Qiao, H.L.; Cheng, A.J. A fast finite difference method for 2D time variable fractional mobile/immobile equation. J. Appl. Math. Comput. 2024, 70, 551–577. [Google Scholar] [CrossRef]
  9. Hu, H.Z.; Chen, Y.P.; Zhou, J.W. Two-grid finite element method for time-fractional nonlinear schrodinger equation. J. Comput. Math. 2024, 42, 1124–1144. [Google Scholar] [CrossRef]
  10. Sheng, Z.H.; Liu, Y.; Li, Y.H. Finite element method combined with time graded meshes for the time-fractional coupled Burgers’ equations. J. Appl. Math. Comput. 2024, 70, 513–533. [Google Scholar] [CrossRef]
  11. Jiao, Y.J.; Li, T.T.; Zhang, Z.Q. Jacobi spectral collocation method of space-fractional Navier-Stokes equations. Appl. Math. Comput. 2025, 488, 129111. [Google Scholar] [CrossRef]
  12. Zhang, X.X.; Wang, J.H.; Wu, Z.S.; Tang, Z.Y.; Zeng, X.Y. Spectral Galerkin methods for Riesz space-fractional convection–diffusion equations. Fractal Fract. 2024, 8, 431. [Google Scholar] [CrossRef]
  13. Gu, Q.L.; Chen, Y.P.; Zhou, J.W.; Huang, J. A fast linearized virtual element method on graded meshes for nonlinear time-fractional diffusion equations. Numer. Algorithms 2024, 97, 1141–1177. [Google Scholar] [CrossRef]
  14. Gu, Q.L.; Chen, Y.P.; Zhou, J.W.; Huang, Y.Q. A two-grid virtual element method for nonlinear variable-order time-fractional diffusion equation on polygonal meshes. Int. J. Comput. Math. 2023, 100, 2124–2139. [Google Scholar] [CrossRef]
  15. Yu, S.S.; Guo, M.; Chen, X.Y.; Qiu, J.L.; Sun, J.Q. Personalized movie recommendations based on a multi-feature attention mechanism with neural networks. Mathematics 2023, 11, 1355. [Google Scholar] [CrossRef]
  16. Li, S.A.; Cao, J.D.; Liu, H.; Huang, C.D. Delay-dependent parameters bifurcation in a fractional neural network via geometric methods. Appl. Math. Comput. 2024, 478, 128812. [Google Scholar] [CrossRef]
  17. Choudhary, K.; DeCost, B.; Chen, C.; Jain, A.; Tavazza, F.; Cohn, R.; Park, C.W.; Choudhary, A.; Agrawal, A.; Billinge, S.J.L.; et al. Recent advances and applications of deep learning methods in materials science. NPJ Comput. Mater. 2022, 8, 59. [Google Scholar] [CrossRef]
  18. Liu, Z.P.; Zhang, Z.M.; Lei, Z.V.; Omura, M.; Wang, R.L.; Gao, S.C. Dendritic deep learning for medical segmentation. IEEE/CAA J. Autom. Sin. 2024, 11, 803–805. [Google Scholar] [CrossRef]
  19. Liu, Y.Q.; Mao, T.; Zhou, D.X. Approximation of functions from Korobov spaces by shallow neural networks. Inf. Sci. 2024, 670, 120573. [Google Scholar] [CrossRef]
  20. Anastassiou, G.A.; Kouloumpou, D. Neural network approximation for time splitting random functions. Mathematics 2023, 11, 2183. [Google Scholar] [CrossRef]
  21. Principe, J.C.; Chen, B.D. Universal approximation with convex optimization: Gimmick or reality? IEEE Comput. Intell. Mag. 2015, 10, 68–77. [Google Scholar] [CrossRef]
  22. Zhang, Z.Z.; Bao, F.; Ju, L.L.; Zhang, G.N. Transferable neural networks for partial differential equations. J. Sci. Comput. 2024, 99, 2. [Google Scholar] [CrossRef]
  23. Li, Y.; Gao, W.; Ying, S.H. RBF-Assisted hybrid neural network for solving partial differential equations. Mathematics 2024, 12, 1617. [Google Scholar] [CrossRef]
  24. Lu, L.; Meng, X.H.; Mao, Z.P.; Karniadakis, G.E. DeepXDE: A deep learning library for solving differential equations. SIAM Rev. 2021, 63, 208–227. [Google Scholar] [CrossRef]
  25. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  26. Cai, S.Z.; Mao, Z.P.; Wang, Z.C.; Yin, M.L.; Karniadakis, G.E. Physics-informed neural networks (PINNs) for fluid mechanics: A review. Acta Mech. Sin. 2021, 37, 1727–1738. [Google Scholar] [CrossRef]
  27. Hou, Q.Z.; Li, Y.X.; Singh, V.P.; Sun, Z.W. Physics-informed neural network for diffusive wave model. J. Appl. Math. Comput. 2024, 637, 131261. [Google Scholar] [CrossRef]
  28. Pang, G.F.; Lu, L.; Karniadakis, G.E. fPINNs: Fractional physics-informed neural networks. SIAM J. Sci. Comput. 2019, 41, A2603–A2626. [Google Scholar] [CrossRef]
  29. Wang, S.P.; Zhang, H.; Jiang, X.Y. Fractional physics-informed neural networks for time-fractional phase field models. Nonlinear Dyn. 2022, 110, 2715–2739. [Google Scholar] [CrossRef]
  30. Ren, H.P.; Meng, X.Y.; Liu, R.R.; Hou, J.; Yu, Y.G. A class of improved fractional physics informed neural networks. Neurocomputing 2023, 562, 126890. [Google Scholar] [CrossRef]
  31. Wu, G.C.; Wu, Z.Q.; Zhu, W. Data-driven discrete fractional chaotic systems, new numerical schemes and deep learning. Chaos 2024, 34, 093144. [Google Scholar] [CrossRef]
  32. Yuan, B.; Wang, H.; Heitor, A.; Chen, X.H. f-PICNN: A physics-informed convolutional neural network for partial differential equations with space-time domain. J. Comput. Phys. 2024, 515, 113284. [Google Scholar] [CrossRef]
  33. Wang, S.P.; Zhang, H.; Jiang, X.Y. Physics-informed neural network algorithm for solving forward and inverse problems of variable-order space-fractional advection—Diffusion equations. Neurocomputing 2023, 535, 64–82. [Google Scholar] [CrossRef]
  34. Zhang, R.F.; Li, M.C.; Albishari, M.; Zheng, F.C.; Lan, Z.Z. Generalized lump solutions, classical lump solutions and rogue waves of the (2+1)-dimensional Caudrey-Dodd-Gibbon-Kotera-Sawada-like equation. Appl. Math. Comput. 2021, 403, 126201. [Google Scholar] [CrossRef]
  35. Zhang, R.F.; Li, M.C.; Cherraf, A.; Vadyala, S.R. The interference wave and the bright and dark soliton for two integro-differential equation by using BNNM. Nonlinear Dyn. 2023, 111, 8637–8646. [Google Scholar] [CrossRef]
  36. Zhang, R.F.; Li, M.C.; Gan, J.Y.; Li, Q.; Lan, Z.Z. Novel trial functions and rogue waves of generalized breaking soliton equation via bilinear neural network method. Chaos Solitons Fractals 2022, 154, 111692. [Google Scholar] [CrossRef]
  37. Jumarie, G. Fractional Partial Differential Equations and Modified Riemann-liouville Derivative New Methods for Solution. J. Appl. Math. Comput. 2007, 8, 31–48. [Google Scholar] [CrossRef]
  38. Mahdy, A.; Marai, G. Fractional complex transform for solving the fractional differential equations. Glob. J. Pure Appl. Math. 2018, 14, 17–37. [Google Scholar]
  39. Song, L.N.; Wang, Q.; Zhang, H.Q. Rational approximation solution of the fractional Sharma—Tasso—Olever equation. J. Comput. Appl. Math. 2009, 224, 210–218. [Google Scholar] [CrossRef]
  40. Zhang, S.; Zhang, H.Q. Fractional sub-equation method and its applications to nonlinear fractional PDEs. Phys. Lett. A 2011, 375, 1069–1073. [Google Scholar] [CrossRef]
  41. Zhang, R.F.; Bilige, S. Bilinear neural network method to obtain the exact analytical solutions of nonlinear partial differential equations and its application to p-gBKP equation. Nonlinear Dyn. 2019, 95, 3041–3048. [Google Scholar] [CrossRef]
Figure 1. Graphical depiction of neuron structure.
Figure 1. Graphical depiction of neuron structure.
Fractalfract 09 00541 g001
Figure 2. Architecture of NNs.
Figure 2. Architecture of NNs.
Fractalfract 09 00541 g002
Figure 3. Overall flow of the proposed analytical method.
Figure 3. Overall flow of the proposed analytical method.
Fractalfract 09 00541 g003
Figure 4. Architectural design of NNs for deriving analytical solutions to the fractional wave equation.
Figure 4. Architectural design of NNs for deriving analytical solutions to the fractional wave equation.
Fractalfract 09 00541 g004
Figure 5. The 3D plot, curve plot, contour plot, and density plot of the exact solution Equation (24).
Figure 5. The 3D plot, curve plot, contour plot, and density plot of the exact solution Equation (24).
Fractalfract 09 00541 g005
Figure 6. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (28).
Figure 6. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (28).
Fractalfract 09 00541 g006
Figure 7. Architectural design of NNs for deriving analytical solutions to the fractional telegraph equation.
Figure 7. Architectural design of NNs for deriving analytical solutions to the fractional telegraph equation.
Fractalfract 09 00541 g007
Figure 8. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (39).
Figure 8. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (39).
Fractalfract 09 00541 g008
Figure 9. Architectural design of NNs for deriving analytical solutions to fractional Sharma–Tasso–Olever equation.
Figure 9. Architectural design of NNs for deriving analytical solutions to fractional Sharma–Tasso–Olever equation.
Fractalfract 09 00541 g009
Figure 10. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (44).
Figure 10. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (44).
Fractalfract 09 00541 g010
Figure 11. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (47).
Figure 11. The 3D plot, curves plot, contour plot, and density plot of the exact solution Equation (47).
Fractalfract 09 00541 g011
Figure 12. Architectural design of NNs for deriving analytical solutions to the fractional biological population model.
Figure 12. Architectural design of NNs for deriving analytical solutions to the fractional biological population model.
Fractalfract 09 00541 g012
Figure 13. The 3D plot, curve plots, contour plot, and density plot of the exact solution Equation (53).
Figure 13. The 3D plot, curve plots, contour plot, and density plot of the exact solution Equation (53).
Fractalfract 09 00541 g013
Figure 14. The 3D plot, curve plots, contour plot, and density plot of the exact solution Equation (56).
Figure 14. The 3D plot, curve plots, contour plot, and density plot of the exact solution Equation (56).
Fractalfract 09 00541 g014
Figure 15. The 3D plot, curve plot, contour plot, and density plot of the exact solution Equation (57).
Figure 15. The 3D plot, curve plot, contour plot, and density plot of the exact solution Equation (57).
Fractalfract 09 00541 g015
Figure 16. The 3D plot, curve plot, contour plot, and density plot of the exact solution Equation (59).
Figure 16. The 3D plot, curve plot, contour plot, and density plot of the exact solution Equation (59).
Fractalfract 09 00541 g016
Figure 17. Calculation results by PINNs for Case I.
Figure 17. Calculation results by PINNs for Case I.
Fractalfract 09 00541 g017
Figure 18. Calculation results by PINNs for Case II.
Figure 18. Calculation results by PINNs for Case II.
Fractalfract 09 00541 g018
Table 1. Architectural configuration of NNs for analytical solution.
Table 1. Architectural configuration of NNs for analytical solution.
Architectural ComponentConfiguration Details
Network DepthNumber of hidden layers
Neuronal distribution across layers
Node CharacteristicsNonlinear activation selection
Bias term implementation
Output ProcessingLinear combination of weighted inputs
Table 2. Computational performance comparison between the two approaches.
Table 2. Computational performance comparison between the two approaches.
Computational Performance IndicatorsPINNsOur Method
Case ICalculation time (s)372.5405 + 0.02100 + 1.03
Mean absolute error0.00041400
Maximum absolute error0.00073360
L 2 relative error0.00015680
Case IICalculation time (s)456.6720 + 0.02400 + 1.17
Mean absolute error0.0001200
Maximum absolute error0.00022190
L 2 relative error0.00041170
Table 3. Comparison between two approaches.
Table 3. Comparison between two approaches.
Our MethodPINNs
FoundationBuilt upon strict mathematical derivationsIntegrates physical laws with deep learning
Result typeAnalytical solutionApproximate solution
Data dependencyNo training data requiredExtensive datasets
Model transparencyClear and logical explanationsLimited interpretability
Computational processDirect computation without iterationsRequires repeated parameter updates
Calculation costLowHigh
PrecisionCompletely accurateWith errors
Optimization algorithmNoneNeed
Modify conditionsNo need to retrainRetraining the NNs
RandomnessNoYes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yuan, S.; Liu, Y.; Yan, L.; Zhang, R.; Wu, S. Neural Networks-Based Analytical Solver for Exact Solutions of Fractional Partial Differential Equations. Fractal Fract. 2025, 9, 541. https://doi.org/10.3390/fractalfract9080541

AMA Style

Yuan S, Liu Y, Yan L, Zhang R, Wu S. Neural Networks-Based Analytical Solver for Exact Solutions of Fractional Partial Differential Equations. Fractal and Fractional. 2025; 9(8):541. https://doi.org/10.3390/fractalfract9080541

Chicago/Turabian Style

Yuan, Shanhao, Yanqin Liu, Limei Yan, Runfa Zhang, and Shunjun Wu. 2025. "Neural Networks-Based Analytical Solver for Exact Solutions of Fractional Partial Differential Equations" Fractal and Fractional 9, no. 8: 541. https://doi.org/10.3390/fractalfract9080541

APA Style

Yuan, S., Liu, Y., Yan, L., Zhang, R., & Wu, S. (2025). Neural Networks-Based Analytical Solver for Exact Solutions of Fractional Partial Differential Equations. Fractal and Fractional, 9(8), 541. https://doi.org/10.3390/fractalfract9080541

Article Metrics

Back to TopTop