Next Article in Journal
Applications Laguerre Polynomials for Families of Bi-Univalent Functions Defined with (p,q)-Wanas Operator
Next Article in Special Issue
The Dynamics of a General Model of the Nonlinear Difference Equation and Its Applications
Previous Article in Journal
Joint Discrete Approximation of Analytic Functions by Shifts of the Riemann Zeta Function Twisted by Gram Points II
Previous Article in Special Issue
Higher-Order Nabla Difference Equations of Arbitrary Order with Forcing, Positive and Negative Terms: Non-Oscillatory Solutions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study of Burgers–Huxley Equation Using Neural Network Method

1
College of Information Engineering, Shanghai Maritime University, Shanghai 201306, China
2
College of Sciences and Arts, Shanghai Maritime University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Axioms 2023, 12(5), 429; https://doi.org/10.3390/axioms12050429
Submission received: 15 March 2023 / Revised: 20 April 2023 / Accepted: 25 April 2023 / Published: 26 April 2023
(This article belongs to the Special Issue Advances in Difference Equations)

Abstract

:
The study of non-linear partial differential equations is a complex task requiring sophisticated methods and techniques. In this context, we propose a neural network approach based on Lie series in Lie groups of differential equations (symmetry) for solving Burgers–Huxley nonlinear partial differential equations, considering initial or boundary value terms in the loss functions. The proposed technique yields closed analytic solutions that possess excellent generalization properties. Our approach differs from existing deep neural networks in that it employs only shallow neural networks. This choice significantly reduces the parameter cost while retaining the dynamic behavior and accuracy of the solution. A thorough comparison with its exact solution was carried out to validate the practicality and effectiveness of our proposed method, using vivid graphics and detailed analysis to present the results.

1. Introduction

Partial differential equations (PDEs) are ubiquitous and fundamental to understanding and modeling the complexities of natural phenomena. From mathematics to physics to economics and beyond, PDEs play a critical role in virtually all fields of engineering and science [1,2,3]. Through their mathematical representation of physical phenomena, PDEs provide a powerful means of gaining insight into complex systems, enabling researchers and engineers to predict behavior and uncover hidden relationships. However, solving PDEs can be a daunting and challenging task. The complexity of these equations often requires sophisticated numerical methods that must balance accuracy and efficiency while solving high-dimensional PDEs. Despite these challenges, PDEs remain a cornerstone of modern science, enabling researchers to unlock discoveries and technological advancements across disciplines.
As numerical and computational techniques continue to rapidly develop, the study of PDEs has become increasingly vital. In recent years, advances in numerical methods and high-performance computing techniques have made it possible to solve complex PDEs more accurately and efficiently than ever before. These new tools can precisely solve specific problems across a broader range of equations while simultaneously computing data faster, reducing the time and cost of solving pending problems. Moreover, these new techniques have allowed researchers to gain deeper insights into the physical meaning behind PDEs, enabling them to revisit natural phenomena from fresh perspectives and explore those that prove challenging to explain by traditional methods. This has led to groundbreaking research discoveries and innovations in various fields of science and engineering.
Machine learning methods [4,5], particularly in the area of artificial neural networks (ANNs) [6,7], have piqued considerable interest in recent years due to their potential to solve differential equations. ANNs are well-known for their exceptional approximation capabilities and have emerged as a promising alternative to traditional algorithms [8]. These methods have a significantly smaller memory footprint and generate numerical solutions that are both closed and continuous over the integration domain without requiring interpolation. ANNs have been applied to differential equations, including ordinary differential equations (ODEs) [9,10], PDEs [11,12], and stochastic differential equations (SDEs) [13,14], making them a valuable tool for researchers and engineers alike. Neural networks have become a powerful and versatile tool for solving differential equations due to their ability to learn intricate mappings from input–output data, further cementing their role as a critical component in the machine learning fields.
In recent years, the application of neural networks in solving differential equations has gained significant attention in the scientific community. One prominent model is the neural ordinary differential equations, which approximates the derivative of an unknown solution using neural networks, parameterizing the derivatives of the hidden states of the network with the help of the differential equation, thus creating a new type of neural network [15]. Another approach is the deep Galerkin method [16], which uses neural networks to approximate the solution of the differential equation in a bid to minimize error. Gorikhovskii et al. [17] introduced a practical approach for solving ODEs using neural networks in the TensorFlow machine-learning framework. In addition, Huang et al. [18] introduce an additive self-attention mechanism to the numerical solution of differential equations based on the dynamical system perspective of the residual neural network.
By utilizing neural network functions to approximate the solutions, neural networks have also been used to solve PDEs. The physics-informed neural network (PINN) method uses the underlying physics of the problem to incorporate constraints into the solution of the neural network, resulting in successful applications to various PDEs such as the Burgers and Poisson equations [19]. Compared to traditional numerical methods, PINNs offer several advantages, including higher accuracy and more efficient computation. Berg et al. [20] introduced a new deep learning-based approach to solve PDEs on complex geometries. They use a feed-forward neural network and an unconstrained gradient-based optimization method to predict PDE solutions. Furthermore, Another exciting development in the field of neural networks and PDEs is the use of convolutional neural networks (CNNs). Ruthotto et al. [21] used a CNN to learn to solve elliptic PDEs and incorporated a residual block structure to improve network performance. Quan et al. [22] presented an innovative approach to addressing the challenge of solving diffusion PDEs, by introducing a novel learning method built on the foundation of the extreme learning machine algorithm. By leveraging this advanced technique, the parameters of the neural network are precisely calculated by solving a linear system of equations. Furthermore, the loss function is ingeniously constructed from three crucial components: the PDE, initial conditions, and boundary conditions. Tang et al. [23] demonstrate through numerical cases that the proposed depth adaptive sampling (DAS-PINNs) method can be used for solving PDEs. Overall, the advancements made in the domain of neural networks have revolutionized how we approach solving complex PDEs in unimaginable ways. These developments suggest that neural networks are a promising tool for solving complex PDEs and that there is great potential for further research and innovation in this area.
This paper proposes a novel approach for solving the Burgers–Huxley equation, which uses a neural network based on the Lie series in the Lie groups of differential equations, adding initial or boundary value terms to the loss function to approximate the solution of the equation by minimization. Slavova et al. [24] constructed a cellular neural network model to study the Burgers–Huxley equation. Shagun et al. [25] employed a feed-forward neural network to solve the Burgers-Huxley equation and investigated the impact of the number of training points on the accuracy of the solution. Kumar et al. [26] proposed a deep learning algorithm based on the deep Galerkin method for solving the Burgers–Huxley equation, which outperformed traditional numerical methods. These studies demonstrate the potential of neural networks in solving differential equations. Nonetheless, it is simple to ignore the underlying nature of these equations, in other words, to fail to capture the nonlinear nature of the equations, which is essential to comprehend the behavior of complex systems. To address this issue, the aim of our proposed method is to approximate the solution of the differential equations by combining the Lie series in Lie groups of differential equations and the power of neural networks. Our proposed method accurately simulates the physical behavior of complicated systems, and the first part of the constructed solution has well captured the nonlinear nature of the equation while reducing the parameter cost of the subsequent neural network and by minimizing the loss function, making the solution converge quickly by introducing initial or boundary value terms required for exact approximation. This work demonstrates the effectiveness of combining neural networks with Lie series to solve differential equations and provides insights into the physical behavior of complex dynamical systems.
The essay is set up as follows. The basic framework and fundamental theory of neural network algorithms based on Lie series in Lie groups of differential equations are introduced in Section 2. The specific steps for the Lie-series-based neural network method to solve the Burgers–Huxley equation are described in Section 3. The method is also applied to the Burgers–Fisher equation and the Huxley equation. Summary and outlook are presented in Section 4.

2. Basic Idea of a Lie-Series-Based Neural Network Algorithm

2.1. Differential Forms and Lie Series Solution

With respect to the Lie group transformation of the parameter ε ,
u * = T ( ε ; u ) G , u * ( 0 ) = u
where G is a Lie group, and ε is a group parameter.
By employing Taylor expansion about neighborhood of ε = 0 ,
u * = T ( ε ; u ) = u + T ( ε ; u ) ε ε = 0 ε + O ε 2 .
Then, u * = u + ε ζ is known as the infinitesimal transformation. D = ζ ( u ) u is called the infinitesimal operator, where ζ ( u ) = T ( ε ; u ) ε ε = 0 .
The following differential equation is given
u = F ( ξ , u ) , u ( 0 ) = u 0
F ( ξ , u ) is a differentiable function, and if (2) is a symmetry of (3), then it has a Lie series solution to the initial value problem (3) and can be written as [27]
u = e ξ D u | ξ = 0

2.2. Algorithm of a Lie-Series-Based Neural Network

The idea of Lie groups is based on the study of continuous symmetry, which at first may seem abstract and complex. However, in the realm of solving differential equations, Lie group methods are a unique approach that goes beyond traditional mathematical techniques. Lie series in the Lie transform groups of differential equations can be used to construct approximate solutions of PDEs and to study their symmetries and other properties. Lie series provide a powerful framework for studying the behavior of differential equations and have many important applications in various fields of science and engineering.
From [28], it is known that
D = D 1 + D 2
The solution of (3) can be written as u = e ξ D u | ξ = 0 = e ξ ( D 1 + D 2 ) u | ξ = 0 .
Theorem 1.
u ¯ ( ξ ; u ) = e ξ D 1 u | ξ = 0 , ξ R n , is the decomposition part of D. The solution of problem (3) belonging to D expanded as follows:
u = u ¯ ( ξ ; u ) + 0 ξ D 2 e ( ξ τ ) D u | u u ¯ ( τ ; u ) d τ
The proof is given below and is detailed in the literature [28].
Proof. 
u = e ξ D u = e ξ D 1 + D 2 u = v = 0 ξ v v ! D 1 v u + v = 1 ξ v v ! D 1 v 1 D 2 u + v = 2 ξ v v ! D 1 v 2 D 2 D u + + v = α ξ v v ! D 1 v α D 2 D α 1 u +
It is known that
ξ v v ! = 0 ξ ( ξ τ ) α 1 ( α 1 ) ! τ v α ( v α ) ! d τ , ( v α 1 , i n t e g e r s )
Equation (7) is rewritten as
u = u ¯ + 0 ξ v = 0 τ v v ! D 1 v D 2 u d τ + 0 ξ ( ξ τ ) v = 0 τ v v ! D 1 v D 2 D u d τ + + 0 ξ ( ξ τ ) α 1 ( α 1 ) ! v = 0 τ v v ! D 1 v D 2 D α 1 u d τ +
From the form of the series solution [27], it follows that
v = 0 τ v v ! D 1 v D 2 D α 1 u = D 2 D α 1 u u u ¯ ( τ ; u )
Hence,
u = u ¯ + α = 1 0 ξ ( ξ τ ) α 1 ( α 1 ) ! D 2 D α 1 u u u ¯ ( τ ; u ) d τ
after commuting the signs of the series and the integral which is allowed within the circle of absolute convergence, the formula
u = u ¯ + 0 ξ D 2 α = 0 ( ξ τ ) α α ! D α u u u ¯ ( τ ; u ) d τ
is obtained, which may also be written as follows:
e ξ D u = e ξ D 1 u + 0 ξ D 2 e ( ξ τ ) D u u u ¯ ( τ ; u ) d τ
  □
The complexities inherent in the integration of the second component, as elucidated by Equation (6), necessitates a sophisticated approach to computation. To tackle this daunting challenge head-on, as elaborated in the reference [29] of our previous work, the functional form of the neural network is utilized to simplify this part and ensure the accuracy of our results.
From [29], u ^ = e ξ D u | ξ = 0 = u ¯ + ξ N ( θ ; ξ ) . The determination of u ¯ from the equation u ¯ = D 1 u ¯ is inspired by the idea of the Lie series solution of the first-order ODE, where the initial value of u ¯ ( 0 ) = u ( 0 ) = u 0 is kept constant throughout the process, ensuring the reliability and truthfulness of our results. N ( θ ; ξ ) is a single output neural network with a single input of ξ , the parameter θ consists of the weight W and the bias b . The Algorithm 1 is described in detail below, as shown in Figure 1.
Algorithm 1: A Lie-series-based neural network algorithm for problem (3)
       Require Determine the operator D according to (3), and solve it with the decomposed part D 1 to obtain u ¯ .
       Begin
          1. Consider a uniformly spaced distribution of discrete points within the initial condition ξ ( = 1 , 2 , , λ ) .
          2. Determining the structure of a neural network. (The number of hidden layers and the number of neurons, the selection of the activation function σ .)
          3. Initialization of the neural networks parameters W , b .
          4. Get u ^ = u ¯ + ξ N ( θ ; ξ ) and substitute back into (3).
          5. Minimize the loss function L ( θ ) .
          6. Update the parameter θ so that u ^ approximates the solution u of problem (3).
       End
In general, the loss function L ( θ ) is defined as follows:
L ( θ ) = L F + L I = 1 λ = 1 λ i = 1 n ξ u ^ i ( ξ , θ ) ξ = ξ F i u ^ 1 , u ^ 2 , , u ^ i 2 2 + 1 2 l = 1 p i = 1 n u ^ i ( ξ , θ ) ξ = ξ l K ( ξ ) ξ = ξ l 2 2
as additional terms with K ( ξ l ) , l = 1 , 2 , , p as initial value or boundary conditions. The L F part of the loss function is derived by substituting the network solution u ^ into the mean squared error generated on both sides of the problem (3). In addition, the mean squared error generated by the network solution u ^ under the initial or boundary value terms are also used to derive the L I component of our loss function. By constructing the components of L F and L I , we can satisfy both the differential equations and the initial values or boundary conditions of the problem under study.
The above algorithm also applies to the system of differential equations d u i d ξ = F i ( u 1 , u 2 , , u n ) , u i ( 0 ) = α i R 1 , i = 1 , 2 , , n , where D = i = 1 n F i ( u i ) u i . For higher-order ODEs or PDEs, the above form can also be transformed with the help of some transformations or calculations.

2.3. The General Structure of the Neural Network

As depicted in Figure 2, our study delves into the complexities of multilayer perceptrons and their unique characteristics, with a particular emphasis on those with a single input unit, m hidden layers of H neurons, a neural network with activation function σ in the hidden layer, and a linear output unit. We present a detailed analysis of this neural network architecture. Specifically, for a given input vector ξ ( = 1 , 2 , , λ ) , the output of the network N = i = 1 H W m + 1 σ ( Z i m ) + b m + 1 , Z i m = j = 1 H w j i m σ ( Z j m 1 ) + b i m , where w j i m is the weight of the j t h neuron in layer m 1 to the i t h neuron in layer m, and b i m is the bias of the i t h neuron in layer m. It can be seen that Z 1 1 = w 11 1 ξ + b 1 1 . In this paper, the activation function σ is chosen tanh ( Z ) = e Z e Z e Z + e Z .

3. Lie-Series-Based Neural Network Algorithm for Solving Burgers Huxley Equation

The generalized Burgers–Huxley equation [30] is a nonlinear PDE that describes the propagation of electrical impulses in excitable media, such as nerve and muscle cells. It is a widely used mathematical framework for modeling intricate dynamical phenomena and has been instrumental in advancing research across multiple domains including physics, biology, economics, and ecology. The equation takes the form
u t + α u δ u x 2 u x 2 = β u 1 u δ η u δ λ
where α , β , λ , η are constants and δ is a positive constant.
When α = 1 , β = 1 , λ = 1 , η = 1 , δ = 1 , the Burgers–Huxley equation is as follows:
u t = 2 u x 2 + u u x + u ( 1 u ) ( u 1 ) , u ( 0 , x ) = 1 2 1 tanh x 4
The exact solution of (12) is u ( t , x ) = 1 2 1 tanh x 4 + 3 t 8 . Using the traveling wave transform ξ = x c t , problem (12) is transformed into an ODE, u + c u + u u + u ( 1 u ) ( u 1 ) = 0 . Naturally, it is transformed into the form of the following system of ODEs
u 1 = u 2 , u 2 = 3 2 u 2 u 1 u 2 u 1 ( 1 u 1 ) ( u 1 1 )
with c = 3 2 , u 1 ( ξ ) = u and initial values u 1 ( 0 ) = 1 2 , u 2 ( 0 ) = 1 8 .
In this study, we address the problem of solving the Burgers–Huxley equation using a Lie-series-based neural network algorithm. The operator D = u 2 u 1 + ( 3 2 u 2 u 1 u 2 u 1 ( 1 u 1 ) ( u 1 1 ) ) u 2 of (13) is chosen as D 1 = u 2 u 1 + 3 2 u 2 u 2 , and the solution of the corresponding initial value problem is u ¯ 1 ( ξ ) = 1 12 7 cosh 3 ξ 2 sinh 3 ξ 2 , u ¯ 2 ( ξ ) = 1 8 cosh 3 ξ 2 1 8 sinh 3 ξ 2 . The solution of this part has been able to capture the nonlinear nature of the equation within a certain range, as shown in Figure 3. To minimize the loss function L ( θ ) , we employ two structurally identical neural networks and boundary value terms, each with 30 neurons in a single hidden layer, and the input ξ ( = 1 , 2 , , 100 ) is 100 training points spaced equally in the interval [ 5 , 3 ] , making u ^ 1 ( ξ ) as close as possible to the exact solution u ( ξ ) of the equation. The generalization ability of the neural network was confirmed in 120 test points at equidistant intervals of ξ 5 , 3.3 . The Lie-series-based neural network algorithm solves the Burgers–Huxley equation model as shown in Figure 4. Furthermore, we demonstrate the ability of neural networks to fit the training and test sets in Figure 5. By plotting the loss function L ( θ ) = L F + L I against the number of iterations in Figure 6, where L F = 1 λ = 1 λ u ^ 1 ( ξ ) u ^ 2 ( ξ ) 2 + u ^ 2 ( ξ ) 3 2 u ^ 2 ( ξ ) + u ^ 1 ( ξ ) u ^ 2 ( ξ ) + u ^ 1 ( ξ ) ( 1 u ^ 1 ( ξ ) ) ( u ^ 1 ( ξ ) 1 ) 2 , and L I = 1 2 u ^ 1 ( 5 ) u ( 5 ) 2 + 1 2 u ^ 2 ( 5 ) u ( 5 ) 2 , λ = 100 . Some 1100 iterations later, L ( θ ) = 3.042 × 10 8 .
We compare the solution u ^ ( t , x ) containing the neural network training and the exact solution u ( t , x ) in the interval t 0 , 1 , x 5 , 2 in the upper panel of Figure 7. Additionally, the lower panel displays the behavior of the solution at t = 0.3 , 0.5 , 0.8 , demonstrating the solitary wave solution of the Burgers–Huxley equation. The contour plots for solution u ^ 1 ( t , x ) and the exact solution u ( t , x ) are shown in Figure 8, further illustrating the accuracy of our proposed algorithm.
To verify the validity and generality of our proposed equation, the method was applied to two classical equations, the Burgers–Fisher, and the Huxley equations. For this purpose, we performed a thorough analysis and obtained strong results that proved the validity of our method. Specifically, when α = 1 , β = 1 , λ = 1 , η = 0 , δ = 1 , the Burgers–Fisher equation is as follows:
u t = 2 u x 2 + u u x + u ( 1 u ) , u ( 0 , x ) = 1 2 1 + tanh x 4
The exact solution of (14) is u ( t , x ) = 1 2 1 + tanh x 4 + 5 t 8 . Similarly, using the traveling wave transform ξ = x c t , problem (14) is transformed into an ODE, u + c u + u u + u ( 1 u ) = 0 , with initial value u ( 0 ) = 1 2 , u ( 0 ) = 1 8 . Transformation of ODEs into the form of a system of differential equations,
u 1 = u 2 , u 2 = 5 2 u 2 u 1 u 2 u 1 ( 1 u 1 )
where u 1 ( ξ ) = u , c = 5 2 , and initial values u 1 ( 0 ) = 1 2 , u 2 ( 0 ) = 1 8 . The operator D = u 2 u 1 + ( 5 2 u 2 u 1 u 2 u 1 ( 1 u 1 ) ) u 2 of (15), D 1 is chosen as u 2 u 1 + 5 2 u 2 u 2 u 1 u 2 , the predicted solution u ^ 1 ( ξ ) = 1 12 e ξ / 2 ( 7 + e 3 ξ / 2 ) + ξ N 1 , u ^ 2 ( ξ ) = 1 24 ( 7 e ξ / 2 ) 1 6 e 2 ξ + ξ N 2 , where the structure of the neural network is a single hidden layer containing 30 neurons with inputs ξ 5 , 2 of equidistant intervals of 100 training points and test points are 120 points of the interval 5 , 2.2 , and the training results are shown in Figure 9. As shown in Figure 10, our method achieves an impressive performance with the loss function L ( θ ) reaches 8.861 × 10 8 in about 700 iterations. This exceptional result again illustrates that the solution of the D 1 part of our proposed method captures the nonlinear nature of the solution, thereby reducing the computational cost associated with additional parameters which are evident from Figure 11. In addition, we provide a three-dimensional representation of the dynamics of the predicted solution u ^ ( t , x ) with the exact solution u ( t , x ) in the interval t 0 , 1 and x 5 , 2 , as shown in Figure 12.
We investigate the Huxley equation under the conditions where α = 0 , β = 1 , λ = 1 , η = 1 , δ = 1 . The equations are as follows:
u t = 2 u x 2 + u ( 1 u ) ( u 1 ) , u ( 0 , x ) = 1 2 1 + tanh x 2 2
The exact solution of (16) is u ( t , x ) = 1 2 1 + tanh 2 x 4 t 4 . Similarly, using the traveling wave transform ξ = x c t , problem (16) is transformed into an ODE, u + c u + u ( 1 u ) ( u 1 ) = 0 . It is transformed into the following differential equation form
u 1 = u 2 , u 2 = 2 2 u 2 u 1 ( 1 u 1 ) ( u 1 1 )
where initial values u 1 ( 0 ) = 1 2 , u 2 ( 0 ) = 1 4 2 , and c = 2 2 , it is clear that u 1 ( ξ ) = u ( ξ ) , u 2 ( ξ ) = u ( ξ ) . In the case of D 1 = u 2 u 1 2 2 u 2 u 2 , the system of differential equations u ¯ 1 = u ¯ 2 , u ¯ 2 = 2 2 u ¯ 2 , the initial values are u ¯ 1 ( 0 ) = 1 2 and u ¯ 2 ( 0 ) = 1 4 2 , this time u ¯ 1 ( ξ ) = 3 4 1 4 e ξ / 2 , u ¯ 2 ( ξ ) = 1 4 2 e ξ / 2 .
For predicting the solution u ^ 1 ( ξ ) and u ^ 2 ( ξ ) , the same neural network with two single hidden layers containing 30 neurons with the same structure is trained by optimization technique Broyden–Fletcher–Goldfarb–Shanno (BFGS) minimizes the Loss function L ( θ ) . The input ξ is the interval [ 2 , 7 ] equidistantly spaced by 100 points. The test set is the 150 points in the interval [ 2 , 7.5 ] . As shown in Figure 13, our proposed method produced excellent predictions for both the trained predicted and exact solutions. The variation of the loss function throughout the process is depicted in Figure 14, and it can be observed that the loss function decreased remarkably during training. Figure 15 shows the dynamics of u ^ 1 ( t , x ) with the exact solution u ( t , x ) , when ξ = x c t is substituted into u ^ 1 ( ξ ) and the predicted solution u ^ 1 ( t , x ) compared with the exact solution u ( t , x ) at t = 0.3 , 0.5 , 0.8 . The contour plot in Figure 16 provides a more visualization of the network solution u ^ 1 ( t , x ) compared to the exact solution u ( t , x ) .

4. Discussion and Conclusions

The exponential growth of information data has resulted in limited data becoming a significant issue in various fields, especially in data-driven applications. Addressing this challenge has become a critical area of research in recent times. To contribute towards finding solutions to this problem, this paper proposes a novel method for resolving the Burgers-Huxley equation using a neural network based on Lie series in Lie groups of differential equations, which is an emerging field with great potential in solving complex problems. To the best of our knowledge, this study represents the first time the Burgers-Huxley equation has been solved using a Lie-series-based neural network algorithm. In physics, engineering, and biology, the Burgers–Huxley equation is a well-known mathematical model that is frequently utilized. Our novel approach offers a unique perspective on solving this equation by adding boundary or initial value items to the loss function, which leads to more accurate predictions and a better understanding of the underlying system. This research opens up new avenues for further exploration of the Lie-series-based neural network algorithm, specifically regarding its applications to other complex models beyond the Burgers–Huxley equation.
In this study, we present a novel method for obtaining a differentiable closed analytical form to provide an effective foundation for further research. The proposed approach is straightforward to use and evaluate. To verify the effectiveness of the suggested method, we applied it to two classic models of the Burgers–Fisher and Huxley equations that have well-known exact solutions. The proposed algorithm exhibits remarkable potential in capturing the nonlinear nature of equations and accelerating the computation process of neural networks. The performance of our method is demonstrated in Figure 3 and Figure 11, which show how the proposed algorithm can capture the nonlinear behavior of the equations more effectively and speed up the computation of subsequent neural networks. To further evaluate the effectiveness of the proposed technique, we plotted the relationship between the loss function and the number of iterations in Figure 6, Figure 10 and Figure 14. Our results indicate that under the influence of the Lie series in Lie groups of differential equations, our algorithm can converge quickly and achieve more precise solutions with fewer data. Moreover, the accuracy of the obtained solutions is significant, and the generalization ability of the neural network is demonstrated by its ability to maintain high accuracy even outside the training domain, as shown in Figure 5, Figure 9 and Figure 13. We compared the performance of each neural network using small parameters (60 weight parameters and 31 bias parameters) with the exact solution to the problem. Our results highlight that the addition of the Lie series in Lie groups of differential equations algorithm remarkably enhances the ability of the neural network to solve a given equation.
Undoubtedly, the proposed method has several limitations that need to be carefully considered. Firstly, the method requires the transformation of PDEs into ODEs before applying the suggested algorithm. Although the results obtained after this transformation are preliminary, they provide useful insights for researchers. Additionally, an inverse transformation must be employed to produce the final solution u ^ ( t , x ) , taking into account the range of values for various variables. The choice of the operator D 1 may also influence the outcomes. Secondly, the current study only addresses nonlinear diffusion issues of the type F ( u ) = α u δ u x + u x x + β u 1 u δ η u δ λ , and the suitability of the technique was assessed via the computation of the loss function. Therefore, the applicability of the method to other types of non-linear PDEs is yet to be investigated, and it might require further adjustments to accommodate such problems. Despite some inherent challenges, our work offers a promising strategy for solving complex mathematical models using neural network algorithms based on Lie series. The computational performance of the proposed algorithm is noteworthy, achieving high solution accuracy at a relatively low time and parameter cost. In light of these findings, it is worth considering the prospect of applying this algorithm to financial modeling, where accurate predictions can have a significant impact.
Moving forward, there is ample scope for extending and improving the proposed algorithm further. Future research could explore how to optimize the performance of the algorithm by addressing its limitations and weaknesses for nonlinear PDE problems. For example, choosing a different neural network framework, CNN or recurrent neural network, etc., may improve the efficiency and accuracy of the method. Additionally, expanding the method’s applicability beyond nonlinear diffusion issues may also yield valuable insights into other areas of mathematical modeling.
In summary, we believe that our work presents an exciting avenue for future research. By building upon our findings and addressing the limitations of the proposed algorithm, we can develop more sophisticated techniques for solving complex mathematical models in finance and other areas. Solving the above problems is the main goal of our next research work.

Author Contributions

Conceptualization, Y.W. and T.C.; methodology, Y.W.; software, Y.W.; validation, Y.W.; writing—original draft preparation, Y.W.; writing—review and editing, Y.W. and T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant number 11571008.

Data Availability Statement

The data used to support the findings of this study are included within the article. The link to the code is https://github.com/yingWWen/Study-of-Burger-Huxley-Equation-using-neural-network-method (accessed on 14 March 2023).

Acknowledgments

The authors thank the support of the National Natural Science Foundation of China [grant number 11571008].

Conflicts of Interest

As far as we know, there are no conflict of interest or financial or other conflicts. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Ockendon, J.R.; Howison, S.; Lacey, A.; Movchan, A. Applied Partial Differential Equations; Oxford University Press on Demand: Oxford, UK, 2003. [Google Scholar]
  2. Mattheij, R.M.; Rienstra, S.W.; Boonkkamp, J.T.T. Partial Differential Equations: Modeling, Analysis, Computation; SIAM: Philadelphia, PA, USA, 2005. [Google Scholar]
  3. Duffy, D.J. Finite Difference Methods in Financial Engineering: A Partial Differential Equation Approach; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  4. Jordan, M.I.; Mitchell, T.M. Machine learning: Trends, perspectives, and prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef]
  5. Mahesh, B. Machine learning algorithms—A review. Int. J. Sci. Res. (IJSR) 2020, 9, 381–386. [Google Scholar]
  6. Yegnanarayana, B. Artificial Neural Networks; PHI Learning Pvt. Ltd.: New Delhi, India, 2009. [Google Scholar]
  7. Zou, J.; Han, Y.; So, S.S. Overview of artificial neural networks. In Artificial Neural Networks: Methods and Applications; Humana Press: Totowa, NJ, USA, 2009; pp. 14–22. [Google Scholar]
  8. Hornik, K. Approximation capabilities of multilayer feedforward networks. Neural Netw. 1991, 4, 251–257. [Google Scholar] [CrossRef]
  9. Lagaris, I.E.; Likas, A.; Fotiadis, D.I. Artificial neural networks for solving ordinary and partial differential equations. IEEE Trans. Neural Netw. 1998, 9, 987–1000. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Chakraverty, S.; Mall, S. Artificial Neural Networks for Engineers and Scientists: Solving Ordinary Differential Equations; CRC Press: Boca Raton, FL, USA, 2017. [Google Scholar]
  11. Yang, L.; Meng, X.; Karniadakis, G.E. B-PINNs: Bayesian physics-informed neural networks for forward and inverse PDE problems with noisy data. J. Comput. Phys. 2021, 425, 109913. [Google Scholar] [CrossRef]
  12. Blechschmidt, J.; Ernst, O.G. Three ways to solve partial differential equations with neural networks–review. GAMM-Mitteilungen 2021, 44, e202100006. [Google Scholar] [CrossRef]
  13. Han, J.; Jentzen, A. Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations. Commun. Math. Stat. 2017, 5, 349–380. [Google Scholar]
  14. Nabian, M.A.; Meidani, H. A deep learning solution approach for high-dimensional random differential equations. Probabilistic Eng. Mech. 2019, 57, 14–25. [Google Scholar] [CrossRef] [Green Version]
  15. Chen, R.T.; Rubanova, Y.; Bettencourt, J.; Duvenaud, D.K. Neural ordinary differential equations. Adv. Neural Inf. Process. Syst. 2018, 31, 6572–6583. [Google Scholar]
  16. Sirignano, J.; Spiliopoulos, K. DGM: A deep learning algorithm for solving partial differential equations. J. Comput. Phys. 2018, 375, 1339–1364. [Google Scholar] [CrossRef] [Green Version]
  17. Gorikhovskii, V.; Evdokimova, T.; Poletansky, V. Neural networks in solving differential equations. J. Phys. Conf. Ser. 2022, 2308, 012008. [Google Scholar] [CrossRef]
  18. Huang, Z.; Liang, M.; Lin, L. On Robust Numerical Solver for ODE via Self-Attention Mechanism. arXiv 2023, arXiv:2302.10184. [Google Scholar]
  19. Lu, L.; Meng, X.; Mao, Z.; Karniadakis, G.E. DeepXDE: A deep learning library for solving differential equations. SIAM Rev. 2021, 63, 208–228. [Google Scholar] [CrossRef]
  20. Berg, J.; Nyström, K. A unified deep artificial neural network approach to partial differential equations in complex geometries. Neurocomputing 2018, 317, 28–41. [Google Scholar] [CrossRef] [Green Version]
  21. Ruthotto, L.; Haber, E. Deep neural networks motivated by partial differential equations. J. Math. Imaging Vis. 2020, 62, 352–364. [Google Scholar] [CrossRef] [Green Version]
  22. Quan, H.D.; Huynh, H.T. Solving partial differential equation based on extreme learning machine. Math. Comput. Simul. 2023, 205, 697–708. [Google Scholar] [CrossRef]
  23. Tang, K.; Wan, X.; Yang, C. DAS-PINNs: A deep adaptive sampling method for solving high-dimensional partial differential equations. J. Comput. Phys. 2023, 476, 111868. [Google Scholar] [CrossRef]
  24. Slavova, A.; Zecc, P. Travelling wave solution of polynomial cellular neural network model for burgers-huxley equation. C. R. l’Acad. Bulg. Sci. 2012, 65, 1335–1342. [Google Scholar]
  25. Panghal, S.; Kumar, M. Approximate analytic solution of Burger Huxley equation using feed-forward artificial neural network. Neural Process Lett. 2021, 53, 2147–2163. [Google Scholar] [CrossRef]
  26. Kumar, H.; Yadav, N.; Nagar, A.K. Numerical solution of Generalized Burger–Huxley & Huxley’s equation using Deep Galerkin neural network method. Eng. Appl. Artif. Intell. 2022, 115, 105289. [Google Scholar]
  27. Olver, P.J. Applications of Lie Groups to Differential Equations; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1993; Volume 107. [Google Scholar]
  28. Gröbner, W.; Knapp, H. Contributions to the Method of Lie Series; Bibliographisches Institut Mannheim: Mannheim, Germany, 1967; Volume 802. [Google Scholar]
  29. Wen, Y.; Chaolu, T.; Wang, X. Solving the initial value problem of ordinary differential equations by Lie group based neural network method. PLoS ONE 2022, 17, e0265992. [Google Scholar] [CrossRef] [PubMed]
  30. Wang, X.; Zhu, Z.; Lu, Y. Solitary wave solutions of the generalised Burgers-Huxley equation. J. Phys. Math. Gen. 1990, 23, 271. [Google Scholar] [CrossRef]
Figure 1. Flow chart of Lie-series-based neural network algorithm.
Figure 1. Flow chart of Lie-series-based neural network algorithm.
Axioms 12 00429 g001
Figure 2. Neural network structure.
Figure 2. Neural network structure.
Axioms 12 00429 g002
Figure 3. Comparison of the u ¯ 1 ( ξ ) solution of the Burgers–Huxley equation with the exact solution u ( ξ ) .
Figure 3. Comparison of the u ¯ 1 ( ξ ) solution of the Burgers–Huxley equation with the exact solution u ( ξ ) .
Axioms 12 00429 g003
Figure 4. Schematic diagram of a Lie series-based neural network algorithm for solving Burgers–Huxley equation.
Figure 4. Schematic diagram of a Lie series-based neural network algorithm for solving Burgers–Huxley equation.
Axioms 12 00429 g004
Figure 5. (Left) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 tanh ξ 4 of (13) in the training set. (Right) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 tanh ξ 4 of (13) in the test set.
Figure 5. (Left) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 tanh ξ 4 of (13) in the training set. (Right) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 tanh ξ 4 of (13) in the test set.
Axioms 12 00429 g005
Figure 6. Curves of Loss function versus number of iterations for Burgers–Huxley equation.
Figure 6. Curves of Loss function versus number of iterations for Burgers–Huxley equation.
Axioms 12 00429 g006
Figure 7. (Top) The true solution u ( t , x ) = 1 2 1 tanh x 4 + 3 t 8 of the Burgers–Huxley equation is on the left, the predicted solution u ^ ( t , x ) is on the right. (Bottom) Comparison of predicted and exact solutions at time t = 0.3 , 0.5 , and 0.8 . (The dashed blue line indicates the exact solution u ( t , x ) , and the solid red line indicates the predicted solution u ^ ( t , x ) ).
Figure 7. (Top) The true solution u ( t , x ) = 1 2 1 tanh x 4 + 3 t 8 of the Burgers–Huxley equation is on the left, the predicted solution u ^ ( t , x ) is on the right. (Bottom) Comparison of predicted and exact solutions at time t = 0.3 , 0.5 , and 0.8 . (The dashed blue line indicates the exact solution u ( t , x ) , and the solid red line indicates the predicted solution u ^ ( t , x ) ).
Axioms 12 00429 g007
Figure 8. Contour plot of the Burgers–Huxley equation with respect to the solution u ^ 1 ( t , x ) and the exact solution u ( t , x ) .
Figure 8. Contour plot of the Burgers–Huxley equation with respect to the solution u ^ 1 ( t , x ) and the exact solution u ( t , x ) .
Axioms 12 00429 g008
Figure 9. (Left) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh ξ 4 of (15) in the training set. (Right) Comparison of u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh ξ 4 of (15) in the test set.
Figure 9. (Left) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh ξ 4 of (15) in the training set. (Right) Comparison of u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh ξ 4 of (15) in the test set.
Axioms 12 00429 g009
Figure 10. Curves of loss function versus number of iterations for Burgers–Fisher equation.
Figure 10. Curves of loss function versus number of iterations for Burgers–Fisher equation.
Axioms 12 00429 g010
Figure 11. Comparison of the u ¯ 1 ( ξ ) solution of the Burgers–Fisher equation with the exact solution u ( ξ ) .
Figure 11. Comparison of the u ¯ 1 ( ξ ) solution of the Burgers–Fisher equation with the exact solution u ( ξ ) .
Axioms 12 00429 g011
Figure 12. (Top) The true solution u ( t , x ) = 1 2 1 + tanh x 4 + 5 t 8 of the Burgers–Fisher equation is on the left, the predicted solution u ^ ( t , x ) is on the right. (Bottom) Comparison of predicted and exact solutions at time t = 0.3 , 0.5 , and 0.8 . (The dashed blue line indicates the exact solution u ( t , x ) , and the solid red line indicates the predicted solution u ^ ( t , x ) ).
Figure 12. (Top) The true solution u ( t , x ) = 1 2 1 + tanh x 4 + 5 t 8 of the Burgers–Fisher equation is on the left, the predicted solution u ^ ( t , x ) is on the right. (Bottom) Comparison of predicted and exact solutions at time t = 0.3 , 0.5 , and 0.8 . (The dashed blue line indicates the exact solution u ( t , x ) , and the solid red line indicates the predicted solution u ^ ( t , x ) ).
Axioms 12 00429 g012
Figure 13. (Left) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh 2 ξ 4 of (17) in the training set. (Right) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh 2 ξ 4 of (17) in the test set.
Figure 13. (Left) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh 2 ξ 4 of (17) in the training set. (Right) Comparison of solution u ^ 1 ( ξ ) with the exact solution u ( ξ ) = 1 2 1 + tanh 2 ξ 4 of (17) in the test set.
Axioms 12 00429 g013
Figure 14. Curves of loss function versus number of iterations for Huxley equation.
Figure 14. Curves of loss function versus number of iterations for Huxley equation.
Axioms 12 00429 g014
Figure 15. (Top) The true solution u ( t , x ) = 1 2 1 + tanh 2 x 4 t 4 of the Huxley equation is on the left, the predicted solution u ^ ( t , x ) is on the right. (Bottom) Comparison of predicted and exact solutions at time t = 0.3 , 0.5 , and 0.8 . (The dashed blue line indicates the exact solution u ( t , x ) , and the solid red line indicates the predicted solution u ^ ( t , x ) ).
Figure 15. (Top) The true solution u ( t , x ) = 1 2 1 + tanh 2 x 4 t 4 of the Huxley equation is on the left, the predicted solution u ^ ( t , x ) is on the right. (Bottom) Comparison of predicted and exact solutions at time t = 0.3 , 0.5 , and 0.8 . (The dashed blue line indicates the exact solution u ( t , x ) , and the solid red line indicates the predicted solution u ^ ( t , x ) ).
Axioms 12 00429 g015
Figure 16. Contour plot of the Huxley equation with respect to the solution u ^ 1 ( t , x ) and the exact solution u ( t , x ) .
Figure 16. Contour plot of the Huxley equation with respect to the solution u ^ 1 ( t , x ) and the exact solution u ( t , x ) .
Axioms 12 00429 g016
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wen, Y.; Chaolu, T. Study of Burgers–Huxley Equation Using Neural Network Method. Axioms 2023, 12, 429. https://doi.org/10.3390/axioms12050429

AMA Style

Wen Y, Chaolu T. Study of Burgers–Huxley Equation Using Neural Network Method. Axioms. 2023; 12(5):429. https://doi.org/10.3390/axioms12050429

Chicago/Turabian Style

Wen, Ying, and Temuer Chaolu. 2023. "Study of Burgers–Huxley Equation Using Neural Network Method" Axioms 12, no. 5: 429. https://doi.org/10.3390/axioms12050429

APA Style

Wen, Y., & Chaolu, T. (2023). Study of Burgers–Huxley Equation Using Neural Network Method. Axioms, 12(5), 429. https://doi.org/10.3390/axioms12050429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop