Open Access
This article is

- freely available
- re-usable

*Algorithms*
**2015**,
*8*(2),
224-233;
https://doi.org/10.3390/a8020224

Article

Numerical Solution of Turbulence Problems by Solving Burgers’ Equation

Institute for Multidisciplinary Mathematics, Polytechnical University of Valencia, Camino de Vera, s/n, 46022 València, Spain

^{*}

Author to whom correspondence should be addressed.

Academic Editor:
Francisco I. Chicharro

Received: 4 April 2015 / Accepted: 30 April 2015 / Published: 8 May 2015

## Abstract

**:**

In this work we generate the numerical solutions of Burgers’ equation by applying the Crank-Nicholson method and different schemes for solving nonlinear systems, instead of using Hopf-Cole transformation to reduce Burgers’ equation into the linear heat equation. The method is analyzed on two test problems in order to check its efficiency on different kinds of initial conditions. Numerical solutions as well as exact solutions for different values of viscosity are calculated, concluding that the numerical results are very close to the exact solution.

Keywords:

Burgers’ equation; nonlinear system of equations; Newton’s scheme; high order iterative method## 1. Introduction

The numerical solution of nonlinear equations and systems is widely used in different branches of Science and Technology, as in the analysis of diffusion phenomena (see, for example, [1,2]), the study of dynamical models of chemical reactors [3], the numerical computing of the dominant eigenvalue of the neutron transport criticality problem (see [4]), the study of radioactive transfer [5], or in the orbit determination of artificial satellites (see [6]), among others.

Burgers’ equation arises in the theory of shock waves, in turbulence problems and in continuous stochastic processes. It has a large variety of applications: modeling of water in unsaturated oil, gas dynamics, heat conduction, elasticity, statics of flow problems, mixing and turbulent diffusion, cosmology, seismology, etc. (see [7] and the references therein).

We consider the one-dimensional Burgers’ equation
with initial condition $u(x,0)=f\left(x\right)$, $a<x<b$, and boundary conditions $u(a,t)={g}_{1}\left(t\right)$, $u(b,t)={g}_{2}\left(t\right)$, $t\ge 0$; where $Re$ is the Reynolds number of the viscous fluid flow problem and $f\left(x\right)$, ${g}_{1}\left(t\right)$ and ${g}_{2}\left(t\right)$ are given functions. Usually ε instead of $\frac{1}{Re}$ is used, for simplicity. Numerical difficulties have been experienced in getting the solution of Burgers’ equation for big values of ε.

$$\frac{\partial u}{\partial t}-\frac{1}{Re}\frac{{\partial}^{2}u}{\partial {x}^{2}}+u\frac{\partial u}{\partial x}=0,\phantom{\rule{2.em}{0ex}}a<x<b,\phantom{\rule{2.em}{0ex}}t>0$$

This problem shows a roughly similar structure to that of Navier-Stokes equations due to the form of the nonlinear convection term and the occurrence of the viscosity term. So, this can be considered as a simplified form of the Navier-Stokes equation. In recent years, many researchers have used different numerical methods specially based on finite difference, finite element boundary techniques and direct variational methods to solve this problem (see, for example, [7,8,9] and the references therein). It has been shown by Hopf and Cole that Burgers’ equation can be solved exactly for an arbitrary initial function $f\left(x\right)$. The transformation
relates $u=u(x,t)$ and $\text{\varphi}=\text{\varphi}(x,t)$, and if ϕ is a solution of the linear diffusion equation
then u is a solution of the quasilinear Burgers’ Equation (1). This transformation let us obtain the exact values of $u(x,t)$ because Equation (3) has a Fourier series solution, however its computational cost is very high and we only use it for comparing the precision of our results (as we will see in the Numerical Section, the integrals of the Fourier coefficients must be calculated). As we can see in [7], some researchers have taken advantage of this transformation and have applied the Crank-Nicolson scheme to the linear Equation (3) in order to obtain the values of $\text{\varphi}(x,t)$ and after it the values of $u(x,t)$. Other authors have used the implicit scheme, where $\frac{\partial u}{\partial t}$ is approximated by backward differences, however, due to its truncation error has order one, the precision of the results decreases.

$$u=-\frac{2\text{\epsilon}}{\text{\varphi}}\frac{\partial \text{\varphi}}{\partial x}$$

$$\frac{\partial \text{\varphi}}{\partial t}=\text{\epsilon}\frac{{\partial}^{2}\text{\varphi}}{\partial {x}^{2}}$$

In this paper, we apply a particular difference scheme called Crank-Nicolson type method, obtained by applying divided differences directly on the nonlinear Equation (1), needing to solve a nonlinear system of equations for obtaining the solution in each value of t. These nonlinear systems are solved by using fixed point iterative methods. In particular, we are going to use the classical Newton’s method of order two,
Traub’s method of order three (see [10])
and the method denoted by M5 (see [11]) of fifth order of convergence, whose expression is reminded in the following
In these iterative expressions, $F\left(x\right)$ denotes the vectorial real function that describes the nonlinear system and ${F}^{\text{'}}\left(x\right)$ is the associated Jacobian matrix.

$$\begin{array}{c}\hfill {x}^{(k+1)}={x}^{\left(k\right)}-{\left[{F}^{\text{'}}\left({x}^{\left(k\right)}\right)\right]}^{-1}F\left({x}^{\left(k\right)}\right)\end{array}$$

$$\begin{array}{c}\hfill {y}^{\left(k\right)}={x}^{\left(k\right)}-{\left[{F}^{\text{'}}\left({x}^{\left(k\right)}\right)\right]}^{-1}F\left({x}^{\left(k\right)}\right),\end{array}$$

$$\begin{array}{c}\hfill {x}^{(k+1)}={y}^{\left(k\right)}-{\left[{F}^{\text{'}}\left({x}^{\left(k\right)}\right)\right]}^{-1}F\left({y}^{\left(k\right)}\right)\end{array}$$

$$\begin{array}{ccc}\hfill {y}^{\left(k\right)}& =& {x}^{\left(k\right)}-{\left[{F}^{\text{'}}\left({x}^{\left(k\right)}\right)\right]}^{-1}F\left({x}^{\left(k\right)}\right),\hfill \end{array}$$

$$\begin{array}{ccc}\hfill {z}^{\left(k\right)}& =& {y}^{\left(k\right)}-5{\left[{F}^{\text{'}}\left({x}^{\left(k\right)}\right)\right]}^{-1}F\left({y}^{\left(k\right)}\right),\hfill \\ \hfill {x}^{(k+1)}& =& {z}^{\left(k\right)}-\frac{1}{5}{\left[{F}^{\text{'}}\left({x}^{\left(k\right)}\right)\right]}^{-1}[-16F\left({y}^{\left(k\right)}\right)+F\left({z}^{\left(k\right)}\right)]\hfill \end{array}$$

The stability of the method is numerically analyzed and found to be unconditionally stable. The presented method has the accuracy of second order in space and time. We test our method by using several values of ε and different initial conditions in order to analyze its stability and consistence, and for doing this it is compared with other known methods and with the exact solution. From these numerical results several interesting conclusions are obtained.

## 2. Development of the Procedure

The easiest way to discretize a partial differential equation problem (like we have on Equation (1)) consist on reducing the continuous domain into an equispaced finite number of points $({x}_{i},{t}_{j})$ located on the nodes of a uniform rectangular mesh. The partial derivatives of u can be estimated now by the following difference quotients
which are called forward differences and have a truncation error of order 1. The partial derivatives of u can also be obtained by
which are backward differences and have also a truncation error of order 1. Finally the partial derivatives of u can also be obtained by the following
which are called central differences and have a truncation error of order 2. In these expressions, ${u}_{i,j}$ denotes the approximated value of the unknown u function on $({x}_{i},{t}_{j})$, $h=\frac{b-a}{n}$ and $k=\frac{{t}_{max}-{t}_{min}}{m}$ are respectively the space and time step on the mesh and n and m are respectively the space and time amount of subintervals to be considered. Then, we consider ${x}_{i+1}={x}_{i}+h$ and ${t}_{j+1}={t}_{j}+k$, where $i=0,1,...,n-1$ and $j=0,1,...,m-1$.

$$\frac{\partial u({x}_{i},{t}_{j})}{\partial x}\approx \frac{{u}_{i+1,j}-{u}_{i,j}}{h},\phantom{\rule{2.em}{0ex}}\frac{\partial u({x}_{i},{t}_{j})}{\partial t}\approx \frac{{u}_{i,j+1}-{u}_{i,j}}{k}$$

$$\frac{\partial u({x}_{i},{t}_{j})}{\partial x}\approx \frac{{u}_{i,j}-{u}_{i-1,j}}{h},\phantom{\rule{2.em}{0ex}}\frac{\partial u({x}_{i},{t}_{j})}{\partial t}\approx \frac{{u}_{i,j}-{u}_{i,j-1}}{k}$$

$$\frac{\partial u({x}_{i},{t}_{j})}{\partial x}\approx \frac{{u}_{i+1,j}-{u}_{i-1,j}}{2h},\phantom{\rule{2.em}{0ex}}\frac{\partial u({x}_{i},{t}_{j})}{\partial t}\approx \frac{{u}_{i,j+1}-{u}_{i,j-1}}{2k}$$

$$\frac{{\partial}^{2}u({x}_{i},{t}_{j})}{\partial {x}^{2}}\approx \frac{{u}_{i+1,j}-2{u}_{i,j}+{u}_{i-1,j}}{{h}^{2}},\phantom{\rule{2.em}{0ex}}\frac{{\partial}^{2}u({x}_{i},{t}_{j})}{\partial {t}^{2}}\approx \frac{{u}_{i,j+1}-2{u}_{i,j}+{u}_{i,j-1}}{{k}^{2}}$$

Let us apply Crank-Nicolson difference scheme, which consists of discretizing Equation (1) in two consecutive time instants of j and average them. In the first instant ${t}_{j}$ we approximate $\frac{\partial u}{\partial t}$ by the forward difference described on Equation (7) and in the second instant ${t}_{j+1}$ we do it by the backward one. At both instants we approximate $\frac{\partial u}{\partial x}$ and $\frac{{\partial}^{2}u}{\partial {x}^{2}}$ by using the central differences (described by Equations (9) and (10) respectively). By means of this procedure we get, for the first instant,
and for the second one,
After averaging both expressions, it results
where $A=\frac{1}{k}$, $B=\frac{1}{4h}$, $C=\frac{\text{\epsilon}}{2{h}^{2}}$ and the range of indices is $i=1,...,n-1$ and $j=0,1,...,m-1$. This expression (which has a truncation error of order 2) results on a m nonlinear systems of $n-1$ unknowns and $n-1$ equations. Each one of these systems comes after considering that we already know the value of ${u}_{0,j}$ and ${u}_{n,j}$, $\forall j$ (from the boundary conditions) and ${u}_{i,0}$, $\forall i$ from the initial condition. Let us notice that the values corresponding to the j subindex have been calculated on the previous systems and the only unknowns will be those with the $j+1$ subindex.

$$\frac{{u}_{i,j+1}-{u}_{i,j}}{k}-\text{\epsilon}\frac{{u}_{i+1,j}-2{u}_{i,j}+{u}_{i-1,j}}{{h}^{2}}+{u}_{i,j}\frac{{u}_{i+1,j}-{u}_{i-1,j}}{2h}=0$$

$$\frac{{u}_{i,j+1}-{u}_{i,j}}{k}-\text{\epsilon}\frac{{u}_{i+1,j+1}-2{u}_{i,j+1}+{u}_{i-1,j+1}}{{h}^{2}}+{u}_{i,j+1}\frac{{u}_{i+1,j+1}-{u}_{i-1,j+1}}{2h}=0$$

$$\begin{array}{c}\hfill B{u}_{i,j+1}{u}_{i+1,j+1}-B{u}_{i,j+1}{u}_{i-1,j+1}-C{u}_{i+1,j+1}+2C{u}_{i,j+1}-C{u}_{i-1,j+1}+A{u}_{i,j+1}=\\ \hfill C{u}_{i+1,j}+(A-2C){u}_{i,j}+C{u}_{i-1,j}+B{u}_{i,j}{u}_{i+1,j}-B{u}_{i,j}{u}_{i-1,j}\end{array}$$

## 3. Numerical Results

We are comparing our results with the exact solution, which can be obtained by the Hopf-Cole transformation mentioned on Equations (2) and (3). For doing this, we first calculate
and after it, by Equation (2), we obtain
where
Our goal is to obtain at least five exact decimals from the exact solution in order to prove the precision of our results, and for that we have seen that it is enough with two summands of the Fourier series, thereby only ${A}_{0}$, ${A}_{1}$ and ${A}_{2}$ will have to be calculated.

$$\text{\varphi}(x,t)={A}_{0}+\sum _{p=1}^{\infty}{A}_{p}{e}^{-{p}^{2}\text{\epsilon}{\pi}^{2}t}cos\left(p\pi x\right)$$

$$u(x,t)=2\pi \text{\epsilon}\frac{{\sum}_{p=1}^{\infty}{A}_{p}{e}^{-{p}^{2}\text{\epsilon}{\pi}^{2}t}p\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}sin\left(p\pi x\right)}{{A}_{0}+{\sum}_{p=1}^{\infty}{A}_{p}{e}^{-{p}^{2}\text{\epsilon}{\pi}^{2}t}cos\left(p\pi x\right)}$$

$${A}_{0}={\int}_{a}^{b}{e}^{-\frac{1}{2\text{\epsilon}}{\int}_{a}^{x}u(\xi ,0)d\xi}dx,$$

$${A}_{p}=2{\int}_{a}^{b}{e}^{-\frac{1}{2\text{\epsilon}}{\int}_{a}^{x}u(\xi ,0)d\xi}cos\left(p\pi x\right)dx,\phantom{\rule{4pt}{0ex}}\phantom{\rule{4pt}{0ex}}p=1,2,...$$

In the following examples we consider that the boundary conditions are zero, so we have that ${g}_{1}\left(t\right)=0$ and ${g}_{2}\left(t\right)=0$, what means that ${u}_{0,j}=0$ and ${u}_{n,j}=0$, $\forall j$. We can compare our results with the exact solution in this case, however our method has been found numerically to be unconditionally stable whatever the boundary conditions are. We use the R2013a version of Matlab with double precision for obtaining both approximated and exact solutions. The stopping criterion used is
and $u(0,t)=u(2,t)=0,t\ge 0$ where $\text{\alpha}=5$ and $\text{\beta}=4$.

$$\parallel {u}^{(k+1)}(x,{t}_{j})-{u}^{\left(k\right)}(x,{t}_{j})\parallel +\parallel F({u}^{(k+1)}(x,{t}_{j}))|<{10}^{-15}$$

**Example 1.**The initial and boundary conditions of Burgers’ equation for this example are
$$u(x,0)=\frac{2\text{\epsilon \beta}\pi sin\left(\pi x\right)}{\text{\alpha}+\text{\beta}cos\left(\pi x\right)},\phantom{\rule{2.em}{0ex}}0\le x\le 2$$

In the following tables we show, for different values of n, m and ε, the average number of iterations by using Newton’, Traub’s and M5 methods. We also show the greatest error, GE, that we obtain for every value of ${x}_{i}$ and ${t}_{j}$, where $i=0,1,...,n-1$ and $j=0,1,...,m-1$. The initial approximation of each nonlinear system to be solved is the solution of the previous one (and for the first system the initial approximation is the exact value of the initial condition). After that, we obtain the matrix of errors as the absolute difference between the approximated solution and the exact solution on each node, finally we call the greatest error as the greatest element from that matrix. While other authors prefer to only compare and give the value of the error in some points of the mesh we have preferred a more rigorous criterium, it consists on comparing all the error values from the matrix of errors and giving the value of GE, in this way we get a better idea about how bad is the worst possible situation and we know for sure that the error of the rest of the points will always be below this GE, the other way we can not know if there is some point that has a greater one.

An interesting result emerges from that error matrix if we consider a sufficiently high upper bound of the studied time. If we obtain the GE between all values of x on each time we observe that the GE increase really fast at first but after the peak is reached it starts to decrease. This result can be seen on the Figure 1.

In Table 1 we set the maximum time of the study, the number of time nodes and the Reynolds constant as ${t}_{max}=1$, $m=100$ and $\text{\epsilon}=\frac{1}{Re}=0.1$ and we try different amount of space nodes. In each case we obtain the already mentioned GE and the average number of iterations (Avg. iter.) needed to solve each nonlinear system by using Newton’, Traub’s and M5 methods.

From these results we observe that, in general, the more space subintervals we take (smaller space step), the less is GE.

In Table 2, we repeat the same process but in this case we set the number of space nodes and try different amount of time nodes. So ${t}_{max}=1$, $n=100$ and $\text{\epsilon}=\frac{1}{Re}=0.1$.

As we can observe, at first it seems that the more time subintervals we take, the less is GE (similarly as it happened with the previous table, where we tried different amount of space subintervals) however there is a limit for this which if we exceed the error starts to increase again. In this case, the minimum error is obtained for $m=20$.

**Figure 1.**Solution (

**a**–

**c**) and error progression curve (

**d**–

**f**) for different times and ε values, with $n=m=40$ and ${t}_{max}=1$.

**Table 1.**Comparison of the GE (greatest error) and methods iterations for different amount of space subintervals.

$n=10$ | $n=20$ | $n=40$ | $n=80$ | $n=160$ | |
---|---|---|---|---|---|

Greatest error | 0.09033 | 0.029932 | 0.0070658 | 0.0017149 | 0.00039376 |

Avg. iter. Newton | 4 | 4 | 4 | 4 | 4 |

Avg. iter. Traub | 3 | 3 | 3 | 3 | 3 |

Avg. iter. M5 | 3 | 3 | 3 | 3 | 3 |

$m=10$ | $m=20$ | $m=40$ | $m=80$ | $m=160$ | |
---|---|---|---|---|---|

Greatest error | 0.0043069 | 0.00015291 | 0.00083542 | 0.0010565 | 0.0011128 |

Avg. iter. Newton | 4.4 | 4.15 | 4 | 4 | 4 |

Avg. iter. Traub | 3.4 | 3.1 | 3 | 3 | 3 |

Avg. iter. M5 | 3 | 3 | 3 | 3 | 3 |

In Table 3, we repeat the process again, however now we study the effect of choosing a small, regular or big value of the Reynolds constant ($Re$), although we actually use $\text{\epsilon}=\frac{1}{Re}$ for simplicity. For doing this, we set the number of space and time nodes and the maximum time of study, so ${t}_{max}=1$, $n=40$ and $m=40$.

$\text{\epsilon}=1$ | $\text{\epsilon}=0.1$ | $\text{\epsilon}=0.05$ | $\text{\epsilon}=0.03$ | $\text{\epsilon}=0.02$ | $\text{\epsilon}=0.01$ | $\text{\epsilon}=0.005$ | |
---|---|---|---|---|---|---|---|

Greatest error | 0.22216 | 0.0068572 | 0.0035207 | 0.0021242 | 0.0014188 | 0.000708 | 0.000296 |

Avg. iter. Newton | 3.9 | 4 | 4 | 4 | 4 | 3.85 | 3 |

Avg. iter. Traub | 3.15 | 3 | 3 | 3 | 3 | 3 | 3 |

Avg. iter. M5 | 2.8 | 3 | 3 | 3 | 3 | 3 | 2 |

First of all, it has to be said that the maximum value of ε that can be used to keep the method working properly will depend of the amount of time subintervals we take, it happens because the bigger is the attenuation, more abrupt are the curves on the time axis. For keeping the error low, a finer mesh will be needed, what can be obtained by increasing the amount of time subintervals. From Table 3 we can numerically see this phenomenon, for a fast attenuation ($\text{\epsilon}=1$) the error is quite big due to the curve is very abrupt (as we said we could decrease the error of this case by taking more time subintervals), however the slower is the attenuation (lower values of ε), lower is GE. If we keep on taking lower values of attenuation (lower ε), the peak of the GE characteristic curve will not be reached and consequently the GE will be even lower (for this example it happens for $\text{\epsilon}=0.01$ and even more for $\text{\epsilon}=0.005$). See Figure 1 for a better understanding of these different situations.

Some conclusions can be obtained from Table 1, Table 2 and Table 3: to use a higher order method (like M5) does not represent a bigger advantage compared with a typical third order method (like Traub), however there is a significant difference between using a third order method compared to a typical second order method (like Newton). The more space subintervals we take, lower is GE; however, it does not happen with the amount of time subintervals, it means that there is an optimum amount of time subintervals that reduces the GE to its minimum but after it the error starts to increase again and the computational cost of having to solve so many systems does not make any sense (due to the increasing number of floating points operations, execution time,...). As we have seen from Table 3, if we want to study a situation with a big attenuation (low $Re$), we will need a higher amount of time subintervals compared to those situations where the attenuation is lower (it means that the computational cost for having to solve more nonlinear systems will be higher), however taking more space subintervals does not seem to help.

**Example 2.**Let us consider now the following initial and boundary conditions for Burgers’ equation

$$\begin{array}{c}\hfill u(x,0)=\left\{\begin{array}{cc}sin\left(\pi x\right),\hfill & 0<x\le 1\hfill \\ -\frac{1}{2}sin\left(\pi x\right),\hfill & 1<x\le 2\hfill \\ 0\hfill & 2<x\le 5\hfill \end{array}\right.\end{array}$$

In Figure 2 we see a plot of the approximated solution by our method at different times for $0\le x\le 5$ and for two different values of ε, and in Table 4 we show the numerical results of this solution for some values of $x,t$ and ε.

**Figure 2.**Solution for different times and ε values, both of them with $n=m=200$ and ${t}_{max}=10$. (

**a**) $\text{\epsilon}=0.1$; (

**b**) $\text{\epsilon}=0.01$.

x | t | $\text{\epsilon}=0.1$ | $\text{\epsilon}=0.01$ |
---|---|---|---|

1.5 | 2 | 0.44533 | 0.63548 |

4 | 0.28961 | 0.34573 | |

6 | 0.20701 | 0.23678 | |

8 | 0.16020 | 0.17999 | |

10 | 0.13041 | 0.14516 | |

3 | 2 | 0.02972 | 0.00000 |

4 | 0.14900 | 0.00470 | |

6 | 0.22312 | 0.47269 | |

8 | 0.22501 | 0.35971 | |

10 | 0.20562 | 0.29021 | |

4.5 | 2 | 0.00001 | 0.00000 |

4 | 0.00145 | 0.00000 | |

6 | 0.01171 | 0.00000 | |

8 | 0.03557 | 0.00000 | |

10 | 0.06242 | 0.02066 |

This results are very similar to those obtained for Mittal and Singhal on their paper (see [12]), where they transform Burgers’ equation into a system of nonlinear ordinary differential equations and solve it by Runge-Kutta-Chebyshev second order method. Although the techniques used to solve the problem are different, the results obtained by Mittal-Singhal and ours, showed in Table 4 and Figure 2 have the same order of magnitude. As we can see in Figure 2 it happens indeed that the lower is ε, slower is the attenuation. Although in this example we have used a piecewise initial condition, and it usually brings more problems, we still have obtained satisfactory results. Since we have considered a quite high upper bound of the studied time ($t=10$) we have had to take a bigger value of m in order to keep a fine mesh size. The results obtained in both Figure 2 and Table 4 are the same regardless if we use Newton’, Traub’s or M5 method.

## 4. Conclusions

As we have seen from Example 1 and Example 2 our method has the same precision than those methods which apply Hopf-Cole transformation, after it obtain the solution of the linearized equation and finally apply the inverse transformation for obtaining the solution of nonlinear Burgers’ equation. Our method is more straightforward due to it does not need to apply the mentioned transformation.

We also have taken advantage of the numerical results for obtaining some conclusions about the importance of the quantity of space and time subintervals for obtaining the solution mesh as well as the dependency between the error and the value of Reynolds number: if we want to decrease the error, we can take more space subintervals, although the computational cost will raise due to then there will be more unknowns per system to be solved. Another way to decrease the error is by taking more time subintervals, however we have seen that there is a limit from where if we keep on taking bigger values of m, the error increases again. At the same time, taking more time subintervals also means to raise the computational cost since then there will be more systems to be solved. We also have seen that choosing bigger values of attenuation (ε) involve bigger errors due to the faster is the attenuation, more abrupt is the curve and worse is the approximation; for solving this, more time subintervals should be taken but there is no need to take more space subintervals.

## Acknowledgments

The authors thank to the anonymous referees for their valuable comments and suggestions. This research was supported by Ministerio de Economía y Competitividad MTM2014-52016-C02-02 and FONDOCYT República Dominicana.

## Author Contributions

The contributions of all of the authors have been similar. All of them have worked together to develop the present manuscript.

## Conflicts of Interest

The authors declare no conflict of interest.

## References

- Lowrie, R.B. A comparison of implicit time integration methods for nonlinear relaxation and diffusion. J. Comput. Phys.
**2004**, 196, 566–590. [Google Scholar] [CrossRef] - An, H.; Mo, Z.; Xu, X.; Liu, X. On choosing a nonlinear initial iterate for solving the 2-D 3-T heat conduction equations. J. Comput. Phys.
**2009**, 228, 3268–3287. [Google Scholar] [CrossRef] - Bruns, D.D.; Bailey, J.E. Nonlinear feedback control for operating a nonisothermal CSTR near an unstable steady state. Chem. Eng. Sci.
**1977**, 32, 257–264. [Google Scholar] [CrossRef] - Willert, J.; Park, H.; Knoll, D.A. A comparison of acceleration methods for solving the neutron transport k-eigenvalue problem. J. Comput. Phys.
**2014**, 274, 681–694. [Google Scholar] [CrossRef] - Ezquerro, J.A.; Gutiérrez, J.M.; Hernández, M.A.; Salanova, M.A. Chebyshev-like methods and quadratic equations. Rev. Anal. Numér. Théor. Approx.
**1999**, 28, 23–35. [Google Scholar] - Andreu, C.; Cambil, N.; Cordero, A.; Torregrosa, J.R. Preliminary Orbit Determination of Artificial Satellites: A Vectorial Sixth-Order Approach. Abstr. Appl. Anal.
**2013**, 2013. [Google Scholar] [CrossRef] - Kadalbajoo, M.K.; Awasthi, A. A numerical method based on Crank-Nicolson scheme for Burgers’ equation. Appl. Math. Comput.
**2006**, 182, 1430–1442. [Google Scholar] [CrossRef] - Hassanien, I.A.; Salama, A.A.; Hosham, H.A. Fourth-order finite difference method for solving Burgers’ equation. Appl. Math. Comput.
**2005**, 170, 781–800. [Google Scholar] [CrossRef] - Wani, S.S.; Thakar, S.H. Crank-Nicolson type method for Burgers’ equation. Int. J. Appl. Phys. Math.
**2013**, 3, 324–328. [Google Scholar] [CrossRef] - Traub, J.F. Iterative Methods for the Solution of Equations; Chelsea Publishing Company: New York, NY, USA, 1982. [Google Scholar]
- Arroyo, V.; Cordero, A.; Torregrosa, J.R. Approximation of artificial satellites preliminary orbits: The efficiency challenge. Math. Comput. Model.
**2011**, 54, 1802–1807. [Google Scholar] [CrossRef] - Mittal, R.C.; Singhal, P. Numerical solution of Burger’s equation. Commun. Numer. Methods Eng.
**1993**, 9, 397–406. [Google Scholar] [CrossRef]

© 2015 by the authors; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/4.0/).