1. Introduction
The grey system theory was originally proposed by Professor Deng [
1,
2], with its core idea centered on integrating deterministic structures (the white part) with empirical inference (the black part) to model and analyze systems characterized by partially known and partially unknown information. Specifically, grey prediction models construct differential equations with well-defined mathematical structures as the white component, while incorporating unknown information inferred from observed data as the black component. This approach enables the discovery of essential system patterns from limited information. The most well-known model in grey system theory is GM(1,1).
Let an original non-negative and uniformly spaced function be
The classical GM(1,1) model employs an accumulated generating operation (AGO) on the original data sequence to reduce randomness and produce a distinctly monotonic increasing 1-AGO sequence, denoted as . Based on this transformed sequence, a first-order differential equation with a step-degree term is formulated. An exact solution is then derived for the equation containing unknown parameters. The least squares approach estimates the unknown parameters, while numerical methods solve the differential equation. Afterward, the derived parameters are substituted into the analytical solution to formulate the grey forecasting equation based on observed data. Ultimately, the inverse accumulated generating operation (IAGO) restores the original dataset, facilitating both reconstruction and future trend prediction.
The 1-AGO sequence
is given as follows
as
By this equation, the 1-AGO transformed sequence maintains an increasing trend, which leads us to express its mathematical form as follows
a denotes the grey developmental coefficient, and
b represents the grey action quantity, both parameters (
a and
b) need to be estimated. By incorporating the initial condition
, the analytical form of the differential equation’s solution can be obtained. The specific form is as follows
Thus, determining the original grey prediction model hinges on computing
a and
b. Hence, performing the integral operation on
, the original differential equation yields the following expression
which is
As we know, the difference between two consecutive terms is
; therefore, the above equation can be rewritten as
For simplicity of expression, we can define . Thus, the information of the first term in the equation has already been provided by the original data. By determining the value of for different values of k, and treating a and b as the least squares coefficients, we can apply the least squares method, and use the formula to obtain the estimated values of a and b.
However, the difficulty lies in the fact that the computation of
depends on the specific values of
a and
b, leading to circular reasoning. Therefore, we consider using numerical methods to approximate the values of
. In GM(1,1), the computation of
relies on linear polynomial interpolation
to obtain the numerical form of
, and is computed using integration methods, with the specific formula as follows
At this point, the numerical solution of the integral is substituted back into Equation (
8). For different values of
k, we assume
and
as the dependent variable, and
a and
b as the least squares fitting coefficients. A least squares fitting equation is established, where the normal equations are as follows
where
Substituting
into Equation (
5), and obtaining the final solution to the differential equation for the 1-AGO sequence, the specific form is as follows
Finally, by subtracting the estimated values of two consecutive terms in the 1-AGO sequence, the estimated values of the original data are derived, as expressed in the following formula
Theoretical analysis indicates that the estimation of the background value has a significant impact on the prediction performance of the model, which has led many researchers to focus their attention on the numerical integration methods for background value calculation. Li and Dai [
3] proposed using a higher-order Newton interpolation polynomial. Tang and Xiang [
4] established a piecewise quadratic interpolation polynomial to resolve this dilemma. Wang et al. [
5], Yang et al. [
6], and Zhu et al. [
7] applied different cubic spline interpolation techniques for high-precision background value estimation. In addition to improving background value estimation, many scholars have also focused on refining the grey action quantity. Xi et al. [
8] innovatively reconstructed the grey action quantity as a multi-parameter hyperbolic delay function. In [
9], a dual-parameter polynomial structure was designed. Both [
10,
11] incorporate polynomial terms into their models to enhance nonlinear modeling capabilities; however, due to computational constraints, the polynomial order is limited to three in both cases. In [
12], a composite grey action quantity with dynamic frequency response was developed, surpassing the traditional linear assumption; however, it introduced a significant computational burden.
The grey models generally used include the GM(1,1) [
13], grey model with the first order equation and N variables (GM(1,N)) [
14], and grey model with Nth order equation and one variable (GM(N,1)) [
15]. At the same time, various optimized algorithms have been derived from grey models, such as unbiased grey model (UGM(1,1)) [
16], fractional time-delayed grey model (FTDGM(1,1)) [
17], fractional grey model (FGM(1,1)) [
18], and the nonlinear Bernoulli model (NGBM(1,1)) [
19]. Among them, NGBM(1,1) has great potential for improvement and optimization. Zhang and Chen [
20] improved modeling accuracy by refining background values and initial conditions. However, during the background value optimization process, while expressing the curve as a nonhomogeneous exponential function, the exponential component of the time response was overlooked. Ding et al. [
21] proposed an improved stepwise Bernoulli model, utilizing the Simpson algorithm to enhance the estimation of
and employing the particle swarm optimization algorithm for parameter computation. Zhou et al. [
22] optimized both the exponent and initial conditions and achieved a certain level of accuracy improvement. In [
23], an innovative application of game theory in the parameter optimization of a novel two-stage Nash NGBM(1,1) was proposed. Zeng et al. [
24] developed a matrix-based NGBM(1,1) for interval sequences, deriving recursive prediction formulas using Cramer’s rule. In [
25], an optimization scheme based on the Fourier series was introduced, first computing error values using NGBM(1,1), followed by Fourier series filtering for secondary computation to achieve high-quality prediction. In [
26], NGBM(1,1) with linearly time-varying parameters was introduced for predicting non-equidistant time series. In [
27], Ma et al. established a multivariate NGBM(1,1), incorporating nonlinear equations and transforming them into a linear form, extending the computational approach of NGBM(1,1). In [
28], the paper focused on the physical background of data, replacing traditional time-series accumulation with physical operators, enabling high-quality prediction in real-world contexts. In [
29], based on the Bernoulli framework, Cheng and Bin proposed a new model in which the grey action quantity is improved to a quadratic polynomial and designed a least squares algorithm for parameter estimation.
However, GM (1,1), based on the Bernoulli equation mentioned above, only includes specified lower-order polynomials asan improvement of the grey action quantity, so the generalization ability of these resulting models is limited. Therefore, this work aims to incorporate unrestricted adaptive-order polynomials into the Bernoulli equation framework to achieve high-quality nonlinear fitting of the data. Additionally, nonlinear least squares estimation is employed instead of the normal equation method to reduce computational load and improve accuracy. The original differential equation is also discretized, eliminating the dependence on analytical solutions for prediction computation, which significantly reduces computational complexity.
In
Section 2, a novel Bernoulli model with polynomial-driven algorithm (BPGM(1,1)) will be established. In
Section 3, numerical examples will be compared and analyzed. In
Section 4, we present our conclusions.
2. Establish BPGM(1,1)
A nonlinear differential equation of the following form is known as a Bernoulli equation
For the given Bernoulli equation, dividing both sides by
yields
By introducing the substitution
, the Bernoulli equation can be transformed into the following form
Based on this, we introduce the Bernoulli equation into the grey prediction model and modify it as follows
Thus, we obtain the improved Bernoulli equation-based grey prediction model (BPGM(1,1)), which is expressed as follows
This Bernoulli equation integrates both the endogenous growth and exogenous driving mechanisms of the system. The polynomial term represents the time-varying external excitation, where the degree of the polynomial reflects the level of nonlinear fluctuations in the external influence. Meanwhile, captures the system’s nonlinear self-feedback.
When
n = 0 and
m = 0, the model reduces to GM(1,1), see the references [
1,
2]. When
m = 0, the model reduces to NGBM(1,1), see [
19].
Therefore, BPGM(1,1) possesses greater flexibility in solution formulation and lays the foundation for high-precision forecasting.
Through a similar derivation and computation process, the analytical solution of the above differential equation can be obtained in the following form
the value of the constant
C can be determined from the initial condition
.
At this point, the key issue lies in estimating the values of the parameters. Traditional methods integrate the original grey prediction equation, estimate the background value, and use the least squares method to solve the normal equation to obtain the estimated parameter values. Unlike traditional approaches, this work employs a nonlinear least squares estimation method to estimate
. This method provides a more robust parameter estimation for handling nonlinear data and data noise. We define the following formula as the residual expression
To solve this nonlinear least squares problem, the Levenberg–Marquardt (LM) algorithm [
30] is introduced. This is a numerical method that combines Gradient Descent and Newton’s Method. The specific formulation is as follows
where
represents the parameter vector,
J is the Jacobian matrix of the residual function,
r is the residual vector, and
is a damping parameter that adjusts the balance between the Gradient Descent and Newton’s Method.
By substituting the estimated parameters into the solution of the differential equation, the predicted values can then be obtained as follows
Since the value of n is unknown beforehand, directly substituting the estimated n obtained from the nonlinear least squares estimation into the equation may lead to cases where n is very close to 1. This would cause to approach 0, creating a risk of exponential explosion when substituted into the exact solution, thereby significantly affecting prediction accuracy.
Moreover, the original analytical solution is highly complex, requiring two nested iterative computations for each predicted value, leading to high computational complexity. Therefore, in this work, we do not use the exact analytical solution. Instead, we discretize the original differential equation into the following form
The equation can be rewritten as
Therefore, substituting the estimated parameters obtained from nonlinear least squares estimation into Equation (
22) improves the data utilization efficiency while reducing computational complexity and avoiding issues such as exponential explosion. Once the
time series is obtained, the original
sequence can be recovered through cumulative subtraction, thereby completing the forecasting process.
At this point, we have fully constructed BPGM(1,1). Summarizing the above discussion, we give the following algorithm 1 for establishing BPGM(1,1).
Algorithm 1 The implementation process of BPGM(1,1) |
Input: original non-negative and uniformly-spaced sequence . Output: grey prediction formula .
- 1:
Compute the 1-AGO sequence by Equation ( 3); - 2:
Set initial values for parameters ; - 3:
Define residual function based on Equation ( 20); - 4:
Solve nonlinear least squares optimization by LM algorithm; - 5:
Extract estimated parameters , , and ; - 6:
Compute the predicted sequence iteratively by Equation ( 24); - 7:
Obtain the predicted values of the original sequence by Equation ( 13).
|
4. Conclusions
In this work, we have improved the classical GM(1,1) through a dual-innovation mechanism, placing the grey prediction model within the broader framework of the Bernoulli equation and employing a data-driven polynomial grey action quantity, thereby formulating BPGM(1,1). Based on the Bernoulli model when , the model exhibits superlinear feedback capability, which can accommodate scenarios such as the Allee effect in population growth within ecological models and autocatalytic explosive growth in chemical reactions. When , the model represents sublinear feedback, making it suitable for fitting saturation growth under resource constraints, as well as sub-growth infection rates in disease transmission. When , the model exhibits inhibitory feedback, allowing it to capture the inverse regulatory effects of policy interventions on growth in economic models. By incorporating a time-varying externally driven adaptive-order polynomial, BPGM(1,1) effectively addresses nonlinear dynamic forecasting problems in complex systems. Moreover, we employ the nonlinear least squares LM algorithm to directly optimize the parameter vector, thereby avoiding the high computational cost associated with traditional algorithms that estimate the background value and solve parameters through matrix inversion. Benefiting from the superior performance of the LM algorithm, the computational complexity is significantly reduced. Furthermore, in the final solving stage, the numerical solution based on first-order explicit Euler discretization is used to replace the analytically derived double-nested summation, which also substantially lowers the time cost of computation. This facilitates the application of BPGM(1,1) to complex real-world scenarios that demand high model dynamism and real-time performance.
Future research could explore integrating other uncertainty modeling methods or concepts from machine learning to further improve prediction accuracy and enhance the model’s ability to generalize to complex scenarios. For instance, it might be valuable to consider adaptively selecting the initial parameters of the iterative process based on the characteristics of the data, to avoid falling into local minima. Additionally, the improved model could be applied to real-time data scenarios in dynamic systems, such as energy demand forecasting or environmental monitoring, to validate its timeliness.