Next Article in Journal
Zeolites Derived from Natural Kaolinite for CO2 Adsorption
Previous Article in Journal
Multi-Step Prediction of Wind Power Based on Hybrid Model with Improved Variational Mode Decomposition and Sequence-to-Sequence Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Hybrid Model Combining Improved VMD and ELM with Extended Maximum Correntropy Criterion for Prediction of Dissolved Gas in Power Transformer Oil

1
NARI Group Corporation (State Grid Electric Power Research Institute), Nanjing 211106, China
2
NARI Technology Co., Ltd., Nanjing 211106, China
3
National Key Laboratory of Risk Defense Technology and Equipment for Power Grid Operation, Nanjing 211106, China
4
School of Electrical Engineering, Xi’an University of Technology, Xi’an 710048, China
*
Authors to whom correspondence should be addressed.
Processes 2024, 12(1), 193; https://doi.org/10.3390/pr12010193
Submission received: 16 December 2023 / Revised: 9 January 2024 / Accepted: 12 January 2024 / Published: 16 January 2024
(This article belongs to the Section Advanced Digital and Other Processes)

Abstract

:
The prediction of dissolved gas change trends in power transformer oil is very important for the diagnosis of transformer faults and ensuring its safe operation. Considering the time series and nonlinear features of the gas change trend, this paper proposes a novel robust extreme learning machine (ELM) model combining an improved data decomposition method for gas content forecasting. Firstly, the original data with nonlinear and sudden change properties will make the forecasting model unstable, and thus an improved variational modal decomposition (IPVMD) method is developed to decompose the original data to obtain the multiple modal dataset, in which the marine predators algorithm (MPA) optimization method is utilized to optimize the free parameters of the VMD. Second, the ELM as an efficient and easily implemented tool is used as the basic model for dissolved gas forecasting. However, the traditional ELM with mean square error (MSE) criterion is sensitive to the non-Gaussian measurement noise (or outliers). In addition, considering the nonlinear non-Gaussian properties of the dissolved gas, a new learning criterion, called extended maximum correntropy criterion (ExMCC), is defined by using an extended kernel function in the correntropy framework, and the ExMCC as a learning criterion is introduced into the ELM to develop a novel robust regression model (called ExMCC-ELM) to improve the ability of ELM to process mutational data. Third, a gas-in-oil prediction scheme is proposed by using the ExMCC-ELM performed on each modal obtained by the proposed IPVMD. Finally, we conducted several simulation studies on the measured data, and the results show that the proposed method has good predictive performance.

1. Introduction

With the rapid development of the power industry, the healthy and stable operation of power transformers as key equipment of the power system is crucial to the safe operation of the entire power system. Therefore, it is essential to proactively forecast the operational status of power transformers and promptly implement appropriate measures to address any anomalies [1]. During the long-term operation of the transformer, the internal insulating oil will deteriorate, causing a small amount of hydrocarbon gas to dissolve in the insulating oil [2]. Generally speaking, with the extension of the operating time, the dissolved gas content will gradually accumulate, and the hydrocarbon gas is mostly flammable gas, which poses a great threat to the safe operation of the transformer [3]. Therefore, monitoring the change of dissolved gas content in oil becomes the basis for ensuring the safe and reliable operation of transformers [4].
In recent years, there has been a growing research focus on the analysis of dissolved gases within transformers. These research methods can be divided into the following categories: dissolved gas prediction method based on the autoregressive integral moving average (ARIMA) model [5], which is implemented based on mathematical-statistical principles through simple parameter estimation and model selection. However, the model requires a large amount of gas data as a sample and has high requirements for data stability, in addition, the model can only capture linear relationships, and the processing ability of nonlinear relationships is poor. However, in practical applications, sample data is lacking and the data fluctuates greatly, so the model is not suitable for practical engineering applications [6]. Dissolved gas prediction methods based on the grey model (GM) [7,8] only require a small amount of gas data to predict gas content over a period in the future. However, this method requires the original data to have a monotonic tendency to change, and it fails when the data oscillates or steeply rises (dip). In recent years, AI methods have been applied in this area, with back-propagation neural networks (BPNNs) [9] being one of the most used gas prediction methods, which has strong learning ability and data fitting ability. However, this method requires a large amount of sample data for model training, and the training time is long, which makes it easy to fall into the local optimal solution [10]. Recurrent neural networks (RNNs) are commonly used in time series data modeling and testing research [11], and their variants LSTM and GRU are developed on its basis [12,13], which can solve the problems of gradient vanishing and gradient explosion existing in the original RNN, and better consider the long-term dependence between data for time series and reduce the impact of information attenuation. However, both RNN and its variant network need to clearly set many time-series feature training data, and have strict requirements for the hidden layer neuron and layer number settings, too few settings make it difficult to achieve the expected results, and too many settings will greatly increase the training time. In the long-term prediction process, errors accumulate and increase prediction errors due to slow network model updates and increased repetitions [14]. Support vector machines (SVMs) and extreme learning machines (ELMs) are suitable for solving small-sample prediction problems [15,16], both of which are robust in processing nonlinear data and can effectively avoid over-fitting problems. Compared with ELM, SVM has more parameters, parameter tuning is more difficult and lacks the ability to capture time series, which is not suitable for modeling long-term time series dependencies. While the above methods have achieved some success in gas forecasting, there are still some shortcomings. On the one hand, when the test data fluctuates sharply, such as steep rises or drops, the network model cannot effectively capture this change process. On the other hand, since the transformer itself operates in a strong magnetic field, the data acquisition and transmission process may be interfered with by noise (pulses). Neural networks lack the ability to handle anomalous data, which may increase prediction errors [17]. Therefore, how to avoid the impact of data abruptness or oscillation on gas prediction deserves further study.
As an efficient method of data processing, modal decomposition (MD) methods [18,19] have been widely used in the field of photovoltaic and wind power prediction. This method can transform the mutation dataset into a smooth dataset in multimodal conditions, effectively reducing the impact of data mutation, but it is rarely used in the prediction of dissolved gas in transformer oil. Riaz et al. used empirical mode decomposition (EMD) technology to extract features from the signal first, and then input the extracted feature signal to SVM for classification processing and finally achieved good classification results [20]. However, the EMD method has certain shortcomings in extracting signal mutation information, which is prone to modal aliasing, resulting in the inability to completely decompose the noise (mutation) components in the signal [21]. Variational modal decomposition (VMD) can eliminate modal aliasing [19], and the frequency characteristics of the decomposed modal components are obvious, the stationarity is high, and the modes are independent of each other, which is conducive to data prediction and processing. The core idea of VMD is to convert the original signal into some non-recursive variational mode calculation process through decomposition, using effective bandwidth iteration and center frequency search to obtain the best decomposition signal [19]. Like the wind power prediction data set, the dissolved gas content in transformer oil often shows the characteristics of poor regularity and high fluctuation. VMD has high decomposition accuracy when processing complex data and can avoid modal aliasing, but it is necessary to set appropriate balanced iterative mode filtering (BIMF) parameters, penalty factors, and tolerance coefficients. Among them, BIMF parameters have the greatest impact on the decomposition accuracy of the model, when the BIMF value is large, it will lead to excessive decomposition and increase the amount of calculation, and when the BIMF value is too small, it will lead to insufficient extraction of the original data features [22]. In addition, although the modal signal obtained by VMD decomposition becomes smooth, there is still a sharp fluctuation (nonlinearity), which makes the machine learning model represented by ELM have a predictive risk. As an effective information theory learning criterion, the maximum correntropy criterion (MCC) [23] of the Gaussian kernel function can capture the high-order statistics of error information, which has potential advantages in dealing with nonlinear problems and suppressing noise (outliers) [24], and its practicability and effectiveness have been verified in the fields of state estimation and target tracking [25,26]. After the above analysis, we need to solve the following problems: First, it is difficult to adjust the parameters of the VMD decomposition method. The second is VMD decomposition, although it decomposes the mutation data into smoothed data in multiple modalities. However, the decomposed data is still a nonlinear fluctuation signal, which is not conducive to the direct use of the machine model.
The contributions of this work are summarized as:
(1)
The dissolved gas content in power transformer oil is characterized by poor regularity and large fluctuations. The VMD method will be used to handle this complex data, which has high decomposition accuracy and can avoid modal aliasing. However, the use of the VMD method requires setting the appropriate number of decomposition layers K and the penalty factor to obtain the optimal decomposition results. To solve this problem, we can introduce the MPA optimization method to adjust the parameters of VMD, to avoid the increase of prediction error caused by insufficient or excessive mode decomposition;
(2)
Given the high computational efficiency of ELM, it has great advantages in handling small sample data and un-modeled nonlinear approximation. However, classical ELM is not suitable for unmodeled nonlinear test systems that lack prior information, nor for systems with large fluctuations in data [27]. To improve the robustness of ELM in processing mutation data, we define a new ExMCC criterion and introduce it into the ELM to develop a novel learning model, which is called ExMCC-ELM;
(3)
Based on the above IPVMD and ExMCC-ELM, we propose a prediction model for dissolved gas in transformer oil. Firstly, the IPVMD decomposition of the dissolved gas sequence data in oil in the transformer is carried out to obtain an easy-to-process stationary data set, and then the ExMCC-ELM model is established for gas data prediction on the dataset. The simulation results show that the robust prediction method based on IVMD-ExMCC-ELM proposed in this paper can accurately predict the gas content under the condition of uncertain gas content abrupt changes and oscillation laws.

2. Improved VMD

2.1. Variational Mode Decomposition (VMD)

VMD is a signal processing method that aims to decompose complex signals into multiple intrinsic mode functions (IMFs) components. Each component has its own center frequency and limited bandwidth, and the sum of all components should be equal to the input signal. The principle of VMD can be described as follows:
(1) A constrained variational problem is constructed to obtain the decomposition results of the signal, where each component corresponds to an intrinsic mode function.
The constrained variational problem can be formulated as follows:
min u k , ω k k t δ t + j π t u k t e j ω k t 2 2
s . t . k u k t = f t
where u k = u 1 , u 2 , u k is the set of k IMF components obtained through VMD decomposition; ω k = ω 1 , ω 2 , ω k is the set of center angular frequencies for each IMF component; δ t is the unit impulse function; t represents the first-order derivative of the function with respect to time t; represents convolution operation.
(2) To solve the constrained variational problem in the VMD method, it can be transformed into an unconstrained variational problem by introducing a quadratic penalty factor α and Lagrange multiplier operator λ . Specifically, introducing the quadratic penalty factor α ensures a higher accuracy of signal reconstruction in the presence of Gaussian noise. Introducing the Lagrange multiplier operator λ ensures that the constraints are satisfied during the solving process. Based on this, the augmented Lagrange function Γ can be derived as follows:
Γ u k , ω k , λ = α k t δ t + j π t u k t e j ω k t 2 2 + f t k u k t 2 2 + λ t , f t k u k t
To further solve the variational problem mentioned above, the alternating direction method of multipliers (ADMM) can be used. In ADMM, a set of variables ( u k n + 1 , ω k n + 1 and λ k n + 1 ) are updated iteratively until convergence is reached to find the saddle point of the augmented Lagrange expression. The corresponding update expressions are as follows:
u ^ k n + 1 ω = f ^ ω i k u ^ i ω + λ ^ ω 2 1 + 2 α ω ω k 2
ω k n + 1 = 0 ω u ^ k ω 2 d ω 0 u ^ k ω 2 d ω
λ ^ n + 1 ω = λ ^ n ω + τ f ^ ω k u ^ k n + 1 ω
To obtain the K IMF components, the iterative update of variables should be performed until the iteration-stopping condition k u ^ k n + 1 u ^ k n 2 2 / k u ^ k n + 1 2 2 < ε is satisfied. Once the iteration-stopping condition is met, the iterative updates can be terminated, and the K IMF components can be obtained. In general, how to select the optimal K is still an important issue for the performance of the VMD, this work will use a novel optimization method to address this problem.

2.2. Marine Predators Algorithm (MPA)

MPA is a new type of meta-heuristic optimization algorithm that is initialized in the same way as most meta-heuristic search algorithms. The initial solution is evenly distributed over the search space on the first test. This approach helps the algorithm better explore the search space to find a better solution.
In the MPA process, optimization is based on different speed ratios of the three stages of the predation-predation cycle. The three-stage prey position update pattern can be described as follows [28]:
(1)
When the rate is relatively high, or the prey moves faster than the predator, the following mathematical model can be used to apply the rule:
When I t e r < 1 3 M a x _ I t e r ,
s t e p s i z e i = R B ( E l i t e i - Pr e y i ) ,   i = 1 , , n
Pr e y i = Pr e y i + P . R s t e p s z e i
where R B is a random vector containing a normal distribution based on Brownian motion; P is a constant equal to 0.5; R is the vector of random numbers in [0, 1]; means multiplying by entries; I t e r is the current number of iterations; M a x _ I t e r is the maximum number of iterations.
(2)
The following model is applied when the rate ratio is relatively low or when predators and prey move at almost the same speed:
When 1 3 M a x _ I t e r < I t e r < 2 3 M a x _ I t e r , for the first half of the population:
s t e p s i z e i = R L ( E l i t e i - R L Pr e y i ) ,   i = 1 , , n / 2
For the second half of the population:
s t e p s i z e i = R B ( R B Pr e y i - Pr e y i ) ,   i = n / 2 , , n
where R L is a vector of random numbers based on the Lévy distribution, indicating Lévy motion, while prey updates its position based on the predator’s movement in Lévy motion; Pr e y i is calculated in the same way as Equation (8).
(3)
When the low rate is faster than the predator than the moving prey, the following model is applied:
When I t e r > 2 3 M A X _ I t e r ,
s t e p s i z e i = R L ( R L E l i t e i Pr e y i ) ,   i = 1 , , n
Pr e y i = E l i t e i + P C F s t e p s z e i
The product of R L and E l i t e i simulates the movement of the predator in the Lévy strategy, while increasing the step size at the E l i t e i position simulates the movement of the predator to help update the predator’s position; and C F = 1 I t e r M a x _ I t e r 2 I t e r M a x _ I t e r is adaptive parameters used to control the predator’s movement step.
(4)
Vortexes and fish gathering facilities (FADs) also have an impact on the behavior of marine predators, and they are often considered locally optimal solutions in the search space. With this in mind, introducing the FADs effect during the simulation process can avoid falling into the local optimal solution. Therefore, we can mathematically represent the FADs effect:
Pr e y i = Pr e y i + C F X min + R ( X max X min ) U r F A D s Pr e y i + F A D s ( 1 r ) + r ( Pr e y r 1 Pr e y r 2 ) r F A D s
where F A D s = 0.2 is F A D the possibility of influencing the optimization process; U is a binary vector containing 0s and 1s; r is a consistent random number in the range [0, 1]; X min and X max are vectors representing the upper and lower bounds of the dimension, respectively; The subscripts r 1 and r 2 are random indexes of the prey matrix.

2.3. MPA-Optimized VMD

In the process of optimizing VMD parameters using MPA, the selection of the appropriate fitness function plays a crucial role in the optimization results. In this paper, the overall orthogonality index IO of the decomposed IMF is selected as the fitness function for parameter optimization. IO reflects the degree of confusion between the orthogonality and decomposition results of the IMF components, with smaller values indicating better orthogonality. In the MPA algorithm, the prey position parameter corresponds to the value of the variable.
IO can be expressed by the following formula:
IO = 1 exp R
where R is the spectral radius of the correlation coefficient matrix between the IMF components, and exp is the natural exponential function.
The correlation coefficient matrix between the various IMF components obtained by the IMF decomposition is denoted as C , and the spectral radius R of the correlation coefficient matrix is defined as:
R = max λ
where λ denotes the modulus of the eigenvalues of the matrix C .
By selecting IO as the fitness function, the MPA algorithm will minimize the IO value by adjusting the value of the parameter, so as to make the IMF components after VMD decomposition more orthogonal, and the decomposition result more structured and interpretable. The prey position parameter is the parameter that needs to be optimized, and the optimal solution to minimize the fitness function (IO value) is found by constantly updating the value of the prey position parameter, and then the optimal number of decomposition layers K and the penalty parameter α. This can improve the decomposition effect of the VMD algorithm and the accuracy of signal processing. The specific process is as follows [29]:
(1)
Set the initial parameters of the VMD and MPA algorithms;
(2)
Initialize the number of predator populations and the number of iterations in the MPA algorithm. Considering the influence of population size and number of iterations on optimization accuracy and computational efficiency, this paper defines the population size as 20 and the number of iterations as 50;
(3)
Initialization produces initial prey, and predators build an elite matrix (predator vector). Note that both predators and prey consider themselves search agents because when predators search for prey, the prey is also looking for food. Calculate the fitness value corresponding to that time;
(4)
The MPA optimization process, as described in Section 2.2;
(5)
After updating the predator’s position, the corresponding fitness is calculated and compared to the previous fitness value. Choose the best fitness position as the top predator position;
(6)
Repeat steps (1–5) until the termination condition is met and output the apex predator position coordinates, which are input into the VMD as the optimal parameters for the decomposition signal.
The flowchart of MPA to optimize VMD is given in Figure 1.
It should be noted that in situations where the input dataset exhibits weak correlation, managing the IMF components can be challenging. One approach to addressing this issue is to employ advanced feature selection methods and incorporating domain knowledge can further enhance the management of IMF components in such challenging scenarios. Additionally, one can try to use them in combination with other IMF components that are more correlated. By synthesizing multiple IMF components, the accuracy and stability of the predictive model can be improved.

3. Extreme Learning Machine with Extended Maximum Correntropy Criterion

3.1. Extended Maximum Correntropy Criterion

Correntropy is an effective measure of the generalized similarity between two random variables. Given two random variables y , the correntropy can be defined as [23]:
V x , y = E κ σ x y
where κ σ represents the kernel function with kernel width; E is the expectation operator. The Gaussian kernel is usually used as a kernel function in (18), expressed as:
κ σ x y = G σ x y = 1 2 π σ 2 exp x y 2 2 σ 2
One can see from (17) that the correntropy only contains a second-order moment of error (SOME) with a single kernel width in the exponent part, which may lead to unsuitable performance for complex environments. Therefore, this paper will design a new extended correntropy consisting of two SOME with two different kernel widths in the exponent part, called extended correntropy (ExC), which is defined as:
V C ( x , y ) = E [ κ C , σ ( x y ) ]
κ C , σ ( x y ) = exp [ ( γ ( x y ) 2 2 σ 1 2 + ( 1 γ ) ( x y ) 2 2 σ 2 2 ) ]
where γ is the parameter that determines the ratio of two kernels; σ 2 and σ 2 are different kernel widths, and when σ 1 = σ 2 or ( γ 0 ( o r 1 ) ), the ExC will degenerate to the original correntropy.
In practice, the probability density function of the two random variables is unknown, and the number of samples { x i , y i } i = 1 N is finite. Therefore, the sample mean estimate in Equation (20) is defined as:
V C * ( x , y ) = 1 N i = 1 N exp γ x i y i 2 2 σ 1 2 + ( 1 γ ) x i y i 2 2 σ 2 2 = 1 N i = 1 N κ C , σ ( x i y i ) = 1 N i = 1 N κ C , σ ( e i )
where N represents the number of samples, x i and represents the i-th element of random variables X and Y .
Figure 2 shows the function curve under different parameters, as can be seen from Figure 2a, when y is constant and σ 2 = 0.5 , the curve gradually becomes flat as σ 1 increases. It can be seen from Figure 2b that when the kernel width value is constant, the function curve tends to be flat with the crease of y , but the overall trend still meets the requirements of convergence and boundedness.
Finally, in conjunction with Figure 2, we summarize the general properties of ExMCC as follows [30]:
Property 1: Symmetry, V C ( x , y ) = V C ( y , x ) .
Property 2: V C ( x , y ) is positive and bounded, 0 < V C ( x , y ) 1 .
Property 3: If 0 γ 1 , then V C ( x , y ) represents a combined second-order statistic that maps the feature space.
Based on the above properties, we can easily conclude that ExC is an extended form of correntropy. Compared with correntropy, when we select the appropriate free parameters in ExC, it is not only more robust to abnormal data, but also has better convergence speed and stability. Similar to correntropy, the EXC is capable of suppressing the interference of non-Gaussian noise due to its inherent robustness to outliers and its ability to capture higher-order statistical moments beyond the mean and variance. Unlike traditional measures such as the MSE, which are sensitive to outliers and assumptions of Gaussianity, the ExC is based on the probability distribution of the data, allowing it to effectively mitigate the impact of non-Gaussian noise by focusing on the underlying statistical structure of the data rather than solely relying on the second-order statistics. This enables ExC to better capture the true underlying signal in the presence of non-Gaussian noise, making it a valuable tool for robust signal processing. Like the maximum correntropy criterion (MCC), the ExC can also be used as a learning criterion in machine learning fields, which will be denoted as extended MCC (ExMCC) in this work.

3.2. Extreme Learning Machine

An extreme learning machine (ELM) is a single hidden layer feed-forward neural network. The number of hidden layer nodes is usually set artificially, while the input weights and biases are determined randomly. In the process of learning and calculation, the weights and biases are not iteratively calculated, and the optimal solution can be calculated when training data is available. Therefore, compared with traditional feedforward networks such as BP, ELM has more advantages in time series data prediction such as fast training speed, strong generalization ability, fewer hyperparameters, and high accuracy. Recently, ELM has been widely used in the forecasting of time series data, and satisfactory prediction results have been obtained. The calculation procedure of the ELM can be described as follows:
Given a set of sample data I = x , y , where x = x 1 , x 2 , , x N T and y = y 1 , y 2 , , y N T represent N-dimensional input and desired output vectors, respectively. Therefore, the hidden layer output of ELM can be expressed as:
h ( x i ) = G ( w i , x i , b i )          = G ( w i x i + b i )
where h i ( ) is the output of the i-th hidden layer node; w i is the weight between the input layer and the hidden layer; b i is the hidden layer bias; G ( ) is the hidden layer activation function, and in this study, sin is used as the activation function.
Then, we further get the ELM output as:
Y = i = 1 L β i h i ( x ) = H ( x ) β
where Y = y ( x 1 ) , y ( x 2 ) , , y ( x N ) T is the output matrix; H ( x ) = h ( x 1 ) , h ( x 2 ) , , h ( x N ) T , ( h ( x ) = h 1 ( x ) , h 2 ( x ) , , h L ( x ) T ) is the hidden layer output matrix; β = β 1 , β 2 , , β N T is the weight between the output layer and the hidden layer.
By solving the Moore-Penrose generalized inverse matrix of H ( x ) , and training the ELM by using the input weights and hidden layer bias, the output weights are obtained:
β = H 1 ( x ) Y
When employing ELM for time series data forecasting, in certain instances, the weight of the final output of ELM may exhibit significant fluctuations due to substantial changes in forecasted data, consequently leading to poor stability of the forecast results. To address this issue, this paper introduces the ExMCC robust criterion, designed to be impervious to abnormal data, thereby enhancing the robustness of ELM in the presence of abnormal data and effectively mitigating the issue of weight fluctuations caused by abnormal data.

3.3. ELM with ExMCC

In this section, we use ExMCC as a learning criterion for ELM to develop a robust ELM model named ExMCC-ELM, so that when the training data is contaminated, the ELM output can be guaranteed to have a stable optimal weight. The detailed derivation process of ExMCC-ELM is as follows.
Based on the theoretical basis of ExMCC, we rewrite Equation (21) as:
J C ( β ) = 1 N i = 1 N exp γ d i h i β i 2 2 σ 1 2 + ( 1 γ ) d i h i β i 2 2 σ 2 2 = 1 N i = 1 N κ C , σ ( D Y ) = 1 N i = 1 N κ C , σ ( e i )
where D = d 1 , d 2 , , d N T represents the desired output; Y = y ( x 1 ) , y ( x 2 ) , , y ( x N ) T represents actual output; e i = d i y i = d i h i β i , d i , h i and represents the i-th elements of D , H and β .
The optimal solution of the weight β is obtained using the maximization cost function J C . Specifically, J C finds the differential with respect to the gradient method and makes it zero.
J C M C β = 0 1 N i = 1 N exp γ d i h i β i 2 2 σ 1 2 + ( 1 γ ) d i h i β i 2 2 σ 2 2 * γ σ 1 2 d i h i β i h i + 1 γ σ 2 2 d i h i β i h i = 0 i = 1 N c ( e i ) * γ σ 1 2 + 1 γ σ 2 2 h i T d i γ σ 1 2 + 1 γ σ 2 2 h i T h i β i = 0 β = i = 1 N c ( e i ) * γ σ 1 2 + 1 γ σ 2 2 h i T h i 1 i = 1 N c ( e i ) * γ σ 1 2 + 1 γ σ 2 2 h i T d i β = H T C H 1 H T C D
c ( e i ) = exp γ d i h i β i 2 2 σ 1 2 + ( 1 γ ) d i h i β i 2 2 σ 2 2
where C = diag c ( e i ) represents the diagonal matrix.
Obviously, Equation (26) is a fixed-point equation β , so the fixed-point iterative method is used to calculate the optimal solution for β . The overall calculation flow of the proposed ExMCC-ELM algorithm is organized in Algorithm 1. In addition, in ExMCC-ELM, it will degenerate into MC-ELM when γ = 0 ( o r 1 ) or σ 1 = σ 2 . Using ExMCC as a learning criterion for ELM can improve ELM robustness by utilizing its ability to model and suppress non-Gaussian noise. Traditional ELM assumes that both input data and noise follow a Gaussian distribution, but in practical applications, non-Gaussian noise such as outliers can often interfere with the data. ExMCC allows for the modeling of non-Gaussian noise. By maximizing the ExC between features and noise, the influence of outliers on the ELM model can be reduced, better adapting to the characteristics of non-Gaussian noise. Therefore, using ExMCC as the learning criterion for ELM can improve ELM robustness in dealing with abnormal data.
Algorithm 1. ExMCC-ELM model. Initialize ExMCC-ELM: σ 1 , σ 2 , λ , β i , b i
Training phase:
  Training input: x = x 1 , x 2 , , x N 1 T ( x i , x i + 1 ) , 0 < i < 250
  Training output: y = y 1 , y 2 , , y N 1 T x i + 2
   Final   output :   β , b i
Testing phase:
  Testing input: β , b i and x = x 251 , x 252 , , x N 2 T ( x i , x i + 1 ) , 250 < i < 300
  Testing output: y = y 1 , y 2 , , y N 2 T x i + 2
  1. For k 1 to M ( M = 10 )
  2. Calculate the error vector e i , based on the initial β i , b i : e i = d i h i β i
  3. Calculate the diagonal matrix C :   C = diag c ( e i )
  4. Calculate the output weight β :   β = H T C H 1 H T C D
  5. Determine whether the loop ends: J C ( β k ) J C ( β k 1 ) < ω , ( ω = 0.00001 )
  6. End For
By using the decomposed data obtained from the VMD with MPA optimization developed in Section 2, one can train the proposed model ExMCC-ELM, and then use the test data to evaluate its prediction capability. Figure 3 gives a diagram to illustrate the complete prediction process by using the proposed method for dissolved gases forecasting.

4. Gas-in-Oil Prediction Scheme via ExMCC-ELM and IPVMD

4.1. Data Construction

Given the large variety of internal gases in power transformers, the connection between individual gases is not tight and the amount of data is lacking. To solve this problem, this paper uses the method of constructing data pairs to predict various gases.
Firstly, the dissolved gas data in the oil ( H 2 , C 2 H 4 , C 2 H 2 , C 2 H 6 , C H 4 ) [31] are used as inputs for IPVMD decomposition, and the decomposition vectors of these five gases were obtained respectively.
The raw dissolved gases in oil data are shown in Figure 4.
To verify whether IPVMD can efficiently extract the temporal information of the gas content sequence. Taking H 2 gas as an example, the IPVMD method was used to perform modal decomposition of the raw H 2 data. Figure 5 illustrates the convergence curve of the MPA-optimized VMD decomposition H 2 process. Figure 6 shows the raw data of H 2 and the five-modal data IMF 1 , IMF 2 , , IMF 5 after IPVMD decomposition and the optimized K is selected as 5 by using the MPA algorithm. It can be seen from Figure 5 that the original strong fluctuation and irregular data after VMD decomposition become smooth in the five modes, which is helpful for training and testing the data of each modality.
Then, the limited data sample I M F k = x 1 , x 2 , , x 300 ( k = 1 : K ), K is the number of modes to be decomposed by the VMD, which is determined by the MPA algorithm) is divided into training data and testing data of the ExMCC-ELM model, respectively, where the first 250 points { x i } i = 1 250 are used as the training set and the last 50 points { x i } i = 251 300 are used as the testing set. Training and testing use data pairs constructed in the form of input and output, respectively.

4.2. MPAVMD-ExMCC-ELM Model Prediction Process

In this paper, the VMD decomposition method is used to decompose the nonstationary timing signal of dissolved gas in transformer oil and convert it into a stationary signal in multi-mode. Secondly, in the face of the uncertainty of signal decomposition, the IO optimization method is further used to optimize the parameters of VMD, and the most satisfactory signal decomposition results have been obtained. Thirdly, the lack of sample data, strong volatility, and poor regularity are not conducive to the characteristics of network model training and testing. This paper introduces ExMCC optimization criteria in the ELM model to overcome the above shortcomings. In this study, the original gas data was divided into training and testing parts, and the
Step 1: Initialize IPVMD and ExMCC-ELM
The exMCC-ELM model was trained with the first 250 sets of data, the last 50 sets of data were tested, and the gas data involved was included H 2 , C 2 H 4 , C 2 H 2 , C 2 H 6 , C H 4 .
The specific steps are as follows:
Step 2: The gas data H 2 , C 2 H 4 , C 2 H 2 , C 2 H 6 , C H 4 was decomposed by IPVMD respectively to obtain its decomposition vector IMF 1 , IMF 2 , , IMF K . K is the number of modes of VMD decomposition, which is determined by the VMD optimization algorithm, and the K value may be different for different gases.
Step 3: The obtained decomposition vector I M F k = x 1 , x 2 , , x 300 ( k = 1 : K ) is used as the training data and test data of the ExMCC-ELM model respectively, where the first 250 points are used as the training set and the last 50 points are used as the test set. Training and testing use data pairs constructed in the form of input and output, respectively. That is, the input and output of the ExMCC-ELM model training process are: x = x 1 , x 2 , , x N 1 T ( x i , x i + 1 ) , 0 < i < 250 , y = y 1 , y 2 , , y N 1 T x i + 2 ; The input and output of the test process are x = x 251 , x 252 , , x N 2 T ( x i , x i + 1 ) , 250 < i < 300 , y k = y 1 , y 2 , , y N 2 T x i + 2 .
Step 4: The prediction results of each decomposition vector IMF 1 , IMF 2 , , IMF K are reconstructed to obtain the prediction results of the raw gas: y = 1 K k = 1 K y k .
In this paper, the performance of the prediction method is evaluated using mean absolute error (MAE), mean squared error (MSE), root mean square error (RMSE), and symmetric mean absolute percentage error (SMAPE), defined as follows:
M A E = 1 N i = 1 N | d i y i | M S E = 1 N i = 1 N d i y i 2 R M S E = 1 N i = 1 N d i y i 2 S M A P E = 100 % N i = 1 N d i y i d i + y i / 2
where N is the sample size; d i is the expected value; y i is the actual predicted value.
The structure diagram of the IPVMD-ExMCC-ELM model for predicting dissolved gases in transformer oil can be seen in Figure 7.

5. Experiments and Analysis

In this study, BP, ELM, IPVMD-ELM, and ExMCC-ELM were used as the control experimental groups to reflect the performance of the proposed method under various test conditions. All simulations in this paper are performed using MATLAB 2022a in i5-8250 U and 1.6 GHz CPUs.

5.1. Prediction and Analysis of H 2

To verify whether IPVMD can effectively extract the time information of the gas content sequence. In this section, taking H 2 gas as an example, the original H 2 data were decomposed using the IPVMD method, and then tested and verified based on ELM and the improved ExMCC-ELM model.
To demonstrate the effectiveness and rationality of the experiment, we use BP, ELM, and ExMCC-ELM as comparative tests to verify the effectiveness of the proposed method based on the gas data in the original state. In this paper, BP is set to the single hidden layer, the number of neurons in the hidden layer is set to 10, and the number of neurons in the hidden layer of ELM, IPVMD-ELM, ExMCC-ELM, and IPVMD-ExMCC-ELM is set to 5, 5, 3, and 3, respectively, to ensure that each method is in the best prediction state. Figure 8 shows the prediction results and errors of various methods for 50-day H 2 data, as shown in the figure, the prediction results based on the original H 2 data have a large error, compared with BP and ELM models, the ELM model based on ExMCC optimization criterion improves the prediction results to a certain extent, because the ExMCC criterion is not sensitive to strongly fluctuating data, indicating that ExMCC is effective for the improvement of ELM model. However, this improvement did not allow the ExMCC-ELM model to maintain its intended performance throughout the process. This suggests serious drawbacks to directly applying raw H 2 data for the next time series forecast. However, based on the H 2 data after IPVMD modal decomposition, this problem can be effectively solved, and the decomposed data can effectively solve the influence of data mutation on the original ELM model, which greatly improves the prediction accuracy.
It can be seen from the data in Table 1 that the MAE and MSE results of IPVMD-ELM are 32.13% and 66.64% lower than those of ELM prediction results, respectively, and the IPVMD-ELM prediction error is relatively stable and there is no huge mutation. The IPVMD-ExMCC-ELM model synthesizes all the characteristics of IPVMD and ExMCC optimization criteria, showing excellent robustness, estimation accuracy, and stability.

5.2. Prediction and Analysis of Other Gases

After the detailed analysis in the previous section, we have obtained the optimal performance of the IPVMD-ExMCC-ELM method in decomposing and predicting the original H 2 gas. However, in power transformer oil, there are usually multiple hydrocarbon gases present, such as CH 4 , C 2 H 2 , C 2 H 4 , and C 2 H 6 , with weak correlations among them. This weak correlation makes it challenging to estimate the individual gas concentrations jointly. Therefore, further validation is needed to assess the applicability of the proposed method for predicting different gas data. Figure 9, Figure 10, Figure 11 and Figure 12 demonstrate the prediction results and errors for several different gases.
Table 2 lists the evaluation indicators for each prediction result, and the data in the table shows that the proposed methods exhibit good predictive performance compared to the IPVMD-ExMCC model, which has higher accuracy. In C 2 H 4 with the highest prediction accuracy, MAE, MSE, and SMAPE were improved by 27.25%, 51.01%, and 30.00%, respectively. In CH 4 with lower prediction accuracy, MAE, MSE, and SMAPE were improved by 38.08%, 44.20%, and 34.48%, respectively. It can be seen that the proposed IPVMD-ExMCC-ELM model can accurately predict the content of dissolved gas in power transformer oil.

5.3. Analysis of the Impact of Different Parameter Settings on the Prediction Effect

In this section, we will investigate the influence of kernel width σ and scale factor γ on the proposed IPVMD-ExMCC-ELM model. As shown in Figure 13, when the scale factor is fixed, the prediction accuracy of the model is weakly affected by different kernel widths. In contrast, the estimation error obtained when σ 1 = σ 2 is the largest, which fully indicates that the proposed ExMCC optimization criterion is better than the original MC. Figure 14 shows that when the core width is fixed, the prediction error increases as the γ increases. This is because a small kernel width is conducive to suppressing the effects of abnormal data or large fluctuations, and this suppression effect gradually decreases as the core width increases. Therefore, the estimation error gradually increases. After the simulation study in this paper, the entropy gravity with a small kernel width should dominate the ExMCC in all experiments, and the predicted effect of the proposed method is the best when σ 1 = 4 , σ 2 = 0.5 , and γ = 0.3 .

6. Conclusions

The power transformer is an important bridge for energy transmission and conversion in the power system, which is of great significance to the safe operation of the system. To realize the prediction of gas in transformer oil, a new IPVMD-ExMCC-ELM prediction model is proposed. Firstly, the IPVMD decomposition method is used to perform modal decomposition of the original gas dataset to obtain a stable data set that is easy to process. Then, the application of ELM in the prediction of dissolved gas concentration in transformer oil is introduced, and the proposed ExMCC optimization criterion is introduced into ELM. The simulation results show that the robust prediction method based on IPVMD-ExMCC-ELM proposed in this paper can accurately predict gas content under the condition of uncertain gas content mutation and oscillation law. The effectiveness and feasibility of the proposed method are verified by simulation analysis. In general, the advantages of IPVMD-ExMCC-ELM proposed in this paper are mainly reflected in the following aspects: Under the evaluation indexes of MAE, MSE, RMSE, and SMAPE, IPVMD-ELM and ExMCC-ELM have higher prediction accuracy than traditional ELM. On the other hand, IPVMD-ExMCC-ELM has higher predictive accuracy than IPVMD-ELM and ExMCC-ELM.

Author Contributions

Conceptualization, G.D. and J.L.; methodology, Z.S. and W.M.; software, Y.G.; validation, C.X. and J.L.; investigation, Z.S. and J.L.; resources, Y.G.; writing—original draft preparation, G.D. and Z.S.; writing—review and editing, G.D. and C.X.; supervision, J.L. and W.M.; project administration, J.L. and W.M.; funding acquisition, J.L. and W.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Science and Technology Project of Nanrui Group, grant number 524609220134, and the National Natural Science Foundation of Chinaunder Grant 61976175.

Data Availability Statement

Data are contained within the article.

Acknowledgments

We would like to express our sincere appreciation to the other members of the laboratory for the help provided in experiments and language editing. In addition, we sincerely thank the editors of this Journal and the anonymous reviewers for their review of our manuscript.

Conflicts of Interest

Authors Gang Du, Zhengming Sheng, Jiaguo Liu, Yiping Gao and Chunqing Xin were employed by the company NARI Group Corporation (State Grid Electric Power Research Institute), and NARI Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhang, X.Y.; Sun, Z. Application of Improved PNN in Transformer Fault Diagnosis. Processes 2023, 11, 474. [Google Scholar] [CrossRef]
  2. Zheng, H.; Shioya, R. A comparison between artificial intelligence method and standard diagnosis methods for power transformer dissolved gas analysis using two public databases. IEEJ Trans. Electr. Electron. Eng. 2020, 15, 1305–1311. [Google Scholar] [CrossRef]
  3. Luo, D.S.; Fang, J.; He, H.Y.; Lee, W.-J.; Zhang, Z.; Zai, H.; Chen, W.; Zhang, K. Prediction for Dissolved Gas in Power Transformer Oil Based on TCN and GCN. IEEE Trans. Ind. Appl. 2022, 58, 7818–7826. [Google Scholar] [CrossRef]
  4. Ali, M.S.; Omar, A.; Jaafar, A.S.A.; Mohamed, S.H. Conventional methods of dissolved gas analysis using oil-immersed power transformer for fault diagnosis: A review. Electr. Pow. Syst. Res. 2023, 216, 109064. [Google Scholar] [CrossRef]
  5. Liu, H.; Zhang, J.; Lian, H. Prediction of the gasses dissolved in transformer oil by sequential learning. High. Volt. Appar. 2019, 55, 193–199. [Google Scholar]
  6. Zhang, W.X.; Zeng, Y.; Li, Y.; Zhang, Z. Prediction of dissolved gas concentration in transformer oil considering data loss scenarios in power system. Energy Rep. 2023, 59, 186–193. [Google Scholar] [CrossRef]
  7. Ma, H.; Lai, H.; Wang, M. Forecasting of dissolved gas in transformer oil by optimized nonlinear grey Bernoulli Markov model. In Proceedings of the 2016 Chinese Control and Decision Conference (CCDC), Yinchuan, China, 28–30 May 2016; pp. 4168–4173. [Google Scholar]
  8. Bagheri, M.; Zollanvari, A.; Nezhivenko, S. Transformer fault condition prognosis using vibration signals over cloud environment. IEEE Access. 2018, 6, 9862–9874. [Google Scholar] [CrossRef]
  9. Ghunem, R.A.; Assaleh, K.; El-hag, A.H. Artificial neural networks with stepwise regression for predicting transformer oil furan content. IEEE Trans. Dielect. Elect. Insul. 2012, 19, 414–420. [Google Scholar] [CrossRef]
  10. Dai, J.; Song, H.; Sheng, G.; Jiang, X. Dissolved gas analysis of insulating oil for power transformer fault diagnosis with deep belief network. IEEE Trans. Dielect. Elect. Insul. 2017, 24, 2828–2835. [Google Scholar] [CrossRef]
  11. Qi, B.; Wang, Y.; Zhang, P.; Li, C.; Wang, H. A novel deep recurrent belief network model for trend prediction of transformer DGA data. IEEE Access 2019, 7, 80069–80078. [Google Scholar] [CrossRef]
  12. Ma, X.; Hu, H.; Shang, Y. A new method for transformer fault prediction based on multifeature enhancement and refined long short-term memory. IEEE Trans. Instrum. Meas. 2021, 70, 2512111. [Google Scholar] [CrossRef]
  13. Su, X.; Wang, Q.; Li, Q. Prediction method for transformer state based on gru network. In Proceedings of the 2020 IEEE/IAS Industrial and Commercial Power System Asia (I CPS Asia), Weihai, China, 13–15 July 2020; pp. 1751–1755. [Google Scholar]
  14. Zhong, M.W.; Cao, Y.F.; He, G.L.; Feng, L.; Tan, Z.; Mo, W.; Fan, J. Dissolved gas in transformer oil forecasting for transformer fault evaluation based on HATT-RLSTM. Electr. Pow. Syst. Res. 2023, 221, 109431. [Google Scholar] [CrossRef]
  15. Yuan, F.; Guo, J.; Xiao, Z.; Zeng, B.; Zhu, W.; Huang, S. An interval forecasting model based on phase space recon-struction and weighted least squares support vector machine for time series of dissolved gas content in transformer oil. Energies 2020, 13, 1687. [Google Scholar] [CrossRef]
  16. Chen, H.C.; Zhang, Y.; Chen, M. Transformer Dissolved Gas Analysis for Highly-Imbalanced Dataset Using Multiclass Sequential Ensembled ELM. IEEE Trans. Dielectr. Electr. Insul. 2023, 30, 2353–2361. [Google Scholar] [CrossRef]
  17. Nanfak, A.; Eke, S.; Meghnefi, F. Hybrid DGA Method for Power Transformer Faults Diagnosis Based on Evolutionary k-Means Clustering and Dissolved Gas Subsets Analysis. IEEE Trans. Dielectr. Electr. Insul. 2023, 30, 2421–2428. [Google Scholar] [CrossRef]
  18. Ran, M.H.; Huang, J.D.; Qian, Q.Y.; Zou, T.; Ji, C. EMD-based gray combined forecasting model–Application to long-term forecasting of wind power generation. Heliyon 2023, 9, e18053. [Google Scholar] [CrossRef] [PubMed]
  19. Yu, M.; Niu, D.X.; Gao, T. A novel framework for ultra-short-term interval wind power prediction based on RF-WOA-VMD and BiGRU optimized by the attention mechanism. Energy 2023, 269, 126738. [Google Scholar] [CrossRef]
  20. Riaz, F.; Hassan, A.; Rehman, S. EMD-based temporal and spectral features for the classification of EEG signals using supervised learning. IEEE Trans. Neural. Syst. Rehabil. Eng. 2016, 24, 28–35. [Google Scholar] [CrossRef]
  21. Guo, W.; Tse, P.W. A novel signal compression method based on optimal ensemble empirical mode decomposition for bearing vibration signals. J. Sound. Vib. 2013, 332, 423–441. [Google Scholar] [CrossRef]
  22. Xiao, H.S.; Li, Q.Q.; Shi, Y.L.; Zhang, T.; Zhang, J. Prediction of Dissolved Gases in Oil for Transformer Based on Grey Theory-Variational Mode Decomposition and Support Vector Machine Improved by NSGA-II. Proc. CSEE 2017, 37, 3643–3653. [Google Scholar]
  23. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties and Applications in Non-Gaussian Signal Processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  24. Chen, B.D.; Liu, X.; Zhao, H.; Principe, J.C. Maximum correntropy Kalman filter. Automatica 2017, 76, 70–77. [Google Scholar] [CrossRef]
  25. Ma, W.T.; Guo, P.; Wang, X.F.; Zhang, Z.; Peng, S.; Chen, B. Robust state of charge estimation for Li-ion batteries based on cubature Kalman filter with generalized maximum correntropy criterion. Energy 2022, 260, 125083. [Google Scholar] [CrossRef]
  26. Liu, X.; Qu, H.; Zhao, J.H.; Yue, P. Maximum correntropy square-root cubature Kalman filter with application to SINS/GPS integrated systems. ISA Trans. 2018, 80, 195–202. [Google Scholar] [CrossRef] [PubMed]
  27. Huang, W.; Zhao, J.; Yu, G.K.; Wong, P.K. Intelligent Vibration Control for Semiactive Suspension Systems Without Prior Knowledge of Dynamical Nonlinear Damper Behaviors Based on Improved Extreme Learning Machine. IEEE/ASME Trans. Mechatron. 2021, 26, 2071–2079. [Google Scholar] [CrossRef]
  28. Wang, N.; Wang, J.S.; Zhu, L.F.; Wang, H.Y.; Wang, G. A novel dynamic clustering method by integrating marine predators algorithm and particle swarm optimization algorithm. IEEE Access 2020, 9, 3557–3569. [Google Scholar] [CrossRef]
  29. Faramarzi, A.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine predators algorithm: A nature-inspired metaheuristic. Expert Syst. Appl. 2020, 152, 113377. [Google Scholar] [CrossRef]
  30. Long, X.Q.; Zhao, H.Q.; Hou, X.Y. A Novel Combinatoric Correntropy Algorithm: Properties and Its Performance Analysis. IEEE Trans. Circuits Syst. II 2022, 69, 5184–5188. [Google Scholar] [CrossRef]
  31. Ju, D. Research on Transformer Fault Diagnosis and Prediction Method Based on Improved Support Vector Machine. Master’s Thesis, Shenyang Agricultural University, Shenyang, China, 2023. [Google Scholar] [CrossRef]
Figure 1. MPA optimizes the VMD process.
Figure 1. MPA optimizes the VMD process.
Processes 12 00193 g001
Figure 2. Function V C * ( x , y ) curve under different parameters: (a) σ value change ( λ = 0.3 ) (b) λ value change ( σ 1 = 2.5 , σ 2 = 0.3 ) .
Figure 2. Function V C * ( x , y ) curve under different parameters: (a) σ value change ( λ = 0.3 ) (b) λ value change ( σ 1 = 2.5 , σ 2 = 0.3 ) .
Processes 12 00193 g002
Figure 3. Predict model via ExMCC -ELM and IPVMD.
Figure 3. Predict model via ExMCC -ELM and IPVMD.
Processes 12 00193 g003
Figure 4. Raw gas-in-oil data.
Figure 4. Raw gas-in-oil data.
Processes 12 00193 g004
Figure 5. MPA optimizes VMD to decompose H 2 convergence curves.
Figure 5. MPA optimizes VMD to decompose H 2 convergence curves.
Processes 12 00193 g005
Figure 6. H 2 raw data versus modal decomposition data.
Figure 6. H 2 raw data versus modal decomposition data.
Processes 12 00193 g006
Figure 7. Structure diagram of IPVMD- ExMCC -ELM prediction model.
Figure 7. Structure diagram of IPVMD- ExMCC -ELM prediction model.
Processes 12 00193 g007
Figure 8. 50-day H 2 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Figure 8. 50-day H 2 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Processes 12 00193 g008
Figure 9. 50-day C 2 H 4 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Figure 9. 50-day C 2 H 4 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Processes 12 00193 g009
Figure 10. 50-day C 2 H 2 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Figure 10. 50-day C 2 H 2 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Processes 12 00193 g010
Figure 11. 50-day C 2 H 6 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Figure 11. 50-day C 2 H 6 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Processes 12 00193 g011
Figure 12. 50-day CH 4 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Figure 12. 50-day CH 4 prediction results of IPVMD-ExMCC-ELM versus other methods: (a) prediction results and (b) prediction error.
Processes 12 00193 g012
Figure 13. Prediction accuracy with different kernel widths σ .
Figure 13. Prediction accuracy with different kernel widths σ .
Processes 12 00193 g013
Figure 14. The prediction results with different γ .
Figure 14. The prediction results with different γ .
Processes 12 00193 g014
Table 1. H 2 prediction results for each method.
Table 1. H 2 prediction results for each method.
BPELMIPVMD-ELMExMCC-ELMIPVMD-ExMCC-ELM
MAE0.27410.23500.15950.14850.0863
MSE0.10270.10820.03610.06330.0147
RMSE0.32040.32890.19000.25170.1211
SMAPE0.01530.01330.00920.00870.0053
Table 2. Prediction results evaluation for each method.
Table 2. Prediction results evaluation for each method.
BPELMIPVMD-ELMExMCC-ELMIPVMD-ExMCC-ELM
CH 4 MAE0.40560.32440.18330.27470.1135
MSE0.27210.24320.04570.17980.0255
RMSE0.52170.49310.21370.42400.1597
SMAPE0.00670.00540.00290.00460.0019
C 2 H 2 MAE0.41430.38460.18070.39940.1196
MSE0.26470.25540.05010.21840.0263
RMSE0.51450.50540.22390.46730.1620
SMAPE0.13150.13740.11180.13100.1028
C 2 H 4 MAE0.27380.24800.09210.22840.0670
MSE0.14530.13160.01490.11480.0073
RMSE0.38110.36280.12220.33890.0855
SMAPE0.02210.02030.00800.01840.0056
C 2 H 6 MAE0.30270.29520.16820.25860.1044
MSE0.19840.20360.06750.16380.0209
RMSE0.44550.45130.25990.40470.1445
SMAPE0.01170.01110.00690.00960.0042
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Du, G.; Sheng, Z.; Liu, J.; Gao, Y.; Xin, C.; Ma, W. A Novel Hybrid Model Combining Improved VMD and ELM with Extended Maximum Correntropy Criterion for Prediction of Dissolved Gas in Power Transformer Oil. Processes 2024, 12, 193. https://doi.org/10.3390/pr12010193

AMA Style

Du G, Sheng Z, Liu J, Gao Y, Xin C, Ma W. A Novel Hybrid Model Combining Improved VMD and ELM with Extended Maximum Correntropy Criterion for Prediction of Dissolved Gas in Power Transformer Oil. Processes. 2024; 12(1):193. https://doi.org/10.3390/pr12010193

Chicago/Turabian Style

Du, Gang, Zhenming Sheng, Jiaguo Liu, Yiping Gao, Chunqing Xin, and Wentao Ma. 2024. "A Novel Hybrid Model Combining Improved VMD and ELM with Extended Maximum Correntropy Criterion for Prediction of Dissolved Gas in Power Transformer Oil" Processes 12, no. 1: 193. https://doi.org/10.3390/pr12010193

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop