Next Article in Journal
The Identification of ECG Signals Using Wavelet Transform and WOA-PNN
Next Article in Special Issue
Blockchain-Based Peer-to-Peer Transactive Energy Management Scheme for Smart Grid System
Previous Article in Journal
Towards Multimodal Equipment to Help in the Diagnosis of COVID-19 Using Machine Learning Algorithms
Previous Article in Special Issue
Energy Load Forecasting Using a Dual-Stage Attention-Based Recurrent Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Smart Grid Stability Prediction Model Using Neural Networks to Handle Missing Inputs

1
Department of Chemical Engineering, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Malaysia
2
Department of Electrical and Electronics Engineering, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Malaysia
3
School of Electrical Engineering, Vellore Institute of Technology, Vellore 632014, India
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(12), 4342; https://doi.org/10.3390/s22124342
Submission received: 28 April 2022 / Revised: 2 June 2022 / Accepted: 3 June 2022 / Published: 8 June 2022
(This article belongs to the Special Issue Resilience Engineering for Smart Energy Systems)

Abstract

:
A smart grid is a modern electricity system enabling a bidirectional flow of communication that works on the notion of demand response. The stability prediction of the smart grid becomes necessary to make it more reliable and improve the efficiency and consistency of the electrical supply. Due to sensor or system failures, missing input data can often occur. It is worth noting that there has been no work conducted to predict the missing input variables in the past. Thus, this paper aims to develop an enhanced forecasting model to predict smart grid stability using neural networks to handle the missing data. Four case studies with missing input data are conducted. The missing data is predicted for each case, and then a model is prepared to predict the stability. The Levenberg–Marquardt algorithm is used to train all the models and the transfer functions used are tansig and purelin in the hidden and output layers, respectively. The model’s performance is evaluated on a four-node star network and is measured in terms of the MSE and R 2 values. The four stability prediction models demonstrate good performances and depict the best training and prediction ability.

1. Introduction

The conventional power grid contains standard power generation units grounded on fossil fuels. With the soaring energy prices, the need for renewable energy sources and climate change, the old power grid is becoming outdated and facing various limitations, such as cybersecurity, privacy and power losses due to one-way communication [1]. This pushes for deploying renewable energy sources to improve sustainability and reliability. A smart grid is a solution to this. The smart grid system is a digital future electricity system that enables a two-way flow of communication, i.e., between the center and the device to the center [2].
This bidirectional communication utilizes advanced computing infrastructure, digital sensing and software capabilities to optimize all the grid components and improve reliability and sustainability. There is a unidirectional flow of energy from the energy provider to the consumer in a traditional grid, and consumers are charged based on their consumption. However, in a smart grid system, the users in the grid can consume, produce, store and trade energy with other users [3]. The smart grid introduces demand response, and the price information is determined as the demand is evaluated with supply and conveyed to the customer.
This paper used the DSGC model to define and relate the price to the grid frequency [2,4]. The mathematical model based on DSGC differential equations seeks to find grid instability for a four-node star architecture [5]. The four-node star architecture consists of a central generation node, the power source and three consumer nodes. The response time of the smart grid users is considered to adjust the consumption/production concerning the price changes.
The model involves real-time pricing, and thus the grid stability has to be maintained with fluctuations in reaction times and electricity price of all users. It is critical to evaluate grid stability as the process is time-critical dynamically. This is because smart grid stability prediction helps increase efficiency through grid optimization, improves electrical supply reliability and consistency and analyses disturbances and fluctuations in energy consumption or production.
Before the utilization of modern techniques to predict smart grid stability, traditional approaches consisted of simulations combining fixed values for one subset and fixed distribution of values for the remaining subset variables [6,7]. The generation of electricity by photovoltaic power is related to the global horizontal irradiance. For the unknown cloud statistics, the irradiance is uncertain for predicting the stability in power generation, causing optical instability in the solar irradiance [4]. Measurement-based methods are another complex and challenging traditional method used to predict power grid stability [8].
Various statistical approaches have been investigated, including autoregressive moving average, Kalman filter and Markov chain model [9], which contribute to insufficient reliability of the grid [10]. Other types of early statistical methods [11] for load forecasting in smart grids have various drawbacks and affect the accuracy of the prediction model. These are built by ineffective, simple regression functions, and thereby do not yield good performance in vast uncertainties [12]. Further, traditional approaches, such as time series analysis, ARMA, ARIMA and Markov models for stability forecasting, exist only in specific operating ranges [13,14].
Additionally, some research involved using conventional parametric methods that include linear regression, auto-regressive moving average and the general exponential methods. Although such models return satisfactory prediction accuracy, they persist with major disadvantages, such as improper response and complex computational problems to meteorological variables and nonlinear electrical load [15]. A probabilistic model was introduced in [16] for power stability.
However, some uncertainties have been observed between regular grid operation and cascading failure operation in the simulation result. Adding on, techniques used for stability assessment require extensive computation time and massive data analysis volume, which makes it tough to obtain a reliable prediction and makes it difficult to take decisions for an operating power system [17].
A few hybrid systems used for dynamic stability prediction have been based on unreliable self-organized maps and responded slowly [6,18]. Another method introduced a situational awareness for stability prediction, a perception of elements for a given time and space in the environment [5]. It was proven that optimized deep-learning models are one of the excellent prediction tools for smart grid stability. Using neural networks for stability prediction has various advantages.
They have multiple training algorithms, do not require significant dataset pre-processing and can produce high accuracy values during training and testing [4]. Further, they can recognize different sets within a whole dataset and give adequate results even when the dataset is incomplete or inaccurate [19]. Finally, the ability to implicitly detect complex nonlinear relationships between independent and dependent variables makes it viable for stability prediction [20].
Comprehensive review work in [21] concluded that most of the works on prediction models using machine learning reported little or no information on the presence and handling of missing data. The missing data is omitted in most models, which is ineffective, affecting their performance. The missing data for the analysis results from many things, such as sensor failure, equipment malfunctions, lost files, etc. This challenges the increasing cost and prediction ability of the proposed models. Thus, there is a need for significant research in handling missing data. On the other hand, predicting the data with neural networks or machine-learning models is more efficient than simply omitting the data or resorting to mean values.
Being motivated by the literature, this paper proposes a novel method to predict the smart grid stability of a four-node star network using a neural network with complete and missing input data, consisting of missing input variables. Thus, the significant contributions of this paper are highlighted as follows:
  • The classic FFNN is designed to predict the stability of the smart grid system of a four-node star network with complete input data.
  • The sub-neural networks are proposed to predict the missing input variables, which are caused due to a sensor, network connection or other system failures. Then, the system’s stability is forecast using these predicted missing input data.
  • The performance of the proposed approach is evaluated in four different case studies in which at least one input variable is missing.
The subsequent sections of the paper are organized as follows: Section 2 presents the comprehensive literature review on smart grid stability prediction using neural networks. Section 3 describes the mathematical modeling and data description of the four-node star network used for the smart grid stability prediction. Section 4 shows the development and performance of the FFNN with complete input data, and Section 5 describes the development and performance evaluation of the FFNN to handle the missing inputs. Finally, Section 6 highlights the conclusions of the proposed work.

2. Literature Review

In this section, an extensive literature survey is conducted on smart grid stability using neural networks. This literature review shows that various neural-network-based techniques have been used for analysis, and the data analyzed in the works are with complete input data. The developed approaches are robust and accurate due to their complex structure that helps classify problems and recognize correlations in raw data and hidden patterns.
A summary of works focused on smart grid stability prediction using various neural networks is highlighted in Table 1. The table contains 55 papers published in the last decade, categorized into publication year, smart grid architecture, neural network type, neural network architecture, activation functions, training algorithms, performance measures and comparison techniques considered for each study. From the research works in Table 1, the year-wise and the publisher-wise contributions to the smart grid’s stability prediction during the last decade are shown in Figure 1. Figure 2 depicts the smart grid architectures identified in the literature survey conducted.
The most popular architectures are IEEE bus systems [6,16,18,22,23] and node network types [4,8,24]. Therefore, in this paper four-node star network was selected to perform the proposed research.
In the analysis, several types of neural networks and hybrid networks were identified, as depicted in Figure 3. Among the most popular neural networks identified are FFNN, which includes the hybridized versions, such as FF-BPNN [25] and FF-DNN [26]. In addition, CNN is another most widely used, with its enhanced and hybrid versions, namely ECNN [27] and CNN-RNN [28]. The hybrid versions of LSTM, including LSTM-RNN [29,30] and LSTM-CNN [31], can also be seen in this literature.
In addition, the performance of DNN for stability prediction was improved by hybridizing with RNN, RL [32], CNN and IRBDNN [33]. On the other hand, optimization algorithms, such as SSA, have also been used with RBFNN to obtain the network’s optimal weights [7]. The hybrid versions of GRU models, such as BiGRU [8] and GRU-RNN [9], have also been used for node networks’ stability prediction. The sub-classification of all these neural-network-based models is also illustrated in Figure 3.
Figure 4 summarizes the various training algorithms and activation functions used in the research work reported in Table 1. The figure concludes that the LM algorithm is the most commonly used for training algorithms, followed by Adam’s optimization algorithm [29,32]. Furthermore, it further depicts that sigmoid, ReLU, tansig and tanh are the most frequently used hidden layer activation functions [30,34]. In contrast, the purelin activation function followed by Sigmoid is most commonly used in the output layer of the neural network [34,35,36].
The significant findings from the literature review on smart grid stability prediction using neural networks are highlighted as follows:
  • No work was conducted to predict stability when there is a missing parameter. Most studies showed that missing data had been either omitted, unreported or replaced with mean/median values.
  • The most popular architectures used for the case studies are IEEE bus systems and node network types (see Figure 2).
  • Among the several types of conventional and hybrid neural networks proposed in the literature, the FFNN and its hybrid versions, such as FF-BPNN and FF-DNN, are widely presented (see Figure 3 and Table 1).
  • The Levenberg–Marquardt algorithm is the most frequently used training algorithm for various networks to predict smart grid stability (see Figure 4).
  • The tansig and purelin activation functions have frequently been used in various networks’ hidden and output layers to predict smart grid stability (see Figure 4).
From the above research gaps, this paper made an effort to develop a forecasting model that handles the missing input data. For the proposed neural-network-based forecasting model, the LM training algorithm was selected as it is one of the fastest backpropagation algorithms and is widely recommended in the literature. The literature proves that effective training necessitates a nonlinear and linear combination of activation functions. Thus, the tansig and purelin activation functions are utilized in the hidden and output layers, respectively.
Further, in our previous work reported in [37], an effort was made to compare the performance of FFNN, cascade and recurrent neural-network-based models for smart grid stability prediction. The work concludes that, for the considered application, the FFNN demonstrated superior performance in terms of the MSE and R 2 values compared to cascade and recurrent neural networks. On the other hand, over the years, researchers have proposed different methodologies and theories for selecting the number of hidden layers and the number of hidden neurons in each hidden layer. As reported in [38], it was concluded that a network with only one hidden layer but sufficient neurons can achieve better performance.
Moreover, this performance can be further improved by adding additional hidden layers. However, the variation in this performance with a multilayer network is minimal. The work reported in [39] concluded the same, stating that a multilayer network has achieved better performance but increased the complexity of the network. Therefore, for the considered application, the FFNN with a single hidden layer was used for all the cases, which improved the performance in predicting smart grid stability.
Table 1. Summary of works focused on forecasting smart grid stability using neural networks.
Table 1. Summary of works focused on forecasting smart grid stability using neural networks.
Ref.YearSmart Grid ArchitectureNeural Network TypeNeural Network ArchitectureActivation FunctionsTraining AlgorithmPerformance MeasuresComparison Techniques
Hidden LayerOutput Layer
[34]2021FFNN2:10:1Tanh, SigmoidLinearLM, BR, SCGMSE, RRTP, SMP, RTP-SMP, GA, ANN, STW
[7]2021Smart grid with photovoltaic and wind turbineSSA-RBFNNRMSESSA-RBFNN with and without RES
[40]2021FFNN3:20:1SigmoidLinearLMMSE, RMSEPV with ANN, Wind with ANN, Hybrid model with ANN
[32]2021DNN-RLLeaky ReLULeaky ReLUAdamMSE
[29]2021LSTM-RNN1:50:50:50:1TanhTanhAdamMAE, RMSE, MAPEGBR, SVM
[4]2021four-node starFFNN24:24:12:1ReLUSigmoidAdam, GDM, NadamAccuracy, Precision, Sensitivity, F-scoreCNN, FNN
[9]2021GRU-RNN3:15:10:1GateCandidateAdaGradRMSE, MAELSSVR, WNN, ELM, SAE, DBN
[41]2021SNN784:400:400:11LIF spike generatorSummation and maximumPrecision, Recall, F-score, AccuracyCNN
[8]2021four-node starLSTM, BiGRU, ELM12:256:128:1, 12:512:256:1, 12:96:30:1Sigmoid SoftplusSigmoid SoftplusAdamRMSE, MAE, R 2 , PICP, PINC, ACEBiGRU, LSTM, XGB, LGBM, ANN
[42]2021Distributed systemsDNN-RLReLUReLUAdamPeak, Mean, Var, PAR, Cost, Computation timeC-DDPG, DPCS, SWAA
[11]2021LSTM, BPNN6:96:48:1, 6:48:24:1, 6:10:1RBFSigmoidAdamMAPE, RMSELSTM, BPNN, MLSTM, ELM, MLR, SVR
[43]2021BPNN3:2:3SigmoidLinearBPRMSE
[14]2021FF-DNNReLUSELUPDNN, Pooling functionFA, MAE, RMSE, SoC, HRSVM, NN-ARIMA, DBN
[44]2021FFNNReLUAlphaBPAccuracy, Precision, Recall, F-scorePSO-KNN, PSO-NN, PSO-DT, PSO-RF
[10]2020CNN-LSTMReLULinearAdamRMSE, MAE, NRMSE, F-scoreARIMA, BPNN, SVM, LSTM, CEEMDAN-ARIMA, CEEMDAN-BPNN, CEEMDAN-SVM
[45]2020RNN, CNNSigmoidTanhAdamArea under the curve, F-score, Precision, Recall, AccuracyLogistic regression, SVM, LSTM
[46]2020NN-LMS24:24:24, 24:96:96:4ReLUReLU
[47]2020NARX-RNN2:5:1SigmoidLinearConjugate gradient with Polak-RibiereNRMSE, RMSE, MAPEARMAX
[48]2020FFNN20:38:1TanhLinearConjugate gradient with Polak-RibiereMSERTEP, LBPP, IBR without ESS
[33]2020IRBDNNRMSE, MAE, MAPEDNN, ARMA, ELM
[30]2020LSTM-RNNSigmoid, Tanh, ReLUAccuracy, Precision, Recall, F-scoreGRU, RNN, LSTM
[22]2020IEEE 14-bus systemCNNReLUSigmoidAdamPrecision, Recall, F-score, Row accuracySVM, LGBM, MLP
[49]2019FF-BPNNGAMSE, Fitness, Accuracy
[50]2019RNN TanhSigmoidBPMAE, RMSE, MAPE, PmeanBPNN, SVM, LSTM, RBF
[28]2019CNN-RNN100:98:49:1ReLUSoftmax MSE, Recall, PTECCCNN, CNN-RNN, LSTM
[51]2019ENN10:1:1GDM and Adaptive LR, LMRMSE, NRMSE, MBE, MAE, R, Forecast skillSimilarity search algorithm, ANN, MLP and ARMA, LSTM
[12]2019FF-DNN, R-DNN2:5:2Sigmoid, Tanh, ReLUSigmoid, Tanh, ReLULMMAPEEnsemble Tree Bagger, Generalized linear regression, Shallow neural networks
[31]2019CNN, LSTM05:10:100ReLUSoftmaxMCC, F-score, Precision, Recall, AccuracyLogistic regression, SVM
[27]2019ECNN32:32:1ReLUSigmoid, SoftmaxAdamMAE, MAPE, MSE, RMSEAdaBoost, MLP, RF
[23]2019IEEE 39-bus New England test systemCNN, LSTMSigmoidTanhGDMAccuracy
[52]2019FFNN76:20:1, 92:20:1, 92:20:1ReLUSigmoidLMMSE, Accuracy, Precision, Recall, F-scoreRF, OneR, JRip, AdaBoost-JRip, SVM and NN (without WOA)
[53]2019ECNNMSE, RMSE, MAE, MAPE
[54]2019FF-DNN, R-DNNSigmoid, Tanh, ReLULinearLMMAPE, Correlation coefficient, NRMSEANN, CNN, CRBM, FF-DNN
[25]2019FF-BPNNReLUGDMMean error, MAD, Percent error, MPE, MAPEClassical forecasting methods
[26]2019FF-DNN1:5:1, 6:5:1SigmoidLinearMAPEDNN-ELM
[55]2018FFNNSigmoidNonlinear and linear networkLMMSE, RMultilayer ANN Models
[56]2018RBF, WRNN7:4:3RBFCompetitiveLMClassification accuracyPooling Neural Network, LM
[13]2018WRNN2:16:16:4RBFRBFRMSE
[57]2017FFNN7:96:48:24:1TanhGaussianDlnet, BPMAPETen state-of-the-art forecasting methods
[58]2017FFNN24:5:1SigmoidSigmoidLMMAPEAFC-STLF, Bi-level, MI-ANN forecast
[59]2017Deep learning based short-term forecasting20:30:25:1ReLUReLURESVM
[24]201710-node networkFFNN, WNN-LQE8:10:1Morlet waveletSigmoidSNRLQE-based WNN, BPNN, ARIMA, Kalman, XCoPred algorithms
[60]2016FFNN3:20:10:3SigmoidLinearLM, BRMSE, RLM, BR
[15]2016FFNN8:10:1SigmoidLinearMAE, MAPE, RMSE, R 2 , MSEGA-MdBP, CGA-MdBP, CGASA-MdBP
[16]2015IEEE 30-bus systemFFNN4:10:1RBFSCG supervised learningMSE, PDF, CDF
[61]2015FFNN10:1:20TanhTanhLVQMean Error, Maximum Error, Success %
[62]2014FFNN7:(10-15):1SigmoidLinearLMR, MAPE
[17]2013FFNNLMMER, MAE, MAPE
[63]2012Microgrid architecture: residential smart house aggregatorBPNN10:1:1TanhLinearLM, SCGSolar insulation and air temperature
[64]2012IEEE 39-bus New England test systemFF-BPNN20:10:5:1TanhSigmoidLM, BRStability
[6]2012IEEE 39-bus New England test systemRBF30:30:9, 30:30:10RBFLinearLMTraining Time, Testing Time, Number of misses, MSE, Classification accuracy %
[18]2011IEEE 39-bus New England test systemRBF36:36:1GaussianLinear Training time, Testing time, Number of misses, MSE, False alarms %, Misses %, Classification accuracy %Traditional NR method
[65]2011Grid-connected PV plantBPNN16:15:7:1SigmoidLinearLMMABE, RMSE, R
[66]2011Medium tension distribution systemRBF33:119:33, 33:129:33RBFLinearMSE, SPREAD
[67]2010BPNN, FFNN8:8:30:1TanhLinearLM, BRMSELM, BR, OSS

3. Mathematical Modeling and Data Description of Four-Node Star Network

In Section 3.1, the mathematical modeling of the four-node star architecture network is developed based on the equations of motion and binding the electricity price to the grid frequency. Then, the description of the generated dataset from the final dynamic equation of DSGC and the correlation analysis between the network parameters are provided.

3.1. Mathematical Modeling and Stability Analysis of Four-Node Star Network

In this section, the mathematical modeling of the four-node star network and the stability analysis are conducted. The central node (center of the “star”) communicates directly with the consumer nodes in a star network topology. The consumer nodes are connected to the central (generation node), enabling bidirectional communication between each node, which helps them to operate at lower power levels. One of the main advantages of star topology is that the networks are independent. In case of failure or errors in one of the consumer nodes, the other consumer nodes are not affected, and the network operates typically.
The network is formed with one power producer in the center (i.e., generation node) and three consumers (i.e., consumer node). Star topologies depend heavily on delay and averaging time. Intermediate delays in a four-node star topology benefit stability, making it a simple, effective and efficient system [4]. From the literature survey conducted, we observed that the star and bus topologies were popular. A conclusion was drawn that star networks used in previous works having similar objectives showed good performance and can achieve the mathematical modeling for the DSGC system. Thus, the four-node star topology was chosen for this work, and the mathematical model of the DSGC system was obtained for the four-node star architecture network given in Figure 5.

3.1.1. Mathematical Modeling

The mathematical model of the DSGC system is obtained for the four-node star architecture network given in Figure 5. The figure shows that the network is formed with one power producer in the center (i.e., Generation Node) and three consumers (i.e., Consumer Node). The mathematical modeling developed with assumptions, such as no uncertainties and external disturbances comprises two parts. The first describes the generator and load dynamics based on equations of motion. The second part is based on binding the electricity price to the grid frequency [4,5,37,68].
The first step in the modeling is applying the energy conservation law. As per the energy conservation law, the power balance equation is given as follows:
P s = P a + P d + P t ,
where P s is the power generated from source.
In (1), P d is the dissipated energy from the turbine, which is proportional to the angular velocity square given as,
P d = K j ( δ ˙ j ( t ) ) 2 ,
where j is the node index (either generator or load), K j is the friction coefficient of j th node and δ j ( t ) is the rotor angle of j th node defined as,
δ j ( t ) = ω t + θ j ( t ) ,
where ω is the grid frequency and θ j is the relative rotor angle.
Similarly, in (1), P a is the accumulated kinetic energy, and P t is the transmitted power given as,
P a = 1 2 M j d dt ( δ ˙ j ( t ) ) 2 ,
P t = m = 1 4 P j m max sin ( δ m δ j ) ,
where M j is the moment of inertia of j th node and P j m max is the maximum capacity of line between j th and m th node.
By substituting (2), (4) and (5) in (1), P j s is obtained as follows:
P j s = 1 2 M j d dt ( δ ˙ j ( t ) ) 2 + K j ( δ ˙ j ( t ) ) 2 m = 1 4 P j m max sin ( δ m δ j )
Now, substituting δ j ( t ) from (3) in (6), d 2 dt 2 θ j ( t ) is obtained as follows:
d 2 dt 2 θ j ( t ) = P j α j d dt θ j ( t ) + m = 1 4 K j m sin ( θ m θ j ) ,
where P j is the generated or consumed power, α j is the damping constant and K j m is the coupling strength between j th and m th nodes. These coefficients are computed as follows:
P j = P j s K j ω 2 M j , α j = 2 K j M j , K j m = P j m max M j ω .
The final step in the modeling is binding the electricity price to the grid frequency ω , allowing consumers to adjust their consumption or production. Thus, the electricity price p j for the j th node is computed as,
p j = p ω c 1 t T j t d dt θ j ( t τ j ) d t ,
where p ω is the electricity price when d θ j / dt = 0 , c 1 is the proportionality coefficient, T j and τ j are the average and reaction times, respectively.
The power consumed or produced P ^ j ( p j ) at price p j is defined as,
P ^ j ( p j ) P j + c j ( p j p ω ) ,
where c j is the coefficient proportional to the elasticity price.
For the four-node star network shown in Figure 5, it is assumed that the algebraic sum of power consumed or generated is equal to zero. Thus, the assumption is given as,
j = 1 4 P j = 0 .
Therefore, the final dynamic equation of DSGC system for the four-node star architecture network is obtained by substituting (7), (9) and (10) in (11) as follows:
d 2 dt 2 θ j ( t ) = P j α j d dt θ j ( t ) + m = 1 4 K j m sin ( θ m θ j ) γ j T j θ j ( t τ j ) θ j ( t τ j T j ) ,
where γ j = c 1 × c j .

3.1.2. Stability Analysis

In the first stage of analyzing the network’s dynamical stability around the grid’s steady-state operation, the fixed points of the network are computed by solving d 2 dt 2 θ j = 0 and d dt θ j = 0 , which are obtained as,
θ j ( t ) , d dt θ j ( t ) = ( θ j , ω j ) .
The above equation shows that the fixed point exists only if the grid has an adequate coupling strength coefficient K j m to transmit the power from the generation nodes to the consumer nodes. Furthermore, in the obtained fixed point, the value of ω j is d dt θ j , which is equal to zero. Thus, the value of ω j = 0 . The fixed points highlight that it only depends on the value of θ j , which must be analyzed to determine the stability.
Next, the Jacobian matrix of the system is obtained to compute the eigenvalues that determine the network’s stability. Thus, the Jacobian matrix J is calculated as,
J = θ j ( d dt θ m ) ω j ( d dt θ m ) θ j ( d dt ω m ) ω j ( d dt ω m ) .
The eigenvalues λ of the above Jacobian matrix determine the network’s stability. The matrix has infinitely many solutions. However, only a finite number of solutions can have a real positive component ( Re ( λ ) 0 ), determining the network’s instability. In addition, the negative real part ( Re ( λ ) < 0 ) indicates stability. Therefore, the network’s stability condition is summarized as follows:
Stability = Stable , if Re ( λ ) < 0 , Unstable , if Re ( λ ) 0 .

3.2. Data Description of Four-Node Star Network

From the differential Equation (12), it is to be noted that the parameters τ j , P j and γ j are the predictive features of the network. The values of these parameters used for the simulation are shown in Table 2 [68]. The value of j ranges from 1 to 4, in which index 1 is the generator node, and the remaining indices (2, 3 and 4) are consumer nodes. Further, the values of simulation constants α j , T j and K j m used in the simulation are given in Table 2. The range of P j ( j ( 2 , 3 , 4 ) ) at the consumer nodes are also shown in Table 2. The value of P 1 at the generating node in the model is computed as,
P 1 = j = 2 4 P j .
The generated dataset contains 60,000 samples for all the 12 predictive variables and one dependent variable, Re ( λ ) . The predictive features are shown in Figure 6a–c, and the dependent variable Re ( λ ) , whose values are the real part of the roots from the dynamic equation of DSGC system in (12), is shown in Figure 6d. The dataset is zoomed in, and the region of the first 300 samples is shown in the figures.

3.3. Correlation Analysis

The Pearson’s correlation matrix between the predictive features of the network ( τ j , P j , γ j ) and dependent variable Re ( λ ) is shown in Figure 7. As reported in [69], the interpretation from the Pearson’s correlation coefficients is given in Table 3. Therefore, from Figure 7 and Table 3, it can be observed that there is a moderate negative correlation coefficient of −0.579 between P 1 and its sum components ( P 2 , P 3 and P 4 ). In addition, there is a weak positive correlation between dependent variable Re ( λ ) , τ j and γ j of around 0.28 and 0.29, respectively. In contrast, there is negligible correlation between dependent variable Re ( λ ) and P j . Furthermore, it is worth highlighting that there is a negligible correlation between the predictive features ( τ j , P j , γ j ) of the network.
As Pearson’s correlation matrix describes the strength and associated direction between the variables, it can be concluded that the relationship between any two predictive features or between predictive features and the dependent variable is not very strong. Therefore, in this analysis, all the parameters are considered for developing and evaluating the performance of the proposed model (refer to Section 4). In addition, only the power parameters have been considered for developing and assessing the performance of the proposed model that handles the missing inputs since there is a moderate correlation between P 1 and its sum components compared to other parameters (refer to Section 5).
A detailed research flowchart of the complete smart grid stability design model is portrayed in Figure 8.

4. Development and Performance Evaluation of Feedforward Neural Network

This section develops and evaluates the performance of an FFNN to predict the stability of the smart grid. The methodology for preparing a prediction model using complete input data is shown in Figure 9. The figure shows that data collection, analysis and pre-processing occur first. The input data is identified for the next step, which includes the prediction model to predict stability using the input data. The dataset used in this study consists of 60,000 samples.
The neural network used to predict stability utilizing the input data is a three-layered FFNN as shown in Figure 10. The first layer in the architecture indicates the input layer, which consists of 12 nodes equivalent to the 12 input parameters τ j , P j , γ j j { 1 , 2 , 3 , 4 } . The number of nodes in the middle layer, i.e., the hidden layer ‘ N h ,’ is 10, which can be calculated using the number of nodes in input layer ‘ N i ’ as follows [70]:
N h = 10 + 1 N i .
The third layer represents the output layer, consisting of 1 node, the output parameter (Stability). The dataset is divided into 80% and 20% for training and testing. The neural network is trained using the Levenberg–Marquardt algorithm. The tansig and purelin activation functions are utilized in the hidden and output layers. The training algorithm and activation functions are chosen as per the results of the comprehensive literature review conducted as shown in Figure 4. The training and testing outputs for the neural network are shown in Figure 11a,b. The neural network performance is measured in terms of R 2 and MSE [71,72,73,74]. The model has achieved an R 2 value of 0.9739 during training and 0.9738 during testing. Additionally, the model achieved MSE values of 0.0077 during both training and testing. Thereby, the accurate performance of the prediction model is depicted as the R 2 , and the MSE values are close to 1 and 0.

5. Development and Performance Evaluation of Feedforward Neural Network to Handle Missing Input

This section develops and evaluates a novel prediction model to handle missing inputs. The flowchart for the methodology adopted for predicting the missing input data is represented in Figure 12. Herein, four cases of missing inputs are taken, described in the flow chart as Case 1, Case 2, Case 3 and Case 4. This flowchart is represented in three stages.
The first stage includes data collection, analysis, pre-processing and defining the missing inputs. A prediction model is prepared using a sub-neural network to handle the missing inputs in the second stage. After the missing input parameters are predicted, the prediction model that handles missing inputs is prepared to predict stability. The primary neural network is an FFNN trained using the Levenberg–Marquardt algorithm for each case.
The tansig and purelin transfer functions are used in the hidden and output layers. The dataset consists of 60,000 samples, out of which 80% are used for training and 20% for testing. The input layer consists of 12 nodes corresponding to the 12 input parameters. The output layer consists of one node corresponding to the one output parameter. The number of nodes in the middle layer, i.e., the hidden layer ‘ N h ’, is 10, which can be calculated using (17).
Standard specifications for each sub-neural model in the four cases are as follows: tansig and purelin transfer functions are used in the hidden and output layers. The dataset, consisting of 60,000 samples, is divided into 80% for training and 20% for testing. The training algorithm used for the sub-neural network is the Levenberg–Marquardt algorithm. Different missing input variables are considered in each layer for each of the four cases, as explained underneath.

5.1. Case 1

One missing input variable is considered in the first case, which will be predicted using a sub-neural network. The sub-neural-network model for this section is an FFNN that consists of three layers (refer to Figure 13). The first input layer consists of three nodes composed of three nodes similar to the three input parameters: accumulated power ( P 2 ), dissipated power ( P 3 ) and transmitted power ( P 4 ). The last layer consists of one output node composed of one output parameter, i.e., source power ( P 1 ). The number of nodes in the hidden layer is 10, computed using (17).
The training and testing outputs for the case 1 sub-neural network is shown in Figure 14a,b for 60,000 samples and zoomed in for 300 samples as shown in the bottom subplot. The neural network performance is measured in terms of R 2 and MSE. The model achieved an R 2 value of 0.9992 during training and testing. Additionally, the model achieved MSE values of 0.0008 during training and 0.0008 during testing.
The primary neural network was trained to predict stability using the predicted output variables. The testing output variables of the sub-neural network are substituted in the primary neural network. Finally, the leading neural network is tested after predicting the missing input variables. The training and testing outputs for the case 1 primary neural network is shown in Figure 15a,b for 60,000 samples and zoomed in for 300 samples as shown in the bottom subplot. The neural network performance is measured in terms of R 2 and MSE. The model achieved an R 2 value of 0.9721 during training and 0.8413 during testing. Additionally, the model achieved MSE values of 0.0080 during training and 0.0085 during testing.

5.2. Case 2

Case 2 involves two missing input variables for stability prediction using FFNN as shown in Figure 16. The network shows that the input layer consists of two nodes relative to the input parameters: the source and transmitted powers (i.e., P 1 and P 4 ). The output layer has two nodes corresponding to accumulated and dissipated power (i.e., P 2 and P 3 ). The number of nodes in the hidden layer is 10 (refer to (17)). In the next step, the testing output variables of the sub-neural network are substituted and trained in the primary neural network model.
Upon prediction of the missing input variables, the primary neural network is tested. The MSE and R 2 performance measures are used to handle the missing data for the prediction model. The sub-neural-network model achieved MSE values of 0.1661 during the training and 0.1667 during testing. The R 2 values are 0.7082 during training and 0.7072 during testing. The sub-neural network’s performances during training and testing for the first 300 samples are shown in Figure 17a,b, respectively.
The next step of case 2 involves training the primary neural network by utilizing the obtained testing output variables from the sub-neural network that is also measured in terms of MSE and R 2 values. The neural network attains an MSE of 0.0077 during training and testing. Furthermore, the R 2 values have obtained 0.9738 during the training and testing phases. The final developed model’s performance for all the 60,000 samples and the first 300 samples of both the phases are represented in Figure 18a,b, respectively. The response in plots of the final leading model showcases the best prediction and tracking ability at both phases. The MSE and R 2 values relative to 0 and 1, respectively, indicate that the final proposed model for this case gives a superior performance.

5.3. Case 3

Further, Case 3 uses two missing input variables for a feedforward sub-neural network for stability prediction as shown in Figure 19. The input layer has two nodes relative to the two input parameters: the source power and accumulated power (i.e., P 1 and P 2 ). The output layer has two nodes corresponding to the two output parameters, dissipated and transmitted powers (i.e., P 3 and P 4 ). The number of nodes in the hidden layer is 10. In the next step, the testing output variables of the sub-neural network are substituted and trained in the primary neural network model. Upon prediction of the missing input variables, the primary neural network is trained. The MSE and R 2 performance measures are used to handle the missing data for the prediction model. The sub-neural-network model achieved MSE values of 0.1659 during the training and 0.1673 during the testing phases.
The R 2 values are found to be 0.7085 and 0.7061 during training and testing. The sub-neural network’s performance during training and testing for the first 300 samples is shown in Figure 20a,b, respectively.
Next, the primary neural network was trained by utilizing the obtained testing output variables from the sub-neural network measured using MSE and R 2 values. The neural network attained an MSE of 0.0083 during training and 0.0082 during testing. Furthermore, the R 2 values obtained were 0.9720 during training and 0.9721 during the testing phases. The final developed model’s performance for the 60,000 samples and the first 300 zoomed-in samples at both stages are represented in Figure 21a,b, respectively. The response in plots of the final leading model showcases the best prediction and tracking ability at both phases. The MSE and R 2 values relative to 0 and 1, respectively, indicate that the final proposed model for this case gives a satisfactory performance.

5.4. Case 4

Finally, the study of case 4 considers one missing input variable that is predicted using a sub-neural network as shown in Figure 22. Here, the first input layer has three nodes representing the input parameters: the source power, accumulated power and dissipated power (i.e., P 1 , P 2 and P 3 ). The last layer, the output layer, is composed of one node corresponding to the transmitted power (i.e., P 4 ). The number of nodes in the hidden layer is 10. The primary neural network was trained to use the predicted output variables to predict the stability after replacement in the primary neural network.
Once the missing input variable is predicted, the primary neural network is trained similarly to previous cases. The model’s performance handling the missing data is measured using R 2 and MSE. The sub-neural-network model achieved MSE values are 0.0001 in during training and testing and an R 2 value of 0.9999 during the training and testing phases. The performance of the sub-neural-network model developed for all the 60,000 samples and the first 300 samples zoomed in is depicted in Figure 23a,b for training and testing, respectively.
Further, the primary neural network was trained similarly to other cases by utilizing the testing output variables from the sub-neural network. The model obtains an MSE of 0.0084 during training and 0.0084 during the testing phase. The R 2 values were 0.9717 during training and 0.9715 during testing. The performance of the final developed neural network having 60,000 samples and the first 300 samples during both training and testing are shown in Figure 24a,b, respectively. The model’s response offers both phases the best training and prediction ability. The MSE values obtained are close to 0, and the R 2 values are relative to 1, highlighting that the proposed model is accurate for stability prediction.

5.5. Summary

Table 4 depicts the performance evaluation of the developed FFNN model using complete input data and the model that handles the missing inputs (Case 1, Case 2, Case 3 and Case 4). The R 2 and MSE results of the sub-neural-network model of case 4, with one missing input variable: transmitted power (i.e., P 4 ), shows the model’s best training and prediction ability in both phases. For all the sub-neural-network models, the R 2 value has achieved at least 70% and 97% for the primary neural network. We noticed that the MSE values obtained are close to 0. Furthermore, the R 2 values are relative to 1, which indicates the excellent performance of all the models.

6. Conclusions

The primary goal of this paper was to tackle the issue of stability prediction when there are missing variables involved. This missing variable could be due to the failure of a sensor, network connection or other system. This paper successfully solved this issue by proposing a novel FFNN model that handles missing inputs. The model’s performance was evaluated on a four-node star network. In this study, four cases of missing input variables were taken.
For each case, a sub-neural network was first prepared to predict the missing variables, and then these predicted values were fed into the primary neural network to predict the stability. Among all four cases, case 4 showed the best performance with an MSE value of 0.0001 and an R 2 value of 0.9999 during training and testing for the sub-neural network. In addition, the primary network showed an MSE value of 0.0084 and an R 2 value of 0.9717 during training and 0.9715 during testing. For all the four cases, the models achieved an MSE close to 0 and an R 2 value close to 1 thus indicating the excellent performance of the prediction models.
However, this work was limited to predicting the power parameter using a sub-neural network because the algebraic sum of the power consumed or generated was assumed to be zero, and uncertainties and disturbances were not considered. Moreover, the reaction time and price elasticity parameters are highly nonlinear in the considered dataset. As a result, this proposed model faces a shortcoming in predicting the missing variables (reaction time and price elasticity) using a sub-neural network that predicts the stability using a primary network. Therefore, extending the proposed model to predict these highly nonlinear input parameters will be addressed in our future work.

Author Contributions

Conceptualization, K.B. and M.B.O.; methodology, K.B.; software, R.M., J.C. and K.R.S.; validation, R.M., J.C. and K.R.S.; formal analysis, K.B., R.M., J.C. and K.R.S.; investigation, K.B.; resources, R.I.; data curation, R.M., J.C. and K.R.S.; writing—original draft preparation, R.M., J.C. and K.R.S.; writing—review and editing, K.B.; visualization, M.B.O.; supervision, R.I.; project administration, M.B.O. and R.I.; and funding acquisition, M.B.O. and R.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Yayasan Universiti Teknologi PETRONAS Fundamental Research Grant (YUTP-FRG) number 015-LCO0166.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this research are available from the UC Irvine Machine Learning Repository (https://archive.ics.uci.edu/ml/datasets/Electrical+Grid+Stability+Simulated+Data+, accessed on 15 February 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AdaBoostAdaptive Boosting
AdaGradAdaptive Gradient algorithm
AdamAdaptive Movement Estimation
ACEAverage Coverage Error
Adaptive LRAdaptive Linear Regression
AFCAccurate and Fast Converging
ANNArtificial Neural Network
ARIMAAutoregressive Integrated Moving Average
ARMAAutoregressive Moving Average
ARMAXAutoregressive-Moving-Average Model With Exogenous Inputs
BiGRUBidirectional Gated Recurrent Unit
BPBack Propagation
BPNNBack Propagation Neural Network
BRBayesian Regularization
C-DDPGCentralized Based Deep Deterministic Policy Gradient
CDFCumulative Distribution Function
CEEMDANComplete Ensemble Empirical Mode Decomposition Adaptive Noise
CGAChaos Search Genetic Algorithm
CGASAChaos Search Genetic Algorithm Furthermore, Simulated Annealing
CNNConvolutional Neural Network
CRBMConvolutional Restricted Boltzmann Machine
DBNDeep Belief Network
DNNDeep Neural Network
DPCSDistributed Power Consumption Scheduling
DSGCDecentral Smart Grid Control
DTDecision Tree
ECNNEnhanced Convolutional Neural Network
ELMExtreme Learning Machine
ENNElman Neural Network
ESSEnergy Storage Systems
FAForecast Accuracy
FFFeedforward
FFNNFeedforward Neural Network
FSForecast Skill
GAGenetic Algorithm
GBRGradient Boosting Regression
GDMGradient Descent Method
GRUGated Recurrent Unit
HRHit Rate
IBRInclining Block Rate
IRBDNNIterative Resblock Based Deep Neural Network
KNNk-Nearest Neighbors
LBPPLoad Based Pricing Policy
LGBMLight Gradient Boosting Machine
LIFLeaky Integrate and Fire Neuron
LMLevenberg–Marquardt
LMSLagrange Multiplier Selection
LQELink Quality Estimation
LSSVRLeast Squares Support Vector Regression
LVQLearning Vector Quantization
MABEMean Absolute Bias Error
MADMedian Absolute Deviation
MAEMean Absolute Error
MAPEMean Absolute Percentage Error
MBEMean Biased Error
MCCMatthews Correlation Coefficient
MdBPModified Back Propagation
MERMean Error Rate
MI-ANNMutual Information Artificial Neural Network
MLPMulti-Layer Perceptron
MLRMulti-variable Linear Regression
MLSTMMultiplicative Long Short-Term Memory
MNEMean Normalized Error
MPEMean Percentage Error
MSEMean Square Error
NadamNesterov-accelerated Adaptive Moment Estimation
NARXNonlinear Autoregressive Network With Exogenous Inputs
NNNeural Network
NRMSENormalized Root Mean Square Error
PARPeak To Average Ratio
PDFProbability Density Function
PDNNPooling Based Deep Neural Network
PICPPrediction Interval Coverage Probability
PINCPrediction Interval Nominal Confidence
PSOParticle Swarm Optimization
PTECCProportion Of Total Energy Classified Correctly
PVPhoto Voltaic
RCorrelation Coefficient
RBFRadial Basis Function
RBFNNBased Radial Basis Function
RDNNRecurrent Deep Neural Network
RERelative Error
ReLURectified Linear Activation Unit
RESRenewable Energy Sources
RFRandom Forest
RLReinforcement Learning
RMSERoot Mean Square Error
RNNRecurrent Neural Network
RTEPReal Time Electrical Pricing
RTPReal Time Price
SAESparse Auto Encoder
SCGScaled Conjugate Gradient
SMPSpot Market Price
SNNSpiking Neural Network
SNRSignal to Noise Ratio
SoCSpeed of Convergence
SPREADSpread of Radial Basis Functions
SSASalp Swam Algorithm
STLFShort Term Load Forecasting
STWSliding Time Window
SVMSupport Vector Machine
SWAASample Weighted Average Approximation
TanhHyperbolic Tangent Function
WNNWavelet Neural Network
WOAWhale Optimization Algorithm
WRNNWavelet Recurrent Neural Network
XGBExtreme Gradient Boosting

References

  1. Gharavi, H.; Ghafurian, R. Smart Grid: The Electric Energy System of the Future; IEEE: Piscataway, NJ, USA, 2011; Volume 99. [Google Scholar]
  2. McLaughlin, K.; Friedberg, I.; Kang, B.; Maynard, P.; Sezer, S.; McWilliams, G. Secure communications in smart grid: Networking and protocols. In Smart Grid Security; Elsevier: Amsterdam, The Netherlands, 2015; pp. 113–148. [Google Scholar]
  3. Rathnayaka, A.D.; Potdar, V.M.; Dillon, T.; Kuruppu, S. Framework to manage multiple goals in community-based energy sharing network in smart grid. Int. J. Electr. Power Energy Syst. 2015, 73, 615–624. [Google Scholar] [CrossRef]
  4. Breviglieri, P.; Erdem, T.; Eken, S. Predicting Smart Grid Stability with Optimized Deep Models. SN Comput. Sci. 2021, 2, 1–12. [Google Scholar] [CrossRef]
  5. Schäfer, B.; Grabow, C.; Auer, S.; Kurths, J.; Witthaut, D.; Timme, M. Taming instabilities in power grid networks by decentralized control. Eur. Phys. J. Spec. Top. 2016, 225, 569–582. [Google Scholar] [CrossRef]
  6. Verma, K.; Niazi, K. Generator coherency determination in a smart grid using artificial neural network. In Proceedings of the 2012 IEEE Power and Energy Society General Meeting, San Diego, CA, USA, 22–26 July 2012; pp. 1–7. [Google Scholar]
  7. Karthikumar, K.; Karthik, K.; Karunanithi, K.; Chandrasekar, P.; Sathyanathan, P.; Prakash, S.V.J. SSA-RBFNN strategy for optimum framework for energy management in Grid-Connected smart grid infrastructure modeling. Mater. Today Proc. 2021. [Google Scholar] [CrossRef]
  8. Massaoudi, M.; Abu-Rub, H.; Refaat, S.S.; Chihi, I.; Oueslati, F.S. Accurate Smart-Grid Stability Forecasting Based on Deep Learning: Point and Interval Estimation Method. In Proceedings of the 2021 IEEE Kansas Power and Energy Conference (KPEC), Manhattan, KS, USA, 19–20 April 2021; pp. 1–6. [Google Scholar]
  9. Xia, M.; Shao, H.; Ma, X.; de Silva, C.W. A Stacked GRU-RNN-based Approach for Predicting Renewable Energy and Electricity Load for Smart Grid Operation. IEEE Trans. Ind. Inform. 2021, 17, 7050–7059. [Google Scholar] [CrossRef]
  10. Gao, B.; Huang, X.; Shi, J.; Tai, Y.; Zhang, J. Hourly forecasting of solar irradiance based on CEEMDAN and multi-strategy CNN-LSTM neural networks. Renew. Energy 2020, 162, 1665–1683. [Google Scholar] [CrossRef]
  11. Li, J.; Deng, D.; Zhao, J.; Cai, D.; Hu, W.; Zhang, M.; Huang, Q. A novel hybrid short-term load forecasting method of smart grid using mlr and lstm neural network. IEEE Trans. Ind. Inform. 2021, 17, 2443–2452. [Google Scholar] [CrossRef]
  12. Mohammad, F.; Kim, Y.C. Energy load forecasting model based on deep neural networks for smart grids. Int. J. Syst. Assur. Eng. Manag. 2020, 11, 824–834. [Google Scholar] [CrossRef]
  13. Capizzi, G.; Sciuto, G.L.; Napoli, C.; Tramontana, E. Advanced and adaptive dispatch for smart grids by means of predictive models. IEEE Trans. Smart Grid 2018, 9, 6684–6691. [Google Scholar] [CrossRef]
  14. Jeyaraj, P.R.; Nadar, E.R.S. Computer-assisted demand-side energy management in residential smart grid employing novel pooling deep learning algorithm. Int. J. Energy Res. 2021, 45, 7961–7973. [Google Scholar] [CrossRef]
  15. Islam, B.; Baharudin, Z.; Nallagownden, P. Development of chaotically improved meta-heuristics and modified BP neural network-based model for electrical energy demand prediction in smart grid. Neural Comput. Appl. 2017, 28, 877–891. [Google Scholar] [CrossRef]
  16. Gupta, S.; Kazi, F.; Wagh, S.; Kambli, R. Neural network based early warning system for an emerging blackout in smart grid power networks. In Intelligent Distributed Computing; Springer: Berlin/Heidelberg, Germany, 2015; pp. 173–183. [Google Scholar]
  17. Neupane, B.; Perera, K.S.; Aung, Z.; Woon, W.L. Artificial neural network-based electricity price forecasting for smart grid deployment. In Proceedings of the 2012 International Conference on Computer Systems and Industrial Informatics, Sharjah, United Arab Emirates, 18–20 December 2012; pp. 1–6. [Google Scholar]
  18. Verma, K.; Niazi, K. Determination of vulnerable machines for online transient security assessment in smart grid using artificial neural network. In Proceedings of the 2011 Annual IEEE India Conference, Yderabad, India, 16–18 December 2011; pp. 1–5. [Google Scholar]
  19. Sakellariou, M.; Ferentinou, M. A study of slope stability prediction using neural networks. Geotech. Geol. Eng. 2005, 23, 419–445. [Google Scholar] [CrossRef]
  20. Tu, J.V. Advantages and disadvantages of using artificial neural networks versus logistic regression for predicting medical outcomes. J. Clin. Epidemiol. 1996, 49, 1225–1231. [Google Scholar] [CrossRef]
  21. Nijman, S.; Leeuwenberg, A.; Beekers, I.; Verkouter, I.; Jacobs, J.; Bots, M.; Asselbergs, F.; Moons, K.; Debray, T. Missing data is poorly handled and reported in prediction model studies using machine learning: A literature review. J. Clin. Epidemiol. 2022, 142, 218–229. [Google Scholar] [CrossRef]
  22. Wang, S.; Bi, S.; Zhang, Y.J.A. Locational detection of the false data injection attack in a smart grid: A multilabel classification approach. IEEE Internet Things J. 2020, 7, 8218–8227. [Google Scholar] [CrossRef]
  23. Niu, X.; Li, J.; Sun, J.; Tomsovic, K. Dynamic detection of false data injection attack in smart grid using deep learning. In Proceedings of the 2019 IEEE Power & Energy Society Innovative Smart Grid Technologies Conference (ISGT), Washington, DC, USA, 18–21 February 2019; pp. 1–6. [Google Scholar]
  24. Sun, W.; Lu, W.; Li, Q.; Chen, L.; Mu, D.; Yuan, X. WNN-LQE: Wavelet-neural-network-based link quality estimation for smart grid WSNs. IEEE Access 2017, 5, 12788–12797. [Google Scholar] [CrossRef]
  25. Ungureanu, S.; Ţopa, V.; Cziker, A. Integrating the industrial consumer into smart grid by load curve forecasting using machine learning. In Proceedings of the 2019 8th International Conference on Modern Power Systems (MPS), Cluj-Napoca, Romania, 21–23 May 2019; pp. 1–9. [Google Scholar]
  26. Alamaniotis, M. Synergism of deep neural network and elm for smart very-short-term load forecasting. In Proceedings of the 2019 IEEE PES Innovative Smart Grid Technologies Europe (ISGT-Europe), Bucharest, Romania, 29 September–2 October 2019; pp. 1–5. [Google Scholar]
  27. Zahid, M.; Ahmed, F.; Javaid, N.; Abbasi, R.A.; Zainab Kazmi, H.S.; Javaid, A.; Bilal, M.; Akbar, M.; Ilahi, M. Electricity price and load forecasting using enhanced convolutional neural network and enhanced support vector regression in smart grids. Electronics 2019, 8, 122. [Google Scholar] [CrossRef] [Green Version]
  28. Çavdar, İ.H.; Faryad, V. New design of a supervised energy disaggregation model based on the deep neural network for a smart grid. Energies 2019, 12, 1217. [Google Scholar] [CrossRef] [Green Version]
  29. Selim, M.; Zhou, R.; Feng, W.; Quinsey, P. Estimating Energy Forecasting Uncertainty for Reliable AI Autonomous Smart Grid Design. Energies 2021, 14, 247. [Google Scholar] [CrossRef]
  30. Alazab, M.; Khan, S.; Krishnan, S.S.R.; Pham, Q.V.; Reddy, M.P.K.; Gadekallu, T.R. A multidirectional LSTM model for predicting the stability of a smart grid. IEEE Access 2020, 8, 85454–85463. [Google Scholar] [CrossRef]
  31. Hasan, M.; Toma, R.N.; Nahid, A.A.; Islam, M.; Kim, J.M. Electricity theft detection in smart grid systems: A CNN-LSTM based approach. Energies 2019, 12, 3310. [Google Scholar] [CrossRef] [Green Version]
  32. Xu, B.; Guo, F.; Zhang, W.A.; Li, G.; Wen, C. E2DNet: An Ensembling Deep Neural Network for Solving Nonconvex Economic Dispatch in Smart Grid. IEEE Trans. Ind. Inform. 2021, 18, 21589379. [Google Scholar]
  33. Hong, Y.; Zhou, Y.; Li, Q.; Xu, W.; Zheng, X. A deep learning method for short-term residential load forecasting in smart grid. IEEE Access 2020, 8, 55785–55797. [Google Scholar] [CrossRef]
  34. Zheng, Y.; Celik, B.; Suryanarayanan, S.; Maciejewski, A.A.; Siegel, H.J.; Hansen, T.M. An aggregator-based resource allocation in the smart grid using an artificial neural network and sliding time window optimization. IET Smart Grid 2021, 4, 612–622. [Google Scholar] [CrossRef]
  35. Bingi, K.; Prusty, B.R. Forecasting models for chaotic fractional-order oscillators using neural networks. Int. J. Appl. Math. Comput. Sci. 2021, 31, 387–398. [Google Scholar] [CrossRef]
  36. Bingi, K.; Prusty, B.R. Chaotic Time Series Prediction Model for Fractional-Order Duffing’s Oscillator. In Proceedings of the 2021 8th International Conference on Smart Computing and Communications (ICSCC), Kochi, India, 1–3 July 2021; pp. 357–361. [Google Scholar]
  37. Bingi, K.; Prusty, B.R. Neural Network-Based Models for Prediction of Smart Grid Stability. In Proceedings of the 2021 Innovations in Power and Advanced Computing Technologies (i-PACT), Kuala Lumpur, Malaysia, 27–29 November 2021; pp. 1–6. [Google Scholar]
  38. Qi, X.; Chen, G.; Li, Y.; Cheng, X.; Li, C. Applying neural-network-based machine learning to additive manufacturing: Current applications, challenges, and future perspectives. Engineering 2019, 5, 721–729. [Google Scholar] [CrossRef]
  39. Sattari, M.A.; Roshani, G.H.; Hanus, R.; Nazemi, E. Applicability of time-domain feature extraction methods and artificial intelligence in two-phase flow meters based on gamma-ray absorption technique. Measurement 2021, 168, 108474. [Google Scholar] [CrossRef]
  40. Chandrasekaran, K.; Selvaraj, J.; Amaladoss, C.R.; Veerapan, L. Hybrid renewable energy based smart grid system for reactive power management and voltage profile enhancement using artificial neural network. Energy Sources Part A Recover. Util. Environ. Eff. 2021, 43, 2419–2442. [Google Scholar] [CrossRef]
  41. Zhou, Z.; Xiang, Y.; Xu, H.; Wang, Y.; Shi, D. Unsupervised Learning for Non-Intrusive Load Monitoring in Smart Grid Based on Spiking Deep Neural Network. J. Mod. Power Syst. Clean Energy 2021, 10, 606–616. [Google Scholar] [CrossRef]
  42. Chung, H.M.; Maharjan, S.; Zhang, Y.; Eliassen, F. Distributed deep reinforcement learning for intelligent load scheduling in residential smart grids. IEEE Trans. Ind. Inform. 2021, 17, 2752–2763. [Google Scholar] [CrossRef]
  43. Cahyono, M.R.A. Design Power Controller for Smart Grid System Based on Internet of Things Devices and Artificial Neural Network. In Proceedings of the 2020 IEEE International Conference on Internet of Things and Intelligence System (IoTaIS), Bali, Indonesia, 27–28 January 2021; pp. 44–48. [Google Scholar]
  44. Khan, S.; Kifayat, K.; Kashif Bashir, A.; Gurtov, A.; Hassan, M. Intelligent intrusion detection system in smart grid using computational intelligence and machine learning. Trans. Emerg. Telecommun. Technol. 2021, 32, e4062. [Google Scholar] [CrossRef]
  45. Ullah, A.; Javaid, N.; Samuel, O.; Imran, M.; Shoaib, M. CNN and GRU based deep neural network for electricity theft detection to secure smart grid. In Proceedings of the 2020 International Wireless Communications and Mobile Computing (IWCMC), Limassol, Cyprus, 15–19 June 2020; pp. 1598–1602. [Google Scholar]
  46. Ruan, G.; Zhong, H.; Wang, J.; Xia, Q.; Kang, C. Neural-network-based Lagrange multiplier selection for distributed demand response in smart grid. Appl. Energy 2020, 264, 114636. [Google Scholar] [CrossRef]
  47. Di Piazza, A.; Di Piazza, M.C.; La Tona, G.; Luna, M. An artificial neural network-based forecasting model of energy-related time series for electrical grid management. Math. Comput. Simul. 2021, 184, 294–305. [Google Scholar] [CrossRef]
  48. Khalid, Z.; Abbas, G.; Awais, M.; Alquthami, T.; Rasheed, M.B. A novel load scheduling mechanism using artificial neural network based customer profiles in smart grid. Energies 2020, 13, 1062. [Google Scholar] [CrossRef] [Green Version]
  49. Fan, L.; Li, J.; Pan, Y.; Wang, S.; Yan, C.; Yao, D. Research and application of smart grid early warning decision platform based on big data analysis. In Proceedings of the 2019 4th International Conference on Intelligent Green Building and Smart Grid (IGBSG), Yichang, China, 6–9 September 2019; pp. 645–648. [Google Scholar]
  50. Li, G.; Wang, H.; Zhang, S.; Xin, J.; Liu, H. Recurrent neural networks based photovoltaic power forecasting approach. Energies 2019, 12, 2538. [Google Scholar] [CrossRef] [Green Version]
  51. Huang, X.; Shi, J.; Gao, B.; Tai, Y.; Chen, Z.; Zhang, J. Forecasting hourly solar irradiance using hybrid wavelet transformation and Elman model in smart grid. IEEE Access 2019, 7, 139909–139923. [Google Scholar] [CrossRef]
  52. Haghnegahdar, L.; Wang, Y. A whale optimization algorithm-trained artificial neural network for smart grid cyber intrusion detection. Neural Comput. Appl. 2020, 32, 9427–9441. [Google Scholar] [CrossRef]
  53. Ahmed, F.; Zahid, M.; Javaid, N.; Khan, A.B.M.; Khan, Z.A.; Murtaza, Z. A deep learning approach towards price forecasting using enhanced convolutional neural network in smart grid. In International Conference on Emerging Internetworking, Data & Web Technologies; Springer: Berlin/Heidelberg, Germany, 2019; pp. 271–283. [Google Scholar]
  54. Duong-Ngoc, H.; Nguyen-Thanh, H.; Nguyen-Minh, T. Short term load forcast using deep learning. In Proceedings of the 2019 Innovations in Power and Advanced Computing Technologies (i-PACT), Vellore, India, 22–23 March 2019; Volume 1, pp. 1–5. [Google Scholar]
  55. Kulkarni, S.N.; Shingare, P. Artificial Neural Network Based Short Term Power Demand Forecast for Smart Grid. In Proceedings of the 2018 IEEE Conference on Technologies for Sustainability (SusTech), Long Beach, CA, USA, 11–13 November 2018; pp. 1–7. [Google Scholar]
  56. Ghasemi, A.A.; Gitizadeh, M. Detection of illegal consumers using pattern classification approach combined with Levenberg-Marquardt method in smart grid. Int. J. Electr. Power Energy Syst. 2018, 99, 363–375. [Google Scholar] [CrossRef]
  57. Vrablecová, P.; Ezzeddine, A.B.; Rozinajová, V.; Šárik, S.; Sangaiah, A.K. Smart grid load forecasting using online support vector regression. Comput. Electr. Eng. 2017, 65, 102–117. [Google Scholar] [CrossRef]
  58. Ahmad, A.; Javaid, N.; Guizani, M.; Alrajeh, N.; Khan, Z.A. An accurate and fast converging short-term load forecasting model for industrial applications in a smart grid. IEEE Trans. Ind. Inform. 2017, 13, 2587–2596. [Google Scholar] [CrossRef]
  59. Li, L.; Ota, K.; Dong, M. Everything is image: CNN-based short-term electrical load forecasting for smart grid. In Proceedings of the 2017 14th International Symposium on Pervasive Systems, Algorithms and Networks & 2017 11th International Conference on Frontier of Computer Science and Technology & 2017 Third International Symposium of Creative Computing (ISPAN-FCST-ISCC), Exeter, UK, 21–23 June 2017; pp. 344–351. [Google Scholar]
  60. Bicer, Y.; Dincer, I.; Aydin, M. Maximizing performance of fuel cell using artificial neural network approach for smart grid applications. Energy 2016, 116, 1205–1217. [Google Scholar] [CrossRef]
  61. Macedo, M.N.; Galo, J.J.; De Almeida, L.; Lima, A.d.C. Demand side management using artificial neural networks in a smart grid environment. Renew. Sustain. Energy Rev. 2015, 41, 128–133. [Google Scholar] [CrossRef]
  62. Muralidharan, S.; Roy, A.; Saxena, N. Stochastic hourly load forecasting for smart grids in korea using narx model. In Proceedings of the 2014 International Conference on Information and Communication Technology Convergence (ICTC), Busan, Korea, 22–24 October 2014; pp. 167–172. [Google Scholar]
  63. Ioakimidis, C.; Eliasstam, H.; Rycerski, P. Solar power forecasting of a residential location as part of a smart grid structure. In Proceedings of the 2012 IEEE Energytech, Cleveland, OH, USA, 29–31 May 2012; pp. 1–6. [Google Scholar]
  64. Hashiesh, F.; Mostafa, H.E.; Khatib, A.R.; Helal, I.; Mansour, M.M. An intelligent wide area synchrophasor based system for predicting and mitigating transient instabilities. IEEE Trans. Smart Grid 2012, 3, 645–652. [Google Scholar] [CrossRef]
  65. Fei, W.; Zengqiang, M.; Shi, S.; Chengcheng, Z. A practical model for single-step power prediction of grid-connected PV plant using artificial neural network. In Proceedings of the 2011 IEEE PES Innovative Smart Grid Technologies, Perth, WA, USA, 13–16 November 2011; pp. 1–4. [Google Scholar]
  66. Qudaih, Y.S.; Mitani, Y. Power distribution system planning for smart grid applications using ANN. Energy Procedia 2011, 12, 3–9. [Google Scholar] [CrossRef] [Green Version]
  67. Zhang, H.T.; Xu, F.Y.; Zhou, L. Artificial neural network for load forecasting in smart grid. In Proceedings of the 2010 International Conference on Machine Learning and Cybernetics, Qingdao, China, 11–14 July 2010; Volume 6, pp. 3200–3205. [Google Scholar]
  68. Arzamasov, V.; Böhm, K.; Jochem, P. Towards concise models of grid stability. In Proceedings of the 2018 IEEE International Conference on Communications, Control, and Computing Technologies for Smart Grids (SmartGridComm), Aalborg, Denmark, 29–31 October 2018; pp. 1–6. [Google Scholar]
  69. Akoglu, H. User’s guide to correlation coefficients. Turk. J. Emerg. Med. 2018, 18, 91–93. [Google Scholar] [CrossRef]
  70. Sheela, K.G.; Deepa, S.N. Review on methods to fix number of hidden neurons in neural networks. Math. Probl. Eng. 2013, 2013. [Google Scholar] [CrossRef] [Green Version]
  71. Bingi, K.; Prusty, B.R.; Panda, K.P.; Panda, G. Time Series Forecasting Model for Chaotic Fractional-Order Rössler System. In Sustainable Energy and Technological Advancements; Springer: Berlin/Heidelberg, Germany, 2022; pp. 799–810. [Google Scholar]
  72. Shaik, N.B.; Pedapati, S.R.; Othman, A.; Bingi, K.; Dzubir, F.A.A. An intelligent model to predict the life condition of crude oil pipelines using artificial neural networks. Neural Comput. Appl. 2021, 33, 14771–14792. [Google Scholar] [CrossRef]
  73. Bingi, K.; Prusty, B.R.; Kumra, A.; Chawla, A. Torque and temperature prediction for permanent magnet synchronous motor using neural networks. In Proceedings of the 2020 3rd International Conference on Energy, Power and Environment: Towards Clean Energy Technologies, Shillong, India, 5–7 March 2021; pp. 1–6. [Google Scholar]
  74. Ramadevi, B.; Bingi, K. Chaotic Time Series Forecasting Approaches Using Machine Learning Techniques: A Review. Symmetry 2022, 14, 955. [Google Scholar] [CrossRef]
Figure 1. Year-wise and the publisher-wise contributions to smart grid’s stability forecasting during the last decade.
Figure 1. Year-wise and the publisher-wise contributions to smart grid’s stability forecasting during the last decade.
Sensors 22 04342 g001
Figure 2. Summary of the smart grid architectures identified in the conducted literature survey.
Figure 2. Summary of the smart grid architectures identified in the conducted literature survey.
Sensors 22 04342 g002
Figure 3. Classification of multiple neural-network-based models developed for smart grid stability prediction.
Figure 3. Classification of multiple neural-network-based models developed for smart grid stability prediction.
Sensors 22 04342 g003
Figure 4. Summary of the various training algorithms and activation functions used in neural-network-based models for smart grid stability prediction.
Figure 4. Summary of the various training algorithms and activation functions used in neural-network-based models for smart grid stability prediction.
Sensors 22 04342 g004
Figure 5. The architecture of the four-node star network.
Figure 5. The architecture of the four-node star network.
Sensors 22 04342 g005
Figure 6. Dataset of predictive and dependent features of four-node star network. (a) Reaction time τ j ; (b) Produced/consumed power P j ; (c) Elasticity coefficient γ j ; (d) Re ( λ ) .
Figure 6. Dataset of predictive and dependent features of four-node star network. (a) Reaction time τ j ; (b) Produced/consumed power P j ; (c) Elasticity coefficient γ j ; (d) Re ( λ ) .
Sensors 22 04342 g006
Figure 7. Pearson’s correlation matrix of the study variables.
Figure 7. Pearson’s correlation matrix of the study variables.
Sensors 22 04342 g007
Figure 8. Research flow diagram for the design of smart grid stability model.
Figure 8. Research flow diagram for the design of smart grid stability model.
Sensors 22 04342 g008
Figure 9. Flow chart of implementation of prediction model with complete input data.
Figure 9. Flow chart of implementation of prediction model with complete input data.
Sensors 22 04342 g009
Figure 10. The architecture of FFNN for predicting smart grid’s stability.
Figure 10. The architecture of FFNN for predicting smart grid’s stability.
Sensors 22 04342 g010
Figure 11. Performance comparison of stability prediction (a) training and (b) testing.
Figure 11. Performance comparison of stability prediction (a) training and (b) testing.
Sensors 22 04342 g011
Figure 12. Flow chart of implementation of prediction model that handles missing input data for the four cases.
Figure 12. Flow chart of implementation of prediction model that handles missing input data for the four cases.
Sensors 22 04342 g012
Figure 13. The architecture of FFNN developed for case 1.
Figure 13. The architecture of FFNN developed for case 1.
Sensors 22 04342 g013
Figure 14. Performance of the sub-neural network for case 1 during (a) training and (b) testing.
Figure 14. Performance of the sub-neural network for case 1 during (a) training and (b) testing.
Sensors 22 04342 g014
Figure 15. Performance of the primary neural network for case 1 during (a) training and (b) testing.
Figure 15. Performance of the primary neural network for case 1 during (a) training and (b) testing.
Sensors 22 04342 g015
Figure 16. The architecture of FFNN developed for case 2.
Figure 16. The architecture of FFNN developed for case 2.
Sensors 22 04342 g016
Figure 17. Performance of the sub-neural network for case 2 during (a) training and (b) testing.
Figure 17. Performance of the sub-neural network for case 2 during (a) training and (b) testing.
Sensors 22 04342 g017
Figure 18. Performance of the primary neural network for case 2 during (a) training and (b) testing.
Figure 18. Performance of the primary neural network for case 2 during (a) training and (b) testing.
Sensors 22 04342 g018
Figure 19. The architecture of FFNN developed for case 3.
Figure 19. The architecture of FFNN developed for case 3.
Sensors 22 04342 g019
Figure 20. Performance of the sub-neural network for case 3 during (a) training and (b) testing.
Figure 20. Performance of the sub-neural network for case 3 during (a) training and (b) testing.
Sensors 22 04342 g020
Figure 21. Performance of the primary neural network for case 3 during (a) training and (b) testing.
Figure 21. Performance of the primary neural network for case 3 during (a) training and (b) testing.
Sensors 22 04342 g021
Figure 22. The architecture of FFNN developed for case 4.
Figure 22. The architecture of FFNN developed for case 4.
Sensors 22 04342 g022
Figure 23. Performance of the sub-neural network for case 4 during (a) training and (b) testing.
Figure 23. Performance of the sub-neural network for case 4 during (a) training and (b) testing.
Sensors 22 04342 g023
Figure 24. Performance of the primary neural network for case 4 during (a) training and (b) testing.
Figure 24. Performance of the primary neural network for case 4 during (a) training and (b) testing.
Sensors 22 04342 g024
Table 2. Predictive features and simulation constants used for data generation of four-node star network.
Table 2. Predictive features and simulation constants used for data generation of four-node star network.
CategoryParameterRange/Value
τ j [ 0.5 , 10 ] s
Predictive features P j [ 2.0 , 0.5 ] s 2
γ j [ 0.05 , 1 ] s 1
α j 0.1 s 1
Simulation constants T j 2 s
K j m 8 s 2
Table 3. Interpretation of Pearson’s correlation coefficients.
Table 3. Interpretation of Pearson’s correlation coefficients.
CoefficientInterpretation
±0.90–±1.00Very strong correlation
±0.70–±0.89Strong correlation
±0.40–±0.69Moderate correlation
±0.10–±0.39Weak correlation
0.00–±0.09Negligible correlation
Table 4. Performance evaluation comparison of the FFNN with complete input data and the FFNN that handles the missing data.
Table 4. Performance evaluation comparison of the FFNN with complete input data and the FFNN that handles the missing data.
CategoryCaseNetworkStageR 2 MSE
With Complete
Input Data
-PrimaryTraining0.97390.0077
Testing0.97380.0077
Model that
Handles
Missing Input
Data
Case 1SubTraining0.99920.0008
Testing0.99920.0008
PrimaryTraining0.97210.0080
Testing0.84130.0085
Case 2SubTraining0.70820.1661
Testing0.70720.1667
PrimaryTraining0.97380.0077
Testing0.97380.0077
Case 3SubTraining0.70850.1659
Testing0.70610.1673
PrimaryTraining0.97200.0083
Testing0.97210.0082
Case 4SubTraining0.99990.0001
Testing0.99990.0001
PrimaryTraining0.97170.0084
Testing0.97150.0084
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Omar, M.B.; Ibrahim, R.; Mantri, R.; Chaudhary, J.; Ram Selvaraj, K.; Bingi, K. Smart Grid Stability Prediction Model Using Neural Networks to Handle Missing Inputs. Sensors 2022, 22, 4342. https://doi.org/10.3390/s22124342

AMA Style

Omar MB, Ibrahim R, Mantri R, Chaudhary J, Ram Selvaraj K, Bingi K. Smart Grid Stability Prediction Model Using Neural Networks to Handle Missing Inputs. Sensors. 2022; 22(12):4342. https://doi.org/10.3390/s22124342

Chicago/Turabian Style

Omar, Madiah Binti, Rosdiazli Ibrahim, Rhea Mantri, Jhanavi Chaudhary, Kaushik Ram Selvaraj, and Kishore Bingi. 2022. "Smart Grid Stability Prediction Model Using Neural Networks to Handle Missing Inputs" Sensors 22, no. 12: 4342. https://doi.org/10.3390/s22124342

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop