Next Article in Journal
Phylogenetic Weighting Does Little to Improve the Accuracy of Evolutionary Coupling Analyses
Previous Article in Journal
Physical-Layer Security Analysis over M-Distributed Fading Channels
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Recognition of Voltage Sag Sources Based on Phase Space Reconstruction and Improved VGG Transfer Learning

1
College of Electrical Engineering, Sichuan University, Chengdu 610065, China
2
China Mobile (Hangzhou) Information Technology Co., Ltd., Hangzhou 311100, China
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(10), 999; https://doi.org/10.3390/e21100999
Submission received: 4 September 2019 / Revised: 30 September 2019 / Accepted: 30 September 2019 / Published: 12 October 2019

Abstract

:
The recognition of the voltage sag sources is the basis for formulating a voltage sag governance plan and clarifying the responsibility for the accident. Aiming at the recognition problem of voltage sag sources, a recognition method of voltage sag sources based on phase space reconstruction and improved Visual Geometry Group (VGG) transfer learning is proposed from the perspective of image classification. Firstly, phase space reconstruction technology is used to transform voltage sag signals, generate reconstruction images of voltage sag, and analyze the intuitive characteristics of different sag sources from reconstruction images. Secondly, combined with the attention mechanism, the standard VGG 16 model is improved to extract the features completely and prevent over-fitting. Finally, VGG transfer learning model uses the idea of transfer learning for training, which improves the efficiency of model training and the recognition accuracy of sag sources. The purpose of the training model is to minimize the cross entropy loss function. The simulation analysis verifies the effectiveness and superiority of the proposed method.

1. Introduction

In recent years, with the widespread use of power electronic devices in power grid and sensitive devices in industrial production, the impact of voltage sag has gradually attracted attention in the electrical field. Accurate recognition of the source of voltage sag can help the timely formulation of the governance plan and the clear division of the responsibilities of both parties in the accident, effectively reducing economic losses and resolving related disputes [1].
At present, the research on the recognition of voltage sag sources falls into two categories: direct methods [2,3,4,5,6,7] and indirect methods [8,9,10,11,12,13,14,15,16,17,18,19]. The direct methods include the RMS method [2,3] and the deep learning method [4,5,6,7]. Indirect methods include two parts: feature extraction and pattern recognition. Common methods for feature extraction include wavelet transform [8,9], Fourier transform [10,11], S transform [12], Hilbert transform [13], and empirical mode decomposition [14], etc. The main methods of pattern recognition include neural network [15,16], support vector machine [17], principal component analysis [18], fuzzy comprehensive evaluation [19], and so on. Among them, the RMS method [2,3] is simple and easy to implement, but it is easy to produce misjudgment for complex sag situations; the deep learning method [4,5,6,7] does not rely on manual extraction of features, but the model training efficiency is low. In the feature extraction process of indirect methods, the mathematical models are mature, the features are clear. However, indirect methods are limited to two-dimensional plane characteristics, and there are still some shortcomings: wavelet transform [8] is based on weak signal interference and good frequency domain characteristics, otherwise the correct rate will be seriously affected; the time-frequency localization property of short-time Fourier transform [10] is poor, so it cannot well reflect the transient characteristics of voltage change. S-transform [12] has low resolution when analyzing signals with harmonic components. In the pattern recognition process of indirect methods, the neural network method [15] needs to manually set the characteristics which need to be extracted in advance and then use the classifier to classify, relying too much on expert experience to discriminate rules; the recognition results of multi-classifier support vector machine are susceptible to thresholds [17]; the principal component analysis method [18] selects the main feature quantity, and although the number of feature quantities is reduced, the information contained in part of the sample data is lost.
In order to overcome the above problems, from the perspective of image classification, firstly, this paper proposes to use phase space reconstruction technology to extract the characteristics of voltage sag and map voltage sag signals into high-dimensional space [20]. The high-dimensional space image of sag keeps the complete information. By highlighting the characteristics reflected in high-dimensional space, it is obvious and easy to recognize different voltage sag sources. Then, the Visual Geometry Group (VGG) 16 model [21] is improved. There are a lot of research achievements about VGG model in the field of image classification, and its classification effect is good, so it is widely used [22,23,24]. Before the VGG model automatically extracts features, the attention mechanism is added to make the feature description more complete, which further prevents over-fitting and improves the generalization ability of the model [25]. Attention mechanism in deep learning simulates the attention model of human brain. At present, it is very popular and widely used in machine translation [26], speech recognition [27], image caption [28], and many other fields. In addition, VGG transfer learning model is constructed by using the idea of transfer learning [29], which can improve the efficiency of network training and the recognition accuracy of voltage sag sources.
The rest of this paper is organized as follows. In Section 2, feature extraction of different kinds of voltage sags is analyzed with phase space reconstruction theory. The theory of VGG modeling with attention mechanism and voltage sag recognition method with transfer learning are presented in Section 3. In Section 4, some simulation data are used to verify the effectiveness of the proposed method. Conclusions are drawn in Section 5.

2. Voltage Sag Source Feature Extraction Based on Phase Space Reconstruction Theory

2.1. Classification of Voltage Sag Sources

At present, there are three main causes of voltage sag: short circuit fault, large induction motor starting, and unloaded transformer energizing.
  • Short circuit fault is the main cause of voltage sag. Different short circuit faults can cause different sags. The voltage sag caused by three phase short circuit fault is equal in three-phase voltage magnitude. The three-phase magnitude of voltage sag caused by other short circuit types is different. Voltage swell may occur while sag occurs in an asymmetric short circuit. At the beginning and end of the voltage sag, the magnitude suddenly changes, and there is no change in the voltage magnitude during the sag.
  • When a large induction motor is starting, it will draw much larger current from the power supply than normal operation. The typical starting current is 5-6 times the rated working current, thus resulting in voltage sag. When the sag occurs, the three-phase voltage drops at the same time, and the sag magnitude is basically the same. There is no sudden change in the recovery process, and it is gradually recovered.
  • Because of the saturation characteristic of the core, the inrush current of transformer when switched on and off is several times the rated current, which will cause voltage sag. The initial phase angle of three-phase voltage always differs by 120 degrees, so the magnitude of three-phase sag is always unbalanced. Large transformers usually need dozens of cycles to recover because of their small resistance and large reactance. In addition, the voltage waveform of sag contains higher harmonics.
For short circuit faults, according to the fault phases, they are divided into seven types: A1a, A1b, A1c, A2ab, A2ac, A2bc, and A3abc which respectively represent A-phase short circuit, B-phase short circuit, C-phase short circuit, A- and B-phase short circuit, A- and C-phase short circuit, B- and C-phase short circuit, and three phase short circuit fault. For the starting of induction motor, it is recorded as B1. For large transformer energizing, it is recorded as C1. So, there are nine types of voltage sag sources.

2.2. Phase Space Reconstruction Theory

Phase space reconstruction theory holds that the development process of any component in a dynamic system implies information of other relevant components, and the original change rule of the system can be extracted and restored by analyzing the time series data of a component.
For one-dimensional time series x = {x1, x2, ..., xN}, according to Takens’ delay time embedding theory [30], one-dimensional time series is extended to high-dimensional space by using two reconstruction parameters (delay time τ and embedding dimension m):
X = [ X 1 X 2 X M ] = [ x 1 x 1 + τ x 1 + ( m 1 ) τ x 2 x 2 + τ x 2 + ( m 1 ) τ x M x M + τ x M + ( m 1 ) τ ] = [ U 1 U 2 U m ]
where column vectors Uk (k = 1, 2, ..., m) represent coordinates of each dimension, line vectors xk (k = 1, 2, ..., M) constitute the phase points in the phase space, X is a matrix of M × m, and Xk is a vector of 1 × m, τ represents delay time which is the sampling interval of time series, M = N − (m − 1)τ. These M points together constitute the phase trajectory of reconstruction phase space from voltage sag time series.
In order to ensure the visibility of the reconstruction images, m can take 1, 2, 3. At the same time, the larger m is, the larger the dimension of phase space is, the more information it contains. Therefore, m = 3 is the most appropriate. When m = 3, Equation (1) becomes
X = [ X 1 X 2 X M ] = [ x 1 x 1 + τ x 1 + 2 τ x 2 x 2 + τ x 2 + 2 τ x M x M + τ x M + 2 τ ] = [ U 1 U 2 U 3 ]
With C-C method [31], we determine τ = 2. So, Equation (2) becomes
X = [ X 1 X 2 X M ] = [ x 1 x 3 x 5 x 2 x 4 x 6 x M x M + 2 x M + 4 ] = [ U 1 U 2 U 3 ]
In this paper, the sampling interval is 1/1000 s, so τ = 2 means 2/1000 s.

2.3. Phase Space Reconstruction of Different Voltage Sag Signals

The basic frequency of the simulation model is set to 50 Hz, and the total simulation time is set to 1 s. The sampling frequency is set to 1 kHz, so the sampling point of voltage signal is 1000. The phase space reconstruction of voltage sag signal caused by single phase short circuit fault, large induction motor starting and unloaded transformer energizing is carried out. The three-dimensional coordinates are recorded as Ux, Uy, and Uz, respectively, which equal to U1, U2, and U3. The reconstructed images are shown in Figure 1, Figure 2 and Figure 3. The x-coordinate of sampling time is adjusted by omitting 0.34 s when no sag occurs to obtain better quality, while the sampling time of reconstructed image is still one second.
From Figure 1, Figure 2 and Figure 3, it can be seen that the attractor of stable sinusoidal waveform is in the form of a limit cycle in phase space, and the size of the limit cycle represents the size of sinusoidal wave. For example, when the single phase short circuit fault occurs in Figure 1, the magnitude of phase A decreases, while there is a very small limit cycle in the corresponding reconstruction image; when the magnitude of phase B and C increases, there is a corresponding larger limit cycle in the reconstruction image. A larger limit cycle can be used to judge the type of fault as A-phase short circuit fault. The voltage changes slowly when the induction motor is starting. So, in Figure 2, many limit cycles are generated, and there is no sudden change between the cycles. The most special one is Figure 3: in addition to limit cycles, there are strange attractors in the reconstruction image of transformer energizing, which are caused by high-order harmonics in voltage signals. Therefore, the identification of voltage sag sources based on phase space reconstruction image has the following main characteristics:
  • The number of limit cycles
  • The size of limit cycles
  • The existence of strange attractors
  • The number of mutation trajectories
In the case of short circuit fault, the duration of voltage sag is shown as overlap of phase trajectory on phase space reconstruction image, which does not affect its identification. The phase space reconstruction of voltage sag retains the complete sag information and is more intuitive and significant than the original signal waveform.

3. Recognition of Voltage Sag Source Based on VGG

3.1. VGG Network Structure

VGG is a deep convolution neural network developed by researchers from the Visual Geometry Group of Oxford University and Google DeepMind Company. VGG 16 is a classical algorithm for image classification [21].
As shown in Figure 4, VGG 16 is simple in structure, consisting of 13 convolution layers, 5 pooling layers, 3 fully connected layers, and 1 softmax output layer. All activation units of hidden layers adopt Rectified Linear Units (ReLU) function. The convolution kernels are 3 × 3 and the pooling kernels are 2 × 2. Among them, VGG 16 uses several convolution layers with smaller convolution kernels (3 × 3), which on one hand can reduce parameters, and on the other hand, it is equivalent to more non-linear mappings, which can increase the fitting ability of the network. The number of channels in the first layer of VGG 16 network is 64, and the number of channels in each layer of VGG16 network is doubled, up to 512 channels. With the increase of the number of channels, more information can be extracted. In addition, the convolution kernel focuses on enlarging the number of channels and the pooling kernel focuses on narrowing the width and height, which make the model deeper and wider in structure, and control the amount of calculation at the same time.
The supervised pre-training of VGG can be divided into two processes: forward propagation and backward propagation. Forward propagation calculates input characteristics at each level as follows:
x ( l ) = f ( w ( l ) x ( l 1 ) + b ( l ) )
where l is the current layer, x(l–1) is the input of the layer, x(l) is the output of the layer, w is the weight, b is the bias, f(·) is the ReLU function. The equation of the ReLU function is
f ( x ) = max ( 0 , x )
Convolution operation is a linear operation. In order to increase the non-linear ability of the neural network and consider the training complexity of the network, ReLU is used as an activation function to perform nonlinear transformation on the feature after linear operation.
When backward propagation occurs, the parameters wij(l) and bi(l) of each layer are updated by batch gradient descent method. The updated formulas are as follows:
{ w i j ( l ) = w i j ( l ) α w i j ( l ) J ( w , b ) b i ( l ) = b i ( l ) α b i ( l ) J ( w , b )
where α is the learning rate, i is the i-th sample, j is the j-th mapping feature, and J(·) is the cross entropy loss function which will be mentioned in Section 3.2.

3.2. Cross Entropy Loss Function

Cross entropy is a concept in information theory [32]. It was originally used to estimate the average coding length. Given two probability distributions p and q, the cross entropy of p expressed by q is
H ( p , q ) = x p ( x ) log q ( x )
where p(x) is often used to describe the true distribution, q(x) is used to describe the distribution of model prediction in machine learning. The smaller the cross entropy H(p,q), the closer the two probability distributions are.
Information Entropy is the expectation of all information quantities:
H ( X ) = x p ( x ) log p ( x )
Relative entropy is also called Kullback–Leibler divergence. If there are two separate probability distributions p(x) and q(x) for the same random variable x, relative entropy can be used to measure the difference between the two distributions.
D K L ( p | | q ) = x p ( x ) log p ( x ) q ( x ) = x p ( x ) log p ( x ) x p ( x ) log q ( x ) = H ( X ) + [ x p ( x ) log q ( x ) ] = H ( p , q ) H ( X )
where the smaller the relative entropy DKL(p||q) is, the closer the two probability distributions are.
Relative entropy should be used to calculate the difference of probability distribution in calculating loss. However, the Equation (9) shows that:
R e l a t i v e   e n t r o p y = C r o s s   e n t r o p y I n f o r m a t i o n   e n t r o p y
Since information entropy describes the amount of information needed to eliminate the uncertainty of p (the true distribution), its value should be minimum and fixed. Then, optimizing and reducing relative entropy is optimizing cross entropy, so it is easy and good to use cross entropy in machine learning.
Cross entropy loss function is as follows and the purpose of training is to minimize it:
J ( w , b ) = 1 m j = 1 m i = 1 n y j i log h w , b ( x j i )
where n is the number of categories, m is the number of samples, yji is the true value of sample j in class i, and hw,b(xji) is the predicted value of sample j in class i.

3.3. Attention Mechanism

Attention mechanism in deep learning simulates the attention model of human brain. At present, it is very popular and widely used in machine translation [26], speech recognition [27], image caption [28], and many other fields. Attention mechanism can select more key information from a large amount of information and suppress irrelevant information, so as to avoid the problem of over-fitting. Channel attention pays more attention to the global characteristics, while spatial attention has more prominent ability to control the local features, so the mixed attention called Convolutional Block Attention Module (CBAM), which combines channel and spatial dimension, can better depict the complete information [25].
When F is input, the attention module performs the following operations shown in Figure 5, Figure 6 and Figure 7:
M c ( F ) = σ ( W 1 ( W 0 ( F avg ) ) + W 1 ( W 0 ( F max ) ) )
F = M c ( F ) F
M s ( F ) = σ ( f 7 × 7 ( [ F avg ; F max ] ) )
F = M s ( F ) F
where Mc(·) denotes attention extraction on channel dimension as shown in Figure 6, and Ms(·) denotes attention extraction on spatial dimension as shown in Figure 7. represents the multiplication of the corresponding position elements of two matrices, W0 and W1 represent the connection weight in the multi-layer perceptron (MLP) model using ReLU activation function, σ represents sigmoid activation function, Favg represents the vector F after average pooling operation and Fmax represents the vector F after maximum pooling operation, F’avg represents the vector F’ after average pooling operation and F’max represents the vector F’ after maximum pooling operation.
In the channel attention module, maximum pooling and average pooling are used simultaneously, and the parameters are shared by the shared MLP. The result of Mc(·) operation can be obtained by summing up the results of fully connected layers after MLP. The input F is multiplied by the result of Mc(·) operation to get the result F’ of channel attention, which is 224 × 224 × 3.
In the spatial attention module, maximum pooling and average pooling are used for F’ simultaneously, and a vector with 2 channels is obtained by concatenating the two vectors of maximum pooling and average pooling. In order to get the importance of different pixels of feature maps on a plane, one convolution core is used to compress 2 channels into 1 channel. So, the convolution operation is carried out by using one 7 × 7 convolution kernel to ensure that the dimension of result of Ms(·) operation is consistent with one channel, which is 224 × 224 × 1.
In the end, the result of Ms(·) operation is multiplied by the result of channel attention F’ to get the final vector of F’’. F’’ is consistent with F in dimension, which is 224 × 224 × 3.
Adding the attention mechanism to the convolution layer, a VGG 16 model based on attention mechanism can be obtained, as shown in Figure 8. Because this paper divides the sources of sag into nine categories, the fully connected layers are fine-tuned to 1 × 1 × 128 and the Softmax layer changes into 1 × 1 × 9.

3.4. Voltage Sag Source Recognition Based on Improved VGG Transfer Learning

3.4.1. Improved VGG Transfer Learning

Knowledge transfer is a hypothesis that can break down the same distributed samples and greatly increase the cross-domain ability of machine learning. The purpose of transfer learning [29] is to properly introduce existing knowledge into new fields, so that machines can acquire the ability to ‘draw inferences from one instance to another’. Based on the idea of transfer learning, the standard VGG 16 model is used to transfer the target task to the identification of voltage sag source, which greatly improves the efficiency and accuracy of training, and reduces the dependence on sample size.
ImageNet is a data set of more than 15 million images, about 22,000 categories [33]. A standard VGG 16 network has been pre-trained on the entire ImageNet dataset, using pre-computed weights. The complete VGG transfer learning model training process sketch is shown in Figure 9. Among them, the improved VGG transfer learning model is a combined model, which consists of standard VGG 16 and attention-based VGG 16 model. Firstly, the standard VGG 16 has been trained separately on the ImageNet dataset and fixed convolution parameters. Then, using VGG 16 based on attention mechanism, the input data set is trained separately, and the attention module and fully connected parameters are obtained. Finally, the standard VGG 16 and the attention-based VGG 16 model are combined for combined training. In this case, the combined training is for fine-tuning the results of separate training.
In the training process of VGG 16 model based on attention mechanism, the voltage sag signal is transformed into reconstruction image according to the phase space reconstruction technology proposed in Section 2. Voltage sag sources are divided into nine categories, namely A1a, A1b, A1c, A2ab, A2ac, A2bc A3abc, B1, and C1, which are used as labels of reconstruction images. Adam algorithm [34] is used to optimize the model parameters and adjust the learning rate adaptively.

3.4.2. Voltage Sag Source Recognition Framework

The proposed framework for voltage sag source recognition based on improved VGG transfer learning is shown in Figure 10. The process is described as follows:
  • Step 1: The historical data of voltage sag are read from the database. With the technology of phase space reconstruction referred in Section 2, historical reconstruction images of labeled different voltage sag sources can be generated.
  • Step 2: As training and testing data sets in this paper, the reconstruction image data in step 1 are input into the improved VGG transfer learning model in Section 3.4 for training and testing. Then, a trained improved VGG transfer learning model can be obtained.
  • Step 3: For the voltage sag signals to be identified, the corresponding reconstruction images are generated which are input into the trained model in step 2. Finally, the results of voltage sag source recognition are achieved.
In addition, the results of step 3 are added to the historical database for updating.

4. Example Analysis

4.1. Data Acquisition

The simulation models shown in Figure 11 are built in MATLAB/SIMULINK, and their electrical parameters are changed to obtain various types of voltage sag signal samples at fault points or access points. For short circuit faults, the duration, short-circuit impedance and line loads are changed, and 1000 groups of samples are set for each type, totaling 7000 groups. For the starting of induction motor, the starting time, the internal parameters of the motor and the line load are changed, totaling 1000 groups. For large transformer energizing, the switching time, transformer capacity and connection mode are changed, totally 1000 groups. Because the actual data will be affected by noise, the original 9000 groups of data are superimposed with 20 dB and 10 dB white Gaussian noise, and finally 27000 groups of sample data are obtained. The formula of signal to noise ratio is
S N R = 10 lg P s P n
where Ps is the power of sag signal, Pn is the power of noise. The unit of SNR is dB. The larger the SNR is, the smaller the noise is.
The basic frequency of the simulation model is set to 50 Hz and the total simulation time is set to 1 s. The sampling frequency is set to 1 kHz, so the sampling point of voltage signal is 1000. Voltage sag signal samples are reconstructed by phase space reconstruction to obtain the sample of voltage sag reconstruction image, and the data of the sag reconstruction image is used as the input data set of the model in this paper.
Four fold cross validation method was used to validate the experiment, that is, 750 groups of samples were selected as training set in each type of sag source, 250 groups of samples were used as test set, and the average of four times of test accuracy was taken as the result.

4.2. Analysis of Noise Immunity for Voltage Sag Phase Space Reconstruction

Because the actual data will be affected by noise, the original 9000 sets of data are superimposed with 20 dB and 10 dB white Gaussian noise. Three samples of phase space reconstruction image of voltage sag in phase A with 20 dB and 10 dB white Gaussian noise are shown in Figure 12.
Due to the strong noise, the phase trajectory is not smooth. Except for the type of motor starting, all the phase space reconstruction images are hardly affected. Therefore, the image recognition of voltage sag sources based on phase space reconstruction has a good ability to overcome the noise.

4.3. Analysis of Attention Mechanism

Attention mechanism can enhance feature extraction, which leads to better classification effect. As can be seen in Figure 13, a sample attention mechanism process of A- and C-phase short circuit type of sag is clearly shown in the form of heat map.
The input feature F is gray scale reconstruction images of phase A, B, and C. The channel attention, spatial attention and refined feature are shown in visual heat maps, where different colors represent different weights. In the channel attention, phase A is 0.51, phase C is 0.49, and phase B is 0.42, that is, phase A and phase C are more important than phase B. In the spatial attention, the two limit cycles which are 0.37 mean that the sag magnitude is noticed. Combining the channel attention and spatial attention, the refined feature F’’ shows the different weights of three phase’s signal. Phase A is about 0.96 for dark red, phase C is about 0.85 for red, which means that they are the effective information for which type it belong to. Phase B is 0.65 for yellow, which plays the least role in classification. These are consistent with the characteristics of A- and C-phase short circuit fault. Through the attention mechanism, the feature extraction process is strengthened and the final classification results can be more accurate.
Attention mechanism can make the model pay more attention to the effective information, ignore the invalid information, and improve the interpretability of the model. The CBAM attention used in this paper is a combination of channel attention and spatial attention, in which channel attention can indicate which is more important to the results in phase A, B, C, and which pixels and regions in images of phase A, B, C are more important. Combining the two methods can make the model pay more attention to the areas that have great influence on the results, ignore the areas of invalid information, and improve the effectiveness and interpretability of the model.

4.4. Analysis of Classification Effect of Feature Vector

Before the output layer, VGG 16 model based on attention mechanism has a fully connected layer with 128 dimensions, which extract 128 final features of sag. These 128 values are characteristic quantities reflecting the characteristics of the reconstructed image itself, which have no practical significance. For example, as shown in Table 1, the feature data within the same class are similar, while the feature data between classes are clearly distinguished.
In order to observe the quality of automatic feature extraction in the proposed model, t-Distributed Stochastic Neighbor Embedding (t-SNE) [35] algorithm can be used to project the extracted feature vectors to three-dimensional space for observation. The three-dimensional projection of the extracted feature vectors can be drawn as shown in Figure 14. It can be seen that the projection boundaries of nine types of voltage sag feature vectors extracted by the proposed model are clear, and the distribution distance between each type is abundant.

4.5. Network Training Process and Contrastive Analysis

• Separate training of VGG 16 model
VGG 16 model is used for separate training, and the pre-trained parameters of ImageNet data set are not used. As shown in Figure 15a, with the increase of iteration times, the loss of the network does not decrease, and the accuracy of the model is less than 50%, the effect is very poor. If VGG 16 model without pre-trained parameters were used to identify voltage sag sources, it would be necessary to increase sample data and training times to achieve good results, which is inefficient and the results are not necessarily ideal. Therefore, it is necessary to introduce the idea of transfer learning to model combined training.
• Combined training of VGG 16 transfer learning model without attention mechanism
VGG 16 without attention mechanism and standard VGG 16 are used to construct the transfer learning model of VGG 16 without attention mechanism, and combined training is carried out. As shown in Figure 15b, with the increase of the number of iterations, the loss of the network decreases gradually, and the accuracy of the generated model increases gradually which tends to be 96%. However, the stability of network loss and model accuracy is poor. Especially in the late iteration period, the over-fitting phenomenon occurs which leads to the large oscillation.
• Combined training of improved VGG transfer learning model
Attention mechanism is added to construct the improved VGG transfer learning model shown in Section 3.4, and combined training is carried out. In Figure 15c, with the increase of iterations, the loss of the network decreases gradually, and the accuracy of the generated model increases gradually, which tends to be 100%. Moreover, in the late iteration period, the attention mechanism restrains the over-fitting phenomenon, so that the stability of network loss and model accuracy is better.
• Training using voltage sag signal image
Without phase space reconstruction, the improved VGG transfer learning model is trained by using voltage sag signal images as the input directly. As shown in Figure 15d, with the increase of iteration times, the loss of the network decreases and the accuracy of the generated model increases gradually, but the convergence speed is slower than that in Figure 15c, and the accuracy is only about 90% which is lower than the result of reconstruction sample by phase space reconstruction technology. This is because the instantaneous waveforms of voltage sag voltage signals are dense, and the difference between the signal images of different sag sources is not obvious in the long-term signal images.

4.6. Result Analysis

In order to evaluate the recognition results of each type of sag source, Accuracy and F1 [36] are selected as the evaluation indexes. Among them, Accuracy is a single index and F1 is a comprehensive index.
A c c u r a c y = C T
P r e c i s i o n = C P
F 1 = 2 × P r e c i s i o n × A c c u r a c y P r e c i s i o n + A c c u r a c y
where C is the number of samples correctly recognized for a certain type of voltage sag sources, T is the number of samples that are true for this type, P is the number of all samples recognized as this type, Accuracy is the ratio of the number of samples correctly recognized for this type to the number of samples that are true for this type, Precision refers to the ratio of the number of samples correctly recognized for this type to the number of all samples recognized as this type.
The identification results of voltage sag sources are shown in Table 2. From Table 2, it can be seen that the average recognition Accuracy and F1 of the reconstruction image of the noise-free voltage sag signal are 100%, which indicates the feasibility of image recognition of voltage sag sources. With the increase of noise, the average recognition Accuracy decreases to 98.3% and the F1 decreases to 95.4%, which still remain at a high level and reflect the anti-noise ability of this method.
From the point of view of practical application, the average recognition Accuracy and F1 of this method are high for the single phase short circuit fault sag source type which accounts for a large proportion in power system. Among them, the average recognition Accuracy and F1 of A1a type is 99.4% and 98.2% respectively. The average recognition Accuracy and F1 of A1b type is 99.2% and 98.2% respectively. The average recognition Accuracy and F1 of A1c type is 99.5% and 97.9% respectively.
The methods of reference [4,5,15], and [17] (hereinafter referred to as method 1, method 2, method 3, and method 4) are compared with the experimental results of voltage sag source identification method based on phase space reconstruction and improved VGG migration learning proposed in this paper, as shown in Table 3 and Table 4.
Table 3 shows that the recognition accuracy of the proposed method for different voltage sag sources is relatively high. This is because the attention mechanism is added in the process of automatic feature extraction, which improves the expression ability and generalization ability of the model. From Table 4, it can be seen that the convergence rate of this method is faster, because the transfer learning method reduces the training cost and improves the training efficiency.

5. Conclusions

In this paper, a method of voltage sag source identification based on phase space reconstruction and improved VGG transfer learning is proposed. The effectiveness and superiority of the proposed method are verified by an example, which is mainly embodied in the following aspects:
  • Voltage sag signal image is reconstructed into phase space image, which not only retains the complete characteristics of sag, but also has more intuitive and concise image features.
  • Attention mechanism is added to VGG model to further automatically extract image features to prevent over-fitting. It has an excellent classification effect and improves the accuracy of model recognition.
  • The idea of transfer learning is introduced to train the network on the basis of other image classification results, which improves the efficiency of network training.
The basis of image recognition is image accuracy. In this paper, only voltage sag signals of one single sampling frequency are identified as voltage sag sources. How to ensure the accuracy of voltage sag reconstruction under different sampling frequencies is the next research direction.

Author Contributions

Conceptualization, Y.P. and X.M.; methodology, Y.P. and X.S.; software, Y.P. and X.S.; validation, Y.P. and X.S.; formal analysis, Y.P. and H.Y.; investigation, Y.P. and H.Y.; resources, Y.P.; data curation, Y.P. and X.M.; writing—original draft preparation, Y.P.; writing—review and editing, Y.P. and X.M.; visualization, Y.P. and X.S.; supervision, X.M.; project administration, H.Y. and X.M.; funding acquisition, H.Y.

Funding

This research was funded by National Natural Science Foundation of China, grant number 51477105.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Saini, M.K.; Kapoor, R.; Beniwal, R.K.; Aggarwal, A. Recognition of voltage sag causes using fractionally delayed biorthogonal wavelet. Trans. Inst. Meas. Control 2019, 41, 2851–2863. [Google Scholar] [CrossRef]
  2. Ding, N.; Cai, W.; Suo, J.; Wang, J.W. Research on voltage sag sources recognition method. Power Syst. Technol. 2008, 32, 55–59. [Google Scholar]
  3. De Santis, M.; Noce, C.; Varilone, P.; Verde, P. Analysis of the origin of measured voltage sags in interconnected networks. Electric Pow. Syst. Res. 2018, 154, 391–400. [Google Scholar] [CrossRef]
  4. Zheng, Z.C.; Wang, H.; Qi, L.H. Recognition method of voltage sag sources based on deep learning models’ fusion. Proc. CSEE 2019, 39, 97–104+324. [Google Scholar]
  5. Mei, F.; Ren, Y.; Wu, Q.L.; Zhang, C.Y.; Pan, Y.; Sha, H.Y.; Zheng, J.Y. Online recognition method for voltage sags based on a deep belief network. Energies 2019, 12, 1. [Google Scholar] [CrossRef]
  6. Li, D.Q.; Mei, F.; Zhang, C.Y.; Sha, H.Y.; Zheng, J.Y. Self-supervised voltage sag source identification method based on CNN. Energies 2019, 12, 1059. [Google Scholar] [CrossRef]
  7. Bagheri, A.; Gu, I.Y.H.; Bollen, M.H.J.; Balouji, E. A robust transform-domain deep convolutional network for voltage dip classification. IEEE Trans. Power Del. 2018, 33, 2794–2802. [Google Scholar] [CrossRef]
  8. Wang, K.X.; Song, Z.X.; Chen, D.G.; Wang, J.H.; Geng, Y.S. Interference source identification of voltage sag in distribution system based on wavelet transform. Proc. CSEE 2003, 23, 29–34+54. [Google Scholar]
  9. Saini, M.; Aggarwal, A. Fractionally delayed Legendre wavelet transform based detection and optimal features based classification of voltage sag causes. J. Renew. Sustain. Energy 2019, 11, 1. [Google Scholar] [CrossRef]
  10. Zhao, F.Z.; Yang, R.G. Voltage sag disturbance detection based on short time fourier transform. Proc. CSEE 2007, 27, 28–34. [Google Scholar]
  11. Xu, Y.H.; Zhao, Y. Identification of power quality disturbance based on short-term fourier transform and disturbance time orientation by singular value decomposition. Power Syst. Technol. 2011, 35, 174–180. [Google Scholar]
  12. Li, P.; Zhang, Q.S.; Zhang, G.L.; Liu, W.; Chen, F.R. Adaptive S transform for feature extraction in voltage sags. Appl. Soft Comput. 2019, 80, 438–449. [Google Scholar] [CrossRef]
  13. Foroughi, A.; Mohammadi, E.; Esmaeili, S. Application of Hilbert-Huang transform and support vector machine for detection and classification of voltage sag sources. Turk. J. Electr. Eng. Comput. Sci. 2014, 22, 1116–1129. [Google Scholar] [CrossRef]
  14. Manjula, M.; Mishra, S.; Sarma, A.V.R.S. Empirical mode decomposition with Hilbert transform for classification of voltage sag causes using probabilistic neural network. Int. J. Electr. Power Ener. Syst. 2013, 44, 597–603. [Google Scholar] [CrossRef]
  15. Jia, Y.; He, Z.Y.; Zhao, J. A method to identify voltage sag sources in distribution network based on wavelet entropy and probability neural network. Power Syst. Technol. 2009, 33, 63–69. [Google Scholar]
  16. Jadhav, V.R.; Patil, A.S. Classification of Voltage sags at Induction Motor by Artificial Neural Network. In Proceedings of the 2015 International Conference on Energy Systems and Applications, Pune, India, 30 October–1 November 2015. [Google Scholar]
  17. Sha, H.Y.; Mei, F.; Zhang, C.Y.; Pan, Y.; Zheng, J.Y. Identification method for voltage sags based on k-means-singular value decomposition and least squares support vector machine. Energies 2019, 12, 1137. [Google Scholar] [CrossRef]
  18. Zhao, Y.; Zhao, C.; Ye, H.; Yu, F. Method to reduce identification feature of different voltage sag disturbance source based on principal component analysis. Power Syst. Protect. Control 2015, 43, 105–110. [Google Scholar]
  19. Li, C.Y.; Yang, J.L.; Xu, Y.H.; Wu, Y.P.; Wei, P.F. Application of Comprehensive Fuzzy Evaluation Method on Recognition of Voltage Sag Disturbance Sources. Power Syst. Technol. 2017, 41, 1022–1028. [Google Scholar]
  20. Rabinovitch, A.; Thieberger, R. Time series analysis of chaotic signals. Physica D 1987, 28, 409–415. [Google Scholar] [CrossRef]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. Gabriel, G.; Adrian, C.; Valery, N. First-stage prostate cancer identification on histopathological images: Hand-driven versus automatic learning. Entropy 2019, 21, 356. [Google Scholar]
  23. Ha, I.; Kim, H.; Park, S.; Kim, H. Image retrieval using BIM and features from pretrained VGG network for indoor localization. Build. Environ. 2018, 140, 23–31. [Google Scholar] [CrossRef]
  24. Matlani, P.; Shrivastava, M. Hybrid Deep VGG-NET Convolutional Classifier for Video Smoke Detection. Comput. Model. Eng. Sci. 2019, 119, 427–458. [Google Scholar] [CrossRef] [Green Version]
  25. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  26. Choi, H.; Cho, K.; Bengio, Y. Fine-grained attention mechanism for neural machine translation. Neurocomputing 2018, 284, 171–176. [Google Scholar] [CrossRef] [Green Version]
  27. Kim, H.G.; Lee, H.; Kim, G.; Oh, S.H.; Lee, S.Y. Rescoring of N-best hypotheses using top-down selective attention for automatic speech recognition. IEEE Signal Proc. Lett. 2018, 25, 199–203. [Google Scholar] [CrossRef]
  28. Kim, B.; Shin, S.; Jung, H. Variational autoencoder-based multiple image captioning using a caption attention map. Appl. Sci. 2019, 9, 13. [Google Scholar] [CrossRef]
  29. Khanna, R.; Oh, D.; Kim, Y. Through-wall remote human voice recognition using doppler radar with transfer learning. IEEE Sens. J. 2019, 19, 4571–4576. [Google Scholar] [CrossRef]
  30. Takens, F. Detecting Strange Atractors in Turbulence; Springer: Berlin, Germany, 1981; pp. 366–381. [Google Scholar]
  31. Kim, H.S.; Eykholt, R.; Salas, J.D. Nonlinear dynamics, delay times and embedding windows. Physica D 1999, 127, 48–60. [Google Scholar] [CrossRef]
  32. Meijia, J.; Ochoa, A.; Mederos, B. Reconstruction of PET images using cross-entropy and field of experts. Entropy 2019, 21, 83. [Google Scholar] [CrossRef]
  33. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.H.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  34. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Laurens, V.D.M. Accelerating t-SNE using Tree-Based Algorithms. J. Mach. Learn. Res. 2014, 15, 3221–3245. [Google Scholar]
  36. Alohaly, M.; Takabi, H.; Blanco, E. Automated extraction of attributes from natural language attribute-based access control (ABAC) policies. Cybersecurity 2019, 2, 2. [Google Scholar] [CrossRef]
Figure 1. Voltage sag signal waveform (left) and corresponding phase space reconstruction image (right) of A-phase short circuit fault in phase (a) A, (b) B, and (c) C.
Figure 1. Voltage sag signal waveform (left) and corresponding phase space reconstruction image (right) of A-phase short circuit fault in phase (a) A, (b) B, and (c) C.
Entropy 21 00999 g001
Figure 2. Voltage sag signal waveform (left) and corresponding phase space reconstruction image (right) of large induction motor starting in phase (a) A, (b) B, and (c) C.
Figure 2. Voltage sag signal waveform (left) and corresponding phase space reconstruction image (right) of large induction motor starting in phase (a) A, (b) B, and (c) C.
Entropy 21 00999 g002
Figure 3. Voltage sag signal waveform (left) and corresponding phase space reconstruction image (right) of unloaded transformer energizing in phase (a) A, (b) B, and (c) C.
Figure 3. Voltage sag signal waveform (left) and corresponding phase space reconstruction image (right) of unloaded transformer energizing in phase (a) A, (b) B, and (c) C.
Entropy 21 00999 g003aEntropy 21 00999 g003b
Figure 4. Standard VGG 16 network structure. 224 × 224 × 3 and so on represent the dimensions of matrix. Convolution + ReLU means ReLU function activation after convolution operation. Fully connected + ReLU means ReLU function activation after fully connected operation.
Figure 4. Standard VGG 16 network structure. 224 × 224 × 3 and so on represent the dimensions of matrix. Convolution + ReLU means ReLU function activation after convolution operation. Fully connected + ReLU means ReLU function activation after fully connected operation.
Entropy 21 00999 g004
Figure 5. The whole attention module. 224 × 224 × 3 and so on represent the dimensions of matrix.
Figure 5. The whole attention module. 224 × 224 × 3 and so on represent the dimensions of matrix.
Entropy 21 00999 g005
Figure 6. Channel attention module. 224 × 224 × 3 and so on represent the dimensions of matrix.
Figure 6. Channel attention module. 224 × 224 × 3 and so on represent the dimensions of matrix.
Entropy 21 00999 g006
Figure 7. Spatial attention module. 224 × 224 × 3 and so on represent the dimensions of matrix.
Figure 7. Spatial attention module. 224 × 224 × 3 and so on represent the dimensions of matrix.
Entropy 21 00999 g007
Figure 8. VGG16 network structure based on attention mechanism. 224 × 224 × 3 and so on represent the dimensions of matrix. Convolution + ReLU means ReLU function activation after convolution operation. Fully connected + ReLU means ReLU function activation after fully connected operation.
Figure 8. VGG16 network structure based on attention mechanism. 224 × 224 × 3 and so on represent the dimensions of matrix. Convolution + ReLU means ReLU function activation after convolution operation. Fully connected + ReLU means ReLU function activation after fully connected operation.
Entropy 21 00999 g008
Figure 9. Training process of improved VGG transfer learning model.
Figure 9. Training process of improved VGG transfer learning model.
Entropy 21 00999 g009
Figure 10. Voltage sag source recognition framework based on improved VGG transfer learning. The sample means an example of Figure 1.
Figure 10. Voltage sag source recognition framework based on improved VGG transfer learning. The sample means an example of Figure 1.
Entropy 21 00999 g010
Figure 11. The simulation models of (a) short circuit fault, (b) large induction motor starting, and (c) unloaded transformer energizing in MATLAB/SIMULINK.
Figure 11. The simulation models of (a) short circuit fault, (b) large induction motor starting, and (c) unloaded transformer energizing in MATLAB/SIMULINK.
Entropy 21 00999 g011aEntropy 21 00999 g011b
Figure 12. Samples of phase space reconstruction of voltage sag with no noise (left), 20 dB (middle) and 10 dB (right) white Gaussian noise in phase A of (a) single phase short circuit fault, (b) large induction motor starting and (c) unloaded transformer energizing.
Figure 12. Samples of phase space reconstruction of voltage sag with no noise (left), 20 dB (middle) and 10 dB (right) white Gaussian noise in phase A of (a) single phase short circuit fault, (b) large induction motor starting and (c) unloaded transformer energizing.
Entropy 21 00999 g012aEntropy 21 00999 g012b
Figure 13. The visualization process of attention mechanism. The channel attention Mc, spatial attention Ms and refined feature F’’ are shown in heat maps.
Figure 13. The visualization process of attention mechanism. The channel attention Mc, spatial attention Ms and refined feature F’’ are shown in heat maps.
Entropy 21 00999 g013
Figure 14. 3D projection for feature extraction based on attention-based VGG 16 model.
Figure 14. 3D projection for feature extraction based on attention-based VGG 16 model.
Entropy 21 00999 g014
Figure 15. (a) VGG 16 model separate training process, (b) VGG 16 transfer learning model without attention mechanism combined training process, (c) improved VGG transfer learning model combined training process and (d) combined training process directly adopting voltage sag signal image.
Figure 15. (a) VGG 16 model separate training process, (b) VGG 16 transfer learning model without attention mechanism combined training process, (c) improved VGG transfer learning model combined training process and (d) combined training process directly adopting voltage sag signal image.
Entropy 21 00999 g015
Table 1. 128-dimensional feature vectors of the features for different types of voltage sags.
Table 1. 128-dimensional feature vectors of the features for different types of voltage sags.
Dimension12345678
A 1a4.5741080.630084005.07992804.9305180
4.5552460.635833005.04529304.8930740
4.4596940.652800004.85616204.7995240
A1b3.6130045.856254000.665880000
3.5235905.643432000.597539000
3.5503565.732181000.459643000
A1c2.3092630.511907005.8530893.5689752.2784460
2.2187580.484885005.7088403.4868152.1932940
2.6734870.980059006.9631944.1161532.5320540
A2ab4.8652301.00690900002.1888550
4.9292751.03781400002.1627570
4.9548881.04124100002.1058830
A2ac3.7293041.305431003.4397870.9345091.0634460
3.7633251.190667003.4630851.1093071.0938390
3.9190071.267426003.7061301.1296991.2406830
A2bc1.4020770.637477000000
1.2281100.477076000000
1.3151130.497585000000
A3abc3.2557970000.7328822.76322600
3.0857770000.6835692.61649800
3.2676280000.7975952.76169400
B103.58484800.1301281.752869000
03.79080500.2302351.670009000
03.98016300.1231131.961738000
C102.755503000000
02.504199000000
02.720874000000
Table 2. The identification results of voltage sag sources.
Table 2. The identification results of voltage sag sources.
Sag Source TypeAccuracy/%F1/%
Noise-Free20 dB10 dBNoise-Free20 dB10 dB
A1a10010098.410098.795.8
A1b10099.698.010098.895.8
A1c10010098.410098.695.1
A2ab10099.899.610096.893.9
A2ac10010098.410095.893.7
A2bc10099.298.410096.995.9
A3abc10010097.210099.393.8
B110010098.810099.696.8
C110010099.210099.996.8
Table 3. Comparison of the recognition accuracy of voltage sag sources.
Table 3. Comparison of the recognition accuracy of voltage sag sources.
Sag Source TypeAccuracy/%
Method 1Method 2Method 3Method 4Proposed Method
A1a99.497.896.593.499.5
A1b99.698.296.393.699.2
A1c98.597.195.294.199.5
A2ab98.997.896.494.299.8
A2ac98.197.696.993.699.6
A2bc99.397.997.193.899.2
A3abc99.298.296.394.599.1
B199.298.397.294.299.6
C199.698.696.893.699.7
Table 4. Comparison of convergence rate of voltage sag source recognition model.
Table 4. Comparison of convergence rate of voltage sag source recognition model.
Method 1Method 2Method 3Method 4Proposed Method
Train times416465614750305

Share and Cite

MDPI and ACS Style

Pu, Y.; Yang, H.; Ma, X.; Sun, X. Recognition of Voltage Sag Sources Based on Phase Space Reconstruction and Improved VGG Transfer Learning. Entropy 2019, 21, 999. https://doi.org/10.3390/e21100999

AMA Style

Pu Y, Yang H, Ma X, Sun X. Recognition of Voltage Sag Sources Based on Phase Space Reconstruction and Improved VGG Transfer Learning. Entropy. 2019; 21(10):999. https://doi.org/10.3390/e21100999

Chicago/Turabian Style

Pu, Yuting, Honggeng Yang, Xiaoyang Ma, and Xiangxun Sun. 2019. "Recognition of Voltage Sag Sources Based on Phase Space Reconstruction and Improved VGG Transfer Learning" Entropy 21, no. 10: 999. https://doi.org/10.3390/e21100999

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop