Next Article in Journal
Collaborative Indoor Access Point Localization Using Autonomous Mobile Robot Swarm
Next Article in Special Issue
Partial Discharge Monitoring on Metal-Enclosed Switchgear with Distributed Non-Contact Sensors
Previous Article in Journal
An Advanced Bio-Inspired PhotoPlethysmoGraphy (PPG) and ECG Pattern Recognition System for Medical Assessment
Previous Article in Special Issue
Survey on the Performance of Source Localization Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Partial Discharge Signals by Combining Adaptive Local Iterative Filtering and Entropy Features

1
Department of Engineering, Glasgow Caledonian University, 70 Cowcaddens Road, Glasgow G4 0BA, UK
2
Institute of Energy and Environment, University of Strathclyde, 204 George Street, Glasgow G1 1XW, UK
3
Innovation Centre for Online Systems, 7 Townsend Business Park, Bere Regis BH20 7LA, UK
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(2), 406; https://doi.org/10.3390/s18020406
Submission received: 9 January 2018 / Revised: 22 January 2018 / Accepted: 26 January 2018 / Published: 31 January 2018
(This article belongs to the Special Issue UHF and RF Sensor Technology for Partial Discharge Detection)

Abstract

:
Electromagnetic Interference (EMI) is a technique for capturing Partial Discharge (PD) signals in High-Voltage (HV) power plant apparatus. EMI signals can be non-stationary which makes their analysis difficult, particularly for pattern recognition applications. This paper elaborates upon a previously developed software condition-monitoring model for improved EMI events classification based on time-frequency signal decomposition and entropy features. The idea of the proposed method is to map multiple discharge source signals captured by EMI and labelled by experts, including PD, from the time domain to a feature space, which aids in the interpretation of subsequent fault information. Here, instead of using only one permutation entropy measure, a more robust measure, called Dispersion Entropy (DE), is added to the feature vector. Multi-Class Support Vector Machine (MCSVM) methods are utilized for classification of the different discharge sources. Results show an improved classification accuracy compared to previously proposed methods. This yields to a successful development of an expert’s knowledge-based intelligent system. Since this method is demonstrated to be successful with real field data, it brings the benefit of possible real-world application for EMI condition monitoring.

1. Introduction

Insulation condition monitoring is essential in High-Voltage (HV) equipment including insulated cables, transformers, large generators and motors which are operating in power plants. This allows early identification of several events related to insulation failure, such as Partial Discharge (PD) and Corona, which could lead to equipment breakdown followed by high maintenance or replacement costs. The process to identifying these insulation breakdown events can be performed through Electro-Magnetic Interference (EMI) diagnostics, which involves data sensing via the EMI technique and data analysis by experts. EMI diagnostics is a recognised technique used to identify the presence of insulation degradation and conductor-related faults in industrial HV machinery [1]. The downsides of an expert’s analysis are the high costs, human time and impracticability for continuous monitoring. In [2] the authors introduced, for the first time, an automatic and continuous condition monitoring solution, based on a pattern recognition approach. The developed model is seen as a transfer of expert knowledge to a software model. The classification results may be considered acceptable but with room for improved methodologies.
Currently, the most popular method for PD classification involves the analysis of Phase Resolved PD (PRPD) patterns [3,4]. To the author’s knowledge, the gaps found in the literature are as follows: (a) most of the proposed PD pattern recognition techniques are tested on simulated or experimental PD data (see [5,6,7,8,9,10,11]); (b) Techniques have addressed only the classification of different PD types; (c) classification using EMI data has not been fully analysed and is currently still under development.
In this paper, we improve the performance of previous EMI discharge sources classification [2,12,13,14,15], which aims to distinguish between a number of events including PD, Corona (C), Noise (N), Process Noise (PN) etc. Our previous method involves a state-of-the-art signal decomposition technique combined with an entropy measurement technique. Adaptive Local Iterative Filtering (ALIF) was used to decompose the EMI signal and obtain a number of time series signals at various frequency components, called Intrinsic Mode Functions (IMFs). This method has the benefit to effectively capture the amplitude and frequency variations in non-stationary signals compared to existing signal decomposition methods such as Empirical Mode Decomposition (EMD) [16]. PD signals are reported to be non-stationary [17], thus the obtained IMFs by ALIF may help to efficiently retrieve the unique time-frequency characteristics of each EMI event. Permutation Entropy (PE) is popular in biomedical application for electrocardiogram-type signals analysis and classification [18]. This type of signals exhibit the same behaviour as some of the captured EMI discharge signals, also referred to as events, and this is the main motivation for using the PE-based measure. The solution in this paper introduces a new feature extraction technique called Dispersion Entropy (DE), which is a modified and improved version of PE [19]. Both PE and DE are implemented to extract the features of each IMF signal, which are subsequently fed into a Multi-Class Support Vector Machine (MCSVM) classifier to distinguish between the different EMI events. The concept of these algorithms will be described later in this paper.
The next section briefly describes the overall process of the proposed solution. Section 3 describes in more detail the EMI monitoring approach used for signal measurement. Section 4 briefly defines the theory of the feature extraction and classification algorithms. Section 5 presents the application of the developed model to the measured data. Section 6 presents the classification results and finally conclusions are provided in the last section.

2. Proposed Solution

Figure 1 outlines the method followed in the proposed classification solution. The data is first captured on power sites from various electrical apparatus. This figure illustrates an example of data sensed at a transformer. The data is captured using the EMI technique where a High Frequency Current Transformer (HFCT) is connected around the neutral earth cable. The data is logged into a device with quasi peak detection facility that provides a frequency spectrum from which the time resolved signals can be retrieved. Each signal is investigated by experts and subsequently labelled according to the event it contains (e.g., PD, C, N etc.) based on past experience of expert forensic confirmation. ALIF is used to decompose each signal into its multiple frequency components resulting in IMF time series from which PE and DE features are extracted. The feature vector is finally fed into an MCSVM classifier along with the relevant labels to train a model that can potentially learn and identify the characteristics of various events. Further details on EMI experimental setup and the measured data will be provided in Section 3 and Section 5 respectively.

3. EMI Measurement Technique

PD appears as a dielectric failure in electrical insulation apparatus due to HV stress; this results in an energy release that can produce EMI. EMI signals are well known to be generated by various types of faults in generators, motors, cables and associated auxiliary equipment [20]. The produced eruption of low energy electromagnetic pulses spreads in all directions from the discharge event. The emitted EMI signal can be sensed and analyzed across a wide frequency range. For instance, they can be sensed in the range as small as 10s of kHz while loose connections in an Isolated Phase Bus (IPB) can produce signals up to 100 MHz and above. EMI can both be radiated and conducted from the discharge site. The conducted energy is detected and measured with a split-core HFCT normally across a frequency range of 10 kHz to 100 MHz. The EMI measurement of these signals follows the CISPR-16-1-1 standard [21] to ensure compatibility of measurements and results interpretation across EMI measurement instruments. The split-core HFCT can be suitably located around the neutral connection conduit of a generator stator winding, at the grounding transformer, or, a motor cable supply conduit. The measurement of the EMI signals emitted is sensitive to various electrical and mechanical defects, including loose or broken stator and rotor bars, slot discharges, winding insulation defects, contamination on the windings, shaft eccentricity, bearing wear, etc. The severity, location and deterioration level of defects can often be measured by the EMI technique long before detection by means of more conventional methods. This is of value to applications that desire a system diagnostics approach. EMI analysis can distinguish between different defects and discharge sources, also referred to as events, and monitors not only activity within, e.g., a generator stator winding but, also, within adjacent auxiliary equipment as well. A number of issues have been identified by EMI techniques with exciters, cables, IPB and transformers, as well as machine stator windings. The resulting EMI spectrum is also unique for each physical location and type of defect within the electrical system [22].

4. Classification Theory

This section introduces the mathematical theory of ALIF, PE, DE feature extraction and MCSVM classification algorithms used in the proposed pattern recognition technique.

4.1. Adaptive Local Iterative Filtering

ALIF is a recursive approach that breaks down a multicomponent signal into a predefined number of IMFs, arranged from high to low frequency “space”, through subtraction of local averages from the signal [16]. The moving averages are calculated through convolution using low-pass filtering. Broadly, ALIF is a joining of EMD [23] with Iterative Filtering (IF) except that the filters are generated adaptively depending on the nature of the signal. For instance, if the signal’s characteristics change rapidly, short filters will be applied. On the contrary, if its characteristics change slowly, longer filters will be required. The filters are designed using Fokker–Planck equations in a Partial Differential Equation (PDE)-based model which is briefly explained in the following section. A complete and detailed explanation can be found in [16]. The mathematical theory of ALIF is described as follows. Let f ( n ) = I M F 1 ( n ) + I M F 2 ( n ) + I M F K ( n ) + R ( n ) be the multicomponent discrete signal which is decomposed into I M F i ( n ) ; i = 1 , 2 , , K ; K N IMFs plus a residual trend R ( n ) . The coefficients of a low pass filter at step j and point n with length 2 l j ( n ) + 1 are denoted as: h j ( n , τ ) R , τ [ l j ( n ) , l j ( n ) ] where l j ( n ) is the variable that defines the filter’s length boundaries. The moving average of the signal calculated in each iteration is defined as:
L j ( f ) ( n ) : = τ = l j ( n ) l j ( n ) f ( n + τ ) · h j ( n , τ ) .
The filtered output of ALIF is then obtained by Equation (2) from which the single IMF is extracted in what is called the inner-iteration.
f j + 1 ( n ) = f j ( n ) L j ( f ) ( n ) .
The algorithm uses a sequence of inner and outer iterations to obtain the IMFs. The inner iteration converges to save f j ( n ) as IMF. Starting with f 1 = f , the operator S 1 , j ( f j ) ( n ) = f j ( n ) L 1 , j ( f j ) ( n ) = f j + 1 ( n ) captures the fluctuations in f j ( n ) . Then, the first IMF is obtained as I M F 1 ( n ) = lim n S 1 , j ( f j ) ( n ) . In real implementations, j is limited according to a Stopping Distance (SD). For each step j, a threshold operator SD is estimated as follows:
D = | | I M F 1 , j I M F 1 , j 1 | | 2 | | I M F 1 , j 1 | | 2 .
The inner iteration is stopped when the D value reaches a particular threshold as recommended in [23,24]. The outer iteration converges when the number of local extreme points in f j ( n ) is at most one, in other words, when the signal is a trend. This results in the residual, calculated as R ( n ) = f I M F 1 I M F K 1 . In practice, another criterion is added to stop the outer iteration which is when the number of desired IMFs is reached.

4.2. Permutation Entropy

PE was developed by Bandt and Pompe [25] to assess the complexity in time series data with more resilience to low frequency artefacts. This makes it suitable for measuring real-world, noisy and chaotic time series signals. PE is derived from Shannon’s entropy theorem which is a measure of information contained in a data set [26]. Bandt and Pompe combined the entropy concept with symbolic dynamics to create a simple, fast to compute, robust and stable measure of regularity in short time series and to overcome classical entropy method limitations, including the requirement of long data sets and high computational cost [27]. Figure 2 illustrates m ! = 6 possible ordinal patterns, for an embedded dimension of m = 3 , that can be mapped to a time sequence data set. Figure 3 demonstrates pattern identification in an example time series, where instances of pattern 1 and 5 are plotted. Note that the time series from 3 to 5 s would be identified as pattern 5 and the time series from 7 to 9 s would be identified as pattern 4. The frequencies of pattern occurrences across the complete time series signal are calculated and presented in a histogram which is normalised in order to obtain the permutation probabilities that are used to compute the PE value.
The PE mathematical theory is described in [28] as follows. First, we denote vectors and scalars by upper and lower case respectively. Based on a given time series { x 1 , x 2 , , x N } , vectors X ( j ) ; j = 1 , 2 , , N ( m 1 ) are constructed as:
X ( j ) = { x ( j ) , x ( j + 1 ) , , x ( j + ( m 1 ) ) } .
The sequence in Equation (4) is arranged to provide components in increasing order as follows:
{ x ( j + ( i 1 1 ) ) x ( j + ( i 2 1 ) ) x ( j + ( i m 1 ) ) } .
If two successive components are equal, i.e., x ( j + ( i 1 ) ) = x ( j + ( i 2 ) ) for i 1 i 2 , then their positions can be rearranged to x ( j + ( i 1 ) ) x ( j + ( i 2 ) ) . Next, a different symbol series S ( l ) is calculated for each time series x ( i ) as:
S ( l ) = ( i 1 , i 2 , , i m )
where l = 1 , 2 , , k ( k = m ! ) . That is, there will be m ! different symbol series or permutations π n . The probability p ( π n m ) of each symbol sequence S ( l ) is then calculated mathematically as:
p ( π n m ) = j N 1 u : t y p e ( u ) = π n ( X n m ) j N 1 u : t y p e ( u ) Π ( X n m ) .
The mapping of the symbol S ( l ) to the ordinal pattern is denoted as t y p e ( . ) and the m ! symbols { π n m } n m ! are denoted as Π . The indicator function 1 A ( u ) of a set A is defined as:
1 A ( u ) = 1 i f u A 0 i f u A
Finally, the value of PE can be estimated using the formula:
P E ( m ) = n = 1 m ! p ( π n m ) · l o g ( p ( π n m ) ) .
PE should be suitable to quantify the characteristics contained in EMI signals since they are by nature non-stationary and complex to analyse.

4.3. Dispersion Entropy

DE was introduced in [19] to overcome (PE) and Sample Entropy (SE) limitations. SE is slow in computation particularly for long time series, whereas PE disregards information of the amplitude values mean and amplitude variations [29]. DE is calculated as follows. Given a time series signal X = { x 1 , x 2 , , x N } , with length of N, let X be mapped to Y = { y 1 , y 2 , , y N } using the Normal Cumulative Distribution Function (NCDF). Next, each y j ; j = 1 , , N is assigned a class from 1 to c linearly as follows.
z j c = r o u n d ( c · y j + 0.5 ) .
This provides N members of the classified time series. Here, other linear or non-linear methods can also be employed. Next, embedding vectors Z i m , c with dimension m and time delay d are created:
Z i m , c = { z i c , z i + d c , , z i + ( m 1 ) d c } ; i = 1 , 2 , , N ( m 1 ) d .
The latter is mapped to a dispersion pattern π v o v 1 v m 1 , among c m possible dispersion patterns, in that v 0 = z i c , v 1 = z i + d c , , v m 1 = z i + ( m 1 ) d c .
The dispersion probability of occurrence for each pattern is then calculated as follows.
p ( π v o v 1 v m 1 ) = i N ( m 1 ) d 1 u : t y p e ( u ) = π v o v 1 v m 1 ( Z i m , c ) N ( m 1 ) d .
Finally, the DE value is obtained based on the Shannon entropy formula as follows.
D E ( m , c , d ) = 1 c m p ( π v o v 1 v m 1 ) · l o g ( p ( π v o v 1 v m 1 ) ) .
The calculated DE value measures the spreading in a time series. This information may be useful to obtain the IMF characteristics.

4.4. Support Vector Machine

SVM is a regression and binary classification technique which was introduced in [30]. The algorithm aims to locate a hyperplane that separates two groups of data features. This method is used in various fields and in a wide range of different applications including text classification, sound recognition, image categorisation, and data classification [31]. SVM is popular for its high-dimensional features implementation while providing high detection accuracy compared to other classification algorithms such as neural networks and random forests [32]. This benefits fault classification by reducing confusion. The separation process of a data set belonging to two different classes is achieved in two main steps:
  • SVM generates an optimal line that separates the two different data features, such that feature clusters of one class are grouped on one side of the feature space and the remaining ones are grouped on the other side. This yields to the SVM model which is used in future data classification.
  • SVM classifies the training data set based on the trained model in the previous step; this is known as the testing phase.
The distance between the hyperplane and the nearest point of each class is called the margin. The wider the margin, the better the separation between classes. The separation is achieved by a kernel function such as linear, quadratic, radial basis or polynomial function. For instance, a polynomial function seeks to draw a non-linear line to first separate the data, then map the data to a linear space as illustrated in Figure 4.
The choice of kernel function implementation depends on the nature of the data. The classic approach to find the suitable kernel function and its parameters for optimum performance is a Grid Search method. This involves multiple training/testing of the MCSVM with all possible kernel functions and their parameters [33]. The mathematical theory of SVM is described as follows. Let x i be the data input and y i the associated labels in that i = 1 , 2 , , L , where L is the number of data samples. It is assumed that the data points belong to two classes “A” and “B”. Each data point is non-linearly mapped to a feature space separated by a hyperplane with the basic geometric equation:
f ( x ) = w × x + b = 0
where b is a scalar and w is a P-dimensional vector. These are the key parameters that determine the hyperplane position. If b = 0 , the hyperplane will pass by the origin. Otherwise, the margin is created or increased. The parallel hyperplanes that separate the two different data classes are defined in Equation (14) for the first class and in Equation (15) for the second one.
w × x + b = 1 .
w × x + b = 1 .
Through geometric calculations, the distance between the hyperplanes or the margin width is 2 | w | . This maximises the margin width to achieve an optimum separation between the two classes. In order to maximise the margin width, | w | should be minimised which brings in the criteria: w × x i + b 1 for the first class or w × x i + b 1 for the second one. This will force the points from each class not to exceed the class hyperplanes. The samples located on the hyperplanes are named support vectors and this is from where the name SVM originates. The hyperplane is obtained as a solution to the optimisation problem in Equation (16), while considering the noise slack variable ζ i which determines the range to which the samples overstep the margin, and the error penalty C which represents the trade-off between maximisation of the margin and classification error during the training phase.
m i n 1 2 | | w | | 2 + C i = 1 M ζ i
S u b j e c t t o y i ( w T × x i + b ) 1 ζ i ζ i 0 , i = 1 , , M
where ζ i denotes the distance between the margin and the data point x i which is in error. The calculation of Equation (16) is simplified and solved through a Lagrangian problem which is explained in more detail in [34]. This introduces an α i parameter which expresses w in solving Equation (16). The solution yields to a non-linear decision function expressed as:
f ( x ) = s i g n ( i , j = 1 M α i y i ( x i x j ) + b ) .
The challenges that may be faced in the SVM learning process of high dimension feature space are data over-fitting and computational errors. Over-fitting can be solved by introducing a kernel function which performs a dot product of the feature space, i.e., K ( x i , x j ) = ( Φ T ( x i ) × Φ ( x j ) ) . The definition of this kernel function can be found in [30] . A non-linear vector function Φ ( x ) = ( ϕ 1 ( x ) , , ϕ l ( x ) ) , where l is the feature space’s dimension, is implemented to reformulate the decision function in Equation (17) to:
f ( x ) = s i g n ( i , j = 1 M α i y i ( Φ T ( x i ) × Φ ( x j ) ) + b ) .
Since SVM is a binary classifier, it cannot handle more than two classes. This is not suitable for classification of multiple discharge sources. For this reason, MCSVM is implemented using the One-Against-One approach where k ( k 1 ) / 2 models are constructed. Each model is trained on two classes, A and B, as a normal binary classification. The testing step is performed through a “Max Win” voting method. If the closest class to the test sample is class A, then the vote for this class is increased by one. The winning class is the one which has the highest number of votes.

5. Application to EMI Data

The data was sensed while the assets were operating using the described EMI method in Section 3. The time-resolved signals were recorded at a sampling rate of 24 kHz by means of a PD Surveyor 200 (PDS200) device which follows the CISPR-16-1-1 standard for EMC type filtering. The PDS200 is used for PD surveying and has the ability to detect and analyse Radio Frequency Interference (RFI) as well as EMI radiation. PDS200 is identical to PDS100 except that it has a lower frequency range that starts from 50 kHz up to 1 GHz. In contrast, PDS100 operates in the range of 50 MHz–1 GHz. The low frequency option makes the PDS200 suitable to use for EMI detection and analysis, unlike PDS100 which is used to detect RFI emission, in the Ultra-High-Frequency (UHF) range, only for RFI surveying by means of antennas [35]. It also provides a frequency spectrum of the recorded signals using the EMI standard quasi peak detection method. “EMI experts” selected the time signals measured at a frequency of interest in the spectrum for further analysis and event identification, then they labelled the event type in each signal. These labels help in training the classification algorithm. Some of these signals were found to contain more than one event and therefore were given a combination of two labels (e.g., PD+A) which is considered as a single label. In this work, the multi-labelled signals are considered as a single class. Each measured signal contains 500 cycles over 10s. The overall data was collected from three different sites which are described as follows.
  • Site 1: The data was measured at the neutral earth cable of a 661 MVA hydrogen/water cooled synchronous generator operating at 23.5 kV, 19 kA, 3 phase, 50 Hz, 3000 RPM, 0.85 lag/0.95 lead power factor and 2 pole. A total of 13 signals were identified to contain E+mPD, C+E, C, N, PN and mPD.
  • Site 2: similar to the previous site, the measurements were taken at the neutral earth of different assets including a General Step-Up (GSU) Transformer operating at 430/15.5 kV, 444/12329 A, 3 phase, 50 Hz, 331 MVA, IPB and Station Transformer (Sta XFMR). The events identified in the GSU transformer are mPD+mA, PD+mA, PD+A, PD and the events identified in the IPB are PD and PN. Finally, the DM event was identified in the Sta XFMR.
  • Site 3: The data was measured at the neutral earth cable of an H2 cooled generator, operating at 294.25 MVA, 15 kV, 0.85 PF, 2 pole and 3000 RPM, from which seven signals were selected with an additional signal selected from an H2 cooled Steam Turbine Generator (STG) operating at 15 kV, 2 pole and 3000 RPM. The labelled events found at the neutral earth cable are PN, PD, NVFD, E and the ones found at the STG are: PD, E+mPD and E+PD.
A fourth scenario of common events between sites is also considered. This data set is a subset of the data collected across the three sites and includes PN, PD and mPD. Note that “m”, stands for a minor event which implies an occurrence at a smaller rate and lower discharge level. Here, mPD is considered as a type of PD, thus it is worth investigating its classification among PD signals observed in the other sites.
The described algorithms in Section 4 were applied, and each signal’s multiple frequency components were obtained using ALIF. Figure 5 and Figure 6 illustrate example PD and PN signals respectively and their decomposed IMFs using ALIF. Next, PE and DE values were calculated, with m = 3 and d = 1 , for each IMF to obtain a 1 × 8 feature vector for each signal. Table 1 presents the PE and DE values calculated for each IMF of the illustrated signals in Figure 5 and Figure 6. Differences in IMFs as well as their PE and DE values are observed between PD and PN signals, thus successful classification should be expected. The feature vectors are finally implemented in MCSVM which would draw boundaries that group the IMFs’ PE and DE values in the feature space. Polynomial kernel functions of second- and third-order were implemented based on the findings of a Grid Search method [33]. The training/testing strategy followed a ten-fold hold-on cross validation method, where the model is trained on 90% of the data and is tested on the remaining 10% in ten iterations; each uses a random different data batch. The total classification accuracy is calculated as the average accuracy resulting from all iterations. Training of SVM implies presenting the training data set and its associated labels to the classifier. In the testing phase the unseen data set is presented to the classifier, which will subsequently predict the data labels. The predicted labels are compared to the true labels that are provided by “EMI experts”, and then the accuracy % is calculated. A potential limitation of the work is that the labels may not be accurate, yet a small number of mislabelled data may occur. However, an experienced expert with modest training is capable of identifying many of the common faults related to EMI signals. Furthermore, one mislabelled instance is not likely to greatly affect the classification accuracy as the SVM algorithm would draw a boundary around the majority of instances belonging to a particular class. The classification process is performed using all the data in each scenario (site 1, 2, 3 and common events between sites) discussed previously.

6. Results and Discussion

Classification accuracy for each case is illustrated in Table 2. Remarkably, 100% classification accuracy is observed in sites 2, 3 and in the common data subset scenario. Lower classification accuracy is observed in site 1. However, it is clear that an improvement in accuracy is achieved for each case, compared to [2]. An improvement of 8% is achieved for site 1 and 3 and an improvement of 14% and 4% is achieved for sites 2 and the common condition case respectively. In order to further explain the performance of the ALIF-PE, DE algorithm in site 1, the confusion matrix between event classifications is presented in Figure 7. This matrix shows the class (columns) in which an event (rows) was classified. The highest confusion (17%) is observed between C and N. This could be due to the presence of noise that has similar characteristics to N in the C signal as noise is almost inevitable in real-world signals, especially in an electrical environment. A confusion of 8% between N and PN is also observed. mPD has minor confusion with C and N of 3% each. The overall findings lead to the conclusion that EMI events can be distinguished regardless of the site or origin of the asset’s collected data. It is important to highlight that classification accuracy may be affected if more events or more data are implemented in the classifier.

7. Conclusions

This work introduces an extended pattern recognition solution for real-world EMI signal classification. The solution involves extraction of relevant time and frequency fingerprints using ALIF plus PE and DE entropy-based features. MCSVM was also used to classify different EMI signals collected in three power system sites. Results demonstrate an improved classification accuracy over the previously proposed approach. The other significant contribution is the successful classification of common sources between different sites. The outcome of this work could potentially be exploited to develop a condition monitoring software which is based on EMI expert system knowledge.

Author Contributions

All authors on this paper contributed to the work. “I.M., G.M., B.G.S. and A.N. conceived and designed the experiments and helped draw conclusions; I.M. and M.H.-N. performed the experiments; I.M., A.N. and M.H.-N. analyzed the data; P.B. contributed reagents/materials/analysis tools; I.M. wrote the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AArcing
ALIFAdaptive Local Iterative Filtering
CCorona
DEDispersion Entropy
DMData Modulation
EExciter
EMIElectro-Magnetic Interference
GSUGeneral Step-Up
HFCTHigh Frequency Current Transformer
HVHigh Voltage
IMFIntrinsic Mode Function
IPBIsolated Phase Bus
MCSVMMulti-Class SVM
NVFDNon Variable Frequency Drive
PDPartial Discharge
PDEPartial Differential Equation
PDS200Partial Discharge Surveyor 200
PEPermutation Entropy
PNProcess Noise
PRPDPhase Resolved PD
SDStopping Distance
STGStep Up Generator
Sta XFMRStation Transformer
SVMSupport Vector Machine
UHFUltra-High-Frequency

References

  1. Timperley, J.E.; Vallejo, J.M. Condition Assessment of Electrical Apparatus With EMI Diagnostics. IEEE Trans. Ind. Appl. 2017, 53, 693–699. [Google Scholar] [CrossRef]
  2. Mitiche, I.; Morison, G.; Hughes-Narborough, M.; Nesbitt, A.; Boreham, P.; Stewart, B.G. Classification of Partial Discharge Signals by Combining Adaptive Local Iterative Filtering and Entropy Features. In Proceedings of the IEEE Conference on Electrical Insulation and Dielectric Phenomena, FortWorth, TX, USA, 22–25 October 2017; pp. 335–338. [Google Scholar]
  3. Álvarez, F.; Garnacho, F.; Khamlichi, F.; Ortego, J. Classification of partial discharge sources by the characterization of the pulses waveform. In Proceedings of the IEEE International Conference on Dielectrics (ICD), Montpellier, France, 3–7 July 2016; pp. 514–519. [Google Scholar]
  4. Altenburger, R.; Heitz, C.; Timmer, J. Analysis of Phase-Resolved Partial Discharge Patterns of Voids Based on a Stochastic Process Approach. J. Phys. D Appl. Phys. 2002, 35, 1149–1163. [Google Scholar] [CrossRef]
  5. Badieu, L.V.; Koltunowicz, W.; Broniecki, U.; Batlle, B. Increased Operation Reliability Through PD Monitoring of Stator Winding. In Proceedings of the 13th INSUCON Conference, Birmingham, UK, 16–18 May 2017; pp. 1–6. [Google Scholar]
  6. Raymond, W.J.K.; Illias, H.A.; Abu Bakar, A.H. Classification of Partial Discharge Measured under Different Levels of Noise Contamination. PLoS ONE 2017, 12, 1–20. [Google Scholar] [CrossRef]
  7. Chatpattananan, V.; Pattanadech, N.; Vicetjindavat, K. PCA-LDA for Partial Discharge Classification on High Voltage Equipment. In Proceedings of the IEEE 8th International Conference on Properties applications of Dielectric Materials, Bali, Indonesia, 26–30 June 2006; pp. 479–481, ISBN 1-4244-0189-5. [Google Scholar]
  8. Pattanadech, N.; Nimsanong, P.; Potivejkul, S.; Yuthagowith, P.; Polmai, S. Partial discharge classification using probabilistic neural network model. In Proceedings of the 18th International Conference on Electrical Machines and Systems, Pattaya, Thailand, 25–28 October 2015; pp. 1176–1180, ISBN 978-1-4799-8805-1. [Google Scholar]
  9. Lalitha, E.M.; Satish, L. Wavelet analysis for classification of multi-source PD patterns. IEEE Trans. Dielectr. Electr. Insul. 2000, 7, 40–47. [Google Scholar] [CrossRef]
  10. Wang, K.; Li, J.; Zhang, S.; Qiu, Y.; Liao, R. Time-frequency features extraction and classification of partial discharge UHF signals. In Proceedings of the International Conference on Information Science, Electronics and Electrical Engineering, Sapporo, Japan, 26–28 April 2014; pp. 1231–1235, ISBN 978-1-4799-3197-2. [Google Scholar]
  11. Hunter, J.A.; Hao, L.; Lewin, P.L.; Evagorou, D.; Kyprianou, A.; Georghiou, G.E. Comparison of two partial discharge classification methods. In Proceedings of the Conference Record of the 2010 IEEE International Symposium on Electrical Insulation, San Diego, CA, USA, 6–9 June 2010; pp. 1–5, ISBN 978-1-4244-6301-5. [Google Scholar]
  12. Albarracín, R.; Robles, G.; Martínez-Tarifa, J.M.; Ardila-Rey, J. Separation of Sources in Radiofrequency Measurements of Partial Discharges using Time-Power Ratios Maps. ISA Trans. 2015, 58, 389–397. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Moore, P.J.; Portugues, I.E.; Glover, I.A. Radiometric Location of Partial Discharge Sources on Energized High-Voltage Plant. IEEE Trans. Power Deliv. 2005, 20, 2264–2272. [Google Scholar] [CrossRef]
  14. Robles, G.; Fresno, J.M.; Martínez-Tarifa, J.M. Separation of Radio-Frequency Sources and Localization of Partial Discharges in Noisy Environments. Sensors 2015, 15, 9882–9898. [Google Scholar] [CrossRef] [PubMed]
  15. Albarracín, R.; Ardila-Rey, J.A.; Mas’ud, A.A. On the Use of Monopole Antennas for Determining the Effect of the Enclosure of a Power Transformer Tank in Partial Discharges Electromagnetic Propagation. Sensors 2016, 16, 148. [Google Scholar] [CrossRef] [PubMed]
  16. Cicone, A.; Liu, J.; Zhou, H. Adaptive local iterative filtering for signal decomposition and instantaneous frequency analysis. Appl. Comput. Harmon. Anal. 2016, 41, 384–411. [Google Scholar] [CrossRef]
  17. Xie, Y.; Tang, J.; Zhou, Q. Feature extraction and recognition of UHF partial discharge signals in GIS based on dual-tree complex wavelet transform. Eur. Trans. Electr. Power 2010, 20, 639–649. [Google Scholar] [CrossRef]
  18. Ravelo-García, A.G.; Navarro-Mesa, J.L.; Casanova-Blancas, U.; Martín-González, S.I.; Quintana-Morales, P.J.; Guerra Moreno, I.; Canino-Rodríguez, J.M.; Hernández-Pérez, E. Application of the Permutation Entropy over the Heart Rate Variability for the Improvement of Electrocardiogram-based Sleep Breathing Pause Detection. Entropy 2015, 17, 914–927. [Google Scholar] [CrossRef]
  19. Rostaghi, M.; Azami, H. Dispersion Entropy: A Measure for Time-Series Analysis. IEEE Signal Process. Lett. 2016, 23, 610–614. [Google Scholar] [CrossRef]
  20. Timperley, J.E.; Vallejo, J.M. Condition assessment of electrical apparatus with EMI diagnostics. In Proceedings of the IEEE Petroleum and Chemical Industry Committee Conference, Houston, TX, USA, 5–7 October 2015; pp. 1–8, ISBN 978-1-4799-8502-9. [Google Scholar]
  21. Specification for Radio Disturbance and Immunity Measurement Apparatus and Methods-Part 1: Radio Disturbance and Immunity Measuring Apparatus, IEC: 2015, CISPR 16-1-1, Part 1-1. Available online: http://www.iec.ch/emc/basic_emc/basic_cispr16.htm (accessed on 26 January 2018).
  22. Timperley, J.E.; Vallejo, J.M.; Nesbitt, A. Trending of EMI data over years and overnight. In Proceedings of the IEEE Electrical Insulation Conference, Philadelphia, PA, USA, 8–11 June 2014; pp. 176–179, ISBN 978-1-4799-2789-0. [Google Scholar]
  23. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The Empirical Mode Decomposition and The Hilbert Spectrum for Nonlinear and Non-Stationary Time Series Analysis. Proc. R. Soc. Lond. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  24. Lin, L.; Wang, Y.; Zhou, H. Iterative Filtering as an Alternative Algorithm for Empirical Mode Decomposition. Adv. Adapt. Data Anal. 2009, 1, 543–560. [Google Scholar] [CrossRef]
  25. Bandt, C.; Pompe, B. Permutation Entropy: A Natural Complexity Measure for Time Series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  26. Rathie, P.; Da Silva, S. Shannon, Levy, and Tsallis. Appl. Math. Sci. 2008, 2, 1359–1363. [Google Scholar]
  27. Riedl, M.; Müller, A.; Wessel, N. Practical Considerations of Permutation Entropy. Eur. Phys. J. Spec. Top. 2013, 222, 249–262. [Google Scholar] [CrossRef]
  28. Yan, R.; Liu, Y.; Gao, R.X. Permutation Entropy: A nonlinear Statistical Measure for Status Characterization of Rotary Machines. Mech. Syst. Signal Process. 2012, 29, 474–484. [Google Scholar] [CrossRef]
  29. Massimiliano, Z.; Luciano, Z.; Osvaldo, R.A.; Papo, D. Permutation Entropy and Its Main Biomedical and Econophysics Applications: A Review. Entropy 2012, 14, 1553–1577. [Google Scholar] [CrossRef]
  30. Vapnik, V. The Nature of Statistical Learning Theory; Springer: New York, NY, USA, 1961. [Google Scholar]
  31. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Cambridge, UK, 2006. [Google Scholar]
  32. Lesniak, J.M.; Hupse, R.; Blanc, R.; Karssemeijer, N.; Székely, G. Comparative Evaluation of Support Vector Machine Classification for Computer Aided Detection of Breast Masses in Mammography. Phys. Med. Biol. 2012, 57, 2560–2574. [Google Scholar] [CrossRef] [PubMed]
  33. Boardman, M.; Trappenberg, T. A Heuristic for Free Parameter Optimization with Support Vector Machines. In Proceedings of the International Joint Conference on Neural Networks, Vancouver, BC, Canada, 16–21 July 2006; pp. 1337–1344, ISBN 978-1-4799-2789-0. [Google Scholar]
  34. Widodo, A.; Yang, B.-S. Support Vector Machine in Machine Condition Monitoring and Fault Diagnosis. Mech. Syst. Signal Process. 2007, 21, 2560–2574. [Google Scholar] [CrossRef]
  35. Robles, R.; Albarracín, R.; Vázquez-Roy, J.L.; Rajo-Iglesias, E.; Martínez-Tarifa, J.M.; Rojas-Moreno, M.V.; Sánchez-Fernández, M.; Ardila-Rey, J. On the use of vivaldi antennas in the detection of partial discharges. In Proceedings of the IEEE International Conference on Solid Dielectrics, Bologna, Italy, 30 June–4 July 2013; pp. 302–305, ISBN 978-1-4799-2789-0. [Google Scholar]
Figure 1. Overall process diagram of the proposed approach from data acquisition to pattern recognition of Electro-Magnetic Interference (EMI) events: Partial Discharge (PD), Corona (C), Process Noise (PN) and Data Modulation (DM).
Figure 1. Overall process diagram of the proposed approach from data acquisition to pattern recognition of Electro-Magnetic Interference (EMI) events: Partial Discharge (PD), Corona (C), Process Noise (PN) and Data Modulation (DM).
Sensors 18 00406 g001
Figure 2. Permutation Entropy (PE) ordinal patterns for m = 3 .
Figure 2. Permutation Entropy (PE) ordinal patterns for m = 3 .
Sensors 18 00406 g002
Figure 3. Example of pattern mapping/identification in a time series.
Figure 3. Example of pattern mapping/identification in a time series.
Sensors 18 00406 g003
Figure 4. Support Vector Machine (SVM) linear space mapping using second-order polynomial kernel function.
Figure 4. Support Vector Machine (SVM) linear space mapping using second-order polynomial kernel function.
Sensors 18 00406 g004
Figure 5. Example PD signal decomposed into Intrinsic Mode Functions (IMFs) using the Adaptive Local Iterative Filtering (ALIF) algorithm.
Figure 5. Example PD signal decomposed into Intrinsic Mode Functions (IMFs) using the Adaptive Local Iterative Filtering (ALIF) algorithm.
Sensors 18 00406 g005
Figure 6. Example PN signal decomposed into IMFs using the ALIF algorithm.
Figure 6. Example PN signal decomposed into IMFs using the ALIF algorithm.
Sensors 18 00406 g006
Figure 7. Confusion matrix of site 1.
Figure 7. Confusion matrix of site 1.
Sensors 18 00406 g007
Table 1. PE and Dispersion Entropy (DE) values for each IMF of the example PD and PN signals.
Table 1. PE and Dispersion Entropy (DE) values for each IMF of the example PD and PN signals.
SignalFeatureIMF1IMF2IMF3IMF4
PD
PE1.751.451.241.06
DE1.131.421.501.34
PN
PE1.791.391.140.95
DE2.011.821.641.39
Table 2. Classification accuracy results.
Table 2. Classification accuracy results.
CaseClassification Accuracy %
Site 191
Site 2100
Site 3100
Common data subset100

Share and Cite

MDPI and ACS Style

Mitiche, I.; Morison, G.; Nesbitt, A.; Hughes-Narborough, M.; Stewart, B.G.; Boreham, P. Classification of Partial Discharge Signals by Combining Adaptive Local Iterative Filtering and Entropy Features. Sensors 2018, 18, 406. https://doi.org/10.3390/s18020406

AMA Style

Mitiche I, Morison G, Nesbitt A, Hughes-Narborough M, Stewart BG, Boreham P. Classification of Partial Discharge Signals by Combining Adaptive Local Iterative Filtering and Entropy Features. Sensors. 2018; 18(2):406. https://doi.org/10.3390/s18020406

Chicago/Turabian Style

Mitiche, Imene, Gordon Morison, Alan Nesbitt, Michael Hughes-Narborough, Brian G. Stewart, and Philip Boreham. 2018. "Classification of Partial Discharge Signals by Combining Adaptive Local Iterative Filtering and Entropy Features" Sensors 18, no. 2: 406. https://doi.org/10.3390/s18020406

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop