Next Article in Journal
Digital Technology, Factor Allocation and Environmental Efficiency of Dairy Farms in China: Based on Carbon Emission Constraint Perspective
Previous Article in Journal
Catalyzing Cooling Tower Efficiency: A Novel Energy Performance Indicator and Functional Unit including Climate and Cooling Demand Normalization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Generic Direct M-SVM Model Improved by Kohonen Map and Dempster–Shafer Theory to Enhance Power Transformers Diagnostic

by
Mounia Hendel
1,
Fethi Meghnefi
2,
Mohamed El Amine Senoussaoui
3,
Issouf Fofana
2,* and
Mostefa Brahami
4
1
LGEM, Ecole Supérieure en Génie Electrique et Energétique d’Oran, Oran 31000, Algeria
2
Canada Research Chair, Tier 1, ViAHT, University Québec, Chicoutimi, QC G7H 2B1, Canada
3
LGPCS, University Mustapha Stambouli of Mascara, Mascara 29000, Algeria
4
ICEPS, Faculty of Electrical Engineering, Djilali Liabes University of Sidi Bel Abbes, Sidi Bel Abbes 22000, Algeria
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(21), 15453; https://doi.org/10.3390/su152115453
Submission received: 25 September 2023 / Revised: 12 October 2023 / Accepted: 18 October 2023 / Published: 30 October 2023

Abstract

:
Many power transformers throughout the world are nearing or have gone beyond their theoretical design life. Since these important assets represent approximately 60% of the cost of the substation, monitoring their condition is necessary. Condition monitoring helps in the decision to perform timely maintenance, to replace equipment or extend its life after evaluating if it is degraded. The challenge is to prolong its residual life as much as possible. Dissolved Gas Analysis (DGA) is a well-established strategy to warn of fault onset and to monitor the transformer’s status. This paper proposes a new intelligent system based on DGA; the aim being, on the one hand, to overcome the conventional method weaknesses; and, on the other hand, to improve the transformer diagnosis efficiency by using a four-step powerful artificial intelligence method. (1) Six descriptor sets were built and then improved by the proposed feature reduction approach. Indeed, these six sets are combined and presented to a Kohonen map (KSOM), to cluster the similar descriptors. An averaging process was then applied to the grouped data, to reduce feature dimensionality and to preserve the complete information. (2) For the first time, four direct Multiclass Support Vector Machines (M-SVM) were introduced on the Generic Model basis; each one received the KSOM outputs. (3) Dempster–Shafer fusion was applied to the nine membership probabilities returned by the four M-SVM, to improve the accuracy and to support decision making. (4) An output post-processing approach was suggested to overcome the contradictory evidence problem. The achieved AUROC and sensitivity average percentages of 98.78–95.19% (p-value < 0.001), respectively, highlight the remarkable proposed system performance, bringing a new insight to DGA analysis.

1. Introduction

Indicators of sustainability focusing on energy are crucial tools used to assess and monitor progress toward guaranteeing electricity delivery to end-users [1]. In the last decades, power grids have been facing growing interest in deploying new and intelligent technologies to obtain improved reliability and availability of power supply. This is important in meeting new challenges due to accelerating urbanization and evolving requirements to ensure smart cities. The smart city concept mostly relies on cameras, sensors and monitoring tools to maintain or support human well-being continuously over time. The data collected are processed and analyzed to improve operational efficiency of major equipment such as power transformers, public safety, life quality, and also ensure efficient electrical installations [2,3]. In this context, accurate monitoring of major assets is essential. Power transformers, which are the most essential and expensive devices in the power transmission and distribution networks, are aging worldwide ahead of their theoretical design life. Their role is to facilitate the transition between the different electrical network levels (production, transport and distribution), while minimizing losses from the Joule effect. Due to their importance in an electrical structure, their reliable operation guarantees a good efficiency to the entire distribution network. In particular, when a power transformer suddenly explodes due to various factors, it may affect the generator output, which causes significant damage to the electric company’s economy and to the user’s property, and may also cause irreversible damage to human safety, especially to maintenance personnel. Thus, it is necessary to ensure an excellent monitoring of the power transformer’s condition in order to avoid sudden catastrophic failures [4,5].
Throughout their operation, power transformers are continuously subjected to different stresses (electrical, thermal and mechanical) and aggressive chemical byproducts. Excessive transformer stresses adversely impact insulating materials (solid insulation and oil), leading to an internal transformer failure [6]. This failure can be observed by analyzing the combustible and non-combustible gases formed and dissolved in the transformer’s oil, such as: Hydrogen ( H 2 ), Oxygen ( O 2 ) , Nitrogen ( N 2 ), Carbon Monoxide ( CO ), Methane ( CH 4 ), Carbon Dioxide ( CO 2 ), Ethylene ( C 2 H 4 ), Ethane ( C 2 H 6 ), Acetylene ( C 2 H 2 ), etc. It is therefore essential to regularly analyze the oil’s condition to detect potential defects. This is done by measuring the dissolved gas concentrations [1,7]. To date, several strategies have been proposed for transformers condition monitoring from oil (Furan analysis, Physico-chemical analyses, Dissolved gas analysis (DGA), etc.). Among these, DGA is the most widely used approach for power transformer diagnosis, since it is non-invasive, very simple to implement, inexpensive and above all effective.
In the last decades, a large number of DGA-based approaches have been proposed. These approaches can be subdivided into two categories: traditional methods and Artificial Intelligence (AI)-based techniques. Traditional methods mainly include: Duval’s Triangle (D-T) and Pentagons [8], Rogers Ratios (R-R) [9], Dornenburg Ratios (D-R) [10], IEC Ratios (IEC-R) [10] and Key Gases (K-G) [11]. However, conventional DGA methods mainly perform diagnosis by coding, which leads to problems such as absolute code limits, missing codes and sensitivity to gas volume fraction fluctuations, etc. For example, the IEC-R precision is of concern for incomplete coding reasons (it is, for example, impossible to differentiate between low- and high-energy discharge). R-R are only effective for detecting thermal faults and are ineffective for detecting other faults. The D-R generally detects three fault categories and is not able to return the fault severity. D-T is mainly based on the use of five triangles, and triangles 4 and 5 are in several cases contradictory [7]. The reason why the different approaches can give contradictory answers for the same sample is unclear, and it is difficult for the expert to prioritize an answer. This can induce misinterpretation, leading to incomplete and/or uncertain classification. Hence, there is a need to find powerful alternatives that are not limited by the programmer’s intelligence and that are able to learn and to improve by examples and face situations never seen before, such as AI alternatives. Indeed, these methods are mainly based on the definition of one/or more decision boundaries based on very powerful mathematical models and take into account a sample global (non-partial) exploration. In this sense, several ingenious studies based on AI approaches have been proposed (separately or combined with conventional methods) [1,7,12,13,14] and have proven their effectiveness in comparison to conventional methods.
In this contribution, a new approach for monitoring the power transformer’s condition is presented. The approach is based on a two-stage hybrid system: a descriptor extraction and construction stage, and a discrimination and outputs merging stage. The block diagram of the proposed Power Transformer Diagnosis (PTD) approach is depicted in Figure 1.
-
In the first stage (Stage-1), a relevant descriptor space is constructed. Six descriptor sets extracted according to six distinct DGA techniques are firstly retained: K-G, IEC-R, R-R, D-R, D-T and Gases Percentage (G-P). These parameters are then combined. This is followed by a redundant parameters elimination phase. After that, a standardization process is considered to formalize all the parameters to the same interval. Finally, a novel descriptor reconstruction stage based on the Kohonen Self-Organizing Maps (KSOM) [15] is proposed to decrease the data model and facilitate the discriminator’s work while keeping all the information.
-
In the second stage (Stage-2), four direct probabilistic Multiclass Support Vector Machines (M-SVM) are implemented for the first time, via a Generic M-SVM Model (GM-SVM) [16]: the Weston and Watkins (WW) model [17], the Crammer and Singer (CS) model [18], the Lee et al. (LLW) model [19], and the Quadratic Loss Multi-Class Support Vector machine (M-SVM2) model [20]. These four models are used as a solution to overcome the weaknesses of the indirect M-SVM models widely used in this problem. Each used M-SVM considers the returned parameters set by Stage-1, and calculates the belonging probabilities to each of the considered nine classes.
-
Then, to approve the final outputs, the Dempster–Shafer (DS) [21,22] fusion is applied to combine the four M-SVM outputs within the beliefs and evidence model framework. Also, this stage proposes an alternative to post-process the after-fusion outputs: the example is considered well-classified if its probability exceeds a decision threshold; otherwise, the example is assigned to the reject class.
Thus, the proposed system has a primary objective to improve the descriptor extraction step with the proposed KSOM parameters reconstruction approach. The latter effectively minimizes the information loss as much as possible, unlike the traditional functionalities selection approaches. The second objective focuses on implementing the four direct M-SVM through the Generic M-SVM Model, with the aim of overcoming the limits of traditional M-SVM based on decomposition methods (minimize the classifier complexity and save execution time). Finally, the proposed system has the final objective of strengthening the final decision making by merging four M-SVM by DS Fusion. After that, an approach to solve the contradictory evidence problem, linked to DS fusion, is proposed. Section 2 describes in detail the motivations and the originality of the approach taken.
The paper is organized as follows. Section 2 exposes the foundations, the motivations and the innovations of the study. Section 3 explains the adopted parameter extraction approaches, as well as the considered data reconstruction approach. Section 4 exposes the theoretical framework of probabilistic direct M-SVM and the theoretical framework of the DS fusion model. Section 5 presents the used database description, the M-SVM hyper parameters selection, plus the performance enumeration and their statistical analysis. The paper is terminated with a conclusion and perspectives in Section 6.

2. Study Motivations and Innovations

One of the key points of a PTD method lies in the dissolved gases representation domain assortment (choice) through the parameter extraction. This is due to the fact that the discriminator generalization strongly depends on the parameter space. Different representations have been exposed in the literature, such as: K-G [7,14,21,22,23,24], IEC-R [14,25,26], Personalized Ratios [14,25], R-R [25,26], Logarithmic Data Transformation [23], G-P [23,26], DGA code [27], Standardized Data [23], D-R [26], D-T [25], etc. Also, several comparative studies between these descriptors [23,24,25,26] have been carried out, but each of these approaches has advantages and disadvantages [7]; it would be more interesting to take the benefits of each one through a combination.
However, PTD applications are often characterized by a low learning vector number, and a large vector dimension necessarily accentuates the curse of dimensionality. Indeed, the phenomenon is induced by the fact that there is a strong dissimilarity between the observations and a great divergence among the training examples. This lack of data density inevitably affects and alters the discriminator generalization, whose foundation is essentially based on statistical significance. The dimensionality reduction appears as the main reflection in the face of the curse of dimensionality. There are two categories of dimensionality reduction approach: the variable selection approaches, which consist of electing a descriptor sample from the global variables set; and the variable transformation approaches, which consist of reconstructing a new descriptor set based on the similarity characteristics of the initial variables. The first method category generates a loss of useful information (partial data exploitation), which makes it suboptimal compared to the second [28,29].
For dimensionality reduction, few previous works have been proposed, and to the best of our knowledge, the work carried out in this context is based on the descriptor selection approaches. For example, the authors in [14] proposed a Practical Swarm Optimization (PSO) algorithm for feature selection. Xie et al. [12] proposed an approach based on relief algorithms for parameter reduction. Finally, one study [30] proposed a parameter selection method based on the Genetic Algorithm (GA). The current paper first exposes the use, for the first time, of a very powerful AI approach: KSOM, for merging and reconstructing six descriptor sets that are the most used in real applications (K-G, IEC-R, R-R, D-R, G-P, D-T); the objective being to reduce the descriptor space and to preserve the data entirety contribution.
The issues related to a classification method choice are also very important. The choice must be related to several parameters, more particularly: the nature of the data (noisy, overlapping, redundant, and/or disordered) and the learning set size and nature (small or large). Generally, in PTD applications one encounters non-linear data, a very small training set (as earlier reported), non-equiprobable categories and a significant correlation between the distinct category examples. In this case, the SVM are more efficient than neural networks and offer remarkable generalization results due to their advantages. A few possibilities can be explored:
-
Embedded in real-time processes (low execution time);
-
Applied to randomly distributed and unknown data;
-
Ensure a global optimum due to the convex optimization principle;
-
Do not suffer from overfitting and over-learning problems;
-
Decrease the curse of dimensionality.
Therefore, several previous studies have proposed SVM-based PTD systems [1,21,30,31,32].
The PTD is a multi-class discrimination problem, but SVM were induced by Cortes and Vapnik [33] to cover two-category discrimination problems. Unlike other machine learning approaches, where switching to the polytomy case is intuitive, SVM in most cases use decomposition methods (one-against-one, one-against-all, etc.). These dichotomous approaches generate significant complexity, which leads to a considerable execution time (final decision making, training algorithm), especially when dealing with large category classification problems. To overcome this problem, researchers are currently working on proposing new direct intuitive models that consider a single M-SVM for unique multi-class learning problem resolution. And to this date, four powerful models have been proposed, namely: the WW model, GS model, LLW model and M-SVM2 model. Geurmeur then proposed an innovative GM-SVM model, which brings together the four direct M-SVM offered to date. This research presents, as a second contribution, the introduction of the four direct M-SVM models in the PTD application via the GM-SVM model.
Also, serval information sources fusion is strongly recommended in discrimination problems, especially if false prediction stakes exert influence in a consistent way, such as in the case of this application. It is indeed much more informative to make a decision knowing that several sources agree. This study fits into the evidence theory context, which takes into consideration ambiguity and vagueness, credibility and the conflict of different independent proofs, and then produces a certain and a confident result. A number of previous works have studied DS fusion, whether to interpret the relationship between different data sources (DGA [14], Ratio methods [28], etc.), or the merging of several discriminator outputs [1,7,31,32], and have highlighted the theory contribution in the PTD problem.
For DS theory, when the classifiers are in perfect agreement or the conflict between them is weak, confidence in the final decision is reinforced by a higher recognition rate. However, if the discriminators are in strong conflict, the final decision is translated into a random result, which is necessarily questioned (one of the major DS law criticisms). To overcome this problem, the authors in [32] proposed to symbolize the reliability differences between the discriminator outputs by weighting coefficients; the latter allow reconstructing the base probabilities according to the attributed priority to the proofs. Nevertheless, the four direct M-SVM models implemented in this contribution offer theoretically the same performances, and each model has its own advantages and disadvantages. It is then practically impossible to prioritize one model. On the contrary, it would be more interesting to exploit the performances of each of them in order to reconstruct a more robust model—hence the interest in the merger. The third contribution of this study consists of the DS fusion of four direct M-SVM and proposition of an output post-processing approach to solve the contradictory evidence problem.
Thus, by implementing the proposed intelligent fault diagnosis system, it is hoped to:
-
Increase the four M-SVM performance by facilitating their classification task, thanks to the proposed descriptors reconstruction approach;
-
Reduce complexity and save execution time by implementing direct M-SVM instead of M-SVM based on decomposition methods;
-
Strengthen M-SVM outputs by translating them into posterior probabilities;
-
Strengthen decision-making, gain sensitivity and minimize false alarms by applying the DS fusion and the rejection class introduction.

3. Features Extraction and Reconstruction Methods

This section defines the adopted methodology for the input model construction (of the four direct M-SVM from DGA samples, which includes the following steps:

3.1. Feature Extraction Approaches

The literature exposes a multitude of interpretative-based DGA approaches to determine the condition of oil-immersed transformers. In this contribution, the following six approaches are considered:

3.1.1. Key Gases (K-G)

The first data set is built from nine type of gases: H 2 , O 2 , N 2 , CO, CH 4 , CO 2 , C 2 H 4 , C 2 H 6 ,   C 2 H 2 . The latter intervene in an effective way for the considered power transformer fault interpretation, according to IEC 60599 standard [23].

3.1.2. Gases Percentage (G-P)

The second set is based on nine gas percentage concentrations (retained in the first set): % H 2 , % O 2 , % N 2 , %CO, % CH 4 , % CO 2 , % C 2 H 4 , % C 2 H 6 , % C 2 H 2 ; relative to the total combustible and non-combustible gases sum (TG): ∑( H 2 , O 2 , N 2 ,   CO, CH 4 , CO 2 , C 2 H 4 , C 2 H 6 , C 2 H 2 ).

3.1.3. IEC Ratios (IEC-R)

The third set of data counts three ratios, which are generated from five key gases as follows:
R 1 = C 2 H 2 C 2 H 4
R 2 = C H 4 H 2
R 3 = C 2 H 4 C 2 H 6

3.1.4. Rogers Ratios (R-R)

The fourth descriptors set that is used considers the same ratios as those of the IEC method, in addition to the following ratio:
R 4 = C 2 H 6 C H 4

3.1.5. Dornenburg Ratios (D-R)

The fifth features set is a vector with four elements R1 and R2 of the IEC method, plus the two following ratios:
R 3 = C 2 H 2 C H 4
R 4 = C 2 H 6 C 2 H 2

3.1.6. Duval’s Triangle (D-T)

The last set of descriptors is expressed as a percentage of three key %CH4, %C2H2, and %C2H4, relative to their total concentration (TS): CH 4 ,   C 2 H 2 ,   C 2 H 4 .   The reader is invited to see reference [28] for more details on the graphing approach via Duval’s Triangle 1.

3.2. Features Combination and Redundant Features Elimination

Once all six features are defined by the selected diagnostic criteria application, a concatenation process is applied to generated vectors. This allows obtaining a global vector made up of 32 descriptors. The redundant parameters are then eliminated in order to produce a final vector composed of 27 parameters; these are considered input features to a KSOM map.

3.3. Features Standardization

For the purpose of an efficient preparation of discriminator inputs, it is generally recommended to implement a standardization process (to ensure optimal data modeling). This consists in using a common scale for all descriptors while preserving an identical general distribution and a similar ratio to those of the original parameters. In this sense, a centered/reduced type standardization process is applied:
x i j ˜ = x i j x j ¯ σ x j
Note that x j ¯ and σ x j represent, respectively, the mean and the standard deviation of the jth descriptor vector parameter.
Note: the standardization phase was carried out before the new dataset reconstruction, to facilitate the Vector Quantification (VQ) process by minimizing data dispersion.

3.4. Kohonen Self-Organizing Map Reduction

This research presents the implementation of an unsupervised learning SOM map (see Definition 1) to project the initial parameters space (of 27 descriptors) onto a reduced parameters space (of nine descriptors) based on the VQ principle (without information loss). The new descriptors are given by the external weights, which connect the map neurons to the input vectors. The learning consists of finding the weights which best cover the input space by applying the illustrated algorithm in Definition 1.
Definition 1
(Kohonen map unsupervised learning problem) [15]. Considering a M dimension learning set, where each learning vector consists of N descriptors. Let x 1 ,   x 2 , , x M be a set of input vectors of N dimension. Let v j = v j 1 ,   v j 2 , , x j N   be an external weights vector that connects neuron j 1 , n × m to an input vector x i ; and let w j l be an internal weights vector (obtained through the “Mexican hat” function) that connects neuron j to all these neighbors l 1 , n × m \ j . The KSOM unsupervised learning occurs in three steps:
-
Step 1: Elect cluster with maximum response.
  • Apply to the map input a learning vector x i .
  • Determine the best matching unit j , whose vector v j is closest to the input x i :
    d j t = m i n 1 < k < n × m i = 1 N x i t v i k t
  • Winning neuron j   activation .
-
Step 2: neighborhood construction around the winning neuron.
-
Step 3: adaptation of winning cluster weights:
v j t + 1 = v j t + α t x i t v i j t
v l t + 1 = v l t + α t x i t v i l t
where ( α < 1 ) is the learning step which decreases according to the iterations t .
-
Back to step 1 until the algorithm converges.
Figure 2 and Figure 3 show these weights at the start and at the end of training. When the algorithm starts, the weights of all the neurons in the map are grouped and initialized to low values. During the iterations, they begin to occupy the input space; the learning ends up by taking into account all of this space. Thus, each neuron represents a part of this space.
To select the retained map size, six n × m , different configurations were evaluated (with n and m 2 , 5 ). Each of these configurations generated fifteen epochs. Each epoch was then evaluated by a MultiLayer Perceptron (MLP) to choose the most efficient among the fifteen (from a generalization rate view point). Figure 4 shows the parameter distribution around the clusters for the six selected configurations. Also, Table 1 illustrates the classification rates relating to these same topologies. It can thus be possible to report the following information:
-
The minimum extreme dimension choice (the number of neurons in the map) is based on the fact that the majority of conventional methods use on average four ratios for decision-making. So a set of four parameters is to be reconstructed.
-
The maximum extreme dimension choice is fixed after dead neurons’ appearance, whose external connections do not represent any parameter; just as it is important to not go beyond the initial parameter number (final set < initial set).
-
The final map choice ( 3 × 3 configuration) among the six selected topologies is based on a trade-off between the neuron number and the maximum generalization rate, plus the dead neurons’ absence.
-
The new vector dimension is equal to the retained final map neuron number (for a total of nine parameters).
-
The value of each new descriptor is obtained by averaging the parameters associated with a given cluster.
Table 2 exposes the new descriptors after learning according to the initial parameters. It can thus be possible to see that some map neurons cover only one input descriptor, for example:
-
Neuron 2 (D2) represents the initial descriptor C 2 H 2 C 2 H 4 . This ratio is used by the three conventional methods: IEC-R, R-R and D-R, which considers it important.
-
Neuron 4 (D4) evokes the initial Dornenburg R3 report, considered very important for detecting the low-energy thermal fault category from other categories.
-
Neuron 6 (D6) is seen as being a support and confirmation element of the partial discharge fault if it is greater than 30%.
Also, other neurons cover several old descriptors, which are generally strongly correlated (information redundancy). For example:
-
Neuron 1 (D1) brings together the two descriptors H 2 , H 2 T G , which model the same information revealing the presence of the partial discharge type defect.
-
Neuron 7 (D7) brings together five descriptors involved in an insulation fault identification.
All of these findings confirm the coherence of the groups obtained after Kohonen map application.

4. Discrimination and Outputs Merging Stage

In this section, the mathematical foundations of direct probabilistic M-SVM, as well as the theoretical framework of the DS merging rule, are described.

4.1. Direct Probabilistic Multiclass Support Vector Machines (M-SVM)

Nowadays, SVM finds applications in various fields, and those because of their advantages. The SVM do not consider multi-category classification natively; they are created to calculate dichotomies. Thus, for multi-class applications, two categories of M-SVM can be considered; the first category (indirect M-SVM) is based on the use of decomposition techniques, while the second category (direct M-SVM) is based on a multivariate affine model and currently has four models: the WW model, GS model, LLW model, and M-SVM2. Geurmeur then devised a Generic GM-SVM model, which is a statistical properties unified representation of the four direct M-SVM. The GM-SVM model is considered in this paper to present in a global way the learning problem of each direct M-SVM.

4.1.1. Generic M-SVM Model

Direct M-SVM belong to the class of kernel learning algorithms and are considered as a direct generalization of binary SVM for the multi-category case (without using the decomposition approaches). Therefore, direct M-SVM operate in a specific function class H k , Q deduced from another function H k .
Definition 2
(Class of functions H k , Q ). Let X = x 1 , x 2 , x n be a set of vectors, where each vector is labelled according to ( y i 1 , Q ) Y , with Q 3 . Let H be a Hilbert space endowed with its scalar product . , . H , and H k , . , . H k be the homologous Reproducing kernel Hilbert Space (RKHS) induced by a positive k kernel [34]. The direct M-SVM class functions associated with the kernel k are represented as follows:
H k , Q = H k 1 Q
knowing that:
1 : be the space of real-valued constant functions on X .
and
  h H k , Q ,     x X ,   h x = h ¯ x + b = h ¯ k , k x , . H k + b k 1 k Q
With:
h ¯ = ( h ¯ k ) 1 k Q H k Q
b = b k 1 k Q R Q
(for more details see the Definitions 1–3 provided in Section 2.1 of [34]).
Definition 3 
(Generic M-SVM model, Section 2.1 Definition 4 in [16]). Let a learning set z m = x i , y i ; i = 1 , m with x i X   and y i   1 , Q   ; let H k , Q be the functions Class of kernel k . The multi-class direct M-SVM discriminator training consists of processing and resolving a quadratic program, which is defined as follows:
min h , ξ { M ξ P P + λ P H k h H k 2
s . t . i 1 , m ,   k 1 , Q \ y i , k 1 h y i x i h k x i k 2 ξ i 1 Q + k i 1 , m ,   k , l 1 , Q \ y i 2 , k 3 ξ i 1 Q + k ξ i 1 Q + l = 0 i 1 , m ,   k 1 , Q \ y i , 2 p ξ i 1 Q + k 0 1 K 1 k = 1 Q h k = 0
where:
The bounded operator P H k : defines an orthogonal projection from H k , Q into H k .
ξ   R Q m Z m   are the slack variables, with ( ξ i 1 Q + y i ) 1 i m .
λ ,   p , K 2   R + *   and ( K 1 , K 3 ) 0 , 1 2 .
The matrix M Q m , Q m rank is equal to Q 1 m ,   and it is a diagonal matrix if
p = 1 .
Let δ   be the Kronecker symbol; so, let us specify the general term of I Q m and M 2 , respectively:
m i j , j l = δ i , j δ k , l 1 δ y i , k .
m i j , j l 2 = s 1 δ y i , k 1 δ y i , l δ i , j ( δ k , l + Q 1 Q 1 )
Based on the Generic model unification, Table 3 illustrates the hyper parameter characteristics relating to each of the four M-SVM models.
For this experiment implementation, an open-source software package (MSVMpack) [35] based on the GM-SVM model is used.

4.1.2. Probabilistic Direct M-SVM

Power transformer automatic diagnosis is a critical and complex task; the discrimination error costs are in most cases asymmetrical, difficult to quantify and differ between experts and transformers. Thus, it is desirable and more reliable to consider classifiers that return a posteriori probability estimates, instead of a deterministic classification, which returns the membership class label (this allows associating a confidence degree with the final decision, for example knowing that the probability that the transformer presents a fault is 80% is better than not having any indicators). These probabilistic estimations can then be joined to other information sources for complex but more credible final decision making. However, SVM are at their basis deterministic discriminators. Nevertheless, it is possible to obtain a M-SVM posterior probability estimator by adapting Platt’s binary case solution [36] to the multi-class case, as follows:
k 1 , Q , h ˜ k = exp h k exp k = 1 Q h k

4.2. Dempster–Shafer (DS) Theory

Developed by Dempster [21] and formalized mathematically by Shafer [22], belief theory (or evidence theory) allows imprecision and uncertainty to be modeled simultaneously based on the belief function ( B e l ) and the plausibility function ( P l ); the latter are deduced from the mass functions ( m ). Below, the Mass, Plausibility and Belief functions, as well as the merging and the final decision rules, are explained.

4.2.1. Mass, Plausibility and Belief Functions

Let = ,   w 1 , w 2 , ,   w N be the set subspaces (focal elements) of the discernment frame. The liveness relative to a proposition A (for a given source S ) can be quantified by a mass function “ m ” (coryance measurement) m :   2 0 , 1 and must respect the two following properties: m = 0   and A m A = 1 .
From the mass function, it is easy to deduce B e l and P l   functions attributed to the focal element A , in accordance with the following two mathematical equations:
  A   2 , B e l A = B A ϕ m B
A 2 , P l A = B A ϕ m B = 1 B e l A ¯
With A ¯   being the opposite event of the proposition A . The quantity B e l A illustrates the belief strength in A , justified by the evidence information taken into consideration. Plausibility is the credibility dual function; its quantity P l A is considered as an upper bound on the belief strength likely to be attributed to A following new data.

4.2.2. Merging Rule

The merging rule can be generally applied at three levels, namely: low level (considering the raw data), intermediate level (descriptor merging) and high level (discriminator merging). In this work, a fusion at the high level is retained.
Masses combination on the basis of DS law is achieved by Dempster’s orthogonal sum; thus, for two masses functions   m 1 and m 2 and for all A     2 Ω , it yields:
m A = m 1 m 2 A = B 1 B 2 = A m 1 B 1 · m 2 B 2 1 K
where K represents the conflict:
K = B 1 B 2 = ϕ m 1 B 1 · m 2 B 2
In this study, probabilistic classifiers are considered, so the masses (of the combination rule) are replaced by posterior probabilities [32].

4.2.3. Final Decision Rule

Instead of providing a final decision on singleton hypotheses, and based on the highest probability rule as the decision criterion, the DS rule returns the final decision based on the overall expert responses.
The DS theory operation principle consists of both approving and consolidating; the proofs answer if they are in agreement or return an unfounded result that does not coincide with the desired target, in a significant conflict event between proofs. To overcome this problem, a post-processing process is proposed; the latter is applied to what is generated after merging outputs. Let X   be the input data space and   Y the output data space, where each observation x i in X is associated to labels of Y with a posterior probability h ˜ k . The proposed d h ˜ rule for post-processing outputs is as follows:
  x X i f     k     1 , Q   :   h ˜ k x 0.70 ,   t h e n   d h ˜ x = k     e l s e   d h ˜ x = X *                                                                          
If one of the outputs is greater or equal to 0.70, the example is considered well classified; otherwise, the example is assigned to the reject class X * . Indeed, it appears more logical and more prudent to reject an example with an uncertain assignment than to force an expert to give an answer through weighting biases.
The decision threshold has been set at 0.7, so that the gap between the dominant class and the next class always exceeds 0.4 = 1 0.7 . This threshold is judged quite sufficient and even severe, so that there is no confusion between classes. In other words, since the dominant class always wins a probability higher than 0.7, the eight other remaining classes only win together a probability lower than 0.3.
Furthermore, establishing the rejection class can return important information on the data quality and uncertainty/certainty (the data are erroneous and inconsistent). Indeed, a rejection is generally observed after the merger in two situations:
-
Two or more classifiers are in conflict and do not agree on the class to which the example belongs, despite the fact that each classifier individually indicates a high probability of membership (the existence or not of a defect). In this case, the result provided by the classifiers lacks precision; and by classifying it as a rejection, the merger calls out the need to seek expert advice.
-
The classifiers entirely agree on a dominant class absence and distribute the output probabilities in a uniform manner across several categories. In this case, the data reliability concerning the example is strongly called into question.

5. Results and Discussion

This section provides details on the used database, the M-SVM hyper parameter selection, the statistical evaluation parameters and the obtained results.

5.1. DGA Training, Validation and Evaluation Data

Dissolved Gas Analysis (DGA) is a technique developed to detect certain categories of incipient failures affecting oil-immersed equipment that cannot be easily detected by other conventional methods. This technique is considered one of the most used diagnostic and preventive monitoring tools today. The DGA database used in this study consists of 252 samples. Firstly, in the data collection part, 148 real mineral oil samples were collected from about 50 power transformers belonging to the Algerian utility “Sonelgaz Transport Electricity (STE)”. The DGA database was interpreted by an expert in the field and subsequently subdivided into eight fault categories: Partial Discharge (PD: 16 samples), Low Energy Discharge (LED: 24 samples), High Energy Discharge (HED: 12 samples), Thermal fault (t < 700 °C) (OH1: 16 samples), Thermal fault (t > 700 °C) (OH2: 16 samples), Cellulose Degradation (CD: 20 samples), Thermal (t > 700 °C) and Cellulose Degradation (OH2-CD: eight samples), Energy Discharge and Cellulose Degradation (ED-CD: 12 samples) plus a healthy samples category (N: 24 samples). Secondly, the lines that give the gas concentrations in an inequalities form (Table 4) are duplicated. A database containing 252 examples is therefore used for discriminator learning, validation and testing.

5.2. M-SVM Hyper Parameters Selection

In this study, four cross-validation levels were carried out. The database was firstly split randomly into four separate sets. Then, for each validation level and for each M-SVM model, two sets were used for training, one set was used for the best hyper parameter selection and one set was used for testing. Thus, all the database samples were considered during the training process and were tested during the evaluation process.
It is, of course, affirmative to validate an approach on all the samples than on a set portion (indeed, it could be that the results obtained with this portion are better than the results, which could be obtained with the other portions).
The appropriate hyper parameter selection is a crucial step to build an efficient classification model. For an M-SVM type model, the hyper parameters to be optimized depend on the selected kernel type (in addition to the penalty parameters). The MSVMpack includes three kernels (linear, polynomial and radial) with which the experts can implement. In this study, the radial kernel, which allowed bringing out a more adequate separation, was used. For an M-SVM model selection (with radial kernel), two regularization parameters must be optimized among a grid of values: the kernel parameter γ and the penalty parameter C . Thus, at each validation level, 255 combination pairs ( C ,   γ ) were evaluated with γ and C included respectively in the intervals 2 4 , 2 3 , , 2 10 and 2 12 , 2 11 , , 2 2 [37].
Then, each combination performance was reported by performing the learning on the training data and the evaluation on the validation data. The best combination was then retained for the final discrimination: the training was always done with the same learning data and the discrimination was based on the test data.

5.3. Statistical Evaluation Parameters

Seven statistical metrics are selected; all are deduced from a confusion matrix, Table 5, as follows:
-
Receiver Operating Characteristic (ROC) curve: is a graphical representation that illustrates the classifier performances (Sensitivity and 1-Specificity) variation at all probability thresholds.
-
Area Under the ROC Curve (AUROC-95\% CI): returns an overall performance estimate for all possible discrimination probability thresholds. A confidence interval is also associated to each measurement.
-
Sensitivity and Specificity: these are two elementary and complementary metrics for the expert-performance evaluation. In fact, they are based on all the confusion matrix elements and constitute the ROC curve base.
The sensitivity or True Positive Rate, TPR = TP TP + FN , predicts the expert’s ability to discern the positive population; while specificity or True Negative Rate, TNR = TN FP + TN , predicts the expert’s ability to discern the negative population.
-
Positive Predictive Value (PPV): returns the truly positive individual proportion within a population classified as positive, PPV = TP TP + FP .
-
Negative Predictive Value (NPV): returns the truly negative individual proportion within a population classified as negative, NPV = TN TN + FN .
-
False Negatives (FN): is a positive population for which the test is negative, F P = 1 T N R .
-
False Positives (FP): is a negative population for which the test is positive   F N = 1 T P R .

5.4. Obtained Results

The statistical performances of the proposed fusion method as well as those obtained by the implemented probabilistic direct M-SVM are shown in Table 6 and Table 7 (for comparison). Also, Figure 5 and Figure 6 reported the ROC curve evolution according to the five experts (for the NC class and the HED class). Focusing on these results, the following findings are reported:
The expert performances increase proportionally with the amount of area under the curve; indeed, the expert whose predictions are correct tends to have an AUROC which is close to 100%; an expert whose predictions are erroneous has an AUROC which is close to 0%; and an expert whose prediction is not informative (random) has an AUCROC which is close to 50%. It can be seen from Table 7 that the obtained AUROC sets after applying DS fusion are superior to those of the four direct M-SVM implemented separately with an average AUROC of 98.78% (p-value < 0.001). This can be explained by the fact that the ROC curve plotting is based on the generalization rate variation at different probability thresholds. For the proposed fusion model, all examples are assigned to their categories with a probability greater than 0.70, while the four direct M-SVM employ an all-or-nothing classification; thus, in the worst case, an M-SVM can assign an example to a class with a probability of 0.12, which adversely affects the area under the curve.
A confidence interval (significant at 95%) has also been associated with each AUROC to predict the area under the curve for the future independent samples (the generalization to a larger population); it was observed that the fusion AUROC upper limits can achieve an accuracy of 100% for all categories, contrary to M-SVM taken separately.
After merging, there is a significant distinction in the TPR and FN overall rates (which are considered to be correct classification rate estimators), just as the TNR and FP overall rates amelioration is underlined.
This is due, on the one hand, to the fact that the fusion takes into consideration each M-SVM advantage for a satisfactory result. Indeed, when analyzing the data reported in Table 6, it is found that it is difficult to privilege one model over another and that the M-SVM are in reality complementary (with slightly higher rates for CS-MSVM and slightly lower rates for WW-MSVM). For example, WW-MSVM tended to classify PD and OH2-CD categories well, while WW-MSVM generalizes the LED and HED classes well.
On the other hand, it should be remembered that a fusion outputs processing process was suggested, and a sample is considered to be well classified only if its membership probability is higher than 70%. Otherwise, it is reassigned to the reject category. Thus, attributing an example to the reject class instead of an arbitrary class favorably influences the results.
The FP and FN numbers (which are seen as indicators of overlapping rates between classes) increase consequently the model performances. In reliable diagnostic problems, it is wiser to abstain and leave the choice to the experts in the field than to make a diagnosis whose consequences are irreversible.
After fusion, the results also return an important distinction of the PPV and NPV rates. This still adds a sure value to the proposed approach effectiveness, since these two statistical parameters are seen as being confidence indicators in the obtained results. For example, let us consider the proposed fusion model. The HED class PPV rate is 95.24%, which means that 95.24% of samples classified as HED are really HED, and that 4.76% of the samples are wrongly classified as HED; the HED class NPV rate is 98.65%, which means that 98.65% of samples classified as not HED are really not HED, and that 1.35% of the samples are wrongly classified as not HED.
According to the last column of Table 6, one can see that the rejected examples number after merging is eight examples out of 252, which represents a low percentage of the order of 3% (despite a very severe decision-making threshold). These results justify on the one hand that in 97% of cases the classifiers agree on the class to which the example belongs (which adds an important confidence degree to the results provided by our approach). On the other hand, the results obtained justify the dominant class presence and therefore validate the data reliability (please see Section 4.2.3).
At the end, a temporal analysis (study of the execution time) has been carried out. This proves the necessity to determine the possibility of introducing the proposed model into embedded processes. The analysis results are summarized in Table 8. Quite reasonable execution times for the four M-SVM and for the proposed fusion model can be observed.
However, a more important learning time was noticed. For learning diagnostic systems, the training phase performed offline and the time it may take is not important. What counts is the system response time in the operating phase. For the present study, the response time is almost instantaneous, which once again proves the efficiency of the proposed fusion approach.

5.5. Comparison with Previous Work Results

It is necessary to consider a comparison between the obtained results with the proposed approach and those obtained by other methods in the literature. In this subsection, a comparative study is proposed according to three parameters: the considered classes number, the used samples number and the overall generalization rate. For this, a few contributions whose treated problem is as similar as possible to that of the paper have been selected.
The authors proposed in their studies [7] a DGA diagnostic approach based on a Clustering Technique, a Cumulative Voting merged Technique (CVT) and a revisited k-nearest neighbors (KNN) algorithm. Xie and coauthors [12] proposed an approach based on relief algorithm for features selection, kernel Linear Discriminant Analysis (LDA) for features redundancy elimination and Relevance Vector machines (RVM) as discriminator. The authors in [14] proposed a system, namely: 2-ADOPT. This system was based on two versions of PSO approach (features selection and ensemble classifiers selection) and DS combination rule. The authors of [25] presented in their studies, an approach based on kernel extreme learning machine (KELM) optimized with the Harris hawks optimization (HHO) algorithm. The authors proposed in their research [27] a comparative study of an MLP network performance in power transformers diagnosis according to five different data sets. The study [30] proposed a new approach based on GA for features reduction and a krill herd (IKH) algorithm optimized SVM for generalization. Abdo et al. exposed in [31] a method based on the application of a normalized data set (C-set) approach, an unsupervised Fuzzy C-means (FCM) clustering algorithm and a SVM. The study [32] proposes a multiple Probabilistic output Algorithms (PA) fusion (RVM, SVM and MLP) by the DS law; the hyper parameters of PA algorithms have been optimized by the PSO approach. From Table 9 it is noted that:
-
The categories number taken into account in this research is high compared to all other studies. Likewise, the present work offers a generalization rate which surpasses the obtained rates from other studies, except for the works [12,14,26]. This can be justified by the fact that the authors considered fewer classes and test examples.
-
The samples number is comparable to other studies with the exception of work [7]. This is justified by the fact that the latter considers a KNN type classifier, which necessarily needs a relatively large training sample for efficient performance. Contrary to this work, M-SVM-type discriminators are implemented, one of the advantages of which is to offer high generalization rates from a reduced learning sample.
Especially since all the statistical tests that take into consideration the sample size crown with expressive areas under the curve (AUROC) and are statistically significant (p-values < 0.001), this testifies that the samples number is sufficient to validate the obtained generalization results.
Moreover, in these previous works, the authors took into account during the test only a limited part of the overall samples, while in the present study the samples were fully tested through the cross-validation application.

6. Conclusions

Accurate prediction and diagnosis of possible incipient faults is an indispensable part of power delivery to end-users, is conducive to sustainable development and has practical significance and demands. In this contribution, a power transformer intelligent diagnostic aid system based on real DGA data obtained from power transformers of a western Algerian electricity and gas company is proposed.
Six descriptor sets were firstly constructed and presented as Kohonen map inputs, for data size reduction and information completeness preservation. DS fusion was afterwards applied to the four Generic M-SVM model outputs for accurate and reliable decision-making. Finally, an output post-processing process was applied to overcome the contradictory evidence problem. It is worthwhile mentioning that the four direct M-SVM models implemented in this study offer theoretically the same performances, and each model has their own advantages and their own disadvantages; that the performance of MSV M was improved thanks to the proposed descriptor reconstruction approach; and that the decision making was strengthened through the application of DS fusion and the introduction of the rejection class.
Eight defect categories were considered in addition to the healthy samples class, and the obtained results highlighted the proposed approach effectiveness in detecting incipient failures. Indeed, the latter archieved an AUROC and sensitivity percentage of 98.78% and 95.19% (p-value < 0.001), respectively. Comparisons between the obtained results with different approaches reported in the literature [7,12,14,25,26,30,31] indicate a higher category number. Only a limited part of the overall samples was taken into account, while the samples were fully tested in the present contribution through the cross-validation application. In addition, the present work offers a generalization rate which surpasses the obtained rates from most studies.
Consequently, this study can be seen as very promising and useful for transformer owners. The results presented can be viewed as a benchmark and a challenge for further research. In particular, the authors’ future research will focus on transformer anomaly prevention and exploration of other DGA data manipulation techniques such as ontologies.

Author Contributions

Conceptualization, M.H.; methodology, M.H.; software, M.H.; validation, M.H. and M.E.A.S.; formal analysis, M.H., M.E.A.S., F.M., M.B. and I.F; investigation, M.H.; resources, M.H.; data curation, M.H.; writing—original draft preparation, M.H.; writing—review and editing, M.H., I.F., F.M., M.B. and M.E.A.S.; supervision, M.B., I.F. and F.M.; project administration, M.B.; funding acquisition, M.H. and M.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been supported by the PRFU-C00L07EP31022010001 project and the General Directorate for Scientific Research and Technological Development (DGRSDT).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The database presented in this study includes real oil samples from around fifty power transformers belonging to the western Algerian electricity and gas company Sonelgaz Transport Electricity (STE). The considered equipment is of various ages and present a transformation ratio whose order is included in [60/10, 60/30, 220/60, 400/220] kV, a frequency of 50 Hz and a nominal power that varies between 10 and 120 MVA.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DGADissolved Gas Analysis
M-SVMDirect Multiclass Support Vector Machines
IAArtificial Intelligence
D-TDuval Triangle
R-RRogers Ratios
D-RDornenburg Ratios
IEC-RIEC Ratios
KGKey Gases
PTDPower Transformers Diagnosis
G-PGases Percentage
TGTotal combustible and non-combustible Gases sum
TSTotal Sum
KSOMKohonen Self-Organizing Maps
GM-SVMGeneric M-SVM Model
WWWeston and Watkins model
CSCrammer and Singer model
LLWLee et al. model
M-SVM2Quadratic Loss Multi-Class Support Vector Machine
DSDempster–Shafer fusion
PSOPractical Swarm Optimization
GAGenetic Algorithm
H2Hydrogen
O2Oxygen
N2Nitrogen
COCarbon Monoxide
CH4Methane
CO2Carbon Dioxide
C2H6Ethane
C2H4Ethylene
C2H2Ethylene
VQVector Quantification process
MSVMpackMSVM software package
BelBelief function
PlPlausibility function
MMass function
ROCReceiver Operating Characteristic curve
AUROCArea Under the ROC Curve
TPRTrue Positive Rate
TNRTrue Negative Rate
PPVPositive Predictive Value
NPVNegative Predictive Value
FNFalse Negatives
FPFalse Positives
TPTrue Positives
TNTrue Negatives
STESonelgaz Transport Electricity
PDPartial Discharge
LEDLow Energy Discharge
HEDHigh Energy Discharge
OH1Thermal fault (t < 700 °C)
OH2Thermal fault (t > 700 °C)
CDCellulose Degradation
OH2-CDThermal (t > 700 °C) and Cellulose Degradation
ED-CDEnergy Discharge and Cellulose Degradation
NHealthy samples
C-SetNormalized data set approach
FCMFuzzy C-means clustering algorithm
KELMKernel extreme learning machine
HHOHarris-Hawks-optimization algorithm
CVTCumulative Voting Technique merged
KNNk-Nearest Neighbors algorithm
LDALinear Discriminant Analysis
RVMRelevance Vector machines
MLPMultilayer Perceptron
PAProbabilistic output Algorithms
IKHKrill herd algorithm

References

  1. Muniz, R.N.; da Costa Júnior, C.T.; Buratto, W.G.; Nied, A.; González, G.V. The Sustainability Concept: A Review Focusing on Energy. Sustainability 2023, 15, 14049. [Google Scholar] [CrossRef]
  2. Zhang, N.; Sun, Q.; Yang, L.; Li, Y. All Authors Event-Triggered Distributed Hybrid Control Scheme for the Integrated Energy System. IEEE Trans. Ind. Inform. 2022, 18, 835–846. [Google Scholar] [CrossRef]
  3. Liu, Y.; Li, Z.; Wu, Q.H.; Zhang, H. Real-time Dispatchable Region of Renewable Generation Constrained by Reactive Power and Voltage Profiles in AC Power Networks. CSEE J. Power Energy Syst. 2020, 6, 681–692. [Google Scholar] [CrossRef]
  4. Muangpratoom, P.; Suriyasakulpong, C.; Maneerot, S.; Vittayakorn, W.; Pattanadech, N. Experimental Study of the Electrical and Physiochemical Properties of Different Types of Crude Palm Oils as Dielectric Insulating Fluids in Transformers. Sustainability 2023, 15, 14269. [Google Scholar] [CrossRef]
  5. Shufali, A.W.; Ankur, S.R.; Shiraz, S.; Obaidur, R.; Shaheen, P.; Shakeb, A.K. Advances in DGA based condition monitoring of transformers: A review. Renew. Sustain. Energy Rev. 2021, 149, 111347. [Google Scholar] [CrossRef]
  6. Abbasi, A.R. Fault detection and diagnosis in power transformers: A comprehensive review and classification of publications and methods. Electr. Power Syst. Res. 2022, 209, 107990. [Google Scholar] [CrossRef]
  7. Mominul, M.I.; Gareth, L.; Sujeewa, N.H. A nearest neighbor clustering approach for incipient fault diagnosis of power transformers. Electr. Eng. 2017, 99, 1109–1119. [Google Scholar] [CrossRef]
  8. Irungu, G.K.; Akumu, A.O.; Munda, A.O. A new fault diagnostic technique in oil-filled electrical equipment; the dual of Duval triangle. IEEE Trans. Dielectr. Electr. Insul. 2016, 23, 3405–3410. [Google Scholar] [CrossRef]
  9. Rogers, R.R. IEEE and IEC codes to interpret incipient faults in transformers, using gas in oil analysis. IEEE Trans. Electr. Insul. 1978, EL-13, 349–354. [Google Scholar] [CrossRef]
  10. Khan, S.A.; Equbal, M.D.; Islam, T.A. Comprehensive comparative study of DGA based ANFIS models. IEEE Trans. Dielectr. Electr. Insul. 2015, 22, 590–596. [Google Scholar] [CrossRef]
  11. Bakar, N.A.; Abu Siada, A.; Islam, S. A review of dissolved gas analysis measurement and interpretation techniques. IEEE Electr. Insul. Mag. 2014, 30, 39–49. [Google Scholar] [CrossRef]
  12. Xie, L.; Zhao, Y.; Yan, K.; Shao, M.; Liu, W.; Lui, D. Interpretation of DGA for Transformer Fault Diagnosis with Step-by-step feature selection and SCA-RVM. In Proceedings of the IEEE 16th Conference on Industrial Electronics and Applications, Chengdu, China, 1–4 August 2021. [Google Scholar] [CrossRef]
  13. Kari, T.; Gao, W.; Zhao, D.; Abiderexiti, K.; Mo, W.; Wang, Y.; Luan, L. Hybrid feature selection approach for power transformer fault diagnosis based on support vector machine and genetic algorithm. Inst. Eng. Technol. 2018, 12, 5672–5680. [Google Scholar] [CrossRef]
  14. Peimankar, A.; Weddell, S.J.; Thahirah, J.; Lapthorn, A.C. Evolutionary Multi-Objective Fault Diagnosis of Power Transformers. Swarm Evol. Comput. 2017, 36, 62–75. [Google Scholar] [CrossRef]
  15. Kohonen, T. Self-Organizing Map; Springer: Berlin/Heidelberg, Germany, 1997; Volume 30. [Google Scholar] [CrossRef]
  16. Guermeur, Y. A generic model of multi-class support vector machine. J. Intell. Inf. Database Syst. 2012, 6, 555–577. Available online: https://dl.acm.org/doi/abs/10.1504/IJIIDS.2012.050094 (accessed on 2 October 2012). [CrossRef]
  17. Weston, J.; Watkins, C. Multi-Class Support Vector Machines; Technical Report CSD-TR-98-04; Royal Holloway, University of London, Department of Computer Science: Egham, UK, 1998. [Google Scholar]
  18. Crammer, K.; Singer, Y. On the algorithmic implementation of multiclass kernel based vector machines. J. Mach. Learn. Res. 2001, 2, 265–292. Available online: https://dl.acm.org/doi/10.5555/944790.944813 (accessed on 2 December 2001).
  19. Lee, Y.; Lin, Y.; Wahba, G. Multicategory support vector machines: Theory and application to the classification of microarray data and satellite radiance data. J. Am. Stat. Assoc. 2004, 99, 67–81. [Google Scholar] [CrossRef]
  20. Guermeur, Y.; Monfrini, E. A quadratic loss multi-class svm for which a radius margin bound applies. Informatica 2011, 22, 73–96. Available online: https://dl.acm.org/doi/10.5555/2019505.2019511 (accessed on 2 January 2011). [CrossRef]
  21. Dempster, A.P. A generalisation of Bayesian inference. J. R. Stat. Soc. 1968, 2, 205–247. [Google Scholar] [CrossRef]
  22. Shafer, G. A Mathematical Theory of Evidence; Princeton University Press: Princeton, NJ, USA, 1976. [Google Scholar] [CrossRef]
  23. Senoussaoui, M.A.; Brahami, M.; Fofana, I. Combining and comparing various machine learning algorithms to improve dissolved gas analysis interpretation. Inst. Eng. Technol. 2018, 12, 3673–3679. [Google Scholar] [CrossRef]
  24. Bacha, K.; Souahlia, S.; Gossa, M. Power transformer fault diagnosis based on dissolved gas analysis by support vector machine. Electr. Power Syst. Res. 2012, 83, 73–79. [Google Scholar] [CrossRef]
  25. Han, X.; Ma, S.; Shi, Z.; An, G.; Du, Z.; Zhao, C. A Novel Power Transformer Fault Diagnosis Model Based on Harris-Hawks-Optimization Algorithm Optimized Kernel Extreme Learning Machine. J. Electr. Eng. Technol. 2022, 17, 1993–2001. [Google Scholar] [CrossRef]
  26. Guardado, J.L.; Naredo, J.L.; Moreno, P.; Fuerte, C.R. A Comparative Study of Neural Network Efficiency in Power Transformers Diagnosis Using Dissolved Gas Analysis. IEEE Trans. Power Deliv. 2001, 16, 643–647. [Google Scholar] [CrossRef]
  27. Senoussaoui, M.A.; Brahami, M.; Bousmaha, I.S. Improved Gas Ratios Models for DGA Interpretation Using Artificial Neural Networks; Association of Computer Electronics and Electrical Engineers: New York City, NY, USA, 2019; pp. 167–175. [Google Scholar]
  28. Mansouri, D.E.K.; Benabdeslem, K. Towards Multi-label Feature Selection by Instance and Label Selections. In Advances in Knowledge Discovery and Data Mining; Springer: Berlin/Heidelberg, Germany, 2021; pp. 233–244. [Google Scholar] [CrossRef]
  29. Zhang, P.; Liu, G.; Gao, W.; Song, J. Multi-label feature selection considering label supplementation. Pattern Recognit. 2021, 120, 108137. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Li, X.; Zheng, H.; Yao, H.; Liu, J.; Zhang, C.; Peng, H.; Jiao, J. A Fault Diagnosis Model of Power Transformers Based on Dissolved Gas Analysis Features Selection and Improved Krill Herd Algorithm Optimized Support Vector Machine. IEEE Access 2019, 7, 102803–102811. [Google Scholar] [CrossRef]
  31. Abdo, A.; Liu, H.; Zhang, H.; Guo, J.; Li, Q. A new model of faults classification in power transformers based on data optimization method. Electr. Power Syst. Res. 2021, 200, 107446. [Google Scholar] [CrossRef]
  32. Hua, Y.; Sun, Y.; Xu, G.; Sun, S.; Wang, E.; Pang, Y. A fault diagnostic method for oil-immersed transformer based on multiple probabilistic output algorithms and improved DS evidence theory. Int. J. Electr. Power Energy Syst. 2022, 137, 107828. [Google Scholar] [CrossRef]
  33. Vapnik, V.N. The Nature of Statistical Learning Theory; Springer: New York City, NY, USA, 1995. [Google Scholar]
  34. Berlinet, A.; Thomas-Agnan, C. Reproducing Kernel Hilbert Spaces in Probability and Statistics; Kluwer Academic Publishers: Boston, MA, USA, 2004. [Google Scholar] [CrossRef]
  35. Lauer, F.; Guermeur, Y. MSVMpack: A multi-class support vector machine package. J. Mach. Learn. Res. 2012, 12, 2269–2272. Available online: https://dl.acm.org/doi/10.5555/1953048.2021073 (accessed on 2 July 2011).
  36. Platt, J.C. Probabilities for SV machines. In Advances in Large Margin Classifiers; Smola, A.J., Bartlett, P.L., Schölkopf, B., Schuurmans, D., Eds.; The MIT Press: Cambridge, MA, USA, 1999; Chapter 5; pp. 61–73. [Google Scholar]
  37. Hsu, C.W.; Lin, C.J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed PTD approach.
Figure 1. Block diagram of the proposed PTD approach.
Sustainability 15 15453 g001
Figure 2. KSOM weight positions at the training start.
Figure 2. KSOM weight positions at the training start.
Sustainability 15 15453 g002
Figure 3. KSOM weight positions at the end of the training.
Figure 3. KSOM weight positions at the end of the training.
Sustainability 15 15453 g003
Figure 4. Parameter distribution around the clusters for the six selected configurations.
Figure 4. Parameter distribution around the clusters for the six selected configurations.
Sustainability 15 15453 g004
Figure 5. ROC curve evolution according to the five experts (class NC).
Figure 5. ROC curve evolution according to the five experts (class NC).
Sustainability 15 15453 g005
Figure 6. ROC curve evolution according to the five experts (class HED).
Figure 6. ROC curve evolution according to the five experts (class HED).
Sustainability 15 15453 g006
Table 1. The six retained configuration classification rates.
Table 1. The six retained configuration classification rates.
Map SizeGeneralization Rate (%)Map SizeGeneralization Rate (%)
2   × 2 Map82.94 4   × 3 Map88.10
3   × 2 Map85.32 4   × 4 Map87.30
3   × 3 Map88.10 4   × 5 Map87.30
Table 2. The new descriptors given by the card after learning.
Table 2. The new descriptors given by the card after learning.
The Nine New DescriptorsInitial Descriptors They Represent
D1 H 2 , H 2 TG
D2 C 2 H 2 C 2 H 4
D3 CH 4 TS , C 2 H 2 TS , C 2 H 4 TS
D4 C 2 H 6 CH 4
D5 C 2 H 2
D6 CH 4 H 2
D7 O 2 ,   N 2 ,   CO ,   CO 2 ,   CO TG  
D8 CO 2 TG
D9 CH 4 ,   C 2 H 4 ,   C 2 H 6 ,   O 2 TG ,   N 2 TG ,   CH 4 TG ,   C 2 H 4 TG ,   C 2 H 6 TG ,   C 2 H 2 TG ,   C 2 H 4 C 2 H 6 ,   C 2 H 2 CH 4 ,   C 2 H 6 C 2 H 2
Table 3. Specifications of the four M-SVM as instances of the GM-SVM model.
Table 3. Specifications of the four M-SVM as instances of the GM-SVM model.
M-SVMMpK1K2K3
WW Model I Q m Z m 1110
CS Model 1 Q 1 I Q m Z m 1111
LLW Model I Q m Z m 10 1 Q 1 0
M-SVM2 M 2 20 1 Q 1 0
Table 4. Duplicating process for an ED-CD defect type.
Table 4. Duplicating process for an ED-CD defect type.
H2COO2N2CO2CH4C2H2C2H4C2H6
12784716,97389,51737261767<13
12784716,97389,517372617670.23
12784716,97389,517372617670.43
12784716,97389,517372617670.63
12784716,97389,517372617670.83
12784716,97389,5173726176713
Table 5. The confusion matrix.
Table 5. The confusion matrix.
Predicted
Considered categoryOther categories
RealityConsidered categoryTrue positive (TP)False negative (FN)
Other categoriesFalse positive (FP)True negative (TN)
Table 6. M-SVM and proposed DS-fusion model performances (all p-value < 0.001).
Table 6. M-SVM and proposed DS-fusion model performances (all p-value < 0.001).
ClassStatistical Parameters (%)WW-MSVMLLW-MSVMMSVM2CS-MSVMFusion
NAUROC0.938
[0.883–0.993]
0.963
[0.927–0.999]
0.984
[0.969–0.999]
0.991
[0.980–1.000]
0.994
[0.985–1.000]
Sensitivity87.590.6293.75100100 (1 Reject)
Specificity97.7398.1898.1898.1898.58
False positive2.271.821.821.821.42
False negative12.59.386.2500
VPP84.8587.8888.2488.8991.18
VPN98.1798.6399.08100100
PDAUROC0.956
[0.901–1.000]
0.970
[0.935–1.000]
0.940
[0.896–1.000]
0.975
[0.940–1.000]
0.985
[0.960–1.000]
Sensitivity90.6287.5087.5090.6293.54 (1 Reject)
Specificity98.6398.6399.09100100
False positive1.371.370.9100
False negative9.3812.5012.509.386.46
VPP90.6390.3293.33100100
VPN98.6498.1998.6498.6599.53
LEDAUROC0.966
[0.937–0.996]
0.959
[0.925–0.993]
0.968
[0.941–0.996]
0.968
[0.938–0.998]
0.977
[0.949–1.000]
Sensitivity87.593.7590.6290.6290.32 (1 Reject)
Specificity98.6399.1098.1898.1899.05
False positive1.370.901.822.820.95
False negative12.56.259.389.389.68
VPP90.3293.7587.8887.8893.33
VPN98.1899.0998.6398.6398.59
HEDAUROC
(95% CI)
0.968
[0.925–1.000]
0.971
[0.943–1.000]
0.938
[0.867–1.000]
0.958
[0.916–1.000]
0.986
[0.970–1.000]
Sensitivity83.3387.583.3379.1686.96 (1 Reject)
Specificity98.2498.2498.6899.1299.54
False positive1.761.761.320.880.46
False negative16.6712.516.6720.8413.04
VPP83.338486.9690.4795.24
VPN98.2598.6898.2597.8498.65
OH1AUROC (95% CI)0.962
[0.931–0.993]
0.971
[0.934–1.000]
1.000
[1.000–1.000]
0.996
[0.989–1.000]
0.999
[0.996–1.000]
Sensitivity89.2892.8510096.4296.42
Specificity98.6610010010099.53
False positive1.340000.47
False negative10.727.1503.583.58
VPP89.2910010010096.43
VPN98.6699.1210099.5699.53
OH2AUROC
(95% CI)
0.965
[0.935–0.995]
0.966
[0.928–1.000]
0.981
[0.955–1.000]
0.994
[0.955–1.000]
0.999
[0.996–1.000]
Sensitivity82.1489.2896.4296.42100 (2 Rejects)
Specificity99.5598.6699.1199.1199.53
False positive0.451.340.890.890.47
False negative17.8610.723.583.580
VPP95.8389.2993.1093.1096.30
VPN97.8198.6699.5599.55100
CDAUROC
(95% CI)
0.981
[0.949–1.000]
1.000
[1.000–1.000]
0.996
[0.989–1.000]
1.000
[1.000–1.000]
1.000
[1.000–1.000]
Sensitivity96.8710096.87100100
Specificity99.55100100100100
False positive0.450000
False negative3.1303.1300
VPP96.88100100100100
VPN99.5510099.55100100
OH2-CDAUROC
(95% CI)
0.975
[0.950–0.999]
0.930
[0.853–1.000]
0.936
[0.851–1.000]
0.966
[0.920–1.000]
0.978
[0.945–1.000]
Sensitivity9580859089.47
Specificity97.8499.1499.1499.14100
False positive2.160.860.860.860
False negative520151010.53
VPP79.1788.8989.4790100
VPN99.5698.2998.7199.1499.12
ED-CDAUROC
(95% CI)
0.965
[0.920–1.000]
0.966
[0.936–0.996]
0.932
[0.847–1.000]
0.959
[0.916–1.000]
0.972
[0.926–1.000]
Sensitivity83.3391.6687.5087.50100 (2 Rejects)
Specificity98.6897.8098.6898.2598.64
False positive1.322.21.321.751.36
False negative16.678.3412.5012.500
VPP86.9681.48848488
VPN98.2599.1198.6898.68100
Table 7. Average performances of M-SVM and DS-fusion models.
Table 7. Average performances of M-SVM and DS-fusion models.
WW-MSVMLLW-MSVMMSVM2CS-MSVMFusion
AUROC (95% CI)
(p-value < 0.001)
96.4096.6296.3997.8698.78
Sensitivity88.3990.3591.2292.3095.19
Specificity98.6198.8699.0199.1099.43
False positive1.391.140.990.900.57
False negative11.619.658.717.74.81
VPP88.5890.6291.4492.7095.61
VPN98.5698.8699.0199.1199.49
Table 8. Average prediction times of M-SVM and DS-fusion models.
Table 8. Average prediction times of M-SVM and DS-fusion models.
WW-MSVMLLW-MSVMMSVM2CS-MSVMFusion
Prediction time (s)0.018
±0.004
0.015
±0.002
0.010
±0.003
0.014
±0.004
0.085
±0.009
Table 9. Previous work results.
Table 9. Previous work results.
StudySample CountCategories NumberClassification Rates (%)
CT-CVT-KNN [7]396793
SCA-RVM [12]135697.07
2-ADOPT [14]101497.94
KELM-HHO [25]118588
MLP [27]1026MLP-MRRMLP-RRMLP-IECRMLP-KGMLP-DR
87.8890.9193.9410090.91
AG-IKH-SVM [30]113585.71
Cset-FCM-SVM [31]177686.11
PA-DS [32]1566RVMSVMMLPDS-Fusion
87.885.382.689.1
Proposed Approach2529WWLLWMSVM2CSFusion
88.3990.3591.2292.3095.19
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hendel, M.; Meghnefi, F.; Senoussaoui, M.E.A.; Fofana, I.; Brahami, M. Using Generic Direct M-SVM Model Improved by Kohonen Map and Dempster–Shafer Theory to Enhance Power Transformers Diagnostic. Sustainability 2023, 15, 15453. https://doi.org/10.3390/su152115453

AMA Style

Hendel M, Meghnefi F, Senoussaoui MEA, Fofana I, Brahami M. Using Generic Direct M-SVM Model Improved by Kohonen Map and Dempster–Shafer Theory to Enhance Power Transformers Diagnostic. Sustainability. 2023; 15(21):15453. https://doi.org/10.3390/su152115453

Chicago/Turabian Style

Hendel, Mounia, Fethi Meghnefi, Mohamed El Amine Senoussaoui, Issouf Fofana, and Mostefa Brahami. 2023. "Using Generic Direct M-SVM Model Improved by Kohonen Map and Dempster–Shafer Theory to Enhance Power Transformers Diagnostic" Sustainability 15, no. 21: 15453. https://doi.org/10.3390/su152115453

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop