1. Introduction
Power transformers are critical components of electrical systems, facilitating the efficient transmission of electricity at various voltage levels [
1,
2]. Their primary function is to transfer electrical energy between circuits while maintaining voltage stability, which is crucial for minimizing losses over long distances. Power systems are evolving with distributed energy resources and smart grids. These changes significantly increase operational demands on transformers, making advanced monitoring and diagnostics essential for reliability and safety. For instance, intensive monitoring can enable aging transformers to remain in service longer, as discussed in [
3,
4] underscoring the need for a comprehensive condition monitoring framework. The implementation of hybrid AI models for health management demonstrates the need for innovative technologies. These models help enhance transformer resilience and performance amid challenges like power flow fluctuations and potential failures.
Since most transformers in use today were put in before 1980, most of them are either nearing or have already outlived their design life [
5,
6]. Given the potentially disastrous consequences of an in-service transformer failure, utilities and other power network stakeholders are at serious risk. The windings of a transformer experience magnetic forces because of the mechanical forces it encounters during faults, switching transients, and other system events. The transformer may experience winding deformation if these forces are greater than its tolerance [
7,
8]. Age-induced insulation degradation reduces clamping pressure, contributing to winding deformation and mechanical damage in transformers.
1.1. Contribution
This study makes significant contributions to the field of transformer diagnostics by addressing gaps in fault detection methodologies.
Develops a machine learning-based framework for classifying transformer faults, including winding deformation, partial discharges, and insulation breakdown;
Incorporates numerical indices and statistical features, enhancing classification accuracy across fault types.
Validates fault detection methods using both simulated and real-world FRA data, ensuring practical applicability.
Demonstrates frequency-independent fault classification accuracy, reducing reliance on specific frequency bands.
1.2. Novelty
This research introduces innovative techniques and methodologies that address limitations in prior transformer fault detection studies.
Employs advanced machine learning models (e.g., Decision Trees, Subspace KNN) to achieve 80–100% fault classification accuracy.
Expands fault detection to include complex types like axial displacement and contamination, which are underexplored in prior studies.
Introduces refined feature extraction techniques to enhance sensitivity and reduce false negatives.
Establishes a comprehensive annotated FRA dataset for diverse fault classification and future research.
The article is organized into distinct sections with Abstract summarizing the study’s focus on transformer fault detection using FRA and machine learning, Introduction (
Section 1) discussing the importance of transformers, challenges like aging infrastructure, and the need for advanced diagnostics;
Section 2 focuses on FRA and Frequency Bands explaining their role in fault detection and machine learning integration;
Section 3: Methodology detailing data preparation, transformer simulation, and classifier training;
Section 4: Results and Discussion presenting performance of various machine learning models in detecting faults and analyzing their accuracy;
Section 5 and
Section 6: Conclusion summarizing key findings, highlighting model accuracy (80–100%), and suggesting future improvements.
2. FRA, Frequency Bands, and Their Impact on Fault Detection
The influence of frequency bands on fault detection in power transformers is profound, as different frequency ranges can reveal distinct diagnostic information about the condition of the equipment. By analyzing resonant frequencies generated by various electromagnetic disturbances, practitioners can identify specific fault signatures that correlate with malfunctions. For instance, low-frequency signals are often associated with electrical anomalies, while higher-frequency signals may indicate physical deterioration such as insulation breakdowns. The integration of advanced machine learning models can further enhance the sensitivity of fault detection systems by leveraging these frequency-specific characteristics. Recent studies underscore the importance of this multidimensional approach, highlighting how diverse algorithms can optimize fault diagnosis across multiple frequency bands. This alignment of machine learning techniques with dedicated frequency analysis creates a more accurate and responsive diagnostic framework, ultimately leading to improved maintenance strategies and reduced downtime for power transformers [
9,
10]. In an offline test known as FRA, a sweep frequency and low voltage are injected into one transformer winding terminal, and the response across the other terminal is measured in relation to the tank [
11]. The FRA response, expressed in decibels (dB), is the ratio of the source voltage, Vi, to the magnitude of the response signal, Vo, as a function of frequency [
12]. The FRA signature is thought of as the transformer’s fingerprint, which can be compared to its earlier signatures to detect any alterations and, consequently, any mechanical deformation. However, transformer fingerprints are not often available, especially for older transformers. Other comparison techniques, including comparing sister transformer signatures or comparing phases, can be used in this situation. FRA signature interpretation is still a highly specialized field that requires skilled professionals to identify the type and likely location of the fault, even if the testing technique has become relatively straightforward due to the development of FRA test equipment. Particularly for young and inexperienced power engineers, this is not always a straightforward task. This work presents an effort to use sensitivity analysis to quantify and categorize transformer FRA signatures. A transformer distributed parameters model is subjected to a thorough simulation analysis to examine and categorize the effects of internal defects on the FRA signature. By suitably altering the transformer distributed parameters, mechanical failures such axial displacement, radial buckling, disk space variation, loss of clamping, bushing, and leakage faults are simulated.
3. Methodology
This section presents the methodology employed for fault detection in power transformers using Frequency Response Analysis (FRA) and machine learning algorithms. The process encompasses data preparation, transformer grouping, and classifier training using various machine learning models. The structure is designed to ensure a logical flow, facilitating a clear understanding of the techniques and their application.
3.1. Transformer Grouping and Fault Types
To conduct fault analysis and classification, transformers were grouped into two categories based on their characteristics and operational status.
3.1.1. Group X Transformers
Group X consists of model transformers intentionally modified to exhibit specific flaws. These models share similar RLC circuit structures with actual transformers but are limited to laboratory research applications. The model transformers’ windings allow for varied connections, enabling simulation of multiple defect scenarios to generate a comprehensive database.
3.1.2. Group Y Transformers
Group Y includes operational transformers prone to faults. Faults are identified by analyzing FRA traces before physically inspecting the transformers. The frequency response data provide insights into potential faults within the RLC circuitry, enabling pre-emptive fault identification.
3.1.3. Fault Classification
Transformer faults were categorized into distinct classes for analysis:
Class A: Thermal aging
Class B: Electrical stress
Class C: Moisture ingress
Class D: Contamination-related faults
3.2. Machine Learning Algorithms for Fault Detection
Various machine learning models were employed to classify transformer faults based on FRA data. The process involved data preparation, feature extraction, classifier training, performance evaluation, and validation.
3.2.1. Data Preparation and Feature Extraction
Data preparation involved cleaning and standardizing raw FRA data. Key features were extracted to highlight patterns relevant to fault classification. Numerical and statistical indices were used to enhance model training and validation.
3.2.2. Classifier Training and Testing
The classifiers were trained on prepared data to learn fault patterns and subsequently tested with separate datasets to evaluate generalization capability.
3.2.3. Decision Tree Algorithms
The six primary processes involved are data preparation, feature generation, classifier training, classifier testing, performance analysis, and validation using specific unseen cases and data. Unsatisfactory performance is often attributed to issues such as inaccurate data preparation, imprecise feature generation, or a lack of diverse traces across classes. To address these challenges, the Decision Tree algorithm suggests redesigning the process stages to enhance overall performance [
25,
26,
27,
28,
29,
30,
31,
32,
33]. The process begins with data preparation, where raw data are cleaned, standardized, and properly formatted for use. This step is critical, as poor-quality data can lead to inaccurate results. Next comes feature generation, which involves extracting meaningful features from the raw data to highlight key patterns. If this step is not performed accurately, the model’s learning capability may be compromised. During the classifier training phase, the Decision Tree algorithm is trained on the prepared data, learning to categorize information based on the selected features. Once trained, the model undergoes classifier testing with a separate dataset to assess its generalization ability, ensuring it can make accurate predictions on unseen data. Performance analysis follows, evaluating the model using metrics such as accuracy, precision, and recall identifying areas for improvement. The validation with unseen cases tests the model in real-world scenarios, confirming its reliability and ability to generalize effectively. In
Figure 1, the diagram illustrates the suggested process for designing and optimizing a Decision Tree (DT) algorithm. The process highlights key stages such as data preparation, feature generation, and ensuring adequate variability in traces across classes, all of which are essential for achieving optimal model performance.
Figure 1 outlines the key stages in developing and optimizing a Decision Tree model for transformer fault detection. The process emphasizes iterative improvements, from data preparation and feature extraction to classifier training and performance evaluation. Feedback loops are incorporated to refine the model based on evaluation results, ensuring robust fault classification.
Unsatisfactory performance can often result from issues at any of these stages, such as inaccurate data preparation, imprecise feature generation, or insufficient variability in traces across classes. In such cases, the Decision Tree algorithm recommends redesigning or adjusting the process stages, such as refining the features, revisiting the data preprocessing steps, or introducing more varied training cases to improve the model’s ability to generalize and perform well.
3.2.4. K-Nearest Neighbors (KNN) Algorithm
The classification performance is stable and the KNN approach is straightforward and simple to apply. However, when the sample is imbalanced and the sample dimension is too large, the algorithm performs poorly in terms of operational efficiency and classification impact. Numerous researchers have made improvements to their data or algorithms in this area [
25,
26,
27,
28,
29,
30,
31,
32,
33].
Figure 2 illustrates the KNN-based Transformer failure diagnosis procedure. In this process, mapping Transformer fault samples using the NCA model’s output findings (metric matrix) is referred to as NCA transformation. In the process, the enhanced NCA method suggested in this research is applied.
Figure 2 depicts the KNN-based diagnostic process for transformer fault detection. FRA samples are mapped using the NCA model, followed by feature transformation and distance-based classification. The enhanced NCA method ensures accurate mapping and robust fault detection. Annotations highlight the role of each stage in optimizing the diagnostic process.
3.2.5. Performance Metrics
The following metrics were used to evaluate classifier performance:
True Positive Rates (TPR) where both the Transformer’s current state of health and its anticipated state of health are positive;
Inaccurate Rates Negative (FNR) where the Transformer is in good condition (positive), a fault is anticipated (negative);
False Discovery Rates (FDR), It is expected that the Transformer will be defective (negative), even though it is faulty (negative);
Positive Predictive Values (PPV), in this case both the Transformer’s current state of health and its anticipated state of health are positive.
This ensures that the system not only identifies faults effectively but also minimizes false alarms and missed detections. Such a comprehensive evaluation is crucial for optimizing transformer maintenance strategies, reducing downtime, and enhancing the overall reliability of power systems. Ultimately, the refined predictive performance contributes to proactive maintenance practices, extending the lifespan of transformers and ensuring uninterrupted power supply.
3.2.6. K-Nearest Neighbors (KNN) Accuracy
KNN has been applied to four models: Fine, Medium, Coarse, Cosine, Cubic, weighted, and Optimizable to compare and obtain high-quality results. The accuracy of each model is displayed in the graph below, and each is discussed in the section that follows.
As can be seen from
Table 2, every KNN model except Coarse KNN, which performed quite poorly at 41.2%, performed admirably. The remaining models all achieved excellent accuracy validation (78.8%). The quality of the results is demonstrated by the average accuracy of 90% achieved by the remaining four models.
3.2.7. Performance Analysis and Dataset Complexity
The high classification accuracy (80–100%) achieved by the Decision Tree and Subspace KNN models suggests that the dataset used in this study may be relatively simple or highly structured. While these models are effective for pattern recognition, their performance can be influenced by data separability and preprocessing techniques. A key factor in assessing model reliability is the statistical complexity of the dataset, including class distributions, feature variations, and potential data imbalances.
To ensure a comprehensive evaluation, confusion matrices were analyzed to identify false positives and false negatives, particularly in challenging fault categories such as contamination, which exhibited a 100% False Negative Rate (FNR). This suggests that while the model is highly effective for certain faults, it struggles with others, highlighting the need for improved feature extraction techniques. Furthermore, traditional accuracy metrics alone are insufficient to assess real-world applicability. Instead, precision, recall, and F1-scores offer a more balanced view of classification effectiveness, especially in cases of class imbalance.
Overfitting risks were also considered, given that Decision Trees are prone to memorizing training data rather than generalizing well. To mitigate this, cross-validation techniques were employed to assess model stability across different test sets. The results indicate that while the classifiers perform well within the given dataset, their ability to handle unseen data remains a critical consideration. Future work should explore more advanced machine learning models, such as ensemble methods or deep learning approaches, to enhance predictive robustness and generalization.
3.3. Theoretical Rationale for the Chosen Models
The selection of machine learning models for fault classification in power transformers was guided by both theoretical considerations and the nature of Frequency Response Analysis (FRA) data. Decision Tree and K-Nearest Neighbors (KNN) models were specifically chosen for their ability to handle the unique challenges presented by FRA data, including its multi-dimensional nature and frequency-dependent characteristics. Each model was evaluated based on its interpretability, computational efficiency, and suitability for fault diagnosis tasks.
Decision Trees were selected for their ability to partition the feature space into hierarchical decision regions. This model excels in handling non-linear relationships, making it ideal for analyzing FRA data, where frequency response patterns can vary significantly between different fault types. The inherent interpretability of Decision Trees also makes them valuable in diagnostics, as experts can trace classifications back to specific feature thresholds, providing insights into the physical nature of the detected faults. For instance, insulation breakdown faults may present distinct response amplitudes at specific frequency bands, which the Decision Tree model can isolate and use for accurate classification.
KNN, an instance-based learning model, was chosen for its simplicity and effectiveness in environments where data points naturally cluster based on fault types. FRA data often exhibits such clustering, with samples from similar fault conditions forming distinct patterns in the feature space. KNN classifies new data points by comparing them to their nearest neighbors, making it highly effective in scenarios where fault signatures share similarities within localized frequency ranges. To address potential challenges, such as the high dimensionality of FRA data and class imbalance, the study employed Neighborhood Component Analysis (NCA) for feature transformation, enhancing KNN’s efficiency by focusing on the most informative dimensions.
4. Results and Discussion
In this section, various Frequency Response Analysis (FRA) measurements have been employed to identify faults related to insulation deterioration. The identified transformer faults include thermal aging, electrical stress, moisture ingress, insulation breakdown, corona discharge, thermal overstress, and contamination.
The evaluation of the winding insulation degradation state is achieved through the application of the different machine learning AI models that have been discussed in the previous section. The proposed method is validated using the real-world Transformer FRA data and the results of experimental findings to demonstrate the capacity of the different machine learning AI models to accurately predict the degree of winding insulation degradation, regardless of the presence of random or continuous irregularities in artificial intelligence training data.
Moreover, the approach’s capacity to identify and precisely evaluate winding insulation deterioration in the face of sporadic or persistent anomalies in the data supports the adaptability and dependability of the AI models.
4.1. Insulation Deterioration in Power Transformer Detection Faults Using Decision Tree Model, Trained with Fine Tree ML Models (0–2000 Hz)
This study used a decision tree model trained with Fine tree ML models to detect insulation deterioration faults in power transformers within the frequency range of 0–2000 Hz.
Figure 3a,b presents the validation confusion matrix and sample-level classification results for detecting insulation deterioration faults using a Decision Tree model. Diagonal elements indicate accurate classifications, with 100% accuracy achieved for Classes A (Thermal Aging), B (Electrical Stress), and C (Moisture Ingress) faults.
The perfect alignment between actual and predicted values across all samples underscores the model’s robustness in diagnosing insulation deterioration. The results validate the effectiveness of feature extraction and classifier training for practical fault detection applications.
Class A faults represent thermal aging faults; using this ML mode, it was 100% predicted correctly, and all the seven samples trained were classified correctly; see
Figure 3a,b below. Class B faults represent Electrical Stress (ES) faults, and using a Fine Tree model, an average accuracy of 100% was obtained, and all faults were detected correctly; seven out of seven samples were detected correctly.
Class C faults represent moisture ingress insulation and deterioration faults. All faults for this class were correctly detected with 100% accuracy, and all six samples classified were predicted correctly.
The diagonal elements in the confusion matrix formed by all the classes proved the evidence for the accuracy of this model and confirmed that 100% predictive accuracy was obtained when the model was trained. Also, the models took a very short time to predict the faults of the 2600s and less than a minute to train them.
The graph below in
Figure 4 shows that the classification of the three fault classes, and the results predicted graphically were correct. A trend of sine wave was observed, also confirming that the model’s predictions were accurate. Thermal aging, class A, was observed to have about four curves. Electrical stress (ES) was observed to have three faults. And lastly, the moisture ingress was observed to also have three peaks.
For both Class A and Class B, the C curve demonstrates a distinct behavior at 1000 Hz, highlighting notable characteristics in the system’s response at this frequency. This curve provides valuable insights into performance variations and potential anomalies in the signal.
In addition to the C curve, the presence of V curves is clearly noticeable at 400 Hz, 100 Hz, and 1000 Hz, as depicted in the graph. These V curves indicate different response patterns across these specific frequencies, reflecting the system’s behavior under varying frequency conditions. The distinct patterns observed in these curves can help in diagnosing performance characteristics and identifying potential issues such as resonance or structural changes within the transformer or system under analysis.
Furthermore,
Figure 4 illustrates an important trend: as the frequency increases, the amplitude of the magnitude also rises. This indicates a frequency-dependent response, where higher frequencies correspond to greater signal magnitudes.
4.2. Insulation Deterioration in Power Transformer Detection Faults Using Ensemble, Trained Subspace KNN ML Model (10–15 MHz)
Class A, faults for Insulation Breakdown were correctly classified with an average of 100 % see
Figure 5a, all 13 samples were correctly classified, see
Figure 5b.
Class B, representing Corona Discharge, was successfully classified with perfect accuracy, achieving 100% across all 9 samples. As shown in
Figure 5a,b, the model demonstrated flawless performance in predicting these faults. Similarly, Class C, which corresponds to thermal overstress, also achieved 100% accuracy, with all five samples correctly identified and classified, reflecting the model’s high efficiency in detecting this type of fault. The perfect prediction rates across these classes indicate that the machine learning model is well-suited for identifying such faults, offering strong reliability in diagnostic assessments.
However, Class D, corresponding to contamination-related faults, exhibited a significant issue in classification. Despite the otherwise accurate performance of the model, this particular class experienced 100% False Negative Rate (FNR), where the single sample for contamination was misclassified. This suggests that the model may struggle to differentiate contamination faults from other classes, possibly due to the specific characteristics of the contamination-induced defects not being adequately captured in the training data. The misclassification of Class D emphasizes the need for further refinement in the model, potentially by incorporating additional data or adjusting the feature extraction process to better address the unique properties of contamination faults.
While the model displayed exceptional accuracy in identifying the majority of fault classes, the misclassification of contamination faults highlights an area for improvement. Future work could focus on enhancing the model’s ability to recognize and distinguish contamination faults, which would help further optimize fault detection capabilities for power transformer diagnostics.
All the four faults classified are graphed in
Figure 6. The trends of curves with V shapes have been observed in all four classes. All classes in the study formed similar shapes in the frequency spectrum, with peaks occurring at the same frequencies but at varying magnitudes. The trend in the pattern was continuous across all fault types, indicating a consistent response of the system to different types of faults. These peak frequencies, which appeared in each class, were indicative of specific fault signatures, although the magnitudes differed depending on the fault class. This uniformity in the shape of the curves across the different classes suggests that the model was effectively capturing the key features of the faults, albeit with slight variations in intensity.
Notably, there was an intersection between Classes D and C at a frequency of 14 MHz. This intersection point suggests a potential overlap or similarity in the frequency response of the faults represented by these classes. It may indicate that the model faces difficulty distinguishing between certain fault types at this frequency, leading to a potential for misclassification or ambiguous fault identification. Further investigation into this intersection could help refine the model, allowing for better differentiation between faults with similar spectral characteristics.
While the general pattern of peaks and their associated magnitudes was consistent across fault classes, the intersection between Classes D and C at 14 MHz highlights an area where further analysis is needed. By focusing on this specific region, adjustments to the model’s training or feature extraction techniques could improve its ability to differentiate faults more clearly at this critical frequency, enhancing overall diagnostic accuracy.
4.3. Insulation Deterioration in Power Transformer Detection Faults Using Decision Tree, Trained with Fine Trees ML Model (10–15 MHz)
Class A faults for Insulation Breakdown were correctly classified with an average of 100% see
Figure 7a, all 13 samples were correctly classified, see
Figure 7b. Class B, representing Corona Discharge, was successfully classified with perfect accuracy, achieving 100% across all 9 samples. As shown in
Figure 7a,b, the model demonstrated flawless performance in predicting these faults. Similarly, Class C, which corresponds to thermal overstress, also achieved 100% accuracy, with all five samples correctly identified and classified, reflecting the model’s high efficiency in detecting this type of fault. The perfect prediction rates across these classes indicate that the machine learning model is well suited for identifying such faults, offering strong reliability in diagnostic assessments.
However, Class D, corresponding to contamination-related faults, exhibited a significant issue in classification. Despite the otherwise accurate performance of the model, this class experienced a 100% False Negative Rate (FNR), where the single sample for contamination was misclassified. This suggests that the model may struggle to differentiate contamination faults from other classes, possibly due to the specific characteristics of the contamination-induced defects not being adequately captured in the training data.
The misclassification of Class D emphasizes the need for further refinement in the model, potentially by incorporating additional data or adjusting the feature extraction process to better address the unique properties of contamination faults.
In conclusion, while the model displayed exceptional accuracy in identifying most fault classes, the misclassification of contamination faults highlights an area for improvement. Future work could focus on enhancing the model’s ability to recognize and distinguish contamination faults, which would help further optimize fault detection capabilities for power transformer diagnostics.
All the four faults classified are graphed below in
Figure 8. The trends of curves with V shapes have been observed in all four classes.
All classes in the study formed similar shapes in the frequency spectrum, with peaks occurring at the same frequencies but at varying magnitudes. The trend in the pattern was continuous across all fault types, indicating a consistent response of the system to different types of faults. These peak frequencies, which appeared in each class, were indicative of specific fault signatures, although the magnitudes differed depending on the fault class. This uniformity in the shape of the curves across the different classes suggests that the model was effectively capturing the key features of the faults, albeit with slight variations in intensity.
Notably, there was an intersection between Classes D and C at a frequency of 14 MHz. This intersection point suggests a potential overlap or similarity in the frequency response of the faults represented by these classes. It may indicate that the model faces difficulty distinguishing between certain fault types at this frequency, leading to a potential for misclassification or ambiguous fault identification. Further investigation into this intersection could help refine the model, allowing for better differentiation between faults with similar spectral characteristics.
4.4. Insulation Deterioration in Power Transformer Detection Faults Using Decision Tree, Trained with Fine Trees ML Model (20–100 MHz)
As seen in
Figure 9, in Class A, representing Insulation Breakdown, the model demonstrated flawless performance, achieving 100% accuracy in fault detection. All eight FRA samples were correctly classified and predicted, underscoring the model’s exceptional ability to accurately identify insulation breakdown faults. This remarkable result can be visually confirmed in
Figure 9a,b below, where the predictions align perfectly with the actual fault data.
For Class B, representing Loss Connections, the model achieved a commendable accuracy rate of 85.7%, with only one misclassified FRA sample out of a total of seven. This yields a True Positive Rate (TPR) of 85.7% and a False Negative Rate (FNR) of 14.3%, showcasing the model’s strong ability to detect loss connections, though with room for improvement.
In Class C, which represents Corrosion faults, the model exhibited flawless performance once again, achieving 100% accuracy. All 12 FRA samples were correctly classified and predicted, confirming the model’s reliability in identifying corrosion-related faults. This outstanding performance is illustrated in
Figure 9a,b, where the accuracy of the predictions is clearly depicted.
For Class D, representing Improper Termination faults, the model correctly detected three out of four FRA samples, achieving a TPR of 75% and an FNR of 25%. While the model performed well, there was a single misclassification, indicating that further refinement could enhance the model’s accuracy in this class.
The diagonal elements formed by the confusion matrix across all classes provide strong evidence of the model’s overall accuracy, with the majority of predictions falling on the diagonal. Notably, the two classes with less than 100% accuracy, Classes B and D, are positioned off-diagonal, highlighting areas for potential improvement. However, the overall pattern of correct classifications emphasizes the robustness of the model in accurately identifying different types of faults across all classes.
All four faults predicted above have been plotted graphically in
Figure 10, providing a clear visual representation of the model’s performance. One of the key observations is the presence of valleys in the graphics for each class, which play a crucial role in the accuracy of the model. The consistent trends of these valleys across all classes suggest that the model is accurately detecting fault signatures, as valleys often correspond to significant changes or features in the data that are indicative of specific faults. The regularity and predictability of these valleys indicate the model’s robustness and the high level of confidence in its classifications. The patterns are clear, providing strong evidence of the model’s ability to detect faults accurately and consistently across different fault types, including insulation breakdown, loose connection faults, corrosion, and improper termination. The presence of these well-defined valleys in the graphical representation confirms the model’s reliability in fault detection, contributing to the overall performance of the system.
All classes, started between the magnitude of −40 dB and −20 dB. All classes formed a similar shape between the frequency band of 20 kHz and 60 kHz. From the frequency of 20 kHz, theC classes, split into groups of two, Class A and D showed positive gradients. Class B and C showed negative gradients, and they continued to drop gradually until reaching a frequency of 100 kHz.
5. Key Innovations and Contributions
This study introduces a novel machine learning-driven diagnostic framework that significantly enhances power transformer fault detection using Frequency Response Analysis (FRA). Unlike previous studies that rely on limited fault types and manual interpretation of FRA signatures, this work leverages advanced machine learning models to achieve automated and highly accurate classification of transformer faults across multiple frequency bands. By systematically analyzing FRA data, the proposed approach overcomes traditional limitations, such as the dependency on expert knowledge and the challenges of noisy environments.
A key contribution of this research is the development of a comprehensive, annotated FRA dataset, enabling supervised learning for diverse fault conditions. This dataset bridges the gap in transformer diagnostics by incorporating underexplored fault types, including axial displacement, core buckling, and contamination-related defects, which have received limited attention in prior studies. The integration of refined feature extraction techniques further enhances classification sensitivity, leading to an observed accuracy improvement of 80–100% across various fault classes.
Performance evaluations demonstrate the superiority of the proposed methodology over conventional FRA interpretation techniques. The Decision Tree and Subspace KNN models outperform traditional methods by reducing false negatives and achieving frequency-independent fault detection, ensuring more reliable transformer condition monitoring. Additionally, the study identifies key frequency markers (7.6 MHz, 8.25 MHz, and 8.7 MHz) that serve as robust diagnostic indicators, reinforcing the practical applicability of the approach.
By minimizing reliance on specific frequency bands and improving fault localization accuracy, this work presents a scalable and adaptable solution for predictive transformer maintenance. Future advancements will focus on deep learning integration and real-world deployment, further enhancing diagnostic precision and transformer health management.
6. Conclusions
Reliable transformer fault detection is crucial for maintaining the stability and efficiency of power systems. This study explored the application of Frequency Response Analysis (FRA) data combined with machine learning models to address challenges in detecting transformer faults such as winding deformation, insulation breakdown, and partial discharges. The research aimed to evaluate the effectiveness of machine learning classifiers, including Decision Trees and Subspace KNN, in fault detection across eight frequency bands (20 kHz to 12 MHz), while identifying performance limitations and opportunities for improvement.
The study achieved fault classification accuracies between 80% and 100%, demonstrating the significant potential of machine learning in predictive maintenance strategies. However, challenges remain in accurately detecting complex fault types, such as contamination-related defects and lamination deformation. These findings highlight the need for refined feature extraction techniques and higher-resolution data to improve diagnostic performance.
Integrating machine learning into transformer diagnostics can minimize downtime and enhance asset management. While current results are promising, further research should explore the integration of these models into real-world diagnostic systems, such as online monitoring frameworks. This would allow for continuous assessment of transformer health and early detection of faults before catastrophic failures occur. Additionally, expanding the dataset to include a broader range of transformer types and operational conditions will help improve model generalization and robustness.
Future research should focus on scaling these models through advanced deep learning techniques, improving generalization across diverse transformer conditions, and addressing data scarcity. Exploring hybrid approaches that combine physics-based modeling with machine learning could further enhance predictive accuracy. Ultimately, these advancements will contribute to the development of more reliable and scalable frameworks, ensuring the longevity and operational efficiency of power transformers globally.