Transformer Core Fault Diagnosis via Current Signal Analysis with Pearson Correlation Feature Selection

: The role of transformers in power distribution is crucial, as their reliable operation is essential for maintaining the electrical grid’s stability. Single-phase transformers are highly versatile, making them suitable for various applications requiring precise voltage control and isolation. In this study, we investigated the fault diagnosis of a 1 kVA single-phase transformer core subjected to induced faults. Our diagnostic approach involved using a combination of advanced signal processing techniques, such as the fast Fourier transform (FFT) and Hilbert transform (HT), to analyze the current signals. Our analysis aimed to differentiate and characterize the unique signatures associated with each fault type, utilizing statistical feature selection based on the Pearson correlation and a machine learning classifier. Our results showed significant improvements in all metrics for the classifier models, particularly the k-nearest neighbor (KNN) algorithm, with 83.89% accuracy and a computational cost of 0.2963 s. For future studies, our focus will be on using deep learning models to improve the effectiveness of the proposed method.


Introduction
Predictive maintenance (PM) is a cutting-edge approach that leverages data-driven methodologies to anticipate potential equipment or machinery failures.This proactive technique enables timely maintenance measures by gathering data from strategically placed sensors or analyzing current and voltage levels, minimizing any unforeseen downtime.PM has become an essential tool for efficient and cost-effective operations because it can predict and prevent potential equipment breakdowns [1,2].The prevalence of data-driven strategies compared to model-based methodologies is often credited to the difficulties of creating and maintaining precise physics-of-failure models.The importance of a datadriven approach when training artificial intelligence (AI) models must be considered in PM.This approach plays a crucial role in unlocking the complete potential of AI-based models, guaranteeing their effectiveness in anticipating and averting equipment malfunctions [3,4].
Condition-based maintenance (CBM) has emerged as a pivotal strategy in ensuring transformers' reliability and optimal performance in modern power systems.Transformers are critical in power distribution, stepping up or down voltage levels for efficient energy transfer.As transformers are subjected to various operational stresses, the early detection of potential faults or deterioration is paramount to prevent catastrophic failures and minimize downtime [5].CBM leverages advanced monitoring and diagnostic techniques, such as real-time data acquisition, signal analysis, and predictive modeling, to assess the health status of transformers.This proactive approach enables timely intervention, reducing maintenance costs, enhancing operational efficiency, and extending the lifespan of transformers.In complex power distribution systems, transformers play a pivotal role in ensuring the efficient transmission of electrical energy.However, these critical components are susceptible to faults compromising reliability and performance.One of the most crucial fault types is related to the transformer core, which forms the heart of its operation.Core faults encompass issues such as insulation degradation, winding deformations, and, most notably, the presence of cracks [6].
During the initial stages of a core fault, the transformer may not be affected significantly.However, the damage can become more severe over time if left unchecked.It is essential to conduct preventive evaluations for possible failures, especially core faults, to ensure a reliable energy supply.This can effectively minimize the risk of further damage to the transformer, resulting in shorter outages and reduced repair costs [7,8].Furthermore, given the high cost of transformers and the challenges associated with their maintenance, early fault detection is of paramount importance to facilitate timely repairs, ultimately reducing the risk of significant breakdowns [9][10][11].
This study investigated a 1 kVA transformer in both healthy and faulty states, utilizing electric current data.Distinct current behavior in these states is a crucial indicator of fault patterns, especially in advanced stages.Detecting faults in raw data can be challenging, particularly during early fault development.To address this, signal processing becomes pivotal for implementing condition monitoring, offering data compression, noise reduction, and pattern recognition.A filter-based statistical feature selection approach, including Pearson correlation, is applied for efficient feature selection in time-domain analysis.This enhances precision and allows a more comprehensive observation of faults through various analyses, such as time-domain, frequency-domain, time-frequency, and Pearson-correlation-based statistical feature selection, contributing to proactive maintenance and improved reliability in power electronics systems [12][13][14][15][16].
This study introduces a novel model for identifying core faults in transformers by leveraging electric current data to assess the health of the transformer core.These contributions collectively advance the comprehension of transformer health assessment, laying the foundation for more effective fault detection methodologies in power systems.The study offers the following significant contributions:

•
We have designed an experimental setup with the aim of collecting current signals, which should serve as a baseline for other researchers in analyzing transformer core fault analysis.

•
We applied the Hilbert transform, a time-domain signal processing technique, to extract the magnitude envelope.This step is critical in improving the interpretation of signal analysis.

•
We have established a comprehensive framework for robust feature engineering, focusing on extracting time-domain statistical features and filter-based Pearson correlation feature selection.

•
We have conducted a comparative analysis in terms of performance evaluation to validate the efficiency of the proposed framework.
The subsequent sections of the paper are organized as follows: Section 2 delves into a review of related works and provides insight into the motivation behind the study, while Section 3 explores the theoretical background of the research.Section 4 presents the system model of the proposed fault diagnosis framework.Section 5 provides details on the testbed setup and collection of data.The detailed experimental results of the study are discussed in Section 6, and the paper concludes in Section 7, where the findings are summarized and future works are discussed.

Motivation and Review of the Related Literature
The basic working principle of a transformer lies in the usage of alternating current.When an alternating current flows to a winding, it generates a magnetic field in the core, thus inducing voltage in the secondary side.This is possible through the concept of electromagnetic induction.The core provides a path for the magnetic flux generated by the primary winding.It is commonly made of laminated thin sheets of electrical steel so that the layers are insulated from each other, which reduces eddy current losses, improving the transformer's efficiency.When a magnetic field changes within a conductor, circulating currents called eddy currents are induced within the core.These currents circulate in closed loops and result in the generation of heat due to the electrical resistance of the material [17,18].As stated in [11], the core fault is one of the common failures in transformers, which is the main motivation of this study.Table 1 summarizes various core faults, including brief descriptions, causes, and effects studied in [19,20], namely, saturation and lamination faults, respectively.In this study, we focus on mechanical damage to the transformer core.Research on data-driven fault detection of transformers' health created a new perspective for researchers as a new approach.However, only a few focused on using machine learning (ML) algorithms to develop predictive models.In [21], the authors presented a method for analyzing core looseness faults by acquiring vibration signals and used different time-frequency analysis methods to compare which method worked well.However, based on the mode of the dataset collected (vibration signal) and the study's goal, the wavelet and empirical mode decomposition performed better than the other techniques under study.Another development in transformer fault analysis is the usage of the feature extraction method on vibration signals based on variational mode decomposition, and Hilbert transform (HT) is applied to obtain the Hilbert spectrum of the signal [22].Interestingly, in [20], detection and classification of lamination faults, namely, edge burr and lamination faults, were analyzed.They extracted average, fundamental, total harmonic distortion (THD), and standard deviation (STD) features from the collected current signal and fed them to SVM, KNN, and decision tree (DT) algorithms.
The authors in [23] addressed the issue of acquiring precise data on vibrations and acoustics by developing a simulation technique for transformer core fault detection based on multi-field coupling.They validated their simulation results through physical experiments.However, no noise reduction technique was implemented in the data collection process, which may lead to the inclusion of unwanted signals.Advanced techniques in machine learning and signal processing have revolutionized the way we approach fault detection and prognosis modeling.The paper [24] presents a remarkable case in which these techniques were applied to develop a prognosis model for a transformer based on vibration signals.In addition, the paper [25] offers an online method that utilizes vibration measurements to distinguish the condition of the transformer core.Notably, the researchers in [12,26] also demonstrated the effectiveness of an electric current and Pearson correlation filter-based statistical feature selection approach in analyzing faults in moving machines such as motors.These findings highlight the potential of advanced signal processing and machine learning techniques to revolutionize the field of fault detection and prognosis modeling.
The current signal is a crucial diagnostic tool in power electronics, especially when identifying faults in components like transformers.By carefully analyzing variations in the current signal, experts can gain valuable insights into potential issues, enabling them to diagnose faults quickly and accurately.This is essential for maintaining the reliability and longevity of power electronic systems.In [27], they presented different machine-learningbased methods for diagnosing faults in electric motors.The first method used short-time Fourier transform (STFT) analysis on stator phase current signals for demagnetization fault diagnosis in permanent magnet synchronous motor (PMSM) drive systems, with high accuracy results obtained using k-nearest neighbors (KNNs) and multilayer perceptron (MLP) models.Similarly, in [28], they explored fault diagnosis in wind turbines (WTs) using electrical signals from the generator of a 20-year-old operating WT.For this, signal analysis techniques such as fast Fourier transform (FFT) and periodogram were employed to compare the effectiveness of spectral analysis methods, demonstrating the feasibility of using current signals for fault detection.Interestingly, in [29], the focus shifted to induction motor (IM) fault diagnosis through the stator current and delved into advanced signal processing tools.It showed the efficacy of spectrogram, scalogram, and Hilbert-Huang transforms in detecting rotor failures under various conditions, offering insights into fault diagnosis during transient operations.Also, in [30], they proposed a novel approach for rotor speed estimation in squirrel cage induction motors.This method employed signal processing techniques on airgap signals measured by a Hall effect sensor and compared the fast Fourier transform and Hilbert transform to traditional stator current methods.The results revealed superior accuracy in estimating rotational speed, especially in scenarios involving broken rotor bars-Hall effect sensors.After comparing FFT and HT with traditional stator current methods, the research indicated superior accuracy in estimating rotational speed, particularly in scenarios involving broken rotor bars.
We have integrated signal processing with intelligent algorithms, employing FFT and Hilbert transform.While deep learning methods like class-aware adversarial multiwavelet convolutional neural networks [31] yield accurate results, signal processing adds value by extracting meaningful features from raw data [32,33].This combination enhances the model's interpretability and performance, bridging the gap between complex signals and effective machine learning outcomes.The goal is to understand the impact of varied preprocessing strategies on model outcomes.

Analysis Concerning Time Domain and Frequency Domain
Analyzing signals is of paramount importance in various scientific and engineering applications, and understanding the time and frequency domains is fundamental to this analysis.In the time domain (TD), examining signals provides insights into their temporal behavior, allowing researchers and engineers to understand how the signal varies with respect to time.This is crucial for assessing the transient response of systems, studying the duration and shape of pulses, and investigating dynamic behaviors such as switching events.TD analysis is essential for evaluating the stability, response time, and overall performance of circuits and systems [34].In the frequency domain (FD), analysis reveals the frequency components present in the signal.This is particularly valuable for characterizing harmonic content, identifying resonances, and assessing the spectral distribution of the current.FD analysis is indispensable in the design and optimization of power distribution systems, as well as in identifying and mitigating issues related to electromagnetic interference and power quality [35].

Overview of Selected Machine Learning Algorithms
This comprehensive overview illuminates the diverse capabilities of these machine learning models in solving real-world problems.The AdaBoost classifier (ABC) is a versatile algorithm for categorization, assembling numerous simple decision-makers to form a robust classifier focusing on correcting past mistakes to mitigate overfitting [36].The k-nearest neighbors (KNNs) algorithm classifies new data points by assessing the labels of their 'k' closest neighbors, excelling in both classification and regression tasks [37][38][39].Logistic regression (LR) is a straightforward method for binary choice scenarios, particularly effective when the relationship between features and target group is uncomplicated [40].The multilayer perceptron (MLP) adapts during learning tasks by adjusting its layers to enhance predictive capabilities, managing non-convex loss functions and local minima fluctuations [41].Stochastic gradient descent (SGD) efficiently trains learning models, rapidly improving accuracy after each example, balancing speed and precision [42].Support vector machines (SVMs) excel in intricate grouping and prediction scenarios, particularly identifying optimal group separation [43].

Fast Fourier Transform
The fast Fourier transform (FFT) is a signal processing technique designed for computing the discrete Fourier transform (DFT).To comprehend the intricacies of DFT, it is essential to first delve into the concept of Fourier transform (FT).FT analyzes a signal in the time domain, breaking it down into a representation that exposes its frequency components.It explains the extent to which each frequency contributes to the original signal.Furthermore, FT within a discrete time domain is referred to as DFT, and FFT is recognized as an algorithm specifically tailored for the rapid computation of a large number of DFTs.The FT of a function, denoted as f (t), is shown below [44,45]: The FFT employs complex exponentials or sinusoids of varying frequencies as its basis functions, effecting a transformation into a distinct set of basis functions.Originally devised as an enhancement to the conventional DFT, the FFT significantly diminishes computational complexity from (N 2 ) to (N log N), rendering it especially beneficial for efficiently processing extensive datasets and real-time applications.Mathematically, the FFT can be succinctly expressed as [46]

Hilbert Transform
The derivation of an analytic signal from a real-valued signal entails the utilization of the Hilbert transform (HT).The resultant analytic signal finds widespread application in signal processing and communication systems, serving diverse purposes such as analyzing frequency content, extracting envelope information, and facilitating phase-sensitive operations [47,48].The HT of a real-valued signal f (t) is given by or in terms of the Cauchy principal value: The analytic signal F H (t), combining the original signal f (t) and its HT, is given by The properties of analytical signal include: • Complex representation: The analytic signal is complex, with both real and imaginary components.The actual component signifies the original signal, while the imaginary component represents the Hilbert transform of the signal.
• A 90-degree phase shift: The positive frequency shifts to a negative 90-degree angle, and the negative frequency shifts to a positive 90-degree angle in the context of HT.This introduces a phase shift of 90 degrees between the original signal and its HT, which is crucial in applications such as demodulation and phase-sensitive analysis.
Additionally, the analytic signal, derived through the HT, provides a representation of the original signal that separates positive and negative frequency components.This property is valuable for analyzing the frequency content of a signal.

•
Enveloping: The envelope of the original signal can be extracted from the magnitude of the analytic signal.The envelope represents the slowly varying magnitude of the signal and is useful in applications such as amplitude modulation.

Proposed Diagnostic Framework
This section presents the process for detecting faults in transformers, as illustrated in Figure 1.The following stages are involved: gathering current dataset from both the healthy and faulty states of the transformer; applying signal processing methods, in particular, Hilbert transform; performing statistical feature extraction in the time domain to extract relevant features; using a Pearson correlation filter-based approach to identify highly correlated features; using the selected features for model training and testing; and lastly, carrying out performance evaluation to confirm the model's effectiveness.It is imperative to represent data in a simplified manner, emphasizing only essential features before inputting them into the model to enhance the speed and accuracy of an ML model.Time-domain statistical features are extracted to capture the relevant aspects of the data.The primary objective of this process is dimensionality reduction while retaining crucial properties or features.Moreover, transforming the raw data into a more concise representation yields several advantages, including improved overall model performance by reducing complexity, decreasing computational time, and mitigating the risk of overfitting.In this study, we employ feature engineering through statistical feature extraction, and the 14 extracted features in the time domain and 13 in the frequency domain are illustrated in Tables 2 and 3.
The linear relationship between two variables can be statistically measured on a scale ranging from −1 to 1.If the value is close to 1, they are highly correlated, and the sign indicates a positive or negative correlation.Negative values indicate a negative correlation, signifying that as one variable increases, the other decreases.Conversely, positive values suggest a positive correlation, meaning that both variables increase or decrease simultaneously.A value of 0 denotes no correlation between the variables.This approach is widely employed in data analysis, statistics, and ML for feature selection and comprehending the relationships between variables [49,50].The formula is presented below: Root Mean Square (RMS) Table 3. Frequency-domain statistical features and formulas [51].

Domain Features Formulas
Frequency-based Mean Frequency Median Frequency (P cum > 0.5), where P cum is the cumulative power spectral density Spectral Entropy Peak Frequency and Frequency (P( f k ) = max(P( f k ))) Peak Amplitude max(P( f k )) Dominant Frequency and Frequency (P( f k ) = max(P( f k ))) Spectral Roll-Off (n = 80%, 90%) Frequency(P cum > n • Total Power), where P cum is the cumulative power spectral density.
Values of X and Y are the features extracted from the HT values to simplify data representation, highlighting only the essential features.Covariance quantifies the extent to which the two variables change.The product of the standard deviations of X and Y is calculated in the denominator.The standard deviation gauges the degree of variation or dispersion from their respective averages.This study selected features exhibiting a Pearson correlation coefficient exceeding 95% to ensure distinctiveness among the remaining features.Subsequently, the retained features undergo concatenation and labeling before being employed in the model training and testing phases.

Experimental Setup and Data Collection
Figure 2a depicts the experimental test setup utilized to acquire the current signals.This setup was designed and executed at the Defense and Reliability Laboratory, Kumoh National Institute of Technology, Republic of Korea.The supply voltage was from a standard convenience outlet, featuring a rating of 220 V at a frequency of 60 Hz.Current measurements were conducted using the national instruments NI 9246 current module, seamlessly interfaced with LABVIEW software version 8.6.1 through the national instruments NI cDAQ-9174.The acquisition of current data occurred on the primary side of the circuit.Concurrently, on the secondary side, an electric fan was connected to function as a motor load for the transformer.The comprehensive circuit diagram is presented in Figure 2b.The NI 9246 specifications are listed as follows: • Three isolated analog input channels were employed, each operating at a simultaneous sample rate of 50 kS/s, ensuring comprehensive data collection.

•
The system offers a broad input range of 22 continuous arms, with a ±30 A peak input range and 24-bit resolution, exclusively for AC signals.In this study, we obtained three datasets representing different conditions of transformers: a healthy state (labeled as HLTY), a state with one hole in the core (labeled as 1HCF), and a state with two holes in the core (labeled as 2HCF).To simulate 1HCF, a 5 mm hole was drilled diagonally through the edge of the core.This was to replicate damage focused on the edge of the transformer.In the 2HCF, an additional 5 mm hole was drilled straight through the core from top to bottom, simulating core damage away from the edge of the transformer.Figure 3 illustrates the actual replication of these faults conducted during our experiment in the laboratory.

Applying Signal Processing Technique
In this study, we employed signal processing techniques to unveil crucial details within the signals that were obscured in the raw data.To assess the efficacy of our proposed model utilizing Hilbert transform (HT) on electric current data, we conducted a comparative analysis using fast Fourier transform (FFT) without employing any signal processing technique.Following the signal processing step, we applied a window size of 25 samples to the data before proceeding with statistical feature extraction.
Figure 4a-c display the electric current data obtained from the modules under three working conditions, HLTY, 1HCF, and 2HCF, represented by green, black, and orange.The data values range from −0.5 to 0.5 in all working conditions.During HLTY, the plot reveals that the maximum current in the circuit can reach −0.5 A to 0.5 A, with a notably cleaner waveform compared to other operating conditions.In the case of 1HCF, the current ranges from −0.4 A to 0.4 A, which is lower than the HLTY condition.The plot exhibits a random pattern with distortions in every cycle.Transitioning to 2HCF, the range of values is relatively similar to the 1HCF condition, varying from −0.4 A to 0.4 A. However, compared to HLTY, the waveform pattern differs.Upon close examination of each plot, it seems that using raw data could potentially aid in identifying core faults in transformers.However, upon closer examination in Figure 4d, which presents the plots for all working conditions, it becomes evident that there is no significant difference when comparing HLTY with the faulty conditions (1HCF and 2HCF).Figure 5 illustrates the FFT plots under various operating conditions, revealing the limitation of FFT in capturing essential changes across all scenarios.Upon observation, the plots in all conditions exhibit minimal variation, indicating that when features are extracted, there is a lack of discriminative information.Figure 6 demonstrates the substantial differences revealed after applying HT to the transformer core dataset, particularly distinguishing between healthy and faulty conditions.However, for the 1HCF and 2HCF plots, the differences may not be obvious, but the next section demonstrates a significant increase in the model's performance.This observation underscores the effectiveness of our proposed signal processing technique's usefulness in analyzing transformers' core health based on current data.Identifying relevant characteristics and patterns in the raw signal proves to be pivotal in the initial stages of our methodology, as these factors significantly impact the overall performance of the ML model.

Correlation Matrix of Extracted and Selected Time-Domain Statistical Features
Figure 7a illustrates the correlation plot of the features extracted from the HT.The red intensity in the plot indicates the strength of the correlation among features, with a gradient from more red showing a stronger correlation to less red, and eventually blue, indicating a weaker correlation.This matrix visually represents the relationships between each feature, providing valuable insights for analysis and utilization.There is a notable correlation between the mean and other features, namely, max, peak-to-peak, RMS, variance, and standard deviation, with correlation coefficients of 0.98, 0.94, 1, 0.97, and 0.97, respectively.Recognizing such high correlations is crucial, as incorporating highly correlated features into the model can significantly and negatively impact its performance.Upon extracting features and generating the correlation matrix, it was evident that the features are highly correlated and could impact the model's performance.To address this, we employed filter-based statistical feature selection.As illustrated in Figure 7b, out of the initially extracted 14 features, only 6 were retained, namely, mean, crest factor, interquartile range, median absolute deviation, 5th percentile, and 25th percentile, after eliminating those with high correlations.The resulting selected features were labeled with the values of 0, 1, and 2 and concatenated into a single data frame.This step further refined the dataset before feeding it into the ML model, enhancing its ability to capture relevant patterns and relationships in the data.

ML Diagnostic Results and Discussion
We need to split the dataset to train and test our model effectively.In our study, 80% of the dataset was allocated for training and 20% for testing, with a total size of 2541 samples.We used six established ML models: ABC, KNN, LR, MLP, SGD, and SVC.A summary of the parameters of the different models is presented in Table 4.The ML models highlight improvements under three conditions: raw data, FFT, and HT.The objective is to evaluate and compare the performance of these models in accurately classifying conditions such as HLTY, 1HCF, and 2HCF.Our study's significance lies in evaluating our proposed model's performance thoroughly.By assessing the effectiveness of our method, we can validate its reliability and demonstrate its ability to identify and classify faults proactively.Ultimately, this will contribute to the overall reliability and efficiency of the transformer.In this study, we have employed classification metrics such as TP (true positive), FP (false positive), TN (true negative), and FN (false negative) to represent the counts of accurately predicted positive instances, inaccurately predicted positive instances, accurately predicted negative instances, and inaccurately predicted negative instances, respectively.We use these metrics along with their formulas and brief descriptions, as described in the references [51,52]: Accuracy: Measures the overall correctness of the model.Accuracy = TP + TN TP + FP + TN + FN (7) Precision: Indicates the accuracy of positive predictions.
Precision = TP TP + FP (8) Recall: Emphasizes the model's ability to capture all positive instances.Recall = TP TP + FN (9) F1 score: Provides a harmonic mean by balancing precision and recall.It is particularly valuable in scenarios with uneven class distribution.
Table 5 presents a comprehensive evaluation of the machine learning model using raw data.LR (logistic regression) is identified as the top-performing model, with an accuracy of 65.23%.It shows the highest metrics, except for computational time, where KNN is the fastest with 0.0142 s.On the other hand, the lowest-performing model is SGD, with an accuracy of 29.08%.Generally, the machine learning models' performance appears unsatisfactory when analyzing raw data, making them ineffective for detecting and classifying core faults.To truly understand the effectiveness of a model, it is crucial to incorporate frequencydomain analysis, mainly using FFT.The resulting comprehensive and insightful perspective provides a detailed comparative analysis that sheds light on the model's performance.Table 6 presents a comprehensive evaluation of the machine learning model with FFT signal processing.The data shows that ABC performs the best among all models, with an accuracy of 61.49%, demonstrating superior performance in all aspects.Meanwhile, KNN delivers the fastest computational time, making it a noteworthy contender.However, the SVC model struggles, with an accuracy of only 33.20%.Despite these findings, it is worth noting that the overall performance of the ML models remain low in frequency-domain analyses, which limits their effectiveness in detecting and classifying core faults.The ML model performance metrics presented in Table 7 provide compelling evidence of the superior performance of the KNN model, which boasts the highest values across all parameters, an accuracy of 83.89%, and a remarkable computational time of 0.0156 s.Our proposed method, utilizing HT and Pearson correlation filter-based feature selection, enhances the performance across all metrics in the ML models, except for time cost, as shown in Figure 8.This compelling figure substantiates the effectiveness of our model, which outperforms all ML models when diagnosing and classifying core faults.The confusion matrix, depicted in Figure 9, provides consistent improvements in TP and a reduction in FP, with the ABC and SGD models predicting the highest number of FP instances (121 and 171, respectively).Conversely, other models demonstrated an increase in TP predictions, evident from the values along the diagonal.Our model's notable enhancement in performance when employing the HT for analyses makes it a compelling choice for any team looking to improve the efficiency and accuracy of core fault diagnosis and classification.

Limitations, Open Issues, and Future Directions
This study integrates signal processing with filter-based Pearson correlation feature selection, employing FFT and HT.Our reliance on comparative analysis between signal processing techniques, such as the HT (proposed), and frequency-domain analysis, notably the FFT, using only raw data, provides valuable insights into their respective efficacy.We acknowledge the limitations of utilizing the HT for fault classification compared to deep learning methodologies.While our study demonstrates the effectiveness of the HT in analyzing current signals and characterizing fault signatures in a single-phase transformer core, it is essential to recognize its inherent constraints.Firstly, the performance of the HT may be influenced by signal noise and variability, particularly in real-world applications, where environmental factors and measurement inaccuracies can impact signal quality.It may reduce the robustness and reliability of fault classification outcomes, potentially leading to misdiagnosis or false alarms [53].Secondly, the HT's effectiveness in capturing complex nonlinear relationships and subtle fault patterns may be limited compared to deep learning methodologies.Deep learning algorithms, such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs), excel in learning intricate features and hierarchical representations from raw data, enabling more nuanced fault classification capabilities [54].Additionally, the reliance on handcrafted features and manual feature selection in HT-based approaches may pose challenges in capturing and leveraging all relevant information in the data.On the other hand, deep learning models can automatically extract discriminative features from raw signals, minimizing the need for manual intervention and potentially enhancing diagnostic accuracy.Furthermore, the scalability and generalizability of HT-based fault classification methods may be limited when faced with diverse fault scenarios or variations in transformer operating conditions.With their adaptability and ability to learn from large and varied datasets, deep learning methodologies offer more significant potential for robust performance across various operating conditions and fault types [55].Moreover, the potential for overfitting in deep learning models should be considered, as it can lead to poor generalization and decreased performance on unseen data.Adequate regularization techniques must be employed to mitigate this risk and ensure the reliability of fault classification results [56][57][58].In scenarios where deep learning models may not be feasible, the HT provides a pragmatic alternative that can yield valuable insights into the system's health.Its simplicity and efficiency make it particularly attractive for real-time or resource-constrained environments, where complex modeling approaches may need to be more practical [59,60].Furthermore, the HT's transparent and intuitive nature facilitates straightforward interpretation of results, making it accessible to a broader range of practitioners without extensive expertise in machine learning or data science.It can be advantageous in fields where practicality and ease of implementation are paramount.Overall, while deep learning models offer potent capabilities for fault diagnosis, the HT approach remains relevant in situations where practical considerations necessitate a more straightforward, accessible solution.It is not a question of one approach being superior to the other but instead selecting the most appropriate method based on the specific requirements and constraints of the application.

Conclusions
This study presents the application of HT as a signal processing technique, utilizing a Pearson correlation-based statistical feature approach for classifying the condition of a transformer's core.The study evaluates the performance of various machine learning models on the transformer core current dataset collected during healthy and faulty conditions.The proposed method is compared under two scenarios: without any signal processing technique and when applying FFT.The results illustrate an improvement in the performance of six ML models, as evidenced by their performance metrics.Our current dataset can serve as a reference for future research on monitoring the transformer's core.For future works, we will look at improving the proposed model's accuracy and effectiveness with deep learning algorithms and vibration signal acquisition.
• Specifically designed to accommodate 1 A/5 A nominal CTs, ensuring compatibility and accuracy during measurements.• Channel-to-earth isolation of up to 300 Vrms and channel-to-channel CAT III isolation of 480 Vrms guarantees safety and accuracy during experimentation.• It has ring lug connectors tailored for up to 10 AWG cables, ensuring secure and reliable connections.• It operates within a wide temperature range, from −40 °C to 70 °C, and is engineered to withstand 5 g vibrations and 50 g shocks, ensuring stability and functionality across varying environmental conditions.

Figure 2 .
Figure 2. (a) Experimental testbed setup for transformer analysis.(b) Circuit diagram of the transformer core setup.

Figure 5 .
Figure 5. FFT of all working conditions.

Figure 6 .
Figure 6.HT of all working conditions.

Figure 8 .
Figure8.Plot of ML models' performance evaluation under three conditions: using raw data, using FFT, and using HT.

Table 1 .
Different types of transformer core fault.

Table 4 .
Machine learning models and parameter values.

Table 5 .
Performance evaluation for raw data.

Table 6 .
Performance evaluation for FFT.

Table 7 .
Performance evaluation for HT.