Next Article in Journal
Sliding Window-Based Randomized K-Fold Dynamic ANN for Next-Day Stock Trend Forecasting
Previous Article in Journal
Risk Assessment of Mud Cake on Shield Cutter Head Based on Modified Analytic Hierarchy Process
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Early Detection of Inter-Turn Short Circuits in Induction Motors Using the Derivative of Stator Current and a Lightweight 1D-ResNet

by
Carlos Javier Morales-Perez
1,
David Camarena-Martinez
2,
Juan Pablo Amezquita-Sanchez
1,
Jose de Jesus Rangel-Magdaleno
3,
Edwards Ernesto Sánchez Ramírez
4 and
Martin Valtierra-Rodriguez
1,*
1
ENAP-Research Group, CA-Sistemas Dinámicos, Facultad de Ingeniería, Universidad Autónoma de Querétaro (UAQ), Campus San Juan del Río, Río Moctezuma 249, Col. San Cayetano, San Juan del Río 76807, Queretaro, Mexico
2
ENAP-Research Group, División de Ingeniería, Universidad de Guanajuato (UG), Campus Irapuato-Salamanca, Carretera Salamanca-Valle de Santiago km 3.5 + 1.8 km, Comunidad de Palo Blanco, Salamanca 36885, Guanajuato, Mexico
3
Digital Systems Group, Coordinación de Electrónica, Instituto Nacional de Astrofísica, Óptica y Electrónica (INAOE), Luis Enrique Erro #1, Sta. María Tonanzintla, San Andrés Cholula 72840, Puebla, Mexico
4
Laboratorio de Procesamiento de Imagenes y Señales, ESIME Zacatenco, Instituto Politécnico Nacional (IPN), Unidad Profesional Adolfo López Mateos, Avenida Luis Enrique Erro S/N, UPALM, Alcaldía Gustavo A. Madero 07738, Mexico City, Mexico
*
Author to whom correspondence should be addressed.
Computation 2025, 13(6), 140; https://doi.org/10.3390/computation13060140
Submission received: 13 May 2025 / Revised: 30 May 2025 / Accepted: 2 June 2025 / Published: 4 June 2025
(This article belongs to the Special Issue Diagnosing Faults with Machine Learning)

Abstract

This work presents a lightweight and practical methodology for detecting inter-turn short-circuit faults in squirrel-cage induction motors under different mechanical load conditions. The proposed approach utilizes a one-dimensional convolutional neural network (1D-CNN) enhanced with residual blocks and trained on differentiated stator current signals obtained under different load mechanical conditions. This preprocessing step enhances fault-related features, enabling improved learning while maintaining the simplicity of a lightweight CNN. The model achieved classification accuracies above 99.16% across all folds in five-fold cross-validation and demonstrated the ability to detect faults involving as few as three short-circuited turns. Comparative experiments with the Multi-Scale 1D-ResNet demonstrate that the proposed method achieves similar or superior performance while significantly reducing training time. These results highlight the model’s suitability for real-time fault detection in embedded and resource-constrained industrial environments.

1. Introduction

Induction motors (IMs) are widely used in various industrial applications due to their robustness, cost-effectiveness, and low maintenance requirements. However, demanding operational conditions often lead to faults in different machine components, with stator faults accounting for approximately 37% of failures and inter-turn short-circuit (ITSC) faults representing 33% of stator failures [1,2]. These faults can significantly shorten the motor’s operational lifespan, reduce efficiency, and, in severe cases, lead to complete machine failure.
Conventional fault detection methods often rely on expert knowledge and manual signal analysis, which can be subjective and time-consuming. However, the automatic detection of stator faults has led researchers to develop various techniques based on different types of signals. One such approach involves methods based on the stator’s thermal behavior [3,4], which detect thermal anomalies caused by inter-turn faults. While this technique has proven effective in controlled environments, it can be influenced by other heat sources, limiting its applicability in real implementations. In addition, flux-based techniques have also been explored for ITSC detection. These can be categorized into two main approaches: air-gap flux analysis [5,6] and stray flux analysis [7,8]. While effective, these techniques present specific implementation challenges. Installing internal sensors may be unsafe due to potential mismatches with the machine’s design and construction. Conversely, externally installed sensors are highly susceptible to electromagnetic noise [9] coming from other machines, increasing the complexity of fault detection. Motor current signature analysis (MCSA) is an alternative to address these limitations. This approach enables fault detection by analyzing fault-related signatures in the motor current signal [10,11,12,13]. MCSA is widely used due to the ease of installing current sensors, which are less invasive than other techniques.
Fault diagnosis can be formulated as a pattern recognition problem. So, Artificial Intelligence (AI)-based approaches have proven effective and hold great potential for fault detection in rotating machinery applications [14]. For instance, Oner et al. [15] introduced a technique leveraging inverter switching statistics and an artificial neural network (ANN) for ITSC fault detection in inverter-fed induction motors. Their approach enables the detection of up to two short-circuited turns (SCTs) under various load conditions, achieving a maximum accuracy of 99.51%. While their method achieves a high accuracy for detecting up to two SCTs, it depends on access to the inverter’s internal signals, which may not be available in typical industrial environments. Another relevant study is that of Gundewar and Kane [16], which proposes a convolutional neural network (CNN)-based technique for ITSC fault detection using color images derived from stator current phase data. Their method attains accuracy of up to 99.38%, with a minimum detectable fault of two SCTs, under different supply frequency and mechanical load conditions. However, their image transformation approach adds significant preprocessing overhead. Similarly, Nazemi et al. [17] utilized the fundamental frequency phasor magnitude and the third harmonic, converting them into three-dimensional (3D) images processed by a two-dimensional (2D) CNN. Their method enables the detection of up to three SCTs with a maximum accuracy of 99.98%. Nevertheless, this approach can be sensitive to noise, and the leakage effects can negatively impact it. In addition, Rengifo et al. [18] proposed extracting diagnostic indicators from the magnitude of the space vector of the stator current signal. These indicators are evaluated using machine learning (ML) techniques, achieving up to 100% accuracy with k-Nearest Neighbor (k-NN) and Support Vector Machine (SVM) classifiers, and detecting incipient faults that affect as little as 4% of a phase. Although they reported high accuracy and detection, their method depends on constructing multiple indicators, which could hinder adaptation to changing operational conditions. Finally, Cardenas-Cornejo et al. [19] proposed an automatic ITSC detection approach based on geometric and optimization-based techniques applied to three-phase currents combined with ML classifiers. Their method achieves an accuracy of 95.3% across 13 fault classes and can detect damage affecting as little as 1.41% of the stator turns. However, the signal modeling and optimization techniques can be complex and computationally intensive. Despite the high classification accuracy reported by these AI-based approaches, extracting meaningful features from raw signals often involves complex signal processing techniques. These include transformations from the time domain to more complex representations, which can substantially increase the volume and dimensionality of the data to be analyzed. Consequently, more elaborate classification models and processing stages are typically required. This stage can be computationally intensive and may require expert domain knowledge, which poses a significant challenge for real-world deployment and scalability. In this manner, early fault identification remains a challenge due to the incipient nature of these failures.
On the other hand, the rise of CNNs across various domains has enabled the development of highly effective classification systems, particularly in image recognition tasks, including ITSC fault detection. However, since the signals involved in ITSC diagnosis are inherently one-dimensional (1D), transforming them into 2D or even 3D representations introduces additional preprocessing steps. This conversion increases the computational load and complicates the CNN architecture and the overall diagnostic approach. On the contrary, works such as those of Zheng et al. [20] and She et al. [21] have used 1D-CNN models to classify 1D signals, avoiding the additional steps described above and reducing the amount of data. The model is based on the Multi-Scale 1D-ResNet, whose structure enables the learning of complex features and achieves a high accuracy rate.
This paper presents a novel technique that employs the derivative of the stator current as a preprocessing stage to enhance the fault-related components associated with ITSC signatures. This signal transformation emphasizes the subtle variations introduced by the fault while preserving the temporal characteristics of the original signal. Unlike previous approaches that require converting one-dimensional signals into 2D or 3D formats, the proposed method operates directly on the 1D signal, significantly reducing preprocessing complexity and computational overhead. A 1D-CNN architecture, inspired by the Multi-Scale 1D-ResNet framework, is then applied for classification. The proposed methodology achieves an accuracy rate from 99.16% up to 100.00% accuracy under different mechanical load conditions, with a minimum detectable fault of three SCTs. Moreover, due to its reduced dimensionality and efficient architecture, the model is well suited for practical deployment in real time or embedded monitoring systems.
The main contributions of this paper are the following: (1) the use of a simple yet effective preprocessing step (the discrete derivative of the stator current) that significantly simplifies the signal processing stage by enhancing fault-related features directly in the time domain; (2) the implementation of a compact convolutional neural network architecture with residual blocks, which improves the learning capacity while maintaining a lightweight model and low computational cost compared to existing approaches; and (3) the detection of only the SCT fault at incipient stages (3 SCTs), regardless of the operative mechanical load.
The structure of this paper is as follows: Section 2 introduces the main concepts related to stator current signal processing explored in this work, including the derivative-based approach, ITSC fault characteristics, and the overview of the Residual Neural Network (ResNet) architecture. Section 3 presents the proposed methodology for ITSC fault detection, detailing the design and configuration of the CNN architecture. Section 4 describes the experimental procedures and presents the obtained results. Section 5 provides an in-depth discussion and analysis of the obtained results. Finally, Section 6 offers concluding remarks and outlines directions for future work.

2. Stator Current Signal Processing

The processing of the stator current signal has emerged as a powerful and non-invasive method for diagnosing electrical faults in induction motors, including ITSC. This signal reflects the internal behavior of the machine and can carry valuable fault-related information, making it a convenient choice for condition monitoring.
Among the most established techniques is MCSA, which focuses on detecting characteristic frequency components associated with specific fault types in the stator current spectrum. Stator winding faults can be grouped into turn-to-turn, phase-to-phase, and phase-to-ground faults. This work handles the ITSC fault, a kind of turn-to-turn fault. In this case, ITSC faults cause asymmetry through short-circuited turns that lead to the appearance of sideband frequencies near the fundamental frequency. These frequencies can be expressed as
f i t s c = f s a p ( 1 s ) ± b
where f i t s c is the fault-related frequency component, f s is the supply frequency, a is a positive integer, p is the number of pole pairs of the motor, s is the slip, and b is an odd index. It is essential to highlight that slips s directly affect the location of the fault-related frequency component, which consequently causes the fault frequency to shift either closer to or farther from the supply frequency f s . This dynamic behavior poses a significant challenge for fault detection, as the fault-related component may become masked by other electrical or mechanical phenomena. Furthermore, the increase in stator current caused by the fault can be easily misinterpreted as a natural consequence of increased mechanical loading, making it difficult to distinguish between healthy and faulty conditions based solely on amplitude variations.
The derivative of the stator current signal can be applied as a preprocessing stage to emphasize the components associated with ITSC faults. This approach enhances the visibility of subtle variations in the signal that are often masked under normal operating conditions. Since the stator current can be modeled as the summation of sinusoidal components, it can be expressed as
x ( t ) = i = 0 k 1 A i s i n ( ω i t + ϕ i )
where A i is the amplitude, ω i = 2 π f i is the angular frequency, and ϕ i is the phase of the i-th sine component. Taking the derivative of this signal with respect to time yields
d d t x ( t ) = i = 0 k 1 ω i A i c o s ( ω i t + ϕ i )
This operation amplifies the signal components (typically associated with fault-related phenomena) at a rate proportional to their angular frequency ω i while preserving the overall structure of the original signal. As a result, the derivative enhances the presence of ITSC-related features, making them more distinguishable in subsequent processing stages. Figure 1 illustrates the effect of applying the derivative on the signal spectrum, emphasizing the enhancement of the f i t s c component. Note how the amplitudes of f i t s c are amplified in the derivative.
However, a well-known drawback of differentiation is its tendency to amplify high-frequency noise along with the fault-related components, potentially complicating the classification task. The proposed method employs a CNN capable of directly learning robust, noise-tolerant features from the derivative signal to address this challenge. The CNN architecture acts as a powerful filter, effectively distinguishing between relevant fault signatures and unwanted noise, thus improving the overall reliability and accuracy of the detection system.

2.1. ResNet Overview

The ResNet architecture was first introduced by He et al. in 2015 [22], and it marked a breakthrough in deep learning by enabling the training of very deep architectures without suffering from the vanishing gradient problem. The key innovation behind ResNet is the use of identity shortcut connections, which allow the network to learn residual functions, i.e., the difference between the input and output of a layer, instead of learning complete transformations (see Figure 2). This seemingly simple idea profoundly impacted deep neural networks’ training stability and performance, leading to state-of-the-art results in various computer vision benchmarks such as ImageNet [23].
While initially developed for image classification using 2D convolutions, the residual learning concept has since been successfully adapted to 1D data, including time-series signals and biomedical waveforms [24,25]. Mathematically, it is expressed as follows:
y = F ( x ) + x
where F ( x ) represents the residual mapping (e.g., a stack of Conv-ReLU-BN layers). Furthermore, the residual architecture contributes to the method’s robustness by reducing the impact of noise (typically amplified during the derivative operation), thus improving the network’s ability to detect subtle ITSC faults under varying mechanical loads.
This CNN architecture comprises different key layers, each contributing uniquely to the network’s learning capacity and generalization performance. These layers are introduced below.

2.1.1. 1D Convolutional Layer

This convolutional layer is a feature extractor that slides a set of learnable filters (kernels) across the input signal along the time axis. For a given 1D input x R N , where N is the number of samples, the output feature map y ( k ) corresponding to the filter k is computed as
y i ( k ) = j = 0 K 1 x i + j w j ( k ) + b ( k )
where K is the kernel size; w j ( k ) , the weights; and b ( k ) , the bias. This process enables the model to detect local patterns, such as abrupt changes or distortions.

2.1.2. Batch Normalization

Batch normalization normalizes the output of each convolutional layer to have zero mean and unit variance, computed over each mini-batch. This operation improves training stability and convergence speed and is a regularizer, reducing the risk of overfitting. Its mathematical expression for an input x is presented as follows:
B N ( x i ) = γ x ^ i + β
where x ^ i = x i μ B σ B 2 + ϵ is the normalization of x i , μ B is the mini-batch mean, σ B 2 is the mini-batch variance, ϵ is an small value to avoid zero division, and γ and β are learning parameters that allow the network to rescale and displace the normalized output as needed.

2.1.3. ReLU Activation Function

The Rectified Linear Unit (ReLU) is a non-linear activation function defined as
R e L U ( x ) = max ( 0 , x )
where x represents an input value. ReLU introduces non-linearity into the network, which is crucial for learning complex representations. In the context of 1D signals, the network can focus on significant features (positive activations), suppressing non-informative or noisy regions (negative values are zeroed out). Additionally, ReLU is computationally efficient and helps mitigate the vanishing gradient problem, which can hamper learning in deep networks.

2.1.4. 1D MaxPooling and Average Pooling Layers

These layers perform temporal downsampling by reducing the dimensionality of the feature maps while retaining the most relevant information.
MaxPooling captures the most dominant feature in each window:
y i = max ( x i ( s ) , x i ( s + 1 ) , , x i ( s + k 1 ) )
where s is the stride, and k is the kernel size. Average pooling computes the average value:
y i = 1 k j = 0 k 1 x i s + j
Pooling layers improve computational efficiency and enhance translation invariance in time.

2.1.5. Softmax Activation Layer

Used in the output layer for classification, the softmax function transforms raw outputs into normalized class probabilities:
softmax ( z i ) = e z i j = 1 K e z j
where z = { z 1 , z 2 , , z K } . This enables the model to express its confidence in each class, with the highest probability corresponding to the predicted class.

3. Materials and Methods

This study employs a lightweight CNN architecture with a few 1D residual blocks. These blocks are inspired explicitly by [20,21] and are taken to process the derivative of the stator current signal. This representation enhances fault-related components while suppressing variations in the baseline signal. Including residual connections facilitates extracting relevant temporal features without requiring a deep or complex network, promoting efficient training and reducing the risk of overfitting. This makes the proposed approach particularly suitable for practical and resource-constrained applications.

3.1. Methodology

The proposed methodology (see Figure 3) begins with the stator current acquisition, and then this signal is derived. The goal of applying the derivative is to emphasize the amplitude variations of the harmonic components, particularly those associated with the ITSC fault signature. These variations evolve with a specific frequency rate, denoted by ω i , making fault detection more straightforward and reducing the need for complex signal processing stages.
For this purpose, the discrete derivative of the stator current signal is computed using the simple backward difference method [26,27], which is mathematically expressed as
d d t x [ n ] = x [ n ] x [ n 1 ] T
where x [ n ] R 1 × N represents the acquired discrete signal of length N, and T denotes the sampling period.
A centering and normalization stage is then applied to set the signal mean to 0 and scale the differentiated signal within the range of −1 to 1, based on the maximum absolute amplitude of the signal. Due to this processing, the division by T in Equation (11) becomes unnecessary. Subsequently, a compact CNN architecture is implemented, as shown in Figure 4.
The implemented lightweight 1D-ResNet architecture is described as follows.

Input Block

The input block is used to capture local temporal features from the differentiated signal. It begins with a 1D convolutional layer consisting of 64 filters and a kernel size of 7, using ‘same’ padding to preserve the input length. The relatively large kernel size enables the network to capture broader temporal patterns, while the higher number of filters facilitates the extraction of rich, low-level features.
L2 regularization with a coefficient of 0.001 is applied to the convolutional kernel to reduce overfitting and enhance generalization [28]. This layer is followed by batch normalization, which stabilizes training by reducing internal covariate shift [29]. The ReLU activation function introduces non-linearity and mitigates the vanishing gradient problem. Finally, a max-pooling layer with a pool size of 2 reduces the temporal resolution, enhancing translation invariance and lowering computational complexity [30].
This block is inspired by the multi-scale 1D-ResNet architecture used in [20,21] and has been adapted for compatibility with the proposed architecture. The implementation is illustrated in Figure 4b, and the configuration is shown in Table 1.

3.2. ResNet Blocks

The core of the architecture includes two residual blocks labeled ResNet_1 and ResNet_2 (Figure 4a). ResNet_1 follows the standard residual structure [22] and consists of two consecutive 1D convolutional layers, each followed by batch normalization and ReLU activation. Max pooling with a pool size of 2 is applied after each convolution to reduce the temporal dimension. The block concludes with an adding layer, combining the original input with the output of the second batch normalization layer, forming the residual connection and ultimately reaching an activation layer. This design facilitates better gradient propagation and improves the learning of complex temporal features.
ResNet_2 adopts a modified shortcut structure, where the addition operation is performed between the first convolutional layer’s output and the second layer’s batch normalization output, rather than between the original input and the output of the second layer. The 1D convolutional layers were constructed with 64 filters, and each kernel had a size of 7. This modification maintains simplicity and improves compatibility with the downstream layers, as shown in Figure 4c. Table 2 and Table 3 summarize the basic configuration used.

Output Block

The output block is designed to perform the final classification by summarizing the extracted temporal features and mapping them to fault categories. An average pooling layer is first applied to reduce the feature map’s dimensionality by computing the mean over the temporal axis. This operation retains the most representative information while reducing computational complexity and minimizing the risk of overfitting.
Finally, a dense layer with M units (corresponding to the number of classes) and a softmax activation function is used to output a probability distribution over the fault classes [31]. The average pooling layer serves as a global feature aggregator, making the model more robust to small temporal shifts and enhancing generalization, particularly when input sequences vary slightly in timing or amplitude. This block organization is depicted in Figure 4e, and its configuration is presented in Table 4.
It is essential to mention that the parameter selection, such as L2 regularization and pool size, follows the standard practices widely adopted in the literature [32,33]. The kernel size and number of filters utilized for the convolutional layers in the ResNet blocks are taken from the recommendations in [20,21,33].

3.3. Experimental Setup

The proposed methodology was validated using a laboratory test bench composed of an IM mechanically coupled to a dynamometer, which emulated the mechanical load during operation. The specifications of the components are summarized as follows:
  • The IM under test was a 2 HP three-phase motor, model 218ET3EM145TW from WEG (Nantong, China), rated at 220 VAC and 60 Hz. The stator winding consisted of 141 turns per phase.
  • A Four-Quadrant Dynamometer, model 8540 from Lab-Volt (Eatontown, NJ, USA), rated at 2 kW, was used to simulate and control the applied mechanical load.
To emulate ITSC faults, one phase coil of the stator was deliberately modified to create controlled SCTs. Eight fault scenarios were generated, corresponding to 0, 3, 5, 10, 15, 20, 30, and 40 shorted turns, as illustrated in Figure 5.
The stator current was measured using three current clamps model i200s from Fluke (Tokyo, Japan), one per phase. The analog current signals were digitized using a National Instruments NI-USB-6211 data acquisition system (DAS) and transmitted to a personal computer for processing and storage. The sampling rate of the DAS was set to 6000 samples per second. A standard motor starter was used to start and stop the motor during experiments.
The overall experimental setup, including the data flow and instrumentation, is presented in Figure 6a, and an image of the physical test bench is shown in Figure 6b.

4. Test and Results

The tests were conducted to record the stator current signal under steady-state conditions for 3.5 s per measurement. This duration was selected to prevent severe and irreversible damage to the stator during fault simulations.
To evaluate the proposed method under varying mechanical loads, the dynamometer was configured to apply torques of 0.00, 2.04, 4.09, and 6.13 Nm, corresponding to 0.00%, 33.33%, 66.66%, and 100.00% of the nominal mechanical load, respectively. Combined with the eight different fault conditions (including the healthy case), a total of 32 distinct operating scenarios were generated.
For each scenario, approximately 70 min (in total) of stator current data were acquired at the sampling rate configured in the DAS. The signals were segmented into one-eighth of a second windows to construct the dataset, corresponding to N = 750 samples per segment in the differentiated signal. This segmentation yielded 18,560 signal segments, forming a comprehensive dataset for training and evaluation purposes.
Figure 7 presents a qualitative comparison between the raw stator current signals (blue) and their corresponding derivatives (orange) for different fault severities, 0, 10, 20, 30, and 40 SCT, considering the configuration and conditions previously described. Figure 7a corresponds to the no-load condition, while Figure 7b shows signals under a mechanical load of 4.09 Nm. As the number of short-circuited turns (SCTs) increases, the discrete derivative of the stator current signal reveals sharper and more localized variations in the signal crest compared to the raw current waveform. These variations are especially prominent in amplitude modulation and dynamic frequency components, which tend to intensify with fault severity. The enhanced clarity of these fault-induced features facilitates the extraction of discriminative patterns, improving the classifier’s sensitivity to subtle incipient faults. Consequently, using the current derivative as input to the CNN enhances the network’s ability to detect ITSC faults, as it emphasizes spectral and temporal patterns often obscured in the original time-domain signal.
A standard 80/20 split was applied to divide the dataset into training and testing subsets. The choice of one-eighth-second segments preserved the spectral characteristics associated with fault signatures while maintaining compatibility with the lightweight architecture of the CNN. In addition, five-fold cross-validation was implemented to reduce selection bias and enhance the reliability of the evaluation [34].
A group of eight classes ( M = 8 ) was defined based on the level of the ITSC fault under study. To ensure that the CNN learned to identify fault-related features independently of mechanical load variations, segments corresponding to the same fault level but different load conditions were grouped into the same class. This approach aims to train the model to focus exclusively on fault characteristics, thereby enabling robust fault detection regardless of variations in load. A summary of the implemented CNN and signal configuration is provided in Table 5. It is important not to confuse the “Input block” described in Table 5 with the input layer of the CNN (embedded within the input block; see Table 1), whose shape is ( 750 , 1 ) .
The model was trained for 100 epochs with a batch size of 20. This batch size was selected as a trade-off between training efficiency, memory usage, and execution time. The AdaMax optimizer (https://keras.io/api/optimizers/adamax/, accessed on 1 June 2025) was employed due to its robustness against noisy gradients, which may arise from the differentiated signal. The results obtained from five-fold cross-validation are presented in Table 6.
The five-fold cross-validation demonstrates that the proposed model consistently achieves high classification accuracy, with test accuracies ranging from 99.16% to 100%. The corresponding training and test losses remain low, indicating effective learning without signs of overfitting. Notably, the fifth fold yielded a perfect classification accuracy of 100%, further highlighting the robustness of the proposed approach.
Figure 8 illustrates the training process in terms of accuracy and loss over 100 epochs. The training curves show rapid convergence, with accuracy stabilizing around epoch 70, and minimal divergence between training and testing performance, confirming the generalization capability of the CNN.
To provide a clearer picture of the classifier’s performance, confusion matrices for the worst and best cross-validation folds are presented in Figure 9. In both cases, the confusion matrices exhibit strong diagonal dominance, confirming the model’s ability to accurately identify all eight ITSC fault classes. In the worst-performing scenario (Figure 9a), a total of 31 signals were misclassified. Specifically, 2 healthy signals under no mechanical load (0.00 Nm) were incorrectly classified as having three SCTs; 1 instance of a five-SCT fault at 2.04 Nm was classified as three SCTs; 18 signals with a three-SCT fault at 4.09 Nm were misclassified as five SCTs; and 10 signals of three SCTs at 6.13 Nm were also incorrectly identified as five SCTs. Among these, the most challenging condition in this fold corresponds to the 4.09 Nm mechanical load, for which the classification accuracy for three SCT faults dropped to 84.48% (98 correctly classified out of 116 signals). However, the few misclassifications in the worst-case scenario are minor and primarily occur between adjacent fault levels, which is expected in real-world fault detection scenarios due to the similarity in signal characteristics, and are associated with the most minor detected fault. These results collectively demonstrate the model’s robustness and its suitability for accurately detecting ITSC faults, regardless of mechanical load variations.
In addition, to further evaluate the performance of the proposed methodology, four testing scenarios were developed to compare the proposed approach with the Multi-Scale 1D-ResNet. Scenario A involves classifying centered and normalized time-series data without applying the derivative, using a segment length of N = 512 , as recommended in the Multi-Scale 1D-ResNet. Scenario B incorporates the proposed preprocessing, including the application of the derivative, but also uses N = 512 . Scenario C uses the same preprocessing as Scenario A but increases the segment length to N = 750 . Finally, Scenario D corresponds to the complete configuration proposed in this work. The results of these scenarios are summarized in Table 7. The results demonstrate that the proposed preprocessing stage, based on the discrete derivative of the stator current, effectively enhances the spectral signatures associated with ITSC faults. As illustrated in Figure 7, this processing generates more pronounced and distinguishable patterns than raw signals, particularly as fault severity and mechanical load increase. These clearer spectral features facilitate early-stage fault detection and improve the CNN classifier’s ability to distinguish between fault levels accurately.
The experiments were implemented in Python 3.10 using TensorFlow 2.10 and the Keras 2.10 library. The computational tests were conducted on a laptop model G15 5511 from Dell equipped with an Intel Core i5-11260H @ 2.60 GHz processor, 16 GB of RAM, and an GPU model GeForce RTX 3050 from NVIDIA with 4 GB of dedicated memory. GPU acceleration was enabled following the official setup guidelines provided by the TensorFlow web page [35].

5. Discussion

Most existing works on ITSC fault detection in induction motors rely on 2D or 3D CNN architectures, which require complex preprocessing and higher computational resources. In contrast, this study proposes a lightweight 1D-CNN model tailored for time-series signals, significantly reducing complexity while maintaining competitive accuracy. The architecture facilitates fast training and is suitable for deployment on non-specialized hardware, making it more practical for real-world applications. Furthermore, the proposed model outperforms or matches the performance of a larger model, the Multi-Scale 1D-ResNet, which inspired this approach.
The results presented in Table 6 and Table 7 and Figure 8 demonstrate the effectiveness of the proposed method for detecting ITSC faults under various mechanical load conditions. The CNN-based model achieved classification accuracies exceeding 99.16% across all five folds of the cross-validation, confirming both its reliability and generalization capacity.
A key element in this performance is the use of differentiated current signals, which enhances fault-related features by emphasizing sharper transitions. Although differentiation may introduce additional noise into the signal, this issue was effectively addressed by employing the AdaMax optimizer. As a robust variant of the Adam algorithm, AdaMax proved remarkably resilient to noisy gradients, contributing to stable and smooth convergence during training.
Another necessary strength of the proposed approach lies in its sensitivity to low-severity faults. The model successfully identified cases with as few as three SCTs, demonstrating its potential for early fault detection. This capability is essential for predictive maintenance applications, where timely intervention can significantly reduce the likelihood of severe motor damage or costly downtime.
Furthermore, the model was designed to generalize across varying mechanical load conditions. By grouping data from different load levels into single classes based on fault severity, the CNN was encouraged to learn features that are strictly fault-related rather than load-dependent. This strategy enhances the model’s robustness and adaptability to real-world industrial environments, where load conditions often fluctuate.
As shown in Table 7, Scenario D of the proposed approach achieved perfect classification accuracy (100.00%) while requiring only 19.68 min of training, substantially less than the 91.12 min needed for the Multi-Scale 1D-ResNet to reach a slightly lower accuracy (99.95%). Notably, even Scenario B of the proposed model, which used derivative preprocessing with shorter signal segments ( N = 512 ), reached an accuracy of 99.95% in just 16.30 min. These findings emphasize not only the effectiveness but also the relatively low complexity of the proposed solution.
Table 8 provides a comparative overview of recent ITSC fault detection methods. The proposed approach combines the stator current derivative with a lightweight 1D-CNN architecture based on ResNet blocks, achieving an accuracy of 100.00% while detecting faults with as few as three short-circuited turns under four different mechanical load conditions. In comparison, the method by Gundewar and Kane [16] reaches 99.38% accuracy for two SCTs using a computationally intensive 3D-CNN and image transformations of phase currents. Similarly, Nazemi et al. [17] employed 2D-CNNs with harmonic features and digital filtering to detect three SCTs with 99.98% accuracy across a wide range of loads. Other works, such as [18,36], relied on classical signal processing techniques that involve more complex preprocessing stages. Also, while innovative in applying quaternion analysis, the method by Cardenas-Cornejo et al. [37] requires six SCTs to reach 99.00% accuracy in no-load conditions. In contrast, the proposed method achieves similar or superior performance with significantly lower complexity, making it a practical and efficient alternative.
It is important to note that the testing environments were developed under steady-state conditions using three-phase, 2 HP induction motors. Nevertheless, the proposed approach is scalable and can be applied to induction motors of different power ratings. This is supported by Equation (1), which indicates that the fault signature is directly related to the spectral characteristics of the stator current, rather than the power of the machine. Future research will explore extending this methodology to scenarios involving voltage fluctuations and the detection of ITSC faults in non-induction motor types, aiming to further assess its generalization capabilities. In addition, the nature of the studied problem, combined with the inherent plasticity of the employed ResNet-based architecture, as well as the use of fine-tuning and transfer learning techniques, opens the possibility of extending the proposed approach to continuous learning applications, as addressed by Wang et al. [38] and Ren et al. [39]. This adaptability suggests strong potential for detecting other types of stator faults and broader applications in industrial rotating machinery, including diagnosing other failure modes and applications in prognosis or adapting to different industrial environment settings, as Chen et al. introduce in [40]. However, this will be explored in future works.

6. Conclusions

This work presented a lightweight CNN and robust methodology for detecting ITSC faults in squirrel-cage induction motors. The proposed approach combines a derivative-based preprocessing of stator current signals with a custom-designed CNN architecture, achieving accurate fault classification under various mechanical load conditions.
The model demonstrated strong performance, achieving an accuracy of over 99.16% across five-fold cross-validation. It also proved capable of detecting incipient faults as subtle as three SCTs, which is critical for early intervention in predictive maintenance schemes. The use of differentiated signals highlighted fault-related spectral features, thereby improving CNN training efficiency.
Compared with the well-established Multi-Scale 1D-ResNet architecture, the proposed method achieved comparable or better accuracy with significantly lower training times, up to 4.6 times faster. This efficiency positions the technique as an excellent candidate for real-time deployment in embedded systems or industrial applications with limited computational resources.
Future work will focus on validating the model with real-time data acquisition systems and extending the framework to handle other types of motor faults, such as rotor bar breakage or eccentricity, or continuous learning applications. Furthermore, a detailed investigation will be conducted to assess the effectiveness of the methodology in detecting even more incipient faults (involving fewer than three short-circuited turns) as well as under non-ideal electrical conditions, such as voltage fluctuations or unbalanced supply conditions. These extensions could improve the model’s generalization capabilities and broaden its applicability in real-world industrial environments.

Author Contributions

Conceptualization and methodology: C.J.M.-P., J.d.J.R.-M., J.P.A.-S. and M.V.-R.; software and validation: C.J.M.-P., D.C.-M. and E.E.S.R.; formal analysis, investigation, resources, and data curation: C.J.M.-P., J.d.J.R.-M. and E.E.S.R.; writing—review and editing: all authors; supervision, project administration, and funding acquisition: D.C.-M., J.P.A.-S. and M.V.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The author, C.J. Morales-Perez, thanks the “Secretaría de Ciencias, Humanidades, Tecnología e Innovación (SECIHTI)—México” for supporting a postdoctoral stay at the “Universidad Autonoma de Queretaro (UAQ, México)”. The authors would like to thank the (SECIHTI)—México and the “Sistema Nacional de Investigadoras e Investigadores (SNII)–SECIHTI–México” for their support in this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sheikh, M.A.; Bakhsh, S.T.; Irfan, M.; Nor, N.b.M.; Nowakowski, G. A Review to Diagnose Faults Related to Three-Phase Industrial Induction Motors. J. Fail. Anal. Prev. 2022, 22, 1546–1557. [Google Scholar] [CrossRef]
  2. Gyftakis, K.N. A Comparative Investigation of Interturn Faults in Induction Motors Suggesting a Novel Transient Diagnostic Method Based on the Goerges Phenomenon. IEEE Trans. Ind. Appl. 2022, 58, 304–313. [Google Scholar] [CrossRef]
  3. Adouni, A.; Cardoso, A.J.M. Thermal Analysis of Low-Power Three-Phase Induction Motors Operating under Voltage Unbalance and Inter-Turn Short Circuit Faults. Machines 2021, 9, 2. [Google Scholar] [CrossRef]
  4. Wu, Y.H.; Liu, M.Y.; Song, H.; Li, C.; Yang, X.L. A Temperature and Magnetic Field-Based Approach for Stator Inter-Turn Fault Detection. IEEE Sens. J. 2022, 22, 17799–17807. [Google Scholar] [CrossRef]
  5. Im, S.H.; Gu, B.G. Study of Induction Motor Inter-Turn Fault Part I: Development of Fault Models with Distorted Flux Representation. Energies 2022, 15, 894. [Google Scholar] [CrossRef]
  6. Ray, S.; Dey, D. Development of a Comprehensive Analytical Model of Induction Motor Under Stator Interturn Faults Incorporating Rotor Slot Harmonics. IEEE Trans. Ind. Electron. 2023, 70, 2037–2047. [Google Scholar] [CrossRef]
  7. Gyftakis, K.N.; Cardoso, A.J.M. Reliable Detection of Stator Interturn Faults of Very Low Severity Level in Induction Motors. IEEE Trans. Ind. Electron. 2021, 68, 3475–3484. [Google Scholar] [CrossRef]
  8. Zorig, A.; Hedayati Kia, S.; Chouder, A.; Rabhi, A. A comparative study for stator winding inter-turn short-circuit fault detection based on harmonic analysis of induction machine signatures. Math. Comput. Simul. 2022, 196, 273–288. [Google Scholar] [CrossRef]
  9. Mazaheri-Tehrani, E.; Faiz, J. Airgap and stray magnetic flux monitoring techniques for fault diagnosis of electrical machines: An overview. IET Electr. Power Appl. 2022, 16, 277–299. [Google Scholar] [CrossRef]
  10. Mejia-Barron, A.; Tapia-Tinoco, G.; Razo-Hernandez, J.R.; Valtierra-Rodriguez, M.; Granados-Lieberman, D. A neural network-based model for MCSA of inter-turn short-circuit faults in induction motors and its power hardware in the loop simulation. Comput. Electr. Eng. 2021, 93, 107234. [Google Scholar] [CrossRef]
  11. Bahgat, B.H.; Elhay, E.A.; Elkholy, M.M. Advanced fault detection technique of three phase induction motor: Comprehensive review. Discov. Electron. 2024, 1, 9. [Google Scholar] [CrossRef]
  12. Agah, G.R.; Rahideh, A.; Faradonbeh, V.Z.; Kia, S.H. Stator Winding Interturn Short-Circuit Fault Modeling and Detection of Squirrel-Cage Induction Motors. IEEE Trans. Transp. Electrif. 2024, 10, 5725–5734. [Google Scholar] [CrossRef]
  13. Ghanbari, T.; Mehraban, A.; Farjah, E. Inter-turn fault detection of induction motors using a method based on spectrogram of motor currents. Measurement 2022, 205, 112180. [Google Scholar] [CrossRef]
  14. Wang, G.; Zhao, Y.; Zhang, J.; Ning, Y. A Novel End-To-End Feature Selection and Diagnosis Method for Rotating Machinery. Sensors 2021, 21, 2056. [Google Scholar] [CrossRef]
  15. Oner, M.U.; Sahin, I.; Keysan, O. Neural Networks Detect Inter-Turn Short Circuit Faults Using Inverter Switching Statistics for a Closed-Loop Controlled Motor Drive. IEEE Trans. Energy Convers. 2023, 38, 2387–2395. [Google Scholar] [CrossRef]
  16. Gundewar, S.K.; Kane, P.V. Sensitive Inter-turn Fault Detection Approach for Induction Motor Under Various Operating Conditions. Arab. J. Sci. Eng. 2023, 48, 10787–10801. [Google Scholar] [CrossRef]
  17. Nazemi, M.; Liang, X.; Haghjoo, F. Convolutional Neural Network-Based Online Stator Inter-Turn Faults Detection for Line-Connected Induction Motors. IEEE Trans. Ind. Appl. 2024, 60, 4693–4707. [Google Scholar] [CrossRef]
  18. Rengifo, J.; Moreira, J.; Vaca-Urbano, F.; Alvarez-Alvarado, M.S. Detection of Inter-Turn Short Circuits in Induction Motors Using the Current Space Vector and Machine Learning Classifiers. Energies 2024, 17, 2241. [Google Scholar] [CrossRef]
  19. Cardenas-Cornejo, J.J.; Almanza-Ojeda, D.L.; Gonzáalez-Parada, A.; Hernandez-Ramirez, V.; Ibarra-Manzano, M.A. Complex Signal Analysis for Inter-Turn Short-Circuits Faults on Induction Motors. IEEE Sens. J. 2025, 25, 13433–13440. [Google Scholar] [CrossRef]
  20. Zheng, X.; She, S.; Xia, Z.; Xiong, L.; Zou, X.; Yu, K.; Guo, R.; Zhu, R.; Zhang, Z.; Yin, W. Analyzing the permeability distribution of multilayered specimens using pulsed eddy-current testing with multi-scale 1D-ResNet. NDT E Int. 2025, 149, 103247. [Google Scholar] [CrossRef]
  21. She, S.; Zheng, X.; Zou, X.; Yu, K.; Shen, J.; Wu, F.; Yin, W. Simultaneous carbon fiber layer thickness and direction measurement and identification using a novel eddy current sensor and simplified multi-scale 1D-ResNet network. Measurement 2025, 242, 115812. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar] [CrossRef]
  23. Banerjee, S.; Paik, J.H. A Deterministic-Probabilistic Approach to Neural Network Pruning. IEEE Trans. Artif. Intell. 2025, 1–10. [Google Scholar] [CrossRef]
  24. Tchatchoua, P.; Graton, G.; Ouladsine, M.; Christaud, J.F. Application of 1D ResNet for Multivariate Fault Detection on Semiconductor Manufacturing Equipment. Sensors 2023, 23, 9099. [Google Scholar] [CrossRef] [PubMed]
  25. Li, W.; Gao, J. Automatic sleep staging by a hybrid model based on deep 1D-ResNet-SE and LSTM with single-channel raw EEG signals. PeerJ Comput. Sci. 2023, 9, e1561. [Google Scholar] [CrossRef]
  26. Lewandowski, M. Estimating the first and second derivatives of discrete audio data. EURASIP J. Audio Speech, Music Process. 2024, 2024, 31. [Google Scholar] [CrossRef]
  27. Lubich, C.; Mansour, D.; Venkataraman, C. Backward difference time discretization of parabolic differential equations on evolving surfaces. IMA J. Numer. Anal. 2013, 33, 1365–1385. [Google Scholar] [CrossRef]
  28. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 27 May 2025).
  29. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar] [CrossRef]
  30. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  31. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006; Volume 4. [Google Scholar]
  32. Kuhn, M.; Johnson, K. Applied predictive modeling; Springer: New York, NY, USA, 2013; Volume 26. [Google Scholar]
  33. Wang, Z.; Yan, W.; Oates, T. Time Series Classification from Scratch with Deep Neural Networks: A Strong Baseline. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017. [Google Scholar]
  34. Marcot, B.G.; Hanea, A.M. What is an optimal value of k in k-fold cross-validation in discrete Bayesian network analysis? Comput. Stat. 2021, 36, 2009–2031. [Google Scholar] [CrossRef]
  35. TensorFlow. Build from Source on WINDOWS. 2021. Available online: https://www.tensorflow.org/install/source_windows (accessed on 1 May 2025).
  36. Saucedo-Dorantes, J.J.; Jaen-Cuellar, A.Y.; Perez-Cruz, A.; Elvira-Ortiz, D.A. Detection of Inter-Turn Short Circuits in Induction Motors under the Start-Up Transient by Means of an Empirical Wavelet Transform and Self-Organizing Map. Machines 2023, 11, 958. [Google Scholar] [CrossRef]
  37. Cardenas-Cornejo, J.J.; Ibarra-Manzano, M.A.; González-Parada, A.; Castro-Sanchez, R.; Almanza-Ojeda, D.L. Classification of inter-turn short-circuit faults in induction motors based on quaternion analysis. Measurement 2023, 222, 113680. [Google Scholar] [CrossRef]
  38. Wang, T.; Liu, H.; Guo, D.; Sun, X.M. Continual Residual Reservoir Computing for Remaining Useful Life Prediction. IEEE Trans. Ind. Inform. 2024, 20, 931–940. [Google Scholar] [CrossRef]
  39. Ren, X.; Qin, Y.; Li, B.; Wang, B.; Yi, X.; Jia, L. A core space gradient projection-based continual learning framework for remaining useful life prediction of machinery under variable operating conditions. Reliab. Eng. Syst. Saf. 2024, 252, 110428. [Google Scholar] [CrossRef]
  40. Chen, Z.; Huang, H.Z.; Deng, Z.; Wu, J. Shrinkage mamba relation network with out-of-distribution data augmentation for rotating machinery fault detection and localization under zero-faulty data. Mech. Syst. Signal Process. 2025, 224, 112145. [Google Scholar] [CrossRef]
Figure 1. Frequency spectra of the stator current and its derivative under healthy and faulty conditions. (a) Spectrum of the healthy stator current signal (blue plot) and the spectrum of the stator current signal with 10 SCT (orange plot). (b) Spectrum of the derivative of the healthy stator current (blue plot) and the spectrum of the derivative of the faulty stator current (orange plot).
Figure 1. Frequency spectra of the stator current and its derivative under healthy and faulty conditions. (a) Spectrum of the healthy stator current signal (blue plot) and the spectrum of the stator current signal with 10 SCT (orange plot). (b) Spectrum of the derivative of the healthy stator current (blue plot) and the spectrum of the derivative of the faulty stator current (orange plot).
Computation 13 00140 g001
Figure 2. ResNet block [22].
Figure 2. ResNet block [22].
Computation 13 00140 g002
Figure 3. Block diagram of the proposed methodology.
Figure 3. Block diagram of the proposed methodology.
Computation 13 00140 g003
Figure 4. Proposed architecture. (a) General architecture overview; (b) input block; (c) ResNet_1 block; (d) ResNet_2 block; and (e) output block.
Figure 4. Proposed architecture. (a) General architecture overview; (b) input block; (c) ResNet_1 block; (d) ResNet_2 block; and (e) output block.
Computation 13 00140 g004
Figure 5. Example of simulated SCT for the ITSC fault detection methodology.
Figure 5. Example of simulated SCT for the ITSC fault detection methodology.
Computation 13 00140 g005
Figure 6. Implemented test bench: (a) block diagram, and (b) real implementation.
Figure 6. Implemented test bench: (a) block diagram, and (b) real implementation.
Computation 13 00140 g006
Figure 7. Comparison between the raw stator current signal (blue) and its derivative (orange) for fault severities of 0, 10, 20, 30, and 40 SCT, respectively. (a) No-load condition; (b) 4.09 Nm mechanical load. Note that the derivative signals were scaled for visual comparison with the raw current signals.
Figure 7. Comparison between the raw stator current signal (blue) and its derivative (orange) for fault severities of 0, 10, 20, 30, and 40 SCT, respectively. (a) No-load condition; (b) 4.09 Nm mechanical load. Note that the derivative signals were scaled for visual comparison with the raw current signals.
Computation 13 00140 g007
Figure 8. Plot result of CNN training process: (a) accuracy and (b) loss.
Figure 8. Plot result of CNN training process: (a) accuracy and (b) loss.
Computation 13 00140 g008
Figure 9. Confusion matrix obtained: (a) worst result and (b) best results.
Figure 9. Confusion matrix obtained: (a) worst result and (b) best results.
Computation 13 00140 g009
Table 1. Summary of input block.
Table 1. Summary of input block.
LayerBasic Configuration
Inputshape = (N, 1)
Conv 1Dfilters = 64, kernel_size = 7, kernel_regularizer = L2 (0.001)
Batch normalization
Activationactivation = ‘relu’
Max pooling 1Dpool_size = 2
Table 2. Summary of ResNet_1 block.
Table 2. Summary of ResNet_1 block.
LayerBasic Configuration
Conv 1Dfilters = 64, kernel_size = 7, padding = ‘same’, kernel_regularizer = L2(0.001)
Batch normalization
Activationactivation = ‘relu’
Conv 1Dfilters = 64, kernel_size = 7, padding = ‘same’, kernel_regularizer = L2 (0.001)
Batch normalization (BN)
AddInput + BN output
Activationactivation = ‘relu’
Table 3. Summary of ResNet_2 block.
Table 3. Summary of ResNet_2 block.
LayerBasic Configuration
Conv 1D (C1D)filters = 128, kernel_size = 7, padding = ‘same’, kernel_regularizer = L2 (0.001)
Batch normalization
Activationactivation = ‘relu’
Conv 1Dfilters = 128, kernel_size = 7, padding = ‘same’, kernel_regularizer = L2 (0.001)
Batch normalization (BN)
AddC1D output + BN output
Activationactivation = ‘relu’
Table 4. Summary of output block.
Table 4. Summary of output block.
LayerBasic Configuration
Average Pooling
Denseunits = M, activation = ‘softmax’
Table 5. Summary of CNN implementation.
Table 5. Summary of CNN implementation.
BlockOutput ShapeNo. of Parameters
Input(372, 64)768
ResNet 1(372, 64)57,984
ResNet 2(372, 128)173,312
Output81032
Table 6. Summary of CNN implementation results for the proposed model.
Table 6. Summary of CNN implementation results for the proposed model.
Cross-ValidationAccuracy (%)Loss (%)
TrainTestTrainTest
199.7699.92 8.100 × 10 3 2.608 × 10 3
299.7199.16 9.212 × 10 3 23.217 × 10 3
399.9599.97 1.981 × 10 3 908.306 × 10 6
499.8699.73 4.112 × 10 3 8.214 × 10 3
599.89100.00 4.076 × 10 3 679.589 × 10 6
Table 7. Comparison results between the proposed model and the Multi-Scaled 1D-ResNet.
Table 7. Comparison results between the proposed model and the Multi-Scaled 1D-ResNet.
ProposalMulti-Scaled 1D-ResNet
ABCDABCD
Accuracy (%)87.6199.9590.17100.0097.28100.0099.8699.95
Training time (min)16.1316.3019.6919.6870.5569.7990.6291.12
Table 8. Comparison of recent ITSC fault detection approaches.
Table 8. Comparison of recent ITSC fault detection approaches.
WorkTechniqueLoad (%)Minimal SCTAccuracy (%)
Proposed(a) Stator current derivative
(b) 1D-CNN with ResNet blocks
0.00, 33.33, 66.66, and 1003100.00
Gundewar and Kane [16](a) Current-to-image transformation
(b) Phase current
(c) 3D-CNN
0, 25, 50, 75, and 100299.38
Nazemi et al. [17](a) Current to image transformation
(b) Fundamental frequency phasor magnitude
(c) Third harmonic component
(d) Digital Fourier filtering
(e) 2D-CNN
0, 5, 15, 20, 30, 40, 50, 60, 70, 80, 90, and 100399.98
Saucedo-Dorantes et al. [36](a) Empirical Wavelet Transform
(b) Self-Organizing Map
Not specified2100.00
Rengifo et al. [18](a) Magnitude of space vector of stator current
(b) Fundamental frequency phasor magnitude
(c) Random Forest
(d) Forward Neural Networks
(e) Recurrent Neural Networks
(f) K-NN and SVM
Not specified4 (%)100.00
Cardenas-Cornejo et al. [37](a) Statistical features
(b) Quaternion analysis
(c) Decision Tree Models
No load699.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Morales-Perez, C.J.; Camarena-Martinez, D.; Amezquita-Sanchez, J.P.; Rangel-Magdaleno, J.d.J.; Ramírez, E.E.S.; Valtierra-Rodriguez, M. Early Detection of Inter-Turn Short Circuits in Induction Motors Using the Derivative of Stator Current and a Lightweight 1D-ResNet. Computation 2025, 13, 140. https://doi.org/10.3390/computation13060140

AMA Style

Morales-Perez CJ, Camarena-Martinez D, Amezquita-Sanchez JP, Rangel-Magdaleno JdJ, Ramírez EES, Valtierra-Rodriguez M. Early Detection of Inter-Turn Short Circuits in Induction Motors Using the Derivative of Stator Current and a Lightweight 1D-ResNet. Computation. 2025; 13(6):140. https://doi.org/10.3390/computation13060140

Chicago/Turabian Style

Morales-Perez, Carlos Javier, David Camarena-Martinez, Juan Pablo Amezquita-Sanchez, Jose de Jesus Rangel-Magdaleno, Edwards Ernesto Sánchez Ramírez, and Martin Valtierra-Rodriguez. 2025. "Early Detection of Inter-Turn Short Circuits in Induction Motors Using the Derivative of Stator Current and a Lightweight 1D-ResNet" Computation 13, no. 6: 140. https://doi.org/10.3390/computation13060140

APA Style

Morales-Perez, C. J., Camarena-Martinez, D., Amezquita-Sanchez, J. P., Rangel-Magdaleno, J. d. J., Ramírez, E. E. S., & Valtierra-Rodriguez, M. (2025). Early Detection of Inter-Turn Short Circuits in Induction Motors Using the Derivative of Stator Current and a Lightweight 1D-ResNet. Computation, 13(6), 140. https://doi.org/10.3390/computation13060140

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop