You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

30 January 2025

Deep Learning Algorithm for Automatic Classification of Power Quality Disturbances

and
1
College of Information Technology, University of Bahrain, Sakhir P.O. Box 32038, Bahrain
2
College of Engineering, University of Bahrain, Sakhir P.O. Box 32038, Bahrain
*
Author to whom correspondence should be addressed.
This article belongs to the Section Computing and Artificial Intelligence

Abstract

Power quality disturbances (PQDs) are major obstacles to maintaining the reliability and stability of electrical systems. This study introduces a new multi-scale deep learning method to classify PQDs, aiming to enhance the accuracy and efficiency of power quality (PQ) analysis and monitoring systems. By combining 1-D convolutional neural networks (CNNs) with an attention mechanism, this approach overcomes the limitations of traditional techniques. Moreover, varying-size convolutional layers allow for the direct learning of complex patterns and features from PQ signals. To address the challenge of limited labeled PQ datasets, this research utilizes an open-source dataset generator to create large-scale datasets with annotated PQDs. Through a comparison with existing models in the field, the superiority of the proposed CNN-based approach is evident, achieving an accuracy level of up to 99.49 % . The results demonstrate promising classification performance in terms of simplicity and accuracy, highlighting the potential of this approach to improve PQ analysis and disturbance identification.

1. Introduction

The increasing reliance on electrical power for various aspects of modern society has made PQ a critical concern. Maintaining high PQ is essential to ensure the reliable operation of electrical devices and minimize the risk of equipment failure or power system instability. PQ events, such as voltage sags, swells, and transients, can have significant negative impacts on the performance and lifespan of electrical equipment. Therefore, early identification and mitigation of PQ issues is crucial.
Traditional methods for detecting, classifying, and localizing PQ events have limitations in terms of accuracy, robustness, and computational efficiency. Recent advancements in machine learning (ML) and artificial intelligence (AI) offer promising solutions to overcome these limitations [1,2]. This research aims to develop and evaluate a novel ML algorithm for the automatic classification of PQ events. By leveraging the power of advanced data-driven techniques, the proposed research seeks to improve the monitoring and maintenance of PQ in modern power systems, ultimately enhancing the reliability and stability of the electrical grid.
PQ, a measure of the consistency and reliability of the voltage, current, and frequency of an electrical power system, is an essential factor in the operation of modern power systems [3,4]. The growing integration of renewable energy sources, proliferation of non-linear loads, and increasing reliance on sensitive electronic devices have resulted in increased susceptibility to PQ issues. PQ events, such as voltage sags, swells, transients, harmonics, and flickers, can adversely affect the performance, safety, and lifespan of electrical equipment, leading to increased energy losses, equipment downtime, and maintenance costs. In extreme cases, these events can even cause power system instability or failure, posing risks to public safety and the economy [5,6,7].
Recent advancements in ML and AI offer promising alternatives to traditional PQ analysis techniques. ML and AI algorithms have demonstrated their potential in various domains, including image recognition, natural language processing, and autonomous systems, owing to their ability to learn complex patterns and adapt to different scenarios [8,9,10]. The application of these advanced data-driven techniques to PQ event detection, classification, and localization has the potential to overcome the limitations of traditional methods and enable more accurate, efficient, and reliable PQ monitoring and maintenance.
The motivation behind this research is to explore and exploit the capabilities of ML and AI algorithms for the automatic detection, classification, and localization of PQ events. By developing and evaluating algorithms that can accurately identify and classify PQ events, this research aims to contribute to the identification and mitigation of PQ issues in modern power systems. This work seeks to enhance the electrical grid’s overall reliability, stability, and efficiency, minimizing the negative impacts of PQ events on equipment, businesses, and consumers. Ultimately, this research aims to push the boundaries of PQ analysis and contribute to the development of more intelligent and resilient power systems for the future.
Hence, this paper introduces a novel deep learning approach for PQDs classification. The main contributions of this study are:
  • This study introduces a novel deep learning approach for classifying PQDs. By integrating CNNs with attention mechanisms, the proposed method aims to improve the accuracy and efficiency of PQ analysis and monitoring systems.
  • Introducing a multi-scale convolutional network that employs multiple convolutional layers with varying kernel sizes, allowing it to effectively capture feature patterns across different scales of PQDs.
  • When comparing the introduced CNN approach with other state-of-the-art methods, the results demonstrate promising classification performance. The introduced deep learning approach balances simplicity and accuracy, showcasing its potential to enhance PQ analysis.
The rest of the article is organized as follows: Section 2 briefly reviews some background knowledge. Section 3 illustrates the proposed deep learning CNN architecture. Section 4 discusses the experimental results. Finally, Section 5 presents the conclusions and future work.

3. Materials and Methods

This study utilizes a mathematical model of PQDs to generate a dataset for training a 1-D deep neural network (DNN). Following the approach outlined in [21], training and testing sets were generated, consisting of 10 types of PQD signals, including pure sine waveforms, sag, swell, interruption, harmonics, impulsive transient, oscillatory transient, flickers, notch, and spikes. The parameter variations of these signals conform to the IEEE-1159 standard [22]. The sampling rate was set to 3200 Hz, a common rate used in electric power recording equipment. Each signal sample consists of 10 cycles at a fundamental frequency of 50 Hz. By randomly varying the constraint parameters, a large number of data samples can be generated that meet the quantity requirements for deep learning training data.
The architecture of the developed CNN is shown in Figure 2 and Table 1.
Figure 2. Proposed CNN structure.
Table 1. Proposed CNN architecture.

3.1. Convolution Layers

This study is based on the utilization of a 1-D CNN model for the classification of power quality disturbances. Specifically, a sliding window approach is employed to segment the signals for input into the 1-D CNN model. This strategy allows for the extraction of local features from the signal data, enabling the model to effectively capture temporal patterns and variations of essential features for accurate classification. Assuming that the sliding window iterates one element at a time (i.e., with a stride of 1), the operation can be executed based on the following definitions:
( I K ) i = u = 1 s I ( i + u 1 ) K ( u )
( I K ) i = u = 1 s I ( i u 1 ) K ( u )
where I is the input of length n and K is the sliding window of length s.
By incorporating the sliding window approach in conjunction with the 1-D CNN architecture, we aim to enhance the model’s ability to analyze and classify power quality disturbances with a focus on temporal dynamics. This approach contributes significantly to the robustness and efficiency of our classification methodology, allowing for a more nuanced understanding of the complex waveforms associated with various PQ issues. The convolutional layers are utilized to perform convolutional operations on the input data sequence, effectively extracting meaningful features essential for subsequent processing. The proposed model employs multiple convolutional layers with varying kernel sizes, each dedicated to detecting specific patterns or features within the input sequence. With a stride of 1, each filter moves horizontally and vertically across the input data, advancing one position at a time. This configuration enables the convolutional layers to apply filters through element-wise multiplications and summations, leading to the generation of feature maps. A detailed overview of each convolutional layer, alongside its corresponding kernel sizes, is presented in Figure 2 and Table 1.
The ReLU activation function is utilized by introducing non-linearity into the proposed model. It is defined mathematically as:
f ( x ) = m a x ( 0 , x )
thus, when x is positive, the output is x. On the other hand, when x is negative, the output is 0. Introducing non-linearity into our model enables the CNN to learn complex patterns. Moreover, ReLU maintains a gradient of 1 for positive inputs, facilitating faster learning in the deep network.

3.2. Max Pooling

Max pooling is a down-sampling technique used in the CNNs to reduce the spatial dimensions of feature maps while retaining the most important information. The operation involves sliding a kernel over the input feature map and taking the maximum value within that window. Thus, max pooling helps decrease computational load and mitigate overfitting. Also, it provides a degree of invariance to small translations in the input data, allowing the model to focus on dominant features regardless of their exact position.

3.3. Batch Normalization

Batch normalization is used here to improve the training speed and stability of the proposed CNN. It normalizes the outputs of the previous layer by adjusting and scaling them based on the mean and variance computed from mini-batches during training. Thus, for each mini-batch, it calculates the mean and variance of the layer’s outputs. Then, it normalizes the outputs using the following equation:
x ^ = x ¯ μ σ 2 + ϵ
where μ is the is the mean, σ 2 is the variance, and ϵ is a small constant added for numerical stability. The final step in batch normalization includes scaling and shifting the normalized output using learnable parameters γ and β :
y = γ x ^ + β

3.4. Channel Attention Mechanism Unit

An integrated attention model emphasizes significant features by assigning weights based on their relevance to the classification task. This adaptive prioritization enables the model to focus on specific aspects of the input data, thereby enhancing classification accuracy. In this work, a channel attention mechanism is utilized, particularly suited for PQD data, which is represented as time-series signals. Given that the temporal characteristics of these signals are crucial, the proposed approach concentrates on specific time intervals to effectively capture relevant patterns associated with various types of disturbances. Thus, the proposed model can better discern subtle variations in the data, leading to more accurate and reliable classifications of PQDs. The channel attention mechanism typically consists of two main steps: squeeze and excitation. The squeeze operation aims to aggregate spatial information into a channel descriptor. This is accomplished using global average pooling, which summarizes the feature maps across spatial dimensions (height and width).
z c = 1 H W i = 1 H j = 1 W F c ( i , j )
where z c is the channel descriptor for channel c, H and W are the height and width of the feature map, and F c ( I , j ) represents the value at position ( i , j ) in channel c.
The excitation operation generates attention weights for each channel based on the channel descriptors obtained from the squeeze step.

3.5. The Fully Connected Layer

The fully connected layer 1 is a layer in the CNN architecture that follows the convolutional layers. It establishes a fully connected network structure by connecting every neuron from the previous layer to the current layer. The input size of fully connected layer 1 is 128, which is typically a flattened vector representing the output of the previous layer. Fully connected layer 1 takes the flattened vector as the input and performs matrix multiplication. Each neuron in fully connected layer 1 receives inputs from all of the neurons in the previous layer. Following the matrix multiplication, a bias term may be added to each neuron. This bias term enables the network to shift its decision boundary, allowing for more flexibility in the model’s predictions. The resulting values are then passed through an activation function, commonly ReLU (rectified linear unit), to introduce non-linearity and enhance the model’s representational capabilities. Furthermore, the input size of fully connected layer 2 is 32 neurons, indicating that it receives a 32-dimensional input from the previous layer. Similarly, fully connected layer 3 has an input size of 16 neurons, meaning it processes a 16-dimensional input. The specific sizes of these layers may vary depending on the network architecture and the requirements of the task at hand.

3.6. The Softmax Layer

The softmax layer is a common choice for the final layer in a classification neural network. It takes the output from the previous layer, typically a fully connected layer, and produces probabilities for each class in a multi-class classification problem. The softmax function operates on each element x i of the input vector and computes the following:
s o f t m a x ( x i ) = e x p ( x i ) j = 1 N e x p ( x j )
Hence, the softmax function exponentiates each element, divides it by the sum of exponentiated values across all elements, and normalizes the result to ensure that the probabilities sum up to 1.

3.7. The Classification Layer

The classification layer is the final layer in a neural network model designed for classification tasks. Its purpose is to map the extracted features from the preceding layers of the network to the corresponding class labels.
The output of the classification layer is a vector of probabilities, where each element represents the probability of the corresponding class. The class with the highest probability is often selected as the predicted class for the input.

3.8. Data Generation

Generating synthetic PQDs includes generating voltage and current waveforms with different types of disturbance [7,21,22]. The disturbances are introduced by manipulating the waveform parameters based on established models and algorithms. Moreover, the generated PQDs are labeled and annotated according to their type and severity. This annotation process ensures that the generated dataset contains accurate labels for the different types of disturbances, which are essential for training and evaluating deep learning models.
Two datasets were generated using [7]. The first dataset consists of PQ signals with random Gaussian noise, where the signal-to-noise ratio (SNR) ranges between 20 and 50 dB. The second dataset is generated without any noise. Each dataset contains a total of 50,000 signals, with 5000 signals being dedicated to each disturbance type. The disturbance types included in the datasets are as follows: normal, flicker, harmonics, oscillatory transient, impulsive transient, interruption, notch, sag, spike, and swell.
The datasets aim to provide a diverse representation of PQDs encountered in real-world scenarios. By including different disturbance types and generating datasets with and without noise, the authors enable the evaluation and comparison of various algorithms and techniques for PQ analysis and classification tasks. Collectively, these datasets facilitate a thorough evaluation of the proposed approach across varied contexts and conditions. To maintain analytical consistency, each dataset is segmented into training (60%), validation (20%), and test (20%) subsets. This partitioning allows for effective model training, unbiased performance assessment during the validation phase, and a final evaluation of the model’s generalization capabilities on unseen data. Such a rigorous approach not only strengthens the validity of the findings but also facilitates meaningful comparisons with existing methodologies, ultimately contributing to the advancement of power quality disturbance classification techniques.
Data augmentation is utilized to expand and diversify the training dataset, enhancing the performance and generalization capabilities of the proposed model. By altering the temporal characteristics of PQD signals through the addition of noise to the time domain data, the proposed model can effectively learn to identify and generalize patterns across varying noise levels and temporal scales, improving its robustness and accuracy in classifying power quality disturbances.

3.9. Experimental Setup and Network Training

The experiments were conducted using a robust hardware setup that includes an AMD CPU, Intel processors, and Nvidia Tesla T4 GPUs. This configuration boasts a total of 24 cores and an impressive 384 GB of combined memory, providing ample resources for intensive computational tasks. The experiments were executed using MATLAB software, ensuring access to the latest features and optimizations for data analysis and algorithm development.
Bayesian optimization, a sophisticated approach to hyperparameter tuning, is systematically employed across all models in this study. The optimization procedure is based on Bayes’ theorem [23], such that for a model Z and observation Y:
P ( z | y ) = ( P ( Y | Z ) P ( z ) ) / P ( z )
where P ( Z | Y ) is the posterior probability of Z given Y, P ( Y | Z ) is the likelihood of Y given Z, P ( Z ) is the prior probability of Z, and P ( Z ) is the marginal probability of Z. This advanced technique leverages probabilistic models to intelligently navigate the complex hyperparameter space, striking an optimal balance between the exploration of new parameter combinations and the exploitation of known high-performing regions. By consistently applying this process to each model, including the proposed method and all comparison models, the aim is to ensure a fair, rigorous, and unbiased evaluation framework. This methodical approach not only maximizes the potential of each model to achieve its peak performance on the given datasets but also mitigates the risk of performance disparities arising from inconsistent tuning practices. Furthermore, this standardized tuning process enhances the reproducibility of the results, provides a solid foundation for future research in this domain, and ensures that any performance differences observed between models are more likely to be inherent to the models themselves rather than artifacts of the tuning process.
Thus, the CNN networks are trained using the Adam optimizer with specific training options, including a maximum of 21 epochs, a mini-batch size of 64, a piecewise learning rate schedule with a drop factor of 0.5, a drop period of 10 epochs, an initial learning rate of 0.01, a gradient threshold of 1, and automatic execution environment selection. Training progress is monitored and visualized using training progress plots. Figure 3 shows the training progress of the proposed CNN. The graph demonstrates the performance improvement of the CNN model over multiple epochs. As the number of epochs increases, the accuracy of the model steadily increases. This indicates that the model is learning and making better predictions as it receives more training data and updates its internal parameters.
Figure 3. Training progress of the proposed CNN.
Moreover, the RNN networks are trained using the “Adam” optimizer with the maximum number of training epochs being set to 60. The number of samples in each mini-batch is set to 64, and a piecewise learning rate schedule is employed. The drop factor is set to 0.5 with a drop period of 10 epochs, an initial learning rate of 0.01, and a gradient threshold of 1. The “Execution Environment” is set to “auto”, enabling the code to automatically utilize available hardware resources.

4. Results and Discussion

The performance of the proposed network undergoes a comprehensive evaluation, pitting it against a diverse array of state-of-the-art architectures. This evaluation encompasses both CNN models and RNN models, providing a broad spectrum of comparison points. Among the CNN models, the evaluation includes a CNN architecture detailed in [7] and ResNet50. This comparison allows for a nuanced understanding of how the proposed network performs against both simpler and more complex CNN architectures. The evaluation extends to RNN models as well, incorporating BiLSTM and GRU models. These RNN variants are adept at capturing temporal dependencies in sequential data, offering a different perspective on the problem compared with CNN-based approaches. This multi-faceted comparison strategy ensures a thorough assessment of the proposed network’s capabilities across various architectural paradigms. It not only benchmarks the network against established models but also provides insights into its relative strengths and potential areas for improvement across different types of neural network architectures.

Quantitative Evaluation

Table 2, Table 3, Table 4, Table 5 and Table 6 present the confusion matrices for the proposed CNN, along with those of the compared state-of-the-art methods on datasets without noise. Additionally, Table 7, Table 8, Table 9, Table 10 and Table 11 display the confusion matrices for the proposed CNN and the state-of-the-art methods on datasets with noise.
Table 2. Confusion matrix of the training data using the proposed CNN architecture, without noise.
Table 3. Confusion matrix of the training data using the CNN architecture in [7], without noise.
Table 4. Confusion matrix of the training data using the ResNet50 architecture, without noise.
Table 5. Confusion matrix of the training data using BiLSTM architecture, without noise.
Table 6. Confusion matrix of the training data using GRU architecture, without noise.
Table 7. Confusion matrix of the training data using the proposed CNN architecture, with noise.
Table 8. Confusion matrix of the training data using the CNN architecture in [7], with noise.
Table 9. Confusion matrix of the training data using the ResNet50 architecture, with noise.
Table 10. Confusion matrix of the training data using BiLSTM architecture, with noise.
Table 11. Confusion matrix of the training data using GRU architecture, with noise.
The performance of the proposed method, as evidenced by the confusion matrix, demonstrates a significant advancement over existing state-of-the-art techniques. This detailed analysis highlights the classification results and also showcases the model’s ability to accurately predict various classes of PQDs. By systematically comparing the confusion matrices of the proposed method with those of leading approaches, the proposed method can effectively evaluate key performance metrics such as precision, recall, and overall accuracy for each class. The results reveal that the introduced method consistently outperforms its competitors, achieving higher precision and recall rates across multiple PQD classes. This indicates that the proposed approach enhances classification accuracy and also improves the model’s reliability in distinguishing between different categories of PQDs.
Table 12, Table 13, Table 14, Table 15 and Table 16 provide detailed accuracy values for the proposed approach and the state-of-the-art methods. The accuracy values presented in these tables offer a comprehensive comparison of the proposed method against its alternatives. The proposed approach demonstrates superior performance across all evaluated classes, particularly evident in the accuracy metric, where it consistently outperforms the competing methods. This indicates a stronger ability to correctly identify and classify instances across various categories.
Table 12. Detailed accuracy of the proposed method.
Table 13. Detailed accuracy of the CNN method in [7].
Table 14. Detailed accuracy of the ResNet50 method.
Table 15. Detailed accuracy of the BiLSTM method.
Table 16. Detailed accuracy of the GRU method.
Furthermore, the overall results illustrate that the proposed method effectively minimizes false positives and false negatives compared with both competing models, highlighting its robustness and reliability in classification tasks. Overall, these results underscore the efficacy of the proposed method in delivering enhanced accuracy and performance relative to existing state-of-the-art techniques.
Figure 4 presents a comprehensive bar chart illustrating the detailed accuracy of the proposed approach and the state-of-the-art methods. This visual representation allows for an easy comparison of the performance metrics across these approaches, highlighting the strengths and weaknesses of each model in terms of classification accuracy. By examining the bar chart, viewers can quickly discern how the proposed method stands out in achieving superior accuracy compared with its counterparts, thereby emphasizing its effectiveness in addressing classification challenges.
Figure 4. Detailed accuracy of the proposed method compared with the state-of-the-art methods [7].
Table 17 provides a comprehensive overview of the average accuracy achieved by the proposed CNN and other state-of-the-art methods. It presents a quantitative measure of how well each method performs in accurately classifying the test data.
Table 17. Testing average accuracy comparison.
On the other hand, Figure 5 visually represents the testing accuracy levels of the proposed CNN and state-of-the-art methods. This graphical representation enables a more intuitive understanding of the performance comparison. By examining both the table and the figure, the proposed CNN performs well compared with the existing state-of-the-art methods in terms of testing accuracy.
Figure 5. Testing average accuracy comparison [7].
The proposed model achieved an accuracy of up to 99.49 % , significantly outperforming traditional methods and even some existing models. This demonstrates the effectiveness of leveraging multi-scale CNNs combined with attention mechanisms to capture complex patterns in PQ signals. The integration of attention mechanisms allows the model to focus on the most relevant features of the input data, enhancing its ability to discern subtle variations in power quality disturbances. This can lead to more reliable classifications, especially in noisy environments. By improving the accuracy and efficiency of PQ analysis, the proposed approach has significant implications for real-world applications. Enhanced classification can lead to better monitoring and maintenance of power systems, ultimately contributing to increased reliability and stability of electrical grids. Moreover, future research could explore the model’s scalability and its performance in real-time applications, where rapid and accurate classification of PQ disturbances is crucial. Additionally, investigating the interpretability of the model’s decisions could provide valuable insights into the specific features it relies on for classification, potentially leading to further improvements in PQ analysis techniques.

5. Conclusions and Future Work

This work introduces a novel deep learning approach for classifying PQDs. The method harnesses deep neural networks’ capabilities to accurately identify and categorize various PQDs, enabling effective analysis and monitoring of electrical systems. By leveraging deep learning algorithms, particularly CNNs enhanced with a channel attention mechanism, the proposed method has demonstrated enhanced classification accuracy for PQDs. The model shows proficiency in learning intricate patterns and features from PQ signals, leading to robust and reliable classification performance. This advancement not only improves disturbance detection precision but also enhances overall electrical system efficiency and stability, laying the groundwork for more effective PQ management and mitigation strategies.
The outcomes from the deep learning approach are promising, highlighting its potential to elevate PQ analysis and monitoring systems. Undetected or misdiagnosed PQ issues can result in operational disruptions, equipment failures, increased downtime, and safety risks. Inaccurate identification also hampers targeted mitigation strategies, undermining PQ quality and potentially causing economic losses.
The proposed deep learning methodology opens the door to further exploration in PQ analysis. Future investigations will involve conducting ablation experiments to assess the specific impact of the channel attention mechanism within the deep learning model. These experiments will provide insights into the mechanism’s contribution to classification accuracy, further refining the model and potentially enhancing its performance. Additionally, ongoing research efforts will focus on continuously improving deep learning architectures, exploring sophisticated ensemble methods, and integrating additional relevant features to push the boundaries of classification performance. Moreover, modifying the structure and training algorithm of the developed CNN will be a priority. This research has the potential to revolutionize power system management by advancing our understanding and capabilities in PQ analysis. It could lead to a more reliable and resilient electrical infrastructure that can meet the evolving demands of modern society.

Author Contributions

Conceptualization, F.A.A. and M.R.Q.; methodology, M.R.Q.; software, F.A.A.; validation, M.R.Q.; formal analysis, F.A.A.; investigation, F.A.A.; resources, M.R.Q.; data curation, M.R.Q.; writing—original draft preparation, F.A.A.; writing—review and editing, F.A.A. and M.R.Q.; visualization, F.A.A. and M.R.Q.; supervision, M.R.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available on GitHub at https://github.com/chachkes247/Power-Quality-Disturbances (accessed in 1 December 2024), as referenced in [7].

Acknowledgments

The experiments presented in this paper were carried out using the facilities of the Benefit Advanced AI and Computing Lab at the University of Bahrain—see https://ailab.uob.edu.bh (accessed on 1 December 2024)—with support from Benefit Bahrain Company—see https://benefit.bh (accessed on 15 November 2024). The authors extend sincere thanks to the Benefit Advanced AI and Computing Lab at the University of Bahrain for providing computational facilities.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Khetarpal, P.; Tripathi, M.M. A critical and comprehensive review on power quality disturbance detection and classification. Sustain. Comput. Inform. Syst. 2020, 28, 100417. [Google Scholar] [CrossRef]
  2. Samanta, I.S.; Panda, S.; Rout, P.K.; Bajaj, M.; Piecha, M.; Blazek, V.; Prokop, L. A Comprehensive Review of Deep-Learning Applications to Power Quality Analysis. Energies 2023, 16, 4406. [Google Scholar] [CrossRef]
  3. Qader, M.R. A Method for Analytical Voltage Sags Prediction in Complex Power Network. Recent Adv. Electr. Electron. Eng. 2018, 11, 303–309. [Google Scholar] [CrossRef]
  4. Thentral, T.T.; Palanisamy, R.; Usha, S.; Bajaj, M.; Zawbaa, H.M.; Kamel, S. Analysis of Power Quality issues of different types of household applications. Energy Rep. 2022, 8, 5370–5386. [Google Scholar] [CrossRef]
  5. Sharma, A.; Rajpurohit, B.S.; Singh, S. A review on economics of power quality: Impact, assessment and mitigation. Renew. Sustain. Energy Rev. 2018, 88, 363–372. [Google Scholar] [CrossRef]
  6. Iturrino Garcia, C.A.; Bindi, M.; Corti, F.; Luchetta, A.; Grasso, F.; Paolucci, L.; Piccirilli, M.C.; Aizenberg, I. Power Quality Analysis Based on Machine Learning Methods for Low-Voltage Electrical Distribution Lines. Energies 2023, 16, 3627. [Google Scholar] [CrossRef]
  7. Machlev, R.; Chachkes, A.; Belikov, J.; Beck, Y.; Levron, Y. Open source dataset generator for power quality disturbances with deep-learning reference classifiers. Electr. Power Syst. Res. 2021, 195, 107152. [Google Scholar] [CrossRef]
  8. Albalooshi, F.A.; Asari, V.K. Fast and robust object region segmentation with self-organized lattice Boltzmann based active contour method. J. Electron. Imaging 2024, 33, 043050. [Google Scholar] [CrossRef]
  9. Liu, Y.; Shen, L.; Zhu, X.; Xie, Y.; He, S. Spectral Data-Driven Prediction of Soil Properties Using LSTM-CNN-Attention Model. Appl. Sci. 2024, 14, 11687. [Google Scholar] [CrossRef]
  10. Albalooshi, F.A. Novel Approach in Vegetation Detection Using Multi-Scale Convolutional Neural Network. Appl. Sci. 2024, 14, 10287. [Google Scholar] [CrossRef]
  11. Wang, S.; Chen, H. A novel deep learning method for the classification of power quality disturbances using deep convolutional neural network. Appl. Energy 2019, 235, 1126–1140. [Google Scholar] [CrossRef]
  12. Jamlus, N.U.I.A.; Shahbudin, S.; Kassim, M. Power Quality Disturbances Classification Analysis Using Residual Neural Network. In Proceedings of the 2022 IEEE 18th International Colloquium on Signal Processing & Applications (CSPA), Selangor, Malaysia, 12 May 2022; pp. 442–447. [Google Scholar] [CrossRef]
  13. Bravo-Rodríguez, J.C.; Torres, F.J.; Borrás, M.D. Hybrid Machine Learning Models for Classifying Power Quality Disturbances: A Comparative Study. Energies 2020, 13, 2761. [Google Scholar] [CrossRef]
  14. Dekhandji, F.Z.; Recioui, A.; Ladada, A.; Moulay Brahim, T.S. Detection and Classification of Power Quality Disturbances Using LSTM. Eng. Proc. 2023, 29, 2. [Google Scholar] [CrossRef]
  15. Gunawan, T.S.; Husodo, B.Y.; Ihsanto, E.; Ramli, K. Power Quality Disturbance Classification Using Deep BiLSTM Architectures with Exponentially Decayed Number of Nodes in the Hidden Layers. In Recent Trends in Mechatronics Towards Industry 4.0; Nasir, A.F.A., Ibrahim, A.N., Ishak, I., Mat Yahya, N., Zakaria, M.A., Abdul Majeed, A.P.P., Eds.; Springer: Singapore, 2022; pp. 725–736. [Google Scholar]
  16. Yiğit, E.; Özkaya, U.; Öztürk, Ş.; Singh, D.; Gritli, H. Automatic detection of power quality disturbance using convolutional neural network structure with gated recurrent unit. Mob. Inf. Syst. 2021, 2021, 7917500. [Google Scholar] [CrossRef]
  17. Khetarpal, P.; Nagpal, N.; Al-Numay, M.S.; Siano, P.; Arya, Y.; Kassarwani, N. Power quality disturbances detection and classification based on deep convolution auto-encoder networks. IEEE Access 2023, 11, 46026–46038. [Google Scholar] [CrossRef]
  18. Topaloglu, I. Deep learning based a new approach for power quality disturbances classification in power transmission system. J. Electr. Eng. Technol. 2023, 18, 77–88. [Google Scholar] [CrossRef]
  19. Beniwal, R.K.; Saini, M.K.; Nayyar, A.; Qureshi, B.; Aggarwal, A. A critical analysis of methodologies for detection and classification of power quality events in smart grid. IEEE Access 2021, 9, 83507–83534. [Google Scholar] [CrossRef]
  20. Ramalingappa, L.; Manjunatha, A. Power quality event classification using complex wavelets phasor models and customized convolution neural network. Int. J. Electr. Comput. Eng. (IJECE) 2022, 12, 22–31. [Google Scholar] [CrossRef]
  21. Khokhar, S.; Mohd Zin, A.A.; Memon, A.P.; Mokhtar, A.S. A new optimal feature selection algorithm for classification of power quality disturbances using discrete wavelet transform and probabilistic neural network. Measurement 2017, 95, 246–259. [Google Scholar] [CrossRef]
  22. IEEE 1159-1995; IEEE Recommended Practice for Monitoring Electric Power Quality. IEEE: Piscataway, NJ, USA, 1995.
  23. Herrle, S.R.; Corbett, E.C., Jr.; Fagan, M.J.; Moore, C.G.; Elnicki, D.M. Bayes’ theorem and the physical examination: Probability assessment and diagnostic decision making. Acad. Med. 2011, 86, 618–627. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.