Next Article in Journal
Collaborative Self-Standby Adjustment for Photovoltaics with Rotor Inertial Power Source Control Strategy in Weak Power Grids
Previous Article in Journal
Numerical Investigations of Inlet Recirculation in a Turbocharger Centrifugal Compressor
Previous Article in Special Issue
A Diffusion–Attention-Enhanced Temporal (DATE-TM) Model: A Multi-Feature-Driven Model for Very-Short-Term Household Load Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable AI-Driven Quantum Deep Neural Network for Fault Location in DC Microgrids

by
Amir Hossein Poursaeed
1 and
Farhad Namdari
2,*
1
Department of Electrical Engineering, Faculty of Engineering, Lorestan University, Khorram Abad 68151-44316, Iran
2
Department of Engineering, Faculty of Environment, Science, and Economy, University of Exeter, Exeter EX4 4QF, UK
*
Author to whom correspondence should be addressed.
Energies 2025, 18(4), 908; https://doi.org/10.3390/en18040908
Submission received: 17 January 2025 / Revised: 6 February 2025 / Accepted: 10 February 2025 / Published: 13 February 2025

Abstract

:
Fault location in DC microgrids (DCMGs) is a critical challenge due to the system’s inherent complexities and the demand for high reliability in modern power systems. This study proposes an explainable artificial intelligence (XAI)-based quantum deep neural network (QDNN) framework to address fault localization challenges in DCMGs. First, voltage signals from the DCMG are collected and analyzed using high-order synchrosqueezing transform to detect traveling waves (TWs) and extract critical fault parameters such as time of arrival, magnitude, and polarity of the first and second TWs. These features are fed into the proposed QDNN model that integrates advanced learning techniques for accurate fault localization. The cumulative distance from the fault point to the bus connecting the DCMG to the power network is considered the output vector. The model uses a combination of deep learning and quantum computing techniques to extract features and improve accuracy. To ensure transparency, an XAI technique called Shapley additive explanations (SHAP) is applied, enabling system operators to identify critical fault features. The SHAP-based explainability framework plays a critical role in translating the model’s predictions into actionable insights, ensuring that the proposed solution is not only accurate but also practically implementable in real-world scenarios. The results demonstrate the QDNN framework’s superior accuracy in fault localization even in noisy environments and with high-resistance faults, independent of voltage levels and DCMG configurations, making it a robust solution for modern power systems.

1. Introduction

1.1. Aim and Scope

The increasing deployment of DC energy sources alongside extensive integration of energy storage systems has made DC microgrids (DCMGs) a significant area of focus in modern power systems. This trend aligns with the global push toward renewable energy and smarter power grids, which require efficient, reliable, and scalable solutions [1]. These systems eliminate challenges inherent in AC systems such as reactive power control, frequency synchronization, and harmonic issues [2]. As the adoption of renewable energy accelerates, the importance of robust fault detection methods has grown significantly. Fault detection is essential not only for ensuring the operational safety and efficiency of DCMGs but also for supporting the broader goals of modern energy systems, such as enabling the integration of renewable energy sources and maintaining grid stability in smart grids. However, DCMGs face unique challenges, making traditional AC protection schemes unsuitable [3]. Faults in these systems, if left undetected, can lead to cascading failures, equipment damage, and costly downtime. These issues directly threaten the viability of renewable-based power systems, highlighting the critical need for advanced fault detection techniques [4].

1.2. Research Background

Traditional protection methods for DCMGs typically rely on overcurrent [5] or overvoltage principles [6]. However, when fault currents are low, such as in high-resistance faults, these methods may fail to detect the issue promptly [7]. Derivative-based methods utilize the first or second derivatives of current and voltage. Nonetheless, they are highly sensitive to noise and require accurate threshold settings, which may not always be reliable under varying network conditions [8]. Differential protection operates based on the difference between incoming and outgoing currents in a protection zone. This method has been widely adopted due to its reduced dependency on fault current direction [9]. However, it heavily relies on communication links, making it vulnerable to delays or failures in the communication channel [10]. Non-unit protection schemes, such as distance protection, have been proposed to address these limitations, but the short length of distribution networks in DCMGs often reduces the accuracy of these methods. Therefore, they require additional adjustments for effective implementation [11].
Recently, traveling wave (TW)-based protection schemes have gained traction for their ability to rapidly detect faults independent of fault current magnitude [12]. These methods leverage high-frequency measurements to capture fault-induced transients propagating at near-light speed, enabling fault detection within less than a millisecond [13]. However, the high cost of equipment and the complexity of setting accurate thresholds for distinguishing fault-related transients remain key challenges [14]. Therefore, feature extraction methods would be a viable solution to resolve these issues; wavelet transform [15], Teager energy operator [16], short-time Fourier transform [17], Hilbert–Huang transform [18], high-order synchrosqueezing transform (HOSST) [19], and mathematical morphology [20] have been used to capture the fault-induced transients from voltage and current signals to be used for the protection of DCMGs. Similarly, machine learning approaches have emerged as a promising alternative for fault detection and classification [21]. These techniques, such as support vector machines [22], decision trees [23], a mixture of both [24], deep neural networks (DNN) [25], k-nearest neighbors (KNN) algorithm [26], and recurrent neural networks (RNN) [27], offer enhanced fault location accuracy and fault type recognition. Nevertheless, they demand extensive datasets and computational resources, increasing implementation complexity and cost [28]. Despite the advancements in fault detection methods, a significant gap remains in providing interpretable insights for system operators. This gap emphasizes the need for methodologies not only to improve fault localization accuracy but also to provide intuitive explanations for model predictions.
To summarize the discussed methods and highlight their relative strengths and limitations, Table 1 compares various fault detection and location techniques in DCMGs.

1.3. Contribution

This study introduces a novel explainable artificial intelligence (XAI)-based quantum deep neural network (QDNN) designed to address the limitations of traditional fault detection methods in DCMGs. The proposed methodology begins with acquiring voltage signals from the DCMG and applying HOSST to detect TWs. Key fault parameters, including the time of arrival (TOA), magnitude, and polarity of the first and second TWs, are extracted and used as inputs to the QDNN model, which combines the strengths of classical convolutional neural networks (CNN) for local pattern recognition and quantum neural networks (QNN) for quantum-enhanced feature extraction. The extracted features are processed further using a bidirectional long short-term memory (BD-LSTM) layer to capture temporal dependencies, with an attention mechanism highlighting the most relevant patterns. The model’s output is the predicted fault location, represented as the cumulative distance from the fault point to the main grid connection point. The integration of Shapley additive explanation (SHAP) values as part of the proposed framework bridges the gap between black-box predictions and interpretable insights. This ensures that the outputs of the QDNN model are not only accurate but also actionable, allowing operators to diagnose faults effectively and make data-driven decisions with confidence. As illustrated in Table 1, the proposed QDNN-based method addresses critical shortcomings of traditional approaches, including limited explainability, reliance on communication links, and poor scalability.
The innovations of the paper are listed as follows:
  • Utilizes HOSST to accurately detect TWs and extract critical fault parameters such as TOA, magnitude, and polarity.
  • Employs a CNN and a QNN to combine outputs for comprehensive data representation.
  • Integrates a BD-LSTM layer to capture temporal dependencies in the fault data from both directions.
  • Applies an attention layer to improve the focus on key data points during fault analysis.
  • Employs the SHAP method that provides transparency by identifying feature importance and justifying model predictions. This improves the interpretability of fault localization predictions, enhances trust and reliability, and ensures that system operators and engineers can understand the reasoning behind the model’s outputs.
  • Ensures the framework’s independence from DCMG configurations and voltage levels.
  • Uses advanced spatiotemporal modeling to transform extracted features into precise fault locations, ensuring reliable accuracy for faults across various distances and configurations.
  • The suggested framework offers a robust, explainable, and scalable solution for fault detection in diverse DCMG configurations, contributing to the reliability of modern power systems.

2. The Proposed Method

The proposed method combines HOSST for feature extraction with a QDNN to detect and locate faults in DCMGs. HOSST captures key fault characteristics, while the QDNN integrates classical and quantum neural networks for enhanced data processing. Temporal patterns are identified using a BD-LSTM and attention mechanism, and XAI ensures transparent and reliable results. The method is adaptable to various DCMG setups.

2.1. Feature Extraction Based on HOSST

Effective feature extraction from current signals plays a pivotal role in detecting faults accurately in DCMGs, especially for TW-based protection. This is achieved through a HOSST, a higher order of the initial version of synchrosqueezing transform (SST), which offers a more concentrated and accurate time–frequency representation, crucial for detecting fault-induced TWs due to capturing high-frequency contents from input signals.
Figure 1 provides a schematic representation of the HOSST process. On the left, the original time-domain signal is shown as a sequence of sampled data points, where each sample corresponds to a specific time step. HOSST then transforms this time-domain signal into a time–frequency representation, illustrated on the right. In the output, each row corresponds to a specific frequency bin, and each column represents a time step. The red-highlighted row indicates the frequency bin with the maximum spectral energy for a given time sample, which is crucial for detecting and analyzing transient events like TWs. This visualization simplifies the interpretation of HOSST results, allowing operators to identify the most relevant frequency components for fault detection in DCMGs. In fact, by concentrating spectral energy into distinct frequency bins, HOSST ensures the preservation of transient fault characteristics, which are often diluted in traditional frequency-domain analysis. This enhanced clarity in feature extraction not only improves the accuracy of fault detection but also simplifies the interpretability of fault dynamics for system operators.
HOSST enhances classical synchrosqueezing by reallocating frequency components using higher-order Taylor series expansions, designed to refine time–frequency location for transient events such as TWs, defined as Equation (1).
T s N t , ω = 1 g * 0 f , S S T S w t , f > ξ S S T S w t , f δ ω ω ~ S N t , f d f
where ω denotes the angular frequency, defined as ω = 2 π f , where f is the frequency of the signal component. This term plays a crucial role in the time–frequency representation by determining the spectral resolution in the synchrosqueezing process; T s N represents the high-order synchrosqueezed time-frequency representation, g * ( 0 ) denotes the normalization factor of the wavelet, S S T S w is the short-time Fourier-based synchrosqueezing transform, ω ~ S N depicts the frequency components derived using high-order derivatives, δ is the Dirac delta function to reallocate the time–frequency energy for higher concentration, and ξ is a threshold parameter used to reduce noise effects by filtering out low-energy components. Finally, the chosen order to capture TWs is considered the third order.
Equation (1) represents the high-order synchrosqueezed time-frequency representation, which refines the localization of transient events such as TWs. By redistributing frequency components using higher-order derivatives, it enhances fault detection accuracy.

2.2. Obtaining the Magnitude and Polarity of TWs

The magnitude of TWs is calculated as the modulus of the complex output from HOSST at the frequency corresponding to the maximum value of the spectral envelope, given by Equation (2).
T W m a g t = T s [ N ] ( t , ω * ) = R e T s [ N ] ( t , ω * ) 2 + I m T s [ N ] ( t , ω * ) 2
Here, ω * is the frequency at which the spectral envelope has its maximum value, and T s [ N ] is the FSSTH output for the selected order N .
The polarity of TWs is determined by the phase angle of the complex HOSST output at ω * . The formula is shown in Equation (3):
T W p o l t = atan 2 I m T s [ N ] ( t , ω * ) R e T s [ N ] ( t , ω * )
This provides the direction of propagation, which is essential for fault localization. Equation (3) calculates the polarity of TWs by determining the phase angle of the complex HOSST output at the dominant frequency. This helps distinguish between wave directions, which is crucial for accurate fault localization in DCMGs.

2.3. Data Preparation

After determining the magnitude and polarity of TWs, the TOA, magnitude, and polarity of the first and second TWs are recorded. These parameters are measured relative to the reference bus, which is the point of connection between the DCMG and the main power grid. Measurements are taken incrementally along the lines, typically meter by meter, to capture accurate fault characteristics. The locations of TWs are recorded cumulatively, meaning the distances are calculated from the reference bus to the farthest fault point. The maximum distance from the reference bus is used as the output vector, representing the fault location in relation to the microgrid connection point. This process ensures precise input data preparation for fault localization and enhances the accuracy of the proposed method.
The data is first shuffled randomly to ensure unbiased distribution. The structured data preparation process ensures that the model is trained on a representative dataset, minimizing bias and enhancing generalizability. This approach not only improves the reliability of the fault localization predictions but also facilitates smoother integration of the model into practical fault management workflows. Therefore, 70% of the data is allocated for training, 10% for validation, and the remaining 20% for testing. These datasets are then fed into the proposed network for training and evaluation.

2.4. Proposed QDNN Architecture

Different layers of the proposed QDNN framework are listed as follows:
  • The input layer, which receives the raw data, which is then split into two paths:
    • The first path processes the data using a CNN.
    • The second path sends the data to a QNN for quantum-based feature extraction.
  • CNN path
    • The first convolutional layer applies 32 filters with a kernel size of 3, using the rectified linear unit (ReLU) activation function to capture local patterns in the data.
    • Batch normalization follows this layer to stabilize the learning process and improve training efficiency.
    • The second convolutional layer uses 64 filters with the same kernel size and activation function to deepen the feature extraction.
    • This is followed by a flatten layer, in which the output from the convolutional layers is flattened into a one-dimensional vector to prepare it for further processing.
  • QNN path
    • The QNN uses Angle Embedding to map the input features into quantum states, representing them as angles in quantum circuits.
    • Entanglement layers are applied to create complex quantum relationships between the features.
The output consists of expectation values derived from the quantum measurements, which represent the quantum-enhanced features.
  • Merging paths, in which the outputs of the CNN and QNN are concatenated to combine the local features extracted by the CNN with the quantum-based features from the QNN.
  • BD-LSTM layer, to which the merged features are passed. It processes sequential dependencies in the data, learning patterns over time from both directions.
  • The attention layer, which focuses on the most important temporal features from the BD-LSTM output, assigns higher weights to these features to enhance the model’s understanding of key data points, and reduces noise and improves the interpretability of predictions. This design choice aligns with the need for fault localization systems to prioritize critical data points while discarding less relevant information.
  • Fully connected layer, into which the output from the attention layer is fed. It is a dense layer with 128 neurons, and ReLU activation further processes the combined features.
  • The output layer, the final layer, is a dense layer with a single neuron that produces the model’s prediction, which could represent a fault location.
The architecture of the proposed network is depicted in Figure 2.

2.5. XAI Integration

In the final stage of the proposed method, an XAI approach is integrated to enhance the interpretability of the model’s predictions for fault location in DCMGs. The goal is to explain how the model makes decisions and highlight the key factors affecting fault detection.
The XAI technique utilized in this study is SHAP, which quantitatively helps explain which factors most influence the model’s predictions. SHAP explains how factors such as the time of arrival, magnitude, and polarity of the TWs contribute to the fault location estimation. The SHAP value, i.e., i , for a feature x i is calculated as Equation (4).
i = S N \ i S ! N S 1 ! N ! f S i f S
Here S is a subset of features excluding x i ,   N represents the full set of input features, f S is the model’s prediction based on the features in S , and f S i denotes the model’s prediction when x i is added to S . Equation (4) represents the SHAP value calculation, which quantifies the contribution of each input feature to the model’s predictions. This allows operators to understand which factors most influence fault location estimation, improving model transparency and interpretability.
After the QDNN predicts the fault location for various scenarios, SHAP decomposes the predictions into contributions from each input feature. For instance, if the predicted fault location deviates significantly, SHAP can reveal whether the variation is due to delays in TW TOAs, changes in magnitude, or polarity shifts. This integration of XAI into the proposed method bridges the gap between model predictions and practical fault analysis in DCMGs. By providing detailed insights into the model’s predictions, XAI empowers system operators to prioritize critical fault parameters, enhancing the reliability and safety of microgrid operations.
The flowchart of the proposed method is shown in Figure 3.

2.6. Evaluation Metrics

To evaluate the performance of the proposed method in fault location estimation, six metrics are employed: root mean squared error (RMSE), mean absolute error (MAE), R-squared (R2), Theil’s U statistic (TUS), Willmott’s index of agreement (WIA), and variance accounted for (VAF). These metrics provide a comprehensive assessment of the model’s accuracy and reliability, as defined in Equations (5)–(10), respectively. Each metric highlights a specific aspect of model performance, offering a comprehensive evaluation [29].
Smaller values of RMSE, MAE, and TUS indicate higher accuracy, while R2, WIA, and VAF values closer to 1 reflect better performance. Specifically, the definition of each one is listed as follows:
  • RMSE measures the average magnitude of error between the predicted and actual values. It penalizes larger errors more heavily, making it a reliable indicator of overall accuracy.
  • MAE calculates the average absolute difference between predicted and actual values, offering an easily interpretable measure of error without penalizing outliers as strongly as RMSE.
  • R2 evaluates the proportion of variance in the actual values that is explained by the predictions. Values closer to 1 indicate a better fit.
  • TUS measures the relative accuracy of predictions compared to a naïve forecasting model. Lower values indicate better performance.
  • WIA reflects how well the model captures the variation in the data. It considers both the error magnitude and the distribution of errors, with values closer to 1 indicating higher agreement between predictions and actual values.
  • VAF quantifies the percentage of variance in the actual data that is explained by the predictions. A higher VAF indicates better performance.
R M S E = 1 n i = 1 n T i P i 2
M A E = 1 n i = 1 n T i P i
R 2 = 1 i = 1 n T i P i 2 i = 1 n T i T ¯ 2
T U S = 1 n i = 1 n T i P i 2 1 n i = 1 n T i T ¯ 2
W I A = 1 i = 1 n T i P i 2 i = 1 n T i T ¯ + P i T ¯ 2
V A F = 1 V a r T i P i V a r T i × 100
where n is the number of samples, T i represents the actual value, P i denotes the predicted value, T ¯ is the mean of actual values, and V a r is the variance. The use of diverse evaluation metrics ensures that the model’s performance is assessed from multiple perspectives, capturing both the accuracy of predictions and the robustness under varying operational conditions. This holistic evaluation framework aligns with the industry’s need for reliable and interpretable fault localization methods.
The overall workflow of the proposed fault detection and localization methodology is illustrated in Figure 4.

3. Results

To demonstrate the effectiveness of the proposed methodology, it was applied to a medium-voltage DCMG, with detailed specifications provided in Table 2 and a schematic representation shown in Figure 5. To evaluate the model’s performance, DC faults were simulated at intervals of 50 m throughout the system. These faults included positive pole-to-ground, negative pole-to-ground, and pole-to-pole scenarios, covering the entire system up to its endpoint.
The generated data were divided into three sets for model evaluation: 70% (896 samples) was allocated for training, 10% (128 samples) for validation, and the remaining 20% (257 samples) was used for testing, in which TOA of first and second TW, the magnitude and the polarity of them, are regarded as the input vector to the learning model. This comprehensive dataset ensures a robust assessment of the proposed method across various fault types and locations.

3.1. Impact of Quantum Feature Extraction on Model Performance

To evaluate the effectiveness of the proposed method, the model was trained using the prepared dataset. The primary objective was to assess the impact of quantum-based feature extraction on fault location accuracy. The integration of quantum feature extraction not only reduces error rates but also highlights the interactions between key features, as evident in the model’s consistent performance across all datasets. This deeper understanding of feature dynamics offers a unique advantage in designing future systems with improved reliability. For this purpose, the system was run under two configurations:
  • The first configuration involved training and testing the system without the quantum feature extraction path. Only the CNN was used to extract features.
  • In the second configuration, the QNN was added alongside the CNN for feature extraction, leveraging quantum-enhanced capabilities.
Table 3 provides a comparison of prediction performance metrics between the proposed QDNN and a standard DNN without quantum enhancements. The table includes six evaluation metrics. The results highlight that the RMSE and MAE values for QDNN are consistently lower than those of the standard DNN across all datasets, indicating reduced prediction errors. The R2, WIA, and VAF values for QDNN are closer to 1, demonstrating the superior accuracy and reliability of the proposed method. Finally, the TUS values for QDNN are smaller, confirming better fault location estimation accuracy.
These results confirm that incorporating quantum feature extraction significantly enhances the model’s performance, especially in reducing error rates and improving prediction accuracy across diverse datasets.

3.2. Principal Component Analysis (PCA) and Feature Extraction Performance of QNN for Fault Localization

The analysis of the QNN-extracted features provides a detailed understanding of its effectiveness in fault location for DCMGs. Table 4 displays the loadings of each input vector component on the first three principal components (PCs) obtained through the PCA projection of QNN-extracted features. These loadings quantify the contributions of input features such as the TOA, magnitude, and polarity of the TWs to each PC. Each loading reflects the contribution of the corresponding feature to the variance captured by the PCs. Higher loadings indicate the greater influence of the feature on the respective PC. For instance, the polarity of the second TW shows the highest loading on PC1, signifying its dominant role in fault localization accuracy.
The results in Table 4 show that specific features, such as the polarity of the second TW and the TOA of the second TW, contribute significantly to PC1 and PC2, respectively. These features capture key aspects of the fault dynamics, enhancing the interpretability of the extracted feature space. The polarity of the first TW also plays a crucial role in refining fault localization through PC3. For PC1, which captures the highest variance (92.86%), the polarity of the second traveling wave exhibits the most significant contribution with a loading of 0.78996, followed by the magnitude of the second TW with a loading of −0.46681. These features play a dominant role in fault location, emphasizing their importance in the QNN’s feature extraction process. PC2, which explains an additional 7.03% of the variance, is heavily influenced by the TOA of the second TW, showing a loading of 0.902457. This highlights the significance of temporal features in capturing the behavior of faults. PC3, accounting for 0.11% of the variance, primarily incorporates the magnitude of the second TW (0.614752) and the polarity of the first TW (0.542118), refining the fault localization accuracy.
Table 5 summarizes the explained variance ratio and cumulative variance for the first three PCs. The cumulative variance reaches 99.99%, demonstrating that almost all critical information from the QNN-extracted features is retained within these three components. This indicates the effectiveness of PCA in reducing the dimensionality of the feature space while preserving the most essential fault-related characteristics.
The results in Table 5 highlight the effectiveness of PCA in dimensionality reduction. The first PC captures 92.86% of the variance, and the first three PCs together retain 99.99% of the variance. This demonstrates that PCA preserves nearly all critical information from the original data while reducing the feature space, ensuring efficient input to the QDNN model. Moreover, Figure 6 provides a visualization of the QNN-extracted features in a 3D space after PCA projection.
The points are color-coded based on the cumulative distance of faults (in meters), illustrating distinct clusters corresponding to different fault scenarios. This clustering demonstrates that the QNN successfully transforms raw input features into a structured feature space, enabling clear differentiation between fault locations. The smooth distribution of data points further validates the robustness of the proposed method in handling variations in fault characteristics.
Finally, the reconstruction error is calculated as 0.000525039, which highlights the negligible loss of information during the PCA transformation. This low error validates the reliability of combining QNN with PCA for feature extraction and dimensionality reduction. The ability to preserve critical information ensures that the reduced feature space is highly representative of the original input data.

3.3. XAI Analysis of DNN and QDNN Models

The integration of XAI methods into the proposed framework provides a deeper understanding of the decision-making process for both the standard DNN and the QDNN. The analysis focuses on comparing the feature importance and interpretability of these two models using SHAP. The SHAP analysis further enables a granular understanding of how each feature impacts the fault localization process. By breaking down complex interactions into intuitive visualizations, the proposed framework bridges the gap between advanced machine learning techniques and practical fault management strategies, empowering engineers to make informed decisions. The results are presented in Figure 7 and Figure 8, where the impact of input features on the fault location predictions is visualized and quantified.
Figure 7a shows the SHAP summary plot for the DNN-based fault location model. The plot highlights the impact of each input feature on the model’s output. As can be seen, the TOA of the first TW has the highest impact on the model’s predictions, with a wide range of SHAP values reflecting its critical role in fault localization.
Furthermore, the magnitude of the first TW and the magnitude of the second TW also contribute significantly, indicating their importance in capturing fault characteristics, while the polarity of both the first and the second TWs has a relatively smaller impact, suggesting these features play a secondary role in the DNN’s decision-making process. Figure 7b provides a bar chart of feature importance derived from the SHAP values. The results confirm that temporal and magnitude-based features dominate the fault location predictions, with the TOA of the first TW and the magnitude of the first TW being the most influential. Similarly, Figure 8a illustrates the SHAP summary plot for the QDNN-based fault location model, where the input features are quantum-enhanced.
As can be seen, QNN Feature 2 exhibits the highest impact on the model’s output, as indicated by its large range of SHAP values. This feature encapsulates quantum-enhanced information related to fault scenarios, which significantly improves interpretability and accuracy. Other QNN-derived features (QNN Features 1, 3, 4, 5, and 6) contribute to the model’s predictions but with relatively smaller impacts compared to QNN Feature 2. The SHAP values for QDNN show more concentrated patterns compared to the DNN, reflecting the QDNN’s enhanced ability to process fault-related information effectively. Figure 8b presents the feature importance bar chart for the QDNN. It confirms that QNN Feature 2 dominates the fault location predictions, with a significantly higher importance score compared to other features, which highlights the advantage of integrating quantum feature extraction in the proposed framework.
Based on Figure 7 and Figure 8, the comparison between DNN and QDNN demonstrates the superiority of the QDNN model in capturing and utilizing critical fault information, which can be summarized as follows:
  • The DNN relies heavily on temporal features like the TOA of the first TW, which, while effective, lack the advanced representation provided by quantum-enhanced features.
  • The QDNN, by contrast, leverages quantum-derived features that encapsulate complex relationships within the fault data, leading to more precise and interpretable predictions.
  • The SHAP analysis highlights that the QDNN’s predictions are more robust, with fewer outliers in SHAP values and a clear dominance of specific quantum features.

4. Discussion

In this section, the proposed QDNN is compared with five baseline models: extreme gradient boost (XGB) regressor, k-nearest neighbors regressor (KNN), recurrent neural network (RNN), support vector regressor (SVR), and multilayer perceptron regressor (MLP). The goal of this comparison is to evaluate the performance of these methods in fault location prediction and to demonstrate the superiority of the proposed QDNN framework.
First, to better understand the aforementioned models, these methods are explained:
  • The XGB is a gradient-boosting framework that builds decision trees sequentially, where each tree attempts to correct the errors of the previous ones. This method is known for its speed and efficiency, which makes it suitable for structured data. In this study, XGB was initialized with a random state of 42 to ensure consistent results during training and testing. The default parameters, such as the learning rate and tree depth, were used for simplicity and stability.
  • The KNN works by predicting an output value based on the average of the closest k points in the feature space. The model relies on the Euclidean distance for determining similarity, and k = 5 was used as the default neighbor count. Normalization of features was performed with minimum and maximum scalers to ensure fair distance calculations across features with different scales.
  • The RNN captures sequential patterns and temporal dependencies in data. It was implemented with a single hidden recurrent layer containing 50 units and ReLU activation for non-linearity. The data were reshaped into a 3D format to account for their sequential nature. Adam optimizer was used for training over 50 epochs, with a batch size of 32, ensuring convergence while maintaining computational efficiency.
  • The SVR maps input features into a higher-dimensional space using a kernel function, allowing it to capture non-linear relationships. The radial basis function kernel was employed to handle the complex patterns inherent in fault data. Parameters such as C = 100 for regularization and ϵ = 0.1 to set the margin of tolerance were selected to balance prediction accuracy and model complexity.
  • The MLP is a neural network with multiple layers that maps input features to outputs through backpropagation. In this study, it was configured with two hidden layers containing 50 and 25 units, respectively. The ReLU activation function was applied for non-linearity, and the Adam solver was used for training. The training process was capped at 1000 iterations to ensure convergence and reduce computational load.
The performance metrics of these models on the training, validation, and test datasets are presented in Table 6, Table 7, and Table 8, respectively. These metrics collectively provide a comprehensive evaluation of prediction accuracy and robustness.
On the training dataset in Table 6, QDNN exhibits the best performance, with the lowest RMSE (2.051224) and MAE (0.629743), as well as near-perfect scores for R2, WIA, and VAF. This indicates that QDNN can capture the underlying fault patterns effectively during training. In contrast, MLP shows the highest error values, suggesting overfitting or difficulty in modeling the data.
For the validation dataset according to Table 7, QDNN maintains its superiority, achieving an RMSE of 5.130047 and MAE of 1.660156, confirming its ability to generalize to unseen data. XGB and KNN also perform reasonably well, though their error rates are higher compared to QDNN. SVR and MLP exhibit significant error rates, highlighting their limitations in handling complex fault characteristics.
Finally, on the test dataset in Table 8, QDNN continues to outperform other methods, with an RMSE of 4.725696 and MAE of 1.490272. This demonstrates its robustness and adaptability to new fault scenarios. While XGB and KNN provide acceptable results, they fall short of the precision offered by QDNN. MLP consistently shows the poorest performance across all datasets, indicating that it struggles to model the non-linear and complex relationships inherent in fault location data.

4.1. Analysis of Correlation Plots for Model Comparison

Figure 9 presents the correlation plots comparing predicted versus actual values for the test dataset across six machine learning models. These plots are crucial for evaluating the alignment of predicted and actual values, with the diagonal line representing perfect correlation.
As can be seen in this figure, the QDNN plot demonstrates exceptional alignment with minimal errors, showcasing its ability to capture non-linear relationships effectively. XGB follows closely, with slightly larger deviations at extreme values, but remains highly reliable for complex patterns. KNN exhibits a noticeable spread, reflecting moderate prediction variability. The RNN plot highlights strong sequential modeling capabilities, though small deviations occur. SVR performs well but shows minor sensitivity to hyperparameters, visible in its scatter. MLP achieves good accuracy, with slight dispersion at higher values indicating occasional prediction inconsistencies.

4.2. Analysis of Receiver Operating Characteristic for Regression (RROC) and Area over the Curve (AOC) Results

RROC is a visualization tool used to assess the performance of regression models by analyzing the trade-off between over-estimation and under-estimation, which considers prediction errors, plotting over-estimation on the x-axis and under-estimation on the y-axis. This method helps evaluate model robustness across different datasets. The AOC represents the integral of the RROC curve, quantifying the total error. Smaller AOC values indicate better model performance as they reflect reduced over- and under-estimation errors [30]. The RROC curves are shown in Figure 10 for six models across three datasets. For the training data, QDNN and XGB exhibit the smallest areas, highlighting their superior performance in minimizing prediction errors during training.
Other models, such as KNN and RNN, show larger areas, indicating more pronounced estimation errors. In the validation dataset, QDNN and XGB maintain smaller AOCs, reflecting their robustness. SVR and MLP also perform well but are slightly less precise compared to QDNN and XGB. For the test data, QDNN and XGB again demonstrate their superiority, with minimal over- and under-estimation errors. MLP and SVR perform competitively, while KNN and RNN present larger AOCs, suggesting room for improvement.
The AOC values for fault location predictions are summarized in Table 9 for the training, validation, and test datasets. For the training data, QDNN achieves the lowest AOC value of 1,681,274.969, followed by XGB with 18,883,454.26, which highlights their strong predictive capability during training. SVR has the largest AOC, reflecting higher error levels. In the validation data, QDNN again achieves the smallest AOC value of 215,381.875, followed closely by XGB. SVR and MLP show competitive performance but with slightly higher AOC values. For the test data, QDNN continues to outperform with the lowest AOC value of 736,879.5625. XGB and SVR also perform strongly, while KNN and RNN exhibit higher AOC values, suggesting potential limitations in generalization. Moreover, the integration of SHAP values into the evaluation process sets the QDNN framework apart from conventional methods. Unlike other models that function as black-box systems, the proposed method offers a transparent and interpretable solution, making it highly suitable for critical applications in fault management and grid stability.

4.3. Evaluation Under Varying Fault Resistances

To thoroughly assess the robustness and reliability of the proposed method, we simulated fault scenarios with varying fault resistances: 10, 50, and 100 Ω. The primary goal of this analysis was to compare the performance of the proposed QDNN-based method against other approaches.
The results, presented in Table 10, clearly demonstrate the superior performance of the proposed method. For instance, at a fault resistance of 10 Ω, the RMSE for our method was 4.83517, while XGB and KNN recorded RMSE values of 22.51716 and 11.202487, respectively, showing a substantial difference in accuracy. This trend persisted as the fault resistance increased to 50 Ω, where the RMSE of our method rose only slightly to 4.972953, compared to much higher values of 24.277341 for XGB and 12.081113 for KNN. Even at the highest fault resistance of 100 Ω, the proposed method maintained a low RMSE of 5.02021, while other methods, such as SVR and MLP, exhibited drastic performance declines, with RMSE values reaching 150.23052 and 401.48540, respectively. The MAE values further support this finding. At 100 Ω, our method recorded an MAE of just 1.920077, compared to 12.112224 for XGB and 6.603113 for KNN. These low error values demonstrate that our method is exceptionally precise in estimating fault locations, even under challenging conditions with high fault resistances.
Metrics like R2, WIA, and VAF highlight the stability and reliability of the proposed approach. For example, the R2 value for our method remained consistently high, ranging from 0.99994 at 10 Ω to 0.99949 at 100 Ω, indicating excellent model fit. In contrast, RNN and SVR showed significant drops in R2, with values falling to 0.99477 and 0.99434, respectively, at 100 Ω. Similarly, WIA and VAF values for our method stayed close to optimal, with VAF recording 0.99949 at 100 Ω, compared to 0.99406 for SVR and 0.99349 for MLP. Interestingly, the TUS metric, which measures temporal uncertainty, also highlights the robustness of the proposed method. At 100 Ω, the TUS for our method was only 0.00365, while XGB and RNN exhibited TUS values of 0.01409 and 0.01879, respectively.
These findings collectively emphasize the adaptability of the proposed QDNN-based method in handling varying fault resistances. The robustness of the QDNN framework under varying fault resistances highlights its potential for deployment in diverse DCMG configurations. By maintaining accuracy across challenging fault conditions, the model ensures that critical faults are detected promptly, reducing the risk of cascading failures in real-world systems.

4.4. Evaluation Under Different Noise Levels

To evaluate the robustness of the proposed method under noisy conditions, fault scenarios were simulated at varying noise levels with signal-to-noise ratios (SNR) of 5, 10, and 20 dB. This analysis compared the proposed QDNN-based method with other advanced techniques, providing a comprehensive overview of each method’s accuracy, error resilience, and adaptability under noise disturbances.
The results, summarized in Table 11, highlight the exceptional performance of the proposed method. For instance, at the most challenging noise level of 5 dB, the RMSE for QDNN was 5.39085, significantly outperforming other methods such as XGB (33.10547), KNN (18.67081), and SVR (200.30736). Even at 10 dB, the RMSE for QDNN remained notably low at 5.05827, compared to 27.58789 for XGB and 38.13404 for RNN. This minimal increase in RMSE demonstrates the proposed method’s strong resistance to noise. Similarly, MAE values for QDNN remained substantially lower than those of competing methods. At 10 dB, the proposed method recorded an MAE of 1.77781, compared to 12.76790 for XGB and 26.73370 for RNN. At 20 dB, the MAE of QDNN was 1.49768, further illustrating its superior precision in fault detection despite the presence of noise.
Metrics such as R2, WIA, and VAF reinforce the stability of the proposed approach. At 20 dB, R2 for QDNN was 0.99998, remaining close to optimal and surpassing methods like MLP (0.96790) and SVR (0.99547). The WIA and VAF values for QDNN followed a similar trend, staying near 1 across all noise levels, indicating minimal degradation in accuracy. For instance, at 5 dB, the WIA for QDNN was 0.99990, while SVR and MLP recorded 0.99684 and 0.97964, respectively. The TUS metric further highlights the robustness of QDNN in handling noise-induced temporal uncertainty. At 5 dB, the TUS for QDNN was 0.00364, significantly lower than 0.02454 for XGB and 0.11993 for SVR. This low TUS value indicates that QDNN can provide reliable fault detection with minimal delay, a critical factor in real-time fault management.
These findings collectively emphasize that the proposed QDNN-based method maintains high accuracy, low error, and robust performance across all noise levels, making it a superior choice for real-world applications where measurement signals are often contaminated by environmental or operational disturbances. By providing consistent fault localization accuracy even in noisy environments, the QDNN framework ensures reliable performance in practical scenarios.

4.5. Explainability Analysis Using SHAP for Varying Fault Resistances and Noise Levels

To demonstrate the explainability and interpretability of the proposed QDNN-based method, SHAP was utilized to analyze the model’s behavior under different fault resistances and noise levels. The SHAP method quantifies the contribution of each feature to the fault location prediction by breaking down the overall prediction into individual feature impacts. This interpretability not only ensures the reliability of the model but also allows system operators to better understand how fault characteristics, such as TOA or polarity, influence the decision-making process. Additionally, the balanced distribution of SHAP values across features ensures that no single feature disproportionately influences the prediction, which is particularly important for avoiding overfitting in real-world scenarios.
The SHAP summary plots for the QDNN model, shown in Figure 11, highlight how feature importance and impact vary across different fault resistances.
At 10 Ω fault resistance, shown in Figure 11a, features like QNN Feature 2 and QNN Feature 6 exhibit the highest SHAP values, indicating their critical role in fault location prediction. The SHAP summary plot reveals that these features significantly influence the model’s output, as their high SHAP values indicate a strong correlation between these features and the fault location results. The plot also demonstrates a balanced distribution of positive and negative SHAP values, showing that the model remains consistent and accurate at this fault resistance level. Importantly, the concentration of SHAP values around key features indicates that the model can effectively prioritize relevant factors over noise, ensuring robust predictions even under low-resistance conditions.
When the fault resistance increases to 50 Ω, as depicted in Figure 11b, the model dynamically shifts its reliance on features. Here, QNN Feature 3 and QNN Feature 5 become more prominent, as evidenced by their increased SHAP values. This indicates the model’s adaptability in re-evaluating the importance of features under different resistance levels. Such behavior reflects a deep learning architecture capable of recognizing subtle changes in fault characteristics and recalibrating its predictive focus accordingly. The separation between high-impact and low-impact features is clearly visible in the SHAP plot, further demonstrating the model’s robustness. By dynamically adjusting its reliance on features, the model ensures that the fault location remains accurate and interpretable, even as the resistance level introduces complexity.
At the highest fault resistance of 100 Ω, shown in Figure 11c, QNN Feature 4 demonstrates a crucial role in maintaining prediction accuracy under these extreme conditions. Although the SHAP plot for this feature indicates a relatively smaller range of values, its consistent importance in the model’s decision-making highlights its resilience and critical role in extracting key fault characteristics. The SHAP values for this feature indicate its critical importance in driving accurate predictions under high-resistance conditions. The SHAP plot at this resistance level shows a wider range of values, which could suggest increased variability in the input–output relationship due to the challenging fault scenario. However, the model maintains a clear prioritization of key features, as evidenced by the consistent dominance of QNN Feature 4. This behavior underscores the model’s resilience and its ability to extract meaningful patterns, even under extreme conditions.
To further enhance the transparency and applicability of our proposed QDNN-based fault location framework, we incorporated it to provide detailed insights into the feature contributions under varying noise levels. In this study, fault scenarios under three distinct SNR levels were simulated. The SHAP summary plots, shown in Figure 12, were utilized to analyze the impact of noise on the contribution of individual features within the QDNN model. These features, derived from the QNN architecture, include critical temporal and spatial parameters extracted for fault detection and localization. At low SNR levels, like 5 and 10 dB illustrated in Figure 12a,b, the SHAP values demonstrate a higher variance, reflecting the model’s adaptability to handling noisy inputs while maintaining reliable predictions. As the noise level decreases, as with 20 dB shown in Figure 12c, the SHAP values become more concentrated, showing reduced uncertainty and a stronger alignment with the most influential features, such as QNN Feature 2 and QNN Feature 4. Across all SNR levels, the QDNN model consistently identifies the same key features as highly influential, showcasing its robustness and noise resilience. This detailed analysis validates the applicability of our method in real-world settings where noise is inevitable, emphasizing its capability to maintain transparency and interpretability under varying conditions.

5. Conclusions

This paper presented an innovative framework combining XAI with QDNN to address the persistent challenges in fault location within DCMGs. The proposed methodology integrated HOSST for TW detection, CNNs for local feature extraction, and QNNs for quantum-enhanced data representation. Additionally, the incorporation of BD-LSTM layers and attention mechanisms enabled effective capture and emphasis of critical temporal patterns in fault signals. The experimental results demonstrated the framework’s exceptional accuracy and robustness across diverse DCMG configurations and voltage levels. The QDNN approach consistently outperformed traditional and modern fault location methods, achieving superior metrics in RMSE, MAE, R2, TUS, WIA, and VAF across training, validation, and test datasets along with superiority in correlation plot, RROC curve, and AOC values; specifically, QDNN achieved an RMSE reduction of up to 78.5% compared to traditional methods such as RNN and SVR, demonstrating its ability to significantly minimize prediction errors. Additionally, the AOC values for QDNN were consistently the lowest across the training, validation, and test datasets. Furthermore, the inclusion of XAI techniques and SHAP analysis provided valuable insights into feature importance, enhancing model transparency and interpretability. By overcoming limitations such as sensitivity to noise, reliance on communication links, and the challenge of high-resistance faults, this study contributes a highly adaptable and efficient solution to DCMG fault detection and location. To validate the robustness of the proposed framework in real-world scenarios, we conducted additional experiments modeling practical challenges by evaluating the model’s performance under varying fault resistances and exploring the impact of noise levels on prediction accuracy, showing that QDNN remains stable even under high-resistance faults and high-noise conditions. These results highlight the proposed framework for hybrid AC/DC microgrids, power distribution networks, and other intelligent energy systems. By combining state-of-the-art quantum computing techniques with XAI, the proposed framework not only enhances fault localization accuracy but also offers unparalleled transparency. This dual focus ensures that the model can be seamlessly adopted in real-world systems, fostering greater trust and reliability among system operators. Future research could focus on optimizing the quantum circuit design and expanding the dataset to further validate the framework under broader operational scenarios.

Author Contributions

Conceptualization, A.H.P. and F.N.; methodology, A.H.P. and F.N.; software, A.H.P.; validation, A.H.P. and F.N.; formal analysis, A.H.P.; investigation, A.H.P. and F.N.; data curation, A.H.P.; writing—original draft preparation, A.H.P.; writing—review and editing, A.H.P. and F.N.; visualization, A.H.P.; supervision, F.N.; project administration, F.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Acknowledgments

For the purpose of open access, the author has applied a Creative Commons Attribution (CC BY) license to any Author Accepted Manuscript version arising from this submission.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DNNDeep neural network
QNNQuantum neural network
QDNNQuantum deep neural network
SSTSynchrosqueezing transform
HOSSTHigh-order synchrosqueezing transform
BD-LSTMBidirectional Long Short-Term Memory
CNNConvolutional neural network
XGBExtreme gradient boosting
KNNK-nearest neighbors
RNNRecurrent neural network
SVRSupport vector regression
MLPMultilayer perceptron
RMSERoot mean squared error
MAEMean absolute error
R2R-squared
WIAWillmott’s index of agreement
TUSTheil’s U-statistic
VAFVariance accounted for
RROCRegression receiver operating characteristic
AOCArea over curve
PCPrincipal component
PCAPrincipal component analysis
TWTraveling wave
XAIExplainable artificial intelligence
SHAPShapley additive explanations
TOATime of arrival
ReLURectified linear unit
PVPhotovoltaic
SNRSignal-to-noise ratio

References

  1. Ardriani, T.; Dahono, P.A.; Rizqiawan, A.; Garnia, E.; Sastya, P.D.; Arofat, A.H.; Ridwan, M. A DC Microgrid System for Powering Remote Areas. Energies 2021, 14, 493. [Google Scholar] [CrossRef]
  2. Ali, S.; Zheng, Z.; Aillerie, M.; Sawicki, J.P.; Péra, M.C.; Hissel, D. A Review of DC Microgrid Energy Management Systems Dedicated to Residential Applications. Energies 2021, 14, 4308. [Google Scholar] [CrossRef]
  3. Srivastava, C.; Tripathy, M. DC Microgrid Protection Issues and Schemes: A Critical Review. Renew. Sustain. Energy Rev. 2021, 151, 111546. [Google Scholar] [CrossRef]
  4. Mishra, M.; Patnaik, B.; Biswal, M.; Hasan, S.; Bansal, R.C. A Systematic Review on DC-Microgrid Protection and Grounding Techniques: Issues, Challenges and Future Perspective. Appl. Energy 2022, 313, 118810. [Google Scholar] [CrossRef]
  5. Sanati, S.; Mosayebi, A.; Kamwa, I. Advanced Rapid Directional Over-Current Protection for DC Microgrids Using K-Means Clustering. IEEE Trans. Power Deliv. 2024, 39, 1088–1099. [Google Scholar] [CrossRef]
  6. Braitor, A.C.; Iovine, A.; Siguerdidjane, H. A Power Consensus Controller with Overvoltage Protection for Meshed DC Microgrids. IFAC-Pap. 2023, 56, 11762–11767. [Google Scholar] [CrossRef]
  7. Poursaeed, A.H.; Doostizadeh, M.; Hossein Beigi Fard, S.; Baharvand, A.H.; Namdari, F. Optimal Coordination of Directional Overcurrent Relays: A Fast and Precise Quadratically Constrained Quadratic Programming Solution Methodology. IET Gener. Transm. Distrib. 2024, 18, 4342–4357. [Google Scholar] [CrossRef]
  8. Anjaiah, K.; Pattnaik, S.R.; Dash, P.K.; Bisoi, R. A Real-Time DC Faults Diagnosis in a DC Ring Microgrid by Using Derivative Current Based Optimal Weighted Broad Learning System. Appl. Soft Comput. 2023, 142, 110334. [Google Scholar] [CrossRef]
  9. Saxena, A.; Sharma, N.K.; Samantaray, S.R. An Enhanced Differential Protection Scheme for LVDC Microgrid. IEEE J. Emerg. Sel. Top. Power Electron. 2022, 10, 2114–2125. [Google Scholar] [CrossRef]
  10. Zhang, W.; Zhang, H.; Zhi, N. A Novel Protection Strategy for DC Microgrid Considering Communication Failure. Energy Rep. 2023, 9, 2035–2044. [Google Scholar] [CrossRef]
  11. Aboelezz, A.M.; Sedhom, B.E.; El-Saadawi, M.M. Pilot Distance Protection Scheme for DC Zonal Shipboard Microgrid. In Proceedings of the 2021 4th International Symposium on Advanced Electrical and Communication Technologies, ISAECT 2021, Alkhobar, Saudi Arabia, 6–8 December 2021. [Google Scholar] [CrossRef]
  12. Poursaeed, A.H.; Namdari, F. High-Speed Algorithm for Fault Detection and Location in DC Microgrids Based on a Novel Time–Frequency Analysis. IET Gener. Transm. Distrib. 2024, 18, 4259–4278. [Google Scholar] [CrossRef]
  13. Banerjee, T.; Miao, Z.; Fan, L. Traveling Wave Based Fault Location Methods: Review and Demonstration. In Proceedings of the 2023 North American Power Symposium, NAPS 2023, Asheville, NC, USA, 15–17 October 2023. [Google Scholar] [CrossRef]
  14. Fedorov, A.; Petrov, V.; Afanasieva, O.; Zlobina, I. Limitations of Traveling Wave Fault Location. In Proceedings of the 2020 Ural Smart Energy Conference, USEC 2020, Ekaterinburg, Russia, 13–15 November 2020; pp. 21–25. [Google Scholar] [CrossRef]
  15. Seo, H.C. Development of New Protection Scheme in DC Microgrid Using Wavelet Transform. Energies 2022, 15, 283. [Google Scholar] [CrossRef]
  16. Rao, G.K.; Jena, P. Unit Protection of DC Microgrid Based on the Teager Energy. In Proceedings of the 2020 IEEE International Conference on Power Electronics, Smart Grid and Renewable Energy, PESGRE 2020, Cochin, India, 2–4 January 2020. [Google Scholar] [CrossRef]
  17. Grcić, I.; Pandžić, H.; Novosel, D. Fault Detection in DC Microgrids Using Short-Time Fourier Transform. Energies 2021, 14, 277. [Google Scholar] [CrossRef]
  18. Pan, P.; Mandal, R.K.; Manohar, M.; Shukla, S.K. An Intelligent Protection Scheme for DC Microgrid Using Hilbert–Huang Transform with Robustness against PV Intermittency and DER Outage. Electr. Eng. 2024, 106, 5967–5985. [Google Scholar] [CrossRef]
  19. Poursaeed, A.H.; Namdari, F. An Ultra-Fast Directional Protection Scheme for DC Microgrids Based on High-Order Synchrosqueezing Transform. Sustain. Energy Technol. Assess. 2023, 59, 103407. [Google Scholar] [CrossRef]
  20. Bayati, N.; Baghaee, H.R.; Hajizadeh, A.; Soltani, M.; Lin, Z. Mathematical Morphology-Based Local Fault Detection in DC Microgrid Clusters. Electr. Power Syst. Res. 2021, 192, 106981. [Google Scholar] [CrossRef]
  21. Almutairy, I.; Alluhaidan, M. Fault Diagnosis Based Approach to Protecting DC Microgrid Using Machine Learning Technique. Procedia Comput. Sci. 2017, 114, 449–456. [Google Scholar] [CrossRef]
  22. Bayati, N.; Balouji, E.; Baghaee, H.R.; Hajizadeh, A.; Soltani, M.; Lin, Z.; Savaghebi, M. Locating High-Impedance Faults in DC Microgrid Clusters Using Support Vector Machines. Appl. Energy 2022, 308, 118338. [Google Scholar] [CrossRef]
  23. Tiwari, S.P.; Koley, E. A Decision Tree-Based Algorithm for Fault Detection and Section Identification of DC Microgrid. In DC Microgrids; John Wiley & Sons: Hoboken, NJ, USA, 2021; pp. 397–420. [Google Scholar] [CrossRef]
  24. Ibrahim, M.H.; Badran, E.A.; Abdel-Rahman, M.H. Detect, Classify, and Locate Faults in DC Microgrids Based on Support Vector Machines and Bagged Trees in the Machine Learning Approach. IEEE Access 2024, 12, 139199–139224. [Google Scholar] [CrossRef]
  25. Jalli, R.K.; Mishra, S.P.; Dash, P.K.; Naik, J. Fault Analysis of Photovoltaic Based DC Microgrid Using Deep Learning Randomized Neural Network. Appl. Soft Comput. 2022, 126, 109314. [Google Scholar] [CrossRef]
  26. Reddy, O.Y.; Chatterjee, S.; Chakraborty, A.K. Bilayered Fault Detection and Classification Scheme for Low-Voltage DC Microgrid with Weighted KNN and Decision Tree. Int. J. Green Energy 2022, 19, 1149–1159. [Google Scholar] [CrossRef]
  27. Prince, S.K.; Affijulla, S.; Panda, G. Affijulla, S.; Panda, G. A DWT-RNN-Assisted Intelligent Differential Protection Scheme for Grid-Tied and Islanded DC Microgrid. In Sustainable Energy and Technological Advancements; Springer: Singapore, 2022; pp. 247–257. [Google Scholar] [CrossRef]
  28. Binqadhi, H.; Hamanah, W.M.; Shafiullah, M.; Alam, M.S.; AlMuhaini, M.M.; Abido, M.A. A Comprehensive Survey on Advancement and Challenges of DC Microgrid Protection. Sustainability 2024, 16, 6008. [Google Scholar] [CrossRef]
  29. Poursaeed, A.H.; Namdari, F. Online Voltage Stability Monitoring and Prediction by Using Support Vector Machine Considering Overcurrent Protection for Transmission Lines. Iran. J. Electr. Electron. Eng. 2020, 16, 325–335. [Google Scholar] [CrossRef]
  30. Poursaeed, A.H.; Namdari, F. Online Transient Stability Assessment Implementing the Weighted Least-Square Support Vector Machine with the Consideration of Protection Relays. Prot. Control Mod. Power Syst. 2025, 10, 1–17. [Google Scholar] [CrossRef]
Figure 1. Schematic representation of HOSST.
Figure 1. Schematic representation of HOSST.
Energies 18 00908 g001
Figure 2. Proposed QDNN Architecture for fault detection and location in DCMG. The architecture consists of two primary processing paths: (1) a CNN-based feature extraction module, responsible for capturing spatial patterns in fault signal characteristics, and (2) a QNN-enhanced quantum feature extraction module, leveraging quantum principles to improve fault location accuracy. The extracted features are then combined and processed by a BD-LSTM layer, which learns sequential dependencies from past and future data points. The attention mechanism further refines the learned features by emphasizing the most relevant temporal characteristics, ensuring robust decision-making under varying fault conditions.
Figure 2. Proposed QDNN Architecture for fault detection and location in DCMG. The architecture consists of two primary processing paths: (1) a CNN-based feature extraction module, responsible for capturing spatial patterns in fault signal characteristics, and (2) a QNN-enhanced quantum feature extraction module, leveraging quantum principles to improve fault location accuracy. The extracted features are then combined and processed by a BD-LSTM layer, which learns sequential dependencies from past and future data points. The attention mechanism further refines the learned features by emphasizing the most relevant temporal characteristics, ensuring robust decision-making under varying fault conditions.
Energies 18 00908 g002
Figure 3. The flowchart of the proposed method.
Figure 3. The flowchart of the proposed method.
Energies 18 00908 g003
Figure 4. Overview of the proposed method.
Figure 4. Overview of the proposed method.
Energies 18 00908 g004
Figure 5. Schematic representation of the case study system, a medium-voltage DCMG.
Figure 5. Schematic representation of the case study system, a medium-voltage DCMG.
Energies 18 00908 g005
Figure 6. Visualization of QNN-extracted features in a 3D PCA-projected space.
Figure 6. Visualization of QNN-extracted features in a 3D PCA-projected space.
Energies 18 00908 g006
Figure 7. SHAP analysis of the DNN-based fault location model, demonstrating the influence of different features on model predictions. (a) The SHAP summary plot illustrates that the TOA of the first TW has the highest impact on the model’s output, with its wide SHAP value range confirming its critical role in fault localization. Other important features include the magnitude of the first and second TWs, while TW polarity has a relatively smaller impact. (b) The feature importance bar chart confirms that the DNN primarily relies on temporal features, which may limit its ability to capture more complex relationships compared to the QDNN.
Figure 7. SHAP analysis of the DNN-based fault location model, demonstrating the influence of different features on model predictions. (a) The SHAP summary plot illustrates that the TOA of the first TW has the highest impact on the model’s output, with its wide SHAP value range confirming its critical role in fault localization. Other important features include the magnitude of the first and second TWs, while TW polarity has a relatively smaller impact. (b) The feature importance bar chart confirms that the DNN primarily relies on temporal features, which may limit its ability to capture more complex relationships compared to the QDNN.
Energies 18 00908 g007
Figure 8. SHAP analysis for the QDNN-based fault location model, highlighting the advantages of quantum-enhanced features. (a) The SHAP summary plot shows that QNN Feature 2 contributes the most to fault location predictions, reflecting the power of quantum feature extraction in capturing complex fault patterns. Other QNN-derived features, including QNN Features 1, 3, 4, 5, and 6, also play a role but with a slightly lower impact. The model exhibits more concentrated SHAP values compared to DNN, indicating its improved fault representation capability. (b) The feature importance bar chart confirms that QDNN prioritizes quantum-derived features over conventional temporal indicators, showcasing its robustness in fault identification across varying conditions.
Figure 8. SHAP analysis for the QDNN-based fault location model, highlighting the advantages of quantum-enhanced features. (a) The SHAP summary plot shows that QNN Feature 2 contributes the most to fault location predictions, reflecting the power of quantum feature extraction in capturing complex fault patterns. Other QNN-derived features, including QNN Features 1, 3, 4, 5, and 6, also play a role but with a slightly lower impact. The model exhibits more concentrated SHAP values compared to DNN, indicating its improved fault representation capability. (b) The feature importance bar chart confirms that QDNN prioritizes quantum-derived features over conventional temporal indicators, showcasing its robustness in fault identification across varying conditions.
Energies 18 00908 g008
Figure 9. Correlation plots comparing predicted versus actual values for the test dataset across six models: (a) QDNN; (b) XGB; (c) KNN; (d) RNN; (e) SVR; (f) MLP.
Figure 9. Correlation plots comparing predicted versus actual values for the test dataset across six models: (a) QDNN; (b) XGB; (c) KNN; (d) RNN; (e) SVR; (f) MLP.
Energies 18 00908 g009aEnergies 18 00908 g009b
Figure 10. RROC curves for various datasets: (a) Training; (b) Validation; (c) Test.
Figure 10. RROC curves for various datasets: (a) Training; (b) Validation; (c) Test.
Energies 18 00908 g010
Figure 11. SHAP summary plots for the QDNN model under varying fault resistances: (a) 10 Ω; (b) 50 Ω; (c) 100 Ω.
Figure 11. SHAP summary plots for the QDNN model under varying fault resistances: (a) 10 Ω; (b) 50 Ω; (c) 100 Ω.
Energies 18 00908 g011
Figure 12. SHAP summary plots for the QDNN model under different noise levels: (a) 5 dB; (b) 10 dB; (c) 20 dB.
Figure 12. SHAP summary plots for the QDNN model under different noise levels: (a) 5 dB; (b) 10 dB; (c) 20 dB.
Energies 18 00908 g012
Table 1. Comparison of protection methods for DCMGs.
Table 1. Comparison of protection methods for DCMGs.
MethodDetection AccuracyFault LocationNoise ResilienceSpeedNetwork LinksScalableComplexityExplainabilityCost EffectiveReal-Time
Over-currentModerate🗶LowModerateNoneLowLowNoneHighLow
Over-voltageModerate🗶ModerateModerateNoneLowLowNoneHighLow
Derivative-BasedHigh🗶LowHighNoneLowModerateNoneModerateModerate
Differential-BasedHigh🗸HighHighHighLowHighLowLowLow
DistanceModerate🗸ModerateModerateLowModerateModerateNoneModerateModerate
TW-BasedVery High🗶ModerateVery HighNoneLowHighNoneLowHigh
Feature ExtractionHighNoHighHighNoneModerateModerateLowModerateModerate
Machine LearningVery High🗸Moderate to HighHighNoneHighVery HighLowModerateModerate
Proposed MethodVery High🗸Very HighUltra-FastNoneVery HighHighVery HighHighVery High
🗸 indicates that the method supports fault location, 🗶 while indicates that it does not.
Table 2. Technical specifications and design aspects of the medium-voltage DCMG.
Table 2. Technical specifications and design aspects of the medium-voltage DCMG.
AspectDetails
Voltage Level±2.5 kV
CablesSingle-core, 1.9/3.3 kV copper conductors
Cable Insulation2.8 mm XLPE insulation, encased in 2.5 mm thick PVC sheathing
Cable Depth0.6 m
Cable SpacingHorizontal spacing of 0.4 m between cables
Cable LengthAs indicated in the system diagram
InterconnectionVoltage source converter (Bus 1) to AC system
Photovoltaic (PV) Units- 4 MW PV at Bus 4, connected via DC/DC boost converter
- 3 MW PV at Bus 12, connected similarly
PV Voltage Regulation5 kV MVDC system regulates the 1 kV PV array voltage using MPPT algorithm
DC Loads- 1 MW at Bus 3 (via DC/DC buck converter)
- 3 MW at Bus 7
- 7 MW at Bus 11
AC Distributed Generator5 MW at Bus 9
GroundingTN-S grounding layout (grounding point at converter midpoints)
Relay PlacementPositioned at initiation and termination points of transmission lines, shown in red in the diagram
Sampling Frequency10 MHz sampling frequency for fault detection
Table 3. Comparison of prediction performance metrics for fault location estimation across training, validation, and test datasets using QDNN and standard DNN.
Table 3. Comparison of prediction performance metrics for fault location estimation across training, validation, and test datasets using QDNN and standard DNN.
DatasetMethodRMSEMAER2TUSWIAVAF
TrainingQDNN2.0512240.6297430.9999990.00116310.999999
DNN2.5193861.0030690.9999980.0014290.9999990.999998
ValidationQDNN5.1300471.6601560.9999920.0028880.9999980.999992
DNN6.0647142.5136720.9999880.0034150.9999970.999988
TestQDNN4.7256961.4902720.9999940.0025150.9999980.999994
DNN7.9370093.4708170.9999820.0042240.9999960.999982
Table 4. Loadings of each input vector component on the first three PCs derived from the 3D PCA projection of features extracted by the QNN.
Table 4. Loadings of each input vector component on the first three PCs derived from the 3D PCA projection of features extracted by the QNN.
Input VectorPC1PC2PC3
Time of Arrival of First TW−0.069180.1976560.369015
Time of Arrival of Second TW−0.320880.902457−0.20946
Magnitude of First TW0.2242670.0647690.314094
Polarity of First TW−0.001630.0004110.542118
Magnitude of Second TW−0.466810.0190440.614752
Polarity of Second TW0.789960.3767560.222461
Table 5. Explained variance ratio and cumulative variance for the first three PCs in the QNN-derived feature space.
Table 5. Explained variance ratio and cumulative variance for the first three PCs in the QNN-derived feature space.
Principal ComponentExplained Variance RatioCumulative Variance
PC10.9285902090.928590209
PC20.070251360.998841569
PC30.0011356760.999977246
Table 6. Performance metrics of different methods for fault location prediction on the training dataset.
Table 6. Performance metrics of different methods for fault location prediction on the training dataset.
MethodRMSEMAER2TUSWIAVAF
QDNN2.0512240.6297430.9999990.00116310.999999
XGB6.8587932.2241690.9999850.0038890.9999960.999985
KNN7.7560462.9129460.9999810.0043980.9999950.999981
RNN26.5823219.982340.9997730.0150720.9999430.999775
SVR100.554546.846660.9967490.0570150.9991510.996753
MLP308.0343252.28680.9694950.1746570.9916630.969503
Table 7. Performance metrics of different methods for fault location prediction on the validation dataset.
Table 7. Performance metrics of different methods for fault location prediction on the validation dataset.
MethodRMSEMAER2TUSWIAVAF
QDNN5.1300471.6601560.9999920.0028880.9999980.999992
XGB19.743238.9848710.9998760.0111160.9999690.999876
KNN10.193444.2968750.9999670.0057390.9999920.999967
RNN25.9739520.758370.9997860.0146240.9999470.999788
SVR99.1982648.187170.9968810.0558510.9991850.996889
MLP314.4502262.02630.9686560.1770440.9914110.968878
Table 8. Performance metrics of different methods for fault location prediction on the test dataset.
Table 8. Performance metrics of different methods for fault location prediction on the test dataset.
MethodRMSEMAER2TUSWIAVAF
QDNN4.7256961.4902720.9999940.0025150.9999980.999994
XGB22.0703110.093520.9998620.0117450.9999650.999863
KNN10.982834.6692610.9999660.0058450.9999910.999966
RNN28.7790721.229750.9997650.0153160.9999410.999772
SVR125.192159.98330.9955610.0666250.998830.995639
MLP334.5712273.90070.9682970.1780520.9912020.968785
Table 9. AOC values for fault location predictions across training, validation, and test datasets using six different methods.
Table 9. AOC values for fault location predictions across training, validation, and test datasets using six different methods.
MetricDatasetQDNNGXBKNNRNNSVRMLP
AOCTraining1,681,274.96918,883,454.2624,092,750280,602,973.34,053,483,35638,076,806,618
Validation215,381.8753,191,692.356850,7505,485,229.94280,391,806804,263,696.5
Test736,879.562515,976,186.373,982,70026,583,528.44508,527,231.83,639,802,700
Table 10. Performance metrics of various methods under different fault resistance levels.
Table 10. Performance metrics of various methods under different fault resistance levels.
MetricsFault Resistance (Ω)QDNNXGBKNNRNNSVRMLP
RMSE04.72569622.0703110.9828328.77907125.1921334.5712
104.83514722.51171611.20248729.354651127.695942341.262624
504.97295324.27734112.08111331.656977137.71131368.02832
1005.0202126.48437213.17939634.534884150.23052401.485440
MAE01.49027210.093524.66926121.22975159.983303273.900710
101.56325310.295394.96264621.65434561.182966279.378714
501.70517511.1028725.13618723.35272565.981631301.290776
1001.92007712.1122246.60311325.47571471.97996328.68084
R200.9999940.9998620.9999660.9997650.9955610.968297
100.9999460.9993210.9994730.9992050.9950630.967813
500.9997430.9973330.9974520.9972340.9930720.965876
1000.9994990.9948460.9949570.9947700.9905830.963456
TUS00.0025150.0117450.0058450.0153160.0666250.178052
100.0027200.0119800.0079620.0156220.0679580.181613
500.0029400.0129210.0084300.0168480.0732880.195857
1000.0035650.0140940.0090140.0183790.0799500.213662
WIA00.9999980.9999650.9999910.9999410.9988300.991202
100.9999480.9994040.9994020.9994710.9983310.990706
500.9997480.9974140.9974640.9970760.9963830.988724
1000.9994980.9949880.9949770.9949900.9938560.986246
VAF00.9999940.9998630.9999660.9997720.9956390.968785
100.9999480.9993240.9994090.9992950.9951410.968301
500.9997410.9973010.9974150.9972110.9931500.966363
1000.9994930.9948650.9949360.9947330.9906610.963941
Table 11. Performance metrics of various methods under different noise levels.
Table 11. Performance metrics of various methods under different noise levels.
MetricsSNR (dB)QDNNXGBKNNRNNSVRMLP
RMSE55.3908533.1054718.6708146.04602200.30736535.31376
105.0582727.5878915.4708838.13404162.44073433.62096
204.7512322.2910211.0999928.95814126.32002337.91645
MAE52.0648015.140286.5379534.2706389.97495409.71101
101.7778112.767905.6031126.7337074.97996341.81084
201.4976810.134394.7120021.4410560.58250275.63961
R250.999930.999120.999470.999020.993630.95891
100.999960.999490.999720.999370.994600.96410
200.999980.999830.999920.999740.995470.96790
TUS50.003640.024540.011920.031180.119930.32249
100.003070.019140.009940.023980.095970.24978
200.002580.012010.006010.015630.067490.17972
WIA50.999900.998160.999110.997820.996840.97964
100.999940.999240.999530.998880.997840.98604
200.999970.999640.999850.999500.998600.99062
VAF50.999960.999230.999420.999020.993640.95893
100.999980.999490.999710.999360.994620.96412
200.999990.999830.999920.999740.995470.96790
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Poursaeed, A.H.; Namdari, F. Explainable AI-Driven Quantum Deep Neural Network for Fault Location in DC Microgrids. Energies 2025, 18, 908. https://doi.org/10.3390/en18040908

AMA Style

Poursaeed AH, Namdari F. Explainable AI-Driven Quantum Deep Neural Network for Fault Location in DC Microgrids. Energies. 2025; 18(4):908. https://doi.org/10.3390/en18040908

Chicago/Turabian Style

Poursaeed, Amir Hossein, and Farhad Namdari. 2025. "Explainable AI-Driven Quantum Deep Neural Network for Fault Location in DC Microgrids" Energies 18, no. 4: 908. https://doi.org/10.3390/en18040908

APA Style

Poursaeed, A. H., & Namdari, F. (2025). Explainable AI-Driven Quantum Deep Neural Network for Fault Location in DC Microgrids. Energies, 18(4), 908. https://doi.org/10.3390/en18040908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop