Next Article in Journal
Alkali-Activated Stabilization of Silt Soil Using Garlic Husk Ash: Mechanical, Microstructural, and Durability Performance
Previous Article in Journal
Study on the Mechanical Behavior of a Large-Segment Fully Prefabricated Subway Station During the Construction Process
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Detection for Multimode Processes Using an Enhanced Gaussian Mixture Model and LC-KSVD Dictionary Learning

1
Key Laboratory of Thermo-Fluid Science and Engineering of MOE, School of Energy and Power Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
Xi’an Thermal Power Research Institute Co., Ltd., Xi’an 710054, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(18), 9943; https://doi.org/10.3390/app15189943
Submission received: 7 August 2025 / Revised: 9 September 2025 / Accepted: 10 September 2025 / Published: 11 September 2025

Abstract

Monitoring multimode industrial processes presents significant challenges due to varying operating conditions, nonlinear dynamics, and mode-dependent feature distributions. This paper proposes a novel process monitoring framework that integrates an enhanced Gaussian Mixture Model (GMM) for mode identification with Label Consistent K-SVD (LC-KSVD) for sparse dictionary learning. The improved GMM employs a parallelized Expectation–Maximization algorithm to achieve accurate and scalable mode partitioning in high-dimensional environments. Subsequently, the LC-KSVD then learns label-consistent, discriminative sparse representations, enabling effective monitoring across modes. The proposed method is evaluated through a simulation study and the widely used Continuous Stirred Tank Heater (CSTH) benchmark. Comparative results with traditional techniques such as LNS-PCA and FGMM demonstrate that the proposed method achieves superior fault detection rates (FDRs) and significantly lower false alarm rates (FARs), even under complex mode transitions and mild fault scenarios. Furthermore, the method also provides interpretable fault isolation through reconstruction-error-guided variable contribution analysis. These findings confirm that the proposed LC-KSVD-based scheme offers a reliable solution for fault detection and isolation in multimode process systems.

1. Introduction

To enhance operational safety and ensure product quality, process monitoring and fault isolation have become critical components in modern industrial systems [1]. Over the past few decades, various techniques have been developed to address these tasks, among which multivariate statistical process monitoring (MSPM) has emerged as a mainstream approach due to its effectiveness in handling high-dimensional process data and uncovering abnormal behavior based on statistical patterns [2,3,4].
Conventional MSPM techniques, such as principal component analysis (PCA) and its variants, typically assume that process data follow a time-invariant and unimodal Gaussian distribution [5,6,7]. However, practical industrial processes often operate under multiple distinct modes due to variations in raw material properties, production specifications, throughput demands, and control strategies [8,9,10]. Under such multimodal conditions, the performance of standard MSPM approaches degrades significantly, as they fail to capture the complex distributional shifts and mode-dependent dynamics.
In response to these limitations, a growing body of research has explored multimode process monitoring methods. For instance, Li et al. proposed a local neighborhood standardization strategy to preprocess multimodal data [11]. Yu et al. adopted a finite Gaussian Mixture Model (GMM) combined with Bayesian inference for mode-wise monitoring [12]. Jiang et al. further integrated GMM with optimal principal component selection to enhance diagnosis capability in multimode processes [13]. More recent studies have extended Bayesian or sparse approaches, such as mode identification with transitional modeling [14], sparse principal component selection coupled with Bayesian inference-based probability [15], and Bayesian reconstruction strategies for specific applications like coal mills [16]. Cui et al. additionally proposed dimensionality-reducing GMM-based reconstruction combined with deep learning models for nonlinear multimode monitoring [17]. These methods demonstrate the effectiveness of combining GMM partitioning with local models or probabilistic inference, yet they still mainly focus on clustering and detection, with limited exploration of discriminative dictionary learning or variable-level fault isolation. Further, Kodamana et al. developed mixtures of probabilistic PCA for modeling and monitoring multimode dynamic processes [18]. Zhang et al. proposed a modified PCA algorithm, designated PCA-EWC, for monitoring multimode processes to overcome catastrophic forgetting of PCA for successive modes [19]. Although these approaches improve upon traditional MSPM in handling multimodal characteristics, they are still fundamentally rooted in variance-preserving dimensionality reduction and assume smooth Gaussian latent structures.
In addition, real-world process data may exhibit sparsity and non-Gaussianity that are poorly modeled by PCA-type methods. Sparse representations not only enhance interpretability but also improve generalization and robustness by emphasizing the most informative features [20,21]. Recent studies have demonstrated that sparse coding techniques offer considerable advantages in high-dimensional and heterogeneous environments [22]. Nevertheless, most current multimode monitoring frameworks do not explicitly exploit sparse structures or address non-Gaussian feature distributions. Furthermore, despite increasing interest in fault detection, few studies have systematically tackled the more challenging problem of fault isolation in multimode systems, where overlapping features and operating conditions complicate variable-wise interpretation. Inspired by developments in computer vision and signal processing [23,24,25,26], where dictionary learning has proven effective in extracting sparse and discriminative representations, this paper proposes a novel process monitoring approach based on label-consistent K-SVD (LC-KSVD) dictionary learning [27,28]. Unlike traditional MSPM models, LC-KSVD embeds supervision directly into the dictionary learning process, enabling the model to construct sparse representations that preserve both structural similarity and class-specific information.
Beyond industrial process monitoring, similar methodological challenges have been widely addressed in the domain of structural health monitoring (SHM). Recent advances have integrated machine learning, deep learning, and metaheuristic optimization with vibration-based analysis to enhance structural damage detection. For instance, hybrid approaches combining particle swarm optimization (PSO), radial basis functions, and emerging algorithms such as the YUKI method have been developed for accurate crack identification in composite beams, demonstrating improved robustness and convergence compared with classical PSO [29]. Other studies have explored optimized neural networks and bio-inspired algorithms for defect prediction in civil and mechanical systems [30,31], as well as deep learning-based frameworks for automated SHM tasks [32]. These developments, while originating in SHM, reflect parallel challenges to those faced in multimode process monitoring, reinforcing the relevance of designing monitoring frameworks that not only achieve high detection accuracy but also maintain interpretability and robustness [33,34].
In this context, the present study focuses on industrial process monitoring and aims to address these parallel challenges by integrating probabilistic mode partitioning, discriminative sparse representation, and interpretable fault isolation into a unified framework. This design ensures that the proposed method remains both accurate and practically interpretable under multimode operating conditions. The proposed framework comprises three major components. First, a parallelized GMM-based mode segmentation method is employed to partition training data into coherent operating regimes. Second, an LC-KSVD model is trained using the labeled multimodal data, and monitoring statistics are derived from the reconstruction error under the learned sparse dictionary. Third, a novel fault isolation strategy based on reconstruction error and missing value estimation is developed, leveraging insights from missing data analysis and reconstruction-based contribution evaluation. Experimental results on both a numerical simulation and the widely studied Continuous Stirred Tank Heater (CSTH) benchmark confirm that the proposed method not only outperforms conventional techniques in fault detection accuracy and false alarm suppression but also provides interpretable and precise fault localization, even under complex multimodal conditions. In summary, the contribution of this work lies in the development of an enhanced multimode monitoring framework that advances beyond existing approaches in several important aspects. GMM is improved through a parallel EM algorithm, which significantly accelerates convergence and makes the method scalable to high-dimensional process data. This improvement is further coupled with LC-KSVD dictionary learning, allowing the monitoring model to capture discriminative and interpretable sparse representations across different operating modes. On this basis, we design a reconstruction-error-guided isolation mechanism that extends sparse dictionary learning from fault detection to variable-level fault localization under multimode conditions. By combining these elements into a unified framework, the proposed method achieves efficient mode partitioning, accurate detection, and interpretable fault isolation.
The remainder of this paper is organized as follows: Section 2 details the methodological foundation, including the enhanced GMM for mode identification and the LC-KSVD algorithm for sparse dictionary learning and monitoring. Section 3 introduces the monitoring and fault isolation procedures under the proposed framework. Section 4 presents two case studies to validate the performance of the proposed method. Finally, Section 5 concludes the paper and discusses potential avenues for future research.

2. Methodology

2.1. Enhanced Gaussian Mixture Model

2.1.1. Gaussian Mixture Model Fundamentals

GMM is a probabilistic model that assumes the observed data are generated from a mixture of several Gaussian distributions, each representing a different latent component. It is widely used for modeling multimode distributions, especially in unsupervised clustering and pattern recognition tasks. Formally, given a d-dimensional data point x R d , the probability density function (PDF) of a GMM with K components is defined as
p ( x ) = k = 1 K π k   N ( x μ k , Σ k ) ,
where π k is the mixture coefficient (or prior probability) of the k-th Gaussian component, satisfying k = 1 K π k = 1 and π k 0 , μ k R d is the mean vector of the k-th component, Σ k R d × d is the covariance matrix of the k-th component, and N ( x μ k , Σ k ) denotes the multivariate Gaussian distribution.
N ( x μ , Σ ) = 1 ( 2 π ) d / 2 | Σ | 1 / 2 exp 1 2 ( x μ ) Σ 1 ( x μ ) .
The parameters { π k , μ k , Σ k } k = 1 K are typically estimated via the Expectation– Maximization (EM) algorithm, which iteratively maximizes the log-likelihood of the observed data
log L ( Θ ) = i = 1 N log k = 1 K π k   N ( x i μ k , Σ k ) ,
where Θ = { π k , μ k , Σ k } k = 1 K represents the set of all model parameters and N is the number of training samples.
GMM is particularly well-suited for modeling complex systems such as thermal processes, where data often exhibit multimodal characteristics due to varying operating conditions. The soft clustering nature of GMM enables probabilistic assignment of each data point to multiple components, offering greater flexibility than hard partitioning methods such as k-means.

2.1.2. Parallel EM Algorithm

The Expectation–Maximization (EM) algorithm [35] is a widely used method for estimating the parameters of GMM. Despite its effectiveness, the traditional EM algorithm faces significant computational challenges when applied to large-scale or real-time industrial process data. Specifically, it suffers from slow convergence and poor scalability due to its inherently sequential structure. To address these limitations, we adopt a parallel EM algorithm that exploits the independence of data samples to accelerate computation by distributing the workload across multiple processors. The parallel EM algorithm maintains the core structure of the EM process but parallelizes the Expectation (E) and Maximization (M) steps, thereby significantly improving computational efficiency.
Given the current estimates of parameters Θ ( t ) = { π k ( t ) , μ k ( t ) , Σ k ( t ) } k = 1 K , the E-step computes the posterior probability (responsibility) that a sample x i belongs to the k-th Gaussian component
γ i k ( t ) = π k ( t )   N ( x i μ k ( t ) , Σ k ( t ) ) j = 1 K π j ( t )   N ( x i μ j ( t ) , Σ j ( t ) ) .
This computation is embarrassingly parallel, as each γ i k ( t ) can be calculated independently across all data samples i 1 , , N and mixture components k 1 , , K .
Once responsibilities are computed, the M-step updates the model parameters based on the following soft assignments:
(1) Update of mixing coefficients
π k ( t + 1 ) = 1 N i = 1 N γ i k ( t ) .
(2) Update of mean vectors
μ k ( t + 1 ) = i = 1 N γ i k ( t ) x i i = 1 N γ i k ( t ) .
(3) Update of covariance matrices
Σ k ( t + 1 ) = i = 1 N γ i k ( t ) x i μ k ( t + 1 ) x i μ k ( t + 1 ) i = 1 N γ i k ( t ) .
These summations over i are also naturally parallelizable. By partitioning the dataset and computing partial sums on different processors, followed by a reduction operation (e.g., MPI Reduce in distributed computing), the updated parameters can be efficiently obtained. The convergence of the algorithm is monitored by evaluating the log-likelihood of the dataset
log L ( Θ ( t ) ) = i = 1 N log k = 1 K π k ( t )   N ( x i μ k ( t ) , Σ k ( t ) ) .
The iteration stops when the relative improvement of the log-likelihood falls below a pre-defined threshold ε
log L ( Θ ( t + 1 ) ) log L ( Θ ( t ) ) log L ( Θ ( t ) ) < ε .
The parallel EM algorithm offers several key advantages over its traditional counterpart, particularly in the context of large-scale and real-time industrial process data. By distributing the computational workload of both the E-step and M-step across multiple processors, the algorithm significantly improves execution speed and scalability, enabling efficient parameter estimation even for high-dimensional datasets. This parallelization reduces the convergence time without compromising accuracy, making it well-suited for scenarios where rapid decision-making is essential, such as fault diagnosis in thermal power systems. Furthermore, the algorithm retains the statistical robustness of conventional EM while overcoming its limitations in handling large volumes of multimodal data. By leveraging this parallel EM framework, the proposed diagnostic model can efficiently and autonomously identify distinct operating modes in thermal processes, providing a reliable basis for subsequent dictionary learning.

2.2. Fault Detection with LC-KSVD Dictionary Learning

2.2.1. K-Singular Value Decomposition

Sparse representation and dictionary learning have been extensively studied for feature extraction and pattern recognition in high-dimensional data. Traditional dictionary learning methods, such as K-SVD, focus primarily on minimizing reconstruction error but do not explicitly incorporate label information, which limits their effectiveness in classification-oriented tasks like fault diagnosis. To address this limitation, the Label-Consistent K-SVD (LC-KSVD) algorithm extends K-SVD by embedding label supervision into the dictionary learning process, thereby improving the discriminative power of the learned sparse codes.
Let Y R m × n denote the training dataset collected from m sensors over n samples. The objective of dictionary learning is to find the optimal dictionary D and corresponding sparse codes X that minimize the reconstruction error, defined by
L ( D , X ) = arg   min D , X Y D X F 2 s . t . X 0 T ,
where D R m × k is the dictionary matrix, X R k × n is the sparse coefficient matrix, and T denotes the maximum allowable number of non-zero elements in each column of X.
This optimization problem can be effectively solved using the K-SVD algorithm [36], which updates dictionary atoms column-by-column. Let d j be the j-th column of D, and let x T j denote the j-th row of X. The objective in Equation (10) can be rewritten as
D , X = arg   min D , X Y D X F 2   = arg   min d j , x T j Y j j d j x T j d j x T j F 2   = arg   min d j , x T j E j d j x T j F 2 ,
where E j denotes the residual matrix computed by removing the contribution of all atoms except d j .
In the dictionary update stage of K-SVD, each atom is refined while keeping the sparsity pattern of the coefficients unchanged. To achieve this, only the data samples that currently use a given atom are considered, and the corresponding residual matrix is constructed by removing the contribution of all other atoms. The optimization problem is then reduced to finding the best rank-one approximation of this residual, which can be efficiently solved through singular value decomposition (SVD). The leading singular vectors and values provide the updated dictionary atom and its associated sparse coefficients. After all atoms are updated in this manner, the sparse coding matrix is recomputed using the Orthogonal Matching Pursuit (OMP) algorithm [37] to maintain consistency.

2.2.2. Label Consistent K-SVD Dictionary Learning

After completing the mode division of the training dataset, the operation data from different modes can be regarded as labeled training samples. To effectively extract the distinctive features of samples from different modes, this section adopts the Label Consistent K-SVD (LC-KSVD) dictionary learning algorithm to construct a state monitoring model based on processed multimode data. The LC-KSVD algorithm, proposed by Jiang et al. [27], performs sparse coding on labeled training samples.
The training samples Y R m × n can be divided into N mutually exclusive subsets according to their labels. For the subset i, its corresponding label is denoted by y i . According to the method in [27], the optimization objective of the LC-KSVD algorithm can be formulated as
min D , W , A , X Y D X F 2 + α Q A X F 2 + β H W X F 2 ,
s . t . x i 0 T ,
where α weights the label-consistency term ( Q A X ) and β weights the classifier fitting term ( H W X ), larger values of α and β enforce stronger discriminability, while overly large values may hurt reconstruction.
To facilitate optimization, the LC-KSVD objective can be reformulated into an augmented dictionary learning problem, which can then be solved using the standard K-SVD procedure. In this reformulation, both the discriminative projection matrices and the dictionary are coupled, and thus appropriate initialization is required prior to the K-SVD iterations. To this end, the parameters associated with label consistency and classification are estimated through ridge regression, which provides closed-form solutions for the initialization. This ensures that the label-consistency term aligns the sparse codes with the predefined class structure, while the classifier term guides the dictionary to be discriminative. Once initialized in this way, the iterative dictionary learning follows the same update scheme as conventional K-SVD, but with the additional constraints that encourage both accurate reconstruction and discriminative power.
Once the initial values of A and W are computed, the initial dictionary D can be obtained. Subsequently, the K-SVD algorithm is applied to solve the reformulated objective. At this point, all parameters of the LC-KSVD model, including A , D , W , have been estimated.
To ensure numerical stability and consistency across dictionary atoms, a normalization step is applied to each parameter as follows:
D = d 1 d 1 2 , , d k d k 2 ,
A = a 1 d 1 2 , , a k d k 2 ,
W = w 1 d 1 2 , , w k d k 2 .
These normalization steps ensure that each atom in the dictionary and associated transformation matrices are scaled consistently, thus enhancing the stability and robustness of the LC-KSVD model.

3. Process Monitoring and Fault Detection Procedure

Define the training data as Y = [ Y 1 , Y 2 , , Y c ] , consisting of c modes, where each Y i R m × n i represents data from the i-th mode with n i samples. The procedure for process monitoring using LC-KSVD follows a workflow analogous to PCA-based monitoring.
LC-KSVD models are trained using multimode normal condition data. A statistical model of normal conditions is constructed by evaluating the monitoring statistic derived from the reconstruction error of the LC-KSVD model. Once the model is built, the statistics of test data can be calculated and used to detect abnormalities based on control limits.
The reconstruction error is defined as
Re   =   y i D x i 2 2 .
To determine whether a test sample contains fault information, control limits for the statistic must be predefined. A kernel density estimator (KDE) [38] is used to approximate the probability density function of the reconstruction error,
f ( x ) = 1 n σ i = 1 n K x x i σ ,
where x is the data under consideration, x i is a sample value from the dataset, σ is the window width, and K ( · ) is the kernel function.
The control limit δ 2 is determined by the α -quantile of the estimated distribution. In summary, the LC-KSVD-based monitoring procedure consists of the following steps:
  • Acquire multimode normal operation data with c modes.
  • Normalize data using the mean and standard deviation of each variable.
  • Train the LC-KSVD model on normal data using the K-SVD algorithm.
  • Construct the reconstruction error set from normal operation data.
  • Determine control limits δ 2 using KDE.
  • Solve the optimization problem
    L ( X ) = arg   min X y new D X 2 2 + λ X 1
    using FISTA [39], where the regularization parameter λ balances reconstruction fidelity and sparsity. Larger values enforce sparser codes, while smaller values allow denser representations.
Then, compute the statistic Re of the new data and determine whether a fault has occurred by comparing it with the threshold δ 2 .
After detecting a fault in a sample, it is essential to isolate the fault source. Inspired by reconstruction-based contribution analysis and missing value analysis in multivariate statistics, we propose a fault isolation strategy.
Assume y r is a faulty sample, and its revised version is given by
y normal = y r Ξ i f i ,
where Ξ i is a direction vector with 1 at the i-th index and 0 elsewhere, and f i is the revised amplitude.
The corresponding reconstruction error is defined as
Re normal = arg   min f i , x i y r Ξ i f i D x i F 2 s . t .   x i 0 T .
If the correct fault direction and amplitude are identified, it is possible to achieve Re normal < δ 2 . For each dimension, we assume the associated variable to be faulty and remove it. If the reconstruction error drops below the threshold, that variable is likely the fault source.
Hence, Equation (19) is reformulated as
Re normal = arg   min x i β ( y r D x i ) F 2   s . t .   x i 0 T ,
where β is a direction vector with the i-th element set to 0 and all others set to 1.
This optimization problem can be solved using the FISTA algorithm. The full fault isolation procedure is summarized in Algorithm 1.
Algorithm 1 Fault Isolation for LC-KSVD Process Monitoring
1:
Initialize: S = { 1 , 2 , , m } , Res = { } , fs = { }
2:
for each j S  do
3:
  Initialize β , where β i = 0 for i = j , β i = 1 otherwise
4:
  Solve Equation (20) using FISTA, add Re normal to set Res
5:
  Calculate min ( Res ) and corresponding index i, add i to fs
6:
  if  Re min < δ 2 or len ( fs ) = m  then return fs
7:
  else
8:
     Update S S fs , reset Res
9:
  end if
10:
end for
Compared with prior multimode sparse coding and GMM–Bayesian methods, which primarily emphasize mode partitioning and global fault detection [12,13,15,16,17], the proposed framework goes a step further by explicitly coupling mode identification with discriminative dictionary learning and extending the monitoring scope to variable-level fault isolation. Specifically, the parallel EM-based GMM ensures scalable and robust mode separation, while LC-KSVD leverages label consistency to construct mode-aware sparse representations that enhance discriminability between normal and faulty conditions. Building upon these representations, a reconstruction-error-guided strategy provides interpretable diagnostics by localizing the contribution of individual variables to detected faults. This joint design addresses the limitations of previous multimode approaches, which often lack either discriminative modeling capability or transparent fault isolation, thereby offering a more comprehensive solution for process monitoring and fault detection.

4. Case Studies

In this section, the performance of the proposed method is evaluated using both a numerical simulation and the Continuous Stirred Tank Heater (CSTH) process. The LC-KSVD method is compared with traditional FGMM [12] and LNS-PCA [40] algorithms. To quantitatively assess the performance of these approaches, two widely used indicators are adopted: the fault detection rate (FDR) and the false alarm rate (FAR), defined as follows:
FDR = # ( Re > δ 2 f 0 ) # ( f 0 ) × 100 % ,
FAR = # ( Re > δ 2 f = 0 ) # ( f = 0 ) × 100 % ,
where # { x } denotes the number of samples in set x, and f is the fault indicator.
In addition to FDR and FAR, we also report the mean squared error (MSE) [41] of reconstruction as a supplementary measure of model accuracy
MSE = 1 N m i = 1 N y i y ^ i 2 2 . = 1 N m Y Y ^ F 2 ,
Since the proposed framework relies on sparse representation and reconstruction, MSE provides a direct evaluation of how well the monitoring model captures normal process behavior. A lower reconstruction error on normal data indicates that deviations caused by faults will yield more distinguishable residuals, thereby enhancing fault detection sensitivity.

4.1. Simulation Case

To demonstrate the effectiveness of the proposed method, a numerical simulation model with five variables is considered. The system is formulated as
x 1 x 2 x 3 x 4 x 5 = 0.5869 0.3796 0.7672 0.0489 0.7910 0.4009 0.5929 0.2168 0.3867 0.7956 s 1 s 2 + e 1 e 2 e 3 e 4 e 5 ,
where e 1 , , e 5 are zero-mean white noise with a standard deviation of 0.01, and s 1 , s 2 are two dependent latent variables (data sources). The process operates in three modes, defined as
Mode   1 :   s 1 N ( 10 , 0.8 ) ,   s 2 N ( 12 , 1.3 ) , Mode   2 :   s 1 U ( 0.6 , 6.5 ) ,   s 2 U ( 0.7 , 20 ) , Mode   3 :   s 1 U ( 5 , 5 ) ,   s 2 N ( 30 , 2.5 )
where N ( μ , σ 2 ) denotes the Gaussian distribution with mean μ and variance σ 2 , and U ( a , b ) represents a uniform distribution within the interval [ a , b ] .
For the training phase, 1200 normal samples (400 per mode) are generated according to Equation (24). These samples are used to train the monitoring models. For testing, 1000 samples are generated for each fault scenario, with the first 400 being normal and the remaining 600 containing faults. The specific fault settings are detailed in Table 1. In the simulation case, the dictionary size and sparsity level were set to 200 and 1, respectively. The regularization parameters in the optimization problem in Equation (18) were selected through cross-validation on the training data, resulting in α = 0.41 , β = 0.32 , and λ = 0.05 . The reconstruction accuracy of the proposed method was evaluated using MSE on the normalized training data. The results show that the three operating modes yield consistently low MSE values of 0.078, 0.082, and 0.075, respectively. These small errors indicate that the learned dictionaries accurately capture the intrinsic structure of each mode, thereby ensuring reliable reconstruction of normal process behavior. Such high reconstruction precision provides a solid basis for distinguishing deviations caused by faults and confirms the suitability of the proposed framework for multimode monitoring.
The fault detection performance was quantitatively assessed using the FDR and FAR defined in Equations (21) and (22), respectively. Comparative results among LC-KSVD, LNS-PCA (SPE and T2 statistics), and FGMM (BIP index) are presented in Table 2 and Table 3. As shown in Table 2, the LC-KSVD approach achieved a perfect FDR of 1.0 in both Fault 1 and Fault 2 and 0.965 in Fault 3. These results demonstrate the superior sensitivity of the proposed method compared to the traditional approaches, particularly under complex multimodal distributions where the performance of PCA-based indices deteriorates significantly. Table 3 further confirms the reliability of LC-KSVD through its low FAR across all scenarios, with values consistently below those of FGMM and LNS-PCA.
The fault detection performance was quantitatively evaluated using the FDR and FAR defined in Equations (21) and (22). Comparative results for LC-KSVD, LNS-PCA (SPE and T2 statistics), FGMM (BIP index) [12], DAE [42], MultiMode PCA [43], and VAE [44] are summarized in Table 2 and Table 3. As shown in Table 2, the proposed LC-KSVD approach achieves the highest sensitivity, with perfect detection rates (FDR = 1.0) for Faults 1 and 2 and 0.965 for Fault 3. In contrast, PCA-based indices show significant performance degradation under multimodal distributions, while the neural network-based and GMM-based methods remain consistently inferior. Table 3 further indicates that LC-KSVD also provides the lowest false alarm rates across all three fault scenarios, maintaining values at or below 0.0125, which is lower than those of all competing methods.
The monitoring curves in Figure 1, Figure 2 and Figure 3 provide a visual confirmation of these quantitative results. After the onset of each fault, the reconstruction-error statistic of LC-KSVD promptly exceeds the control limit δ 2 , with clearer separability between normal and faulty samples than the baseline approaches. These results collectively demonstrate that the proposed LC-KSVD framework not only improves detection sensitivity but also achieves more reliable fault discrimination under complex multimode operating conditions. Moreover, Figure 4 and Figure 5 compare the original and reconstructed feature space representations for Faults 1 and 3, respectively. The reconstruction preserves essential features while filtering noise, thereby enhancing robustness. In addition to fault detection, the capability of the proposed method to isolate the source of the fault was also evaluated. Figure 6 presents the fault isolation result for Fault 2 using the reconstruction-error-based variable contribution analysis. The contribution values of each process variable are computed after fault occurrence. These results indicate that the proposed LC-KSVD model accurately attributes the fault to the correct variable, even in the presence of multimodal variation. The result further confirms that the reconstruction-error-guided fault localization strategy is effective and interpretable.
In addition to fault detection, fault isolation performance is depicted in Figure 7, Figure 8 and Figure 9, where the reconstruction-error-based fault contribution method accurately identified the faulty variables. In particular, the proposed method exhibited strong discrimination capability even for mild or slowly evolving faults such as Fault 3, where other methods tend to fail due to insufficient feature sensitivity.
In summary, the simulation case study confirms that the LC-KSVD-based monitoring scheme not only achieves high detection sensitivity and low false alarm rates across various multimode conditions but also effectively localizes fault variables through its reconstruction-error-guided isolation strategy. These results collectively verify the applicability and superiority of the method for complex industrial process monitoring scenarios.

4.2. Continuous Stirred Tank Heater Process Case

In this section, the proposed LC-KSVD-based method is applied to the Continuous Stirred Tank Heater (CSTH) process. The CSTH simulation platform, originally proposed by Thornhill [45], is a widely used benchmark in process monitoring research due to its realistic modeling of heat and volumetric balances. It has been extensively adopted for performance comparison of various fault detection and isolation techniques. As illustrated in Figure 10, hot and cold water streams are fed into a stirred vessel. The mixture is heated to a target temperature using steam and then discharged from the bottom of the tank. The process is governed by multiple Proportional-Integral (PI) control loops. The manipulated variables include hot water flow rate, cold water flow rate, and steam valve position, while the measured outputs comprise liquid level, temperature, and the flow rates of hot and cold water.
Detailed mode configurations used in this study are presented in Table 4, and three experimental fault cases designed for the CSTH process are described in Table 5. During the training stage, 800 normal samples are collected for each mode to construct the monitoring model, with an equal number of samples per mode. For testing, 400 samples are collected for each case, where the first 100 samples represent normal operation and the remaining 300 samples contain faults. In this case study, the dictionary size is set to 200, and the sparsity level is constrained to 1 to ensure compact and interpretable sparse representations. Similarly, the parameters of the proposed method were determined using cross-validation, with α = 0.28 , β = 0.17 , and λ = 0.06 . For the training set, the reconstruction performance was further assessed across four operating modes. The proposed method achieved MSE values of 0.092, 0.054, 0.083, and 0.074, respectively. These results demonstrate that the framework can maintain consistently low reconstruction errors across different modes, even in the presence of complex dynamics. The particularly small MSE in Mode 2 (0.054) highlights the ability of the model to capture mode-specific characteristics with high fidelity. Overall, the low reconstruction errors across all modes confirm the robustness and precision of the proposed approach in modeling multimode industrial processes, which directly contributes to its superior fault detection capability.
The fault detection performance for the CSTH benchmark is quantitatively evaluated in terms of FDR and FAR, as reported in Table 6 and Table 7. The proposed LC-KSVD method achieves perfect detection in both scenarios (FDR = 1.0, FAR = 0), clearly outperforming all baseline approaches. For Fault 1, which corresponds to a multiplicative deviation in the temperature measurement under Mode 3, LNS-PCA (SPE) and MultiMode PCA reach FDRs of 0.887 and 0.900, respectively, while DAE and VAE perform slightly better with 0.923 and 0.943. In contrast, FGMM exhibits very poor sensitivity with an FDR of only 0.107. LC-KSVD, however, consistently detects all faulty samples without false alarms. For Fault 2, which involves a level bias across Modes 1 and 2, most methods yield relatively high detection rates (e.g., LNS-PCA(T2) and FGMM both achieve 1.0, while DAE and VAE reach 0.863 and 0.910, respectively), but several of them suffer from reduced performance under certain modes, and only LC-KSVD maintains both perfect FDR and zero FAR. The monitoring results in Figure 11 and Figure 12 provide further evidence of these advantages. After the fault onset at sample 101, the reconstruction-error statistic of LC-KSVD promptly and distinctly exceeds the control limit, whereas other methods either respond more slowly, show larger fluctuations, or produce false alarms. The reconstruction-error-based monitoring statistic generated by LC-KSVD clearly exceeds the control limit after fault onset while remaining below the threshold during normal operation. This indicates high sensitivity with minimal over-detection. Figure 13 and Figure 14 show the comparison between original and reconstructed signals under fault conditions. The LC-KSVD reconstruction effectively retains dominant process features while filtering out noise and irrelevant variation, thereby improving the clarity and interpretability of the monitoring signals. Figure 15 and Figure 16 demonstrate the fault isolation results for both cases. The proposed reconstruction-error-guided contribution analysis successfully identifies the faulty variables (temperature and level, respectively), even under strong mode-dependent process behavior. This case confirms that the LC-KSVD framework not only detects faults accurately but also provides valuable interpretability through its sparse representation and variable contribution structure.
For Fault 3, which introduces a random disturbance to the cold-water flow under Modes 1 and 3, the performance gap among methods becomes more evident. As shown in Table 6, LC-KSVD again achieves perfect detection (FDR = 1.0, FAR = 0), whereas the other approaches display noticeable degradation. For example, LNS-PCA (SPE) and MultiMode PCA (T2) report FDRs of 0.580 and 0.673, respectively, while DAE and VAE improve sensitivity to 0.860 and 0.930 but still fall short of LC-KSVD. FGMM shows moderate performance with 0.850 FDR but suffers from higher false alarms. The monitoring curves in Figure 17 confirm that LC-KSVD promptly captures the fault onset at sample 101 with a clear and stable separation from the control limit, whereas alternative methods exhibit delayed or fluctuating responses. Moreover, Figure 18 highlights that the reconstruction step in LC-KSVD enhances detectability relative to the original signals, while Figure 19 shows that the proposed contribution analysis correctly isolates the cold-water flow variable as faulty.

5. Conclusions

This study presents an effective multimode process monitoring framework that combines a parallelized GMM for mode separation with LC-KSVD dictionary learning for fault detection and isolation. The proposed method addresses key challenges in industrial settings, including high dimensionality, overlapping modes, and subtle fault signatures. By leveraging the strengths of probabilistic mode clustering and supervised sparse representation, the method not only improves monitoring accuracy but also enhances interpretability through contribution-based fault localization. Extensive experiments on both simulated and real-world CSTH systems confirm the superiority of the proposed approach in terms of detection sensitivity, false alarm suppression, and fault isolation accuracy. Compared with conventional methods, the LC-KSVD framework achieves near-perfect fault detection rates while maintaining zero or near-zero false alarms across all tested scenarios. Additionally, the reconstruction-error-based analysis enables precise identification of faulty variables under multimodal dynamics. Overall, the proposed monitoring strategy demonstrates strong potential for deployment in real industrial applications requiring robust and explainable fault diagnosis under diverse and evolving operating conditions.
Although the proposed framework demonstrates strong performance in both the numerical and CSTH case studies, several limitations should be acknowledged. As a supervised dictionary learning method, LC-KSVD may be sensitive to mislabeled training data, which could compromise the discriminative power of the learned representations. In addition, while the parallel EM and sparse coding steps improve efficiency, the scalability to very high-dimensional processes with hundreds of sensors or very long time-series data still requires further validation. Future research will focus on incorporating robust or semi-supervised strategies to alleviate the impact of noisy labels, conducting more systematic evaluations on gradual and non-stationary faults, and exploring hybrid schemes that integrate LC-KSVD with deep learning architectures to enhance scalability and representation power in large-scale industrial applications.

Author Contributions

Methodology, D.Z., K.H. and Q.D.; Validation, D.Z. and K.H.; Formal analysis, Q.D.; Investigation, K.H.; Writing—original draft, D.Z. and Shengshan Bi; Writing—review & editing, D.Z. and S.B. All authors have read and agreed to the published version of the manuscript.

Funding

The authors appreciate the financial support by the Natural Science Basic Research Program of Shaanxi (No. 2023-JC-YB-389), the Key R&D Program of Shanxi Province (S2023-YF-LLRH-QCYK-0067), and the National Key R&D Program (2022YFB4100700).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

Author Kang He was employed by the Xi’an Thermal Power Research Institute. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Iqbal, R.; Maniak, T.; Doctor, F.; Karyotis, C. Fault detection and isolation in industrial processes using deep learning approaches. IEEE Trans. Ind. Inform. 2019, 15, 3077–3084. [Google Scholar] [CrossRef]
  2. Fan, W.; Zhu, Q.; Ren, S.; Xu, B.; Si, F. Multivariate temporal process monitoring with graph-based predictable feature analysis. Can. J. Chem. Eng. 2023, 101, 909–924. [Google Scholar] [CrossRef]
  3. Fan, W.; Zhang, L.; Zhou, Y.; Yu, C.; Li, Y.; Lv, N. Semi-Supervised Probabilistic Predictable Feature Analysis for Concurrent Process-Quality Monitoring of a Thermal Power Plant. IEEE Trans. Instrum. Meas. 2025, 74, 1–12. [Google Scholar] [CrossRef]
  4. An, J.; Weng, Q.; Ren, S.; Si, F. Multi-mode process fault diagnosis method integrating GMM with reconstruction-based PCA. Can. J. Chem. Eng. 2025. [Google Scholar] [CrossRef]
  5. De Ketelaere, B.; Hubert, M.; Schmitt, E. Overview of PCA-based statistical process-monitoring methods for time-dependent, high-dimensional data. J. Qual. Technol. 2015, 47, 318–335. [Google Scholar] [CrossRef]
  6. Zhu, Q.; Liu, Q.; Qin, S.J. Concurrent quality and process monitoring with canonical correlation analysis. J. Process Control 2017, 60, 95–103. [Google Scholar] [CrossRef]
  7. Zhu, Q. Latent variable regression for supervised modeling and monitoring. IEEE/CAA J. Autom. Sinica 2020, 7, 800–811. [Google Scholar] [CrossRef]
  8. Fan, W.; Ren, S.; Yu, C.; Yu, H.; Wang, P.; Si, F. A mixture of probabilistic predictable feature analysis for multi-mode dynamic process monitoring. J. Taiwan Inst. Chem. Eng. 2023, 143, 104635. [Google Scholar] [CrossRef]
  9. Wang, F.; Tan, S.; Peng, J.; Chang, Y. Process monitoring based on mode identification for multi-mode process with transitions. Chemom. Intell. Lab. Syst. 2012, 110, 144–155. [Google Scholar] [CrossRef]
  10. Du, W.; Fan, Y.; Zhang, Y. Multimode process monitoring based on data-driven method. J. Frankl. Inst. 2017, 354, 2613–2627. [Google Scholar] [CrossRef]
  11. Li, Y.; Yang, D. Local component based principal component analysis model for multimode process monitoring. Chin. J. Chem. Eng. 2021, 34, 116–124. [Google Scholar] [CrossRef]
  12. Yu, J.; Qin, S.J. Multimode process monitoring with Bayesian inference-based finite Gaussian mixture models. AIChE J. 2008, 54, 1811–1829. [Google Scholar] [CrossRef]
  13. Jiang, Q.; Huang, B.; Yan, X. GMM and optimal principal components-based Bayesian method for multimode fault diagnosis. Comput. Chem. Eng. 2016, 84, 338–349. [Google Scholar] [CrossRef]
  14. Tan, S.; Wang, F.; Peng, J.; Chang, Y.; Wang, S. Multimode process monitoring based on mode identification. Ind. Eng. Chem. Res. 2012, 51, 374–388. [Google Scholar] [CrossRef]
  15. Jiang, X.; Zhao, H.; Jin, B. Multimode Process Monitoring Based on Sparse Principal Component Selection and Bayesian Inference-Based Probability. Math. Probl. Eng. 2015, 2015, 465372. [Google Scholar] [CrossRef]
  16. Fan, W.; Ren, S.; Zhu, Q.; Jia, Z.; Bai, D.; Si, F. A novel multi-mode Bayesian method for the process monitoring and fault diagnosis of coal mills. IEEE Access 2021, 9, 22914–22926. [Google Scholar] [CrossRef]
  17. Cui, Y.; Fan, W.; Zhou, Y. Dimensionality reducing Gaussian mixture-based reconstruction for fault detection in multimode processes. Can. J. Chem. Eng. 2024, 102, 4267–4280. [Google Scholar] [CrossRef]
  18. Kodamana, H.; Raveendran, R.; Huang, B. Mixtures of probabilistic PCA with common structure latent bases for process monitoring. IEEE Trans. Control Syst. Technol. 2017, 27, 838–846. [Google Scholar] [CrossRef]
  19. Zhang, J.; Zhou, D.; Chen, M. Monitoring multimode processes: A modified PCA algorithm with continual learning ability. J. Process Control 2021, 103, 76–86. [Google Scholar] [CrossRef]
  20. Xu, S.; Muselet, D.; Trémeau, A. Sparse coding and normalization for deep Fisher score representation. Comput. Vis. Image Underst. 2022, 220, 103436. [Google Scholar] [CrossRef]
  21. Xing, X.; Xie, R.; Zhong, W. Model-based sparse coding beyond Gaussian independent model. Comput. Stat. Data Anal. 2022, 166, 107336. [Google Scholar] [CrossRef]
  22. Ma, H.; Li, S.; Lu, J.; Gong, S.; Yu, T. Impulsive wavelet based probability sparse coding model for bearing fault diagnosis. Measurement 2022, 194, 110969. [Google Scholar] [CrossRef]
  23. Liu, F.; Xu, X.; Xing, X.; Guo, K.; Wang, L. Simple-action-guided dictionary learning for complex action recognition. Neurocomputing 2022, 501, 387–396. [Google Scholar] [CrossRef]
  24. Zhang, L.; Tang, J.; Li, G.; Chen, W. Audio magnetotelluric denoising via variational mode decomposition and adaptive dictionary learning. J. Appl. Geophys. 2022, 204, 104748. [Google Scholar] [CrossRef]
  25. Wang, S.x.; Ge, H.w.; Gou, J.p.; Ou, W.h.; Yin, H.f.; Su, S.z. Fusion of effective dimension reduction and discriminative dictionary learning for high-dimensional classification. Comput. Electr. Eng. 2022, 99, 107684. [Google Scholar] [CrossRef]
  26. Guo, Y.; Sun, H.; Hao, S. Adaptive dictionary and structure learning for unsupervised feature selection. Inf. Process. Manag. 2022, 59, 102931. [Google Scholar] [CrossRef]
  27. Jiang, Z.; Lin, Z.; Davis, L.S. Label consistent K-SVD: Learning a discriminative dictionary for recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2651–2664. [Google Scholar] [CrossRef] [PubMed]
  28. Kariwala, V.; Odiowei, P.E.; Cao, Y.; Chen, T. A branch and bound method for isolation of faulty variables through missing variable analysis. J. Process Control 2010, 20, 1198–1206. [Google Scholar] [CrossRef]
  29. Khatir, A.; Capozucca, R.; Khatir, S.; Magagnini, E.; Benaissa, B.; Le Thanh, C.; Wahab, M.A. A new hybrid PSO-YUKI for double cracks identification in CFRP cantilever beam. Compos. Struct. 2023, 311, 116803. [Google Scholar] [CrossRef]
  30. Khatir, A.; Capozucca, R.; Magagnini, E.; Oulad Brahim, A.; Osmani, A.; Khatir, S.; Abualigah, L. Advancing structural integrity prediction with optimized neural network and vibration analysis. J. Struct. Integr. Maint. 2024, 9, 2390258. [Google Scholar] [CrossRef]
  31. Khatir, A.; Brahim, A.O.; Magagnini, E. An efficient computational system for defect prediction through neural network and bio-inspired algorithms. HCMCOU J. Sci. Adv. Comput. Struct. 2024, 14, 66–80. [Google Scholar] [CrossRef]
  32. Khatir, A.; Capozucca, R.; Khatir, S.; Magagnini, E.; Le Thanh, C.; Riahi, M.K. Advancements and emerging trends in integrating machine learning and deep learning for SHM in mechanical and civil engineering: A comprehensive review. J. Braz. Soc. Mech. Sci. Eng. 2025, 47, 1–34. [Google Scholar] [CrossRef]
  33. Mansouri, A.; Tiachacht, S.; Ait-Aider, H.; Khatir, S.; Khatir, A.; Cuong-Le, T. A novel Optimization-Based Damage Detection in Beam Systems Using Advanced Algorithms for Joint-Induced Structural Vibrations. J. Vib. Eng. Technol. 2025, 13, 1–30. [Google Scholar] [CrossRef]
  34. Yu, W.; Huang, B.; Xiao, G.; Zhang, C. A Variational Bayesian Inference-Based Robust Dissimilarity Analytics Model for Industrial Fault Detection. IEEE Trans. Syst. Man, Cybern. Syst. 2025, 55, 3275–3286. [Google Scholar] [CrossRef]
  35. Fan, W.; Jin, Y.; Yu, C.; Zhou, Y. PSINDy: Probabilistic sparse identification of nonlinear dynamics for temporal process modeling and fault detection. J. Taiwan Inst. Chem. Eng. 2023, 153, 105238. [Google Scholar] [CrossRef]
  36. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  37. Pati, Y.C.; Rezaiifar, R.; Krishnaprasad, P.S. Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition. In Proceedings of the 27th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 1–3 November 1993; pp. 40–44. [Google Scholar]
  38. Fan, W.; Ji, Y.; Wen, H.; Liu, X.; Yu, H.; Yu, C.; Meng, Q. Concurrent quality and process monitoring with a probabilistic sparse nonlinear dynamic method. Control. Eng. Pract. 2025, 165, 106551. [Google Scholar] [CrossRef]
  39. Xu, Q.; Yang, D.; Tan, J.; Sawatzky, A.; Anastasio, M.A. Accelerated fast iterative shrinkage thresholding algorithms for sparsity-regularized cone-beam CT image reconstruction. Med. Phys. 2016, 43, 1849–1872. [Google Scholar] [CrossRef]
  40. Ma, H.; Hu, Y.; Shi, H. A novel local neighborhood standardization strategy and its application in fault detection of multimode processes. Chemom. Intell. Lab. Syst. 2012, 118, 287–300. [Google Scholar] [CrossRef]
  41. Xu, B.; Zhu, Q. An efficient dynamic auto-regressive CCA for time series imputation with irregular sampling. IEEE Trans. Autom. Sci. Eng. 2022, 21, 442–451. [Google Scholar] [CrossRef]
  42. Qian, J.; Song, Z.; Yao, Y.; Zhu, Z.; Zhang, X. A review on autoencoder based representation learning for fault detection and diagnosis in industrial processes. Chemom. Intell. Lab. Syst. 2022, 231, 104711. [Google Scholar] [CrossRef]
  43. Xu, X.; Xie, L.; Wang, S. Multimode process monitoring with PCA mixture model. Comput. Electr. Eng. 2014, 40, 2101–2112. [Google Scholar] [CrossRef]
  44. Lee, S.; Kwak, M.; Tsui, K.L.; Kim, S.B. Process monitoring using variational autoencoder for high-dimensional nonlinear processes. Eng. Appl. Artif. Intell. 2019, 83, 13–27. [Google Scholar] [CrossRef]
  45. Thornhill, N.F.; Patwardhan, S.C.; Shah, S.L. A continuous stirred tank heater simulation model with applications. J. Process Control 2008, 18, 347–360. [Google Scholar] [CrossRef]
Figure 1. Monitoring results of Fault 1 in the numerical case, comparing the proposed method with baseline approaches, where LC-KSVD detects the fault immediately after sample 400.
Figure 1. Monitoring results of Fault 1 in the numerical case, comparing the proposed method with baseline approaches, where LC-KSVD detects the fault immediately after sample 400.
Applsci 15 09943 g001
Figure 2. Monitoring results of Fault 2 in the numerical case, showing that the proposed method achieves clearer detection after the fault at sample 400 than the baselines.
Figure 2. Monitoring results of Fault 2 in the numerical case, showing that the proposed method achieves clearer detection after the fault at sample 400 than the baselines.
Applsci 15 09943 g002
Figure 3. Monitoring results of Fault 3 in the numerical case, where the gradual drift after sample 400 is more distinctly captured by the proposed method compared with others.
Figure 3. Monitoring results of Fault 3 in the numerical case, where the gradual drift after sample 400 is more distinctly captured by the proposed method compared with others.
Applsci 15 09943 g003
Figure 4. Comparison of monitoring indices for Fault 1, where reconstruction enhances the separation between normal and faulty samples.
Figure 4. Comparison of monitoring indices for Fault 1, where reconstruction enhances the separation between normal and faulty samples.
Applsci 15 09943 g004
Figure 5. Comparison of monitoring indices for Fault 3, showing reconstruction strengthens the detection of gradual drift.
Figure 5. Comparison of monitoring indices for Fault 3, showing reconstruction strengthens the detection of gradual drift.
Applsci 15 09943 g005
Figure 6. Comparison of monitoring indices for Fault 2, with reconstruction producing clearer threshold crossings.
Figure 6. Comparison of monitoring indices for Fault 2, with reconstruction producing clearer threshold crossings.
Applsci 15 09943 g006
Figure 7. Fault isolation results for Fault 1 correctly identifying variable x 1 as the faulty contributor.
Figure 7. Fault isolation results for Fault 1 correctly identifying variable x 1 as the faulty contributor.
Applsci 15 09943 g007
Figure 8. Fault isolation results for Fault 2, where variable x 3 is identified as faulty.
Figure 8. Fault isolation results for Fault 2, where variable x 3 is identified as faulty.
Applsci 15 09943 g008
Figure 9. Fault isolation results for Fault 3, correctly attributing the fault to variable x 5 .
Figure 9. Fault isolation results for Fault 3, correctly attributing the fault to variable x 5 .
Applsci 15 09943 g009
Figure 10. Schematic of the CSTH process.
Figure 10. Schematic of the CSTH process.
Applsci 15 09943 g010
Figure 11. Monitoring results of Fault 1 in CSTH, comparing the proposed method with baselines, where LC-KSVD promptly detects the fault at sample 101.
Figure 11. Monitoring results of Fault 1 in CSTH, comparing the proposed method with baselines, where LC-KSVD promptly detects the fault at sample 101.
Applsci 15 09943 g011
Figure 12. Monitoring results of Fault 2 in CSTH, showing superior detection of the fault after sample 101 compared with other methods.
Figure 12. Monitoring results of Fault 2 in CSTH, showing superior detection of the fault after sample 101 compared with other methods.
Applsci 15 09943 g012
Figure 13. Comparison of monitoring indices for Fault 1 in CSTH, with reconstruction providing clearer fault separation.
Figure 13. Comparison of monitoring indices for Fault 1 in CSTH, with reconstruction providing clearer fault separation.
Applsci 15 09943 g013
Figure 14. Comparison of monitoring indices for Fault 2 in CSTH, where reconstruction enhances detectability relative to original data.
Figure 14. Comparison of monitoring indices for Fault 2 in CSTH, where reconstruction enhances detectability relative to original data.
Applsci 15 09943 g014
Figure 15. Fault isolation results for Fault 1 in CSTH, identifying temperature as the faulty variable.
Figure 15. Fault isolation results for Fault 1 in CSTH, identifying temperature as the faulty variable.
Applsci 15 09943 g015
Figure 16. Fault isolation results for Fault 2 in CSTH, correctly isolating the level variable as faulty.
Figure 16. Fault isolation results for Fault 2 in CSTH, correctly isolating the level variable as faulty.
Applsci 15 09943 g016
Figure 17. Monitoring results of Fault 3 in CSTH, showing superior detection of the fault after sample 101 compared with other methods.
Figure 17. Monitoring results of Fault 3 in CSTH, showing superior detection of the fault after sample 101 compared with other methods.
Applsci 15 09943 g017
Figure 18. Comparison of monitoring indices for Fault 3 in CSTH, where reconstruction enhances detectability relative to original data.
Figure 18. Comparison of monitoring indices for Fault 3 in CSTH, where reconstruction enhances detectability relative to original data.
Applsci 15 09943 g018
Figure 19. Fault isolation results for Fault 3 in CSTH, correctly isolating the level variable as faulty.
Figure 19. Fault isolation results for Fault 3 in CSTH, correctly isolating the level variable as faulty.
Applsci 15 09943 g019
Table 1. Faults description in the numerical case.
Table 1. Faults description in the numerical case.
Fault NumberOperating ModeDescriptionSamples Number
Fault 1Mode1Normal Data1–200
Mode2Normal Data200–400
Mode2set deviation in x 1 : x 1 = x 1 + 0.15 400–1000
Fault 2Mode2Normal Data1–200
Mode3Normal Data200–400
Mode3set deviation in x 3 : x 3 = x 3 + 0.35 400–1000
Fault 3Mode3Normal Data1–200
Mode1Normal Data200–400
Mode1set deviation in x 5 : x 5 = x 5 + 0.005 ( i 400 ) 400–1000
Table 2. FDRS of three faults for the numerical case.
Table 2. FDRS of three faults for the numerical case.
FaultLNS-PCAFGMMDAEMultiMode PCAVAELC-KSVD
no. SPE T2 BIP SPE SPE T2 SPE Re
10.016700.04160.9530.810.740.961
20.9650.590.41830.980.9730.880.9931
30.75830.68330.85830.920.90.8730.9430.965
Table 3. FARS of three faults for the numerical case.
Table 3. FARS of three faults for the numerical case.
FaultLNS-PCAFGMMDAEMultiMode PCAVAELC-KSVD
no. SPE T2 BIP SPE SPE T2 SPE Re
10.010.00250.050.00750.0150.0830.060.0025
20.060.01750.01750.050.01250.02750.030.0125
30.001750.00170.020.00750.010.01250.00750.0075
Table 4. Modes of the CSTH process.
Table 4. Modes of the CSTH process.
ModeLevelTemperatureHot Water Valve
Mode112.010.55.5
Mode216.010.55.0
Mode39.010.54.5
Mode412.013.56.0
Table 5. Three kinds of cases in the CSTH process.
Table 5. Three kinds of cases in the CSTH process.
VariableTrain Data ModeTest Data Fault
TemperatureMode3A multiplicative fault is imposed with 1.008 in mode 3
Mode4
LevelMode1A bias fault is added with −0.5 in mode 2
Mode2
TemperatureMode1A random disturbance between −0.5 and 0.5 to the Cold Water flow in mode 1
Mode3
Table 6. FDRs of two faults for CSTH.
Table 6. FDRs of two faults for CSTH.
FaultLNS-PCAFGMMDAEMultimode PCAVAELC-KSVD
no.SPET2BIPSPESPET2SPERe
10.8870.01670.1070.9230.90.860.9431
20.453110.8630.72310.911
30.580.420.850.860.780.6730.931
Table 7. FARs of two faults for CSTH.
Table 7. FARs of two faults for CSTH.
FaultLNS-PCAFGMMDAEMultimode PCAVAELC-KSVD
no.SPET2BIPSPESPET2SPERe
10.0030000000
20.0030000000
30.002500.00400000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, D.; He, K.; Duan, Q.; Bi, S. Fault Detection for Multimode Processes Using an Enhanced Gaussian Mixture Model and LC-KSVD Dictionary Learning. Appl. Sci. 2025, 15, 9943. https://doi.org/10.3390/app15189943

AMA Style

Zhou D, He K, Duan Q, Bi S. Fault Detection for Multimode Processes Using an Enhanced Gaussian Mixture Model and LC-KSVD Dictionary Learning. Applied Sciences. 2025; 15(18):9943. https://doi.org/10.3390/app15189943

Chicago/Turabian Style

Zhou, Dongyang, Kang He, Qing Duan, and Shengshan Bi. 2025. "Fault Detection for Multimode Processes Using an Enhanced Gaussian Mixture Model and LC-KSVD Dictionary Learning" Applied Sciences 15, no. 18: 9943. https://doi.org/10.3390/app15189943

APA Style

Zhou, D., He, K., Duan, Q., & Bi, S. (2025). Fault Detection for Multimode Processes Using an Enhanced Gaussian Mixture Model and LC-KSVD Dictionary Learning. Applied Sciences, 15(18), 9943. https://doi.org/10.3390/app15189943

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop