1. Introduction
Complex process systems serve as the core foundation of modern manufacturing, with their operational efficiency and safety directly impacting a company’s economic performance and production risk. As automation and intelligent technologies continue to advance, the demands on the stability and reliability of complex process systems are also increasing [
1]. At the same time, the rapid development of modern industry has led to increasingly complex and precise process systems [
2], which are inherently more sensitive to disturbances and thus more prone to anomalies. In practical operation, factors such as equipment conditions, environmental changes, or control deviations may lead to process anomalies, which can in turn cause product quality fluctuations, production interruptions, or even serious safety incidents [
3]. These anomalies not only reduce overall production efficiency but may also result in substantial economic losses and personnel hazards. Therefore, identifying and responding to abnormal states in complex process systems has become a critical issue for ensuring safe and efficient production.
Complex process systems generate large volumes of multivariate time series data during operation, capturing the dynamic evolution of key process parameters and reflecting complex process behaviors as well as inter-variable couplings. Under normal conditions, these time series typically exhibit stable and regular patterns [
4]. However, when anomalies occur within the process, certain variables may experience abrupt changes, trend shifts, or disruptions in their inherent correlations. Therefore, by modeling and analyzing the time series data of complex process systems, anomaly detection can be effectively performed, contributing to improved efficiency and enhanced safety in industrial production [
5].
In the field of anomaly detection for multivariate time series in complex process systems, principal component analysis and its variants have been widely applied for modeling data correlations and extracting features. As complex process systems become increasingly complex, anomaly detection tasks face dual challenges: the growing volume of time series data and the nonlinear interdependencies among variables. Simple principal component analysis (PCA)-based methods are no longer sufficient to meet the safety requirements of modern complex process systems. To address this, Arena et al. introduced a method that combines Monte Carlo-based preprocessing with PCA [
6], enabling effective anomaly detection in photovoltaic cell production environments. Subsequently, to address the difficulty of interpreting distributed correlation structures in complex industrial systems, Theisen et al. proposed a decentralized monitoring framework based on sparse PCA [
7], which significantly improves interpretability and reveals inter-variable relationships. Li et al. developed a dynamic reconstruction PCA method that preserves dynamic information in time series data [
8], thereby enhancing the sensitivity and accuracy of anomaly detection and addressing the limitations of traditional PCA in capturing temporal dynamics. Meanwhile, Zhang et al. proposed a generalized Gaussian distribution-based interval PCA method that incorporates uncertainty in complex process systems [
9], improving detection robustness under fluctuating conditions. Similarly, in response to uncertainty in industrial settings, Kini et al. combined multi-scale PCA with a Kantorovich distance-based scheme to further enhance fault detection performance [
10]. These PCA-based improvements have significantly advanced the modeling of nonlinear dependencies and enhanced the robustness of fault detection in a complex industrial process.
Due to the strong coupling between variables in multivariate time series and the complexity of feature evolution, the difficulty of anomaly detection is significantly increased. To address such issues, Zhou et al. introduced the theory of information granulation and constructed a granular Markov model for multivariate time series anomaly detection by integrating fuzzy C-means clustering [
11]. Subsequently, Liu et al. further combined fuzzy information granules with a Markov random walk strategy by constructing a fuzzy granular distance matrix and iteratively computing the stationary distribution to more effectively detect anomalies in multivariate time series [
12]. Ouyang and Zhang proposed an anomaly detection method that integrates information granule reconstruction and fuzzy rule reasoning mechanisms [
13], where the optimization of granular hierarchy and the enhancement of rule expression improved the accuracy of anomaly detection. To further explore the latent feature information in time series, Guo et al. proposed an interpretable time series anomaly pattern detection method based on granular computing [
14]. This method constructs semantic time series through trend-based granulation and semantic abstraction to achieve anomaly detection. Du et al. proposed a time series anomaly detection method for the sintering process using rectangular information granules [
15], which are constructed from the original series and its first-order difference to identify anomalies via granular similarity analysis. Other studies, such as the method proposed by Huang et al., extract time series information by constructing information granules through a sliding window and perform anomaly detection in complex process systems in combination with PCA [
16]. These granulation-based methods provide a powerful means of capturing structural and semantic patterns in complex time series, thereby improving detection accuracy and interpretability.
In addition to statistical and granulation-based approaches, metaheuristic optimization algorithms, particularly particle swarm optimization (PSO), have attracted significant attention in the field of intelligent monitoring. PSO is a population-based optimization method inspired by the social behavior of birds and fishes. Rezaee Jordehi and Jasni [
17] provided a comprehensive survey on parameter selection in PSO, emphasizing that its performance is highly dependent on the proper configuration of inertia weights, acceleration coefficients, and velocity limits. To enhance model generalization in safety-critical environments, Wang et al. [
18] introduced a hybrid fault diagnosis framework that combines PSO with support vector machines, and successfully applied it to nuclear power plant fault detection. Beyond traditional classification models, Rezaee Jordehi [
19] developed a mixed binary–continuous PSO algorithm for unit commitment in microgrids, demonstrating the versatility of PSO in handling uncertainty and nonlinear optimization problems. In the context of equipment fault diagnosis, Li et al. [
20] applied a PSO-optimized SVM to mechanical fault detection of high-voltage circuit breakers, achieving higher diagnostic accuracy compared with standard support vector machine approaches. More recently, Yu et al. [
21] integrated PSO with deep residual networks for vibration signal analysis in pipeline robots, showing that PSO can also be leveraged to tune deep learning architectures for anomaly detection. These studies collectively indicate that PSO not only plays a crucial role in parameter optimization, but also provides a powerful tool for improving anomaly detection performance in complex industrial systems.
The aforementioned studies indicate that PCA and its variants have demonstrated promising capabilities in anomaly detection for multivariate time series, especially by modeling correlations among variables and reducing noise through linear dimensionality reduction. These methods have been successfully applied in various industrial scenarios to identify abnormal behaviors and enhance detection accuracy. However, traditional PCA-based approaches are fundamentally limited by their global linear assumptions and often struggle to capture the underlying nonlinear dynamics and local variations in complex process systems. In parallel, information granulation methods have gained increasing attention for their ability to abstract raw time series data into meaningful granular representations. By capturing local trends, multi-scale structures, and uncertainty characteristics, granulation-based techniques have shown strong potential in enhancing the interpretability and robustness of anomaly detection. Despite these advantages, most existing granulation frameworks rely heavily on predefined rules or static clustering strategies, which restrict their flexibility and expressiveness in modeling global, nonlinear relationships embedded in high-dimensional and dynamic industrial datasets.
Traditional PCA has limited efficiency in compressing large-scale data, mainly due to its global linear assumption. On the other hand, information granulation provides compact local abstractions of time series but lacks the ability to capture nonlinear dependencies among variables. To overcome these complementary limitations, this study proposes a method that integrates information granulation with kernel principal component analysis (KPCA) for industrial process anomaly detection. In the proposed framework, information granulation is first applied to reduce dimensionality and preserve key local dynamics, after which KPCA with a radial basis function (RBF) kernel is used to extract nonlinear correlations in the high-dimensional space. This combination achieves a balanced trade-off: granulation ensures feature compactness and interpretability, while KPCA enhances nonlinear dependency modeling and robustness. As a result, the integration of the two approaches becomes both necessary and effective for anomaly detection in complex process systems.
As illustrated in
Figure 1, the proposed method follows a two-phase workflow comprising offline modeling and online detection. In the offline modeling phase, historical process data are first processed through information granulation, which structurally abstracts raw time series into compact and meaningful representations. These granulated training data are then used to construct a KPCA model, where the RBF kernel enables the capture of nonlinear inter-variable dependencies in a high-dimensional feature space. Control thresholds are also computed based on the distribution of monitoring statistics derived from the training set. In the online monitoring phase, real-time data are continuously granulated in the same manner, and the granulated test data are projected onto the KPCA feature space using the pretrained model. Monitoring statistics are then computed and compared against the thresholds. If any statistical values exceed the thresholds, the process is flagged as an anomaly; otherwise, it is considered to be operating normally. The synergy between information granulation and KPCA not only achieves a balance between feature abstraction and structural preservation, but also enhances the model’s accuracy and adaptability in identifying diverse anomaly patterns and multivariate dependencies in a complex industrial process. The main contributions of this work are as follows:
- (1)
The principle of justifiable granularity (PJG) is applied to transform raw industrial time series into compact and interpretable granular features, effectively preserving local dynamic characteristics.
- (2)
By introducing the nonlinear mapping capability of KPCA on PJG-based granular features, the method captures nonlinear correlations that cannot be represented by linear PCA, thereby overcoming the limitations of traditional PCA in modeling granulated features.
- (3)
PJG enables structural compression, which effectively alleviates the susceptibility of KPCA to redundant features and noise when directly applied to raw high-dimensional data, thus improving interpretability, robustness, and detection performance.
2. Information Granulation Enhanced KPCA for Industrial Anomaly Detection
Complex process systems often generate high-dimensional and complex time series data, making anomaly detection challenging due to nonlinear variable interactions and redundant information. To address these issues, this study introduces an integrated approach combining information granulation and KPCA. The granulation process abstracts raw data into compact representations, while KPCA enables effective nonlinear feature extraction and anomaly detection.
2.1. Information Granulation Scheme
With the increasing structural complexity of modern industrial systems and the explosive growth of generated data, information granulation has emerged as an effective approach for data modeling and abstraction, enabling structural compression and simplification while preserving essential information features.
To begin with, a sliding window mechanism is introduced to segment the input sequence into local fragments. Each segment, denoted as
, captures the short-term dynamic behaviors present in the data. This localized representation allows for more targeted granule construction. Let the subseries be defined as
where
m is the starting index of the subseries, and
n represents the window length.
denotes the value at time
t, and
is a local segment consisting of
n consecutive data points from the raw time series.
To generate an effective information granule from
, two factors are considered: the number of elements within the granule interval and the compactness of that interval [
22]. This is modeled by an objective function
, as follows:
where
denotes the candidate granule for the segment
, bounded by the lower and upper boundaries
and
. The function
represents the cardinality within the specified set. The term
encourages inclusion of more points within the granule, while
penalizes intervals with a wide range to ensure compactness.
To guide the granulation process, a statistical representative value of the segment
, denoted as
, is computed. Typically,
refers to the median value of the segment, serving as a reference point to split the interval symmetrically or asymmetrically. The objective function
, which balances the density of data points and the proximity to the representative value
, is defined as
Subsequently, the particle swarm optimization (PSO) algorithm [
23] is employed to search for the optimal lower and upper boundaries that maximize the objective function
. PSO is selected because it offers high computational efficiency, fast convergence, and strong global search ability, which are particularly suitable for handling the nonlinear and potentially non-convex nature of
. These characteristics ensure stable and reliable determination of granulation boundaries across different process conditions, as follows:
where
and
represent the optimal lower and upper bounds for the granule interval, respectively. Once the boundaries are obtained, they define the optimal interval granule for the segment
. The complete process is illustrated in
Figure 2.
2.2. KPCA Anomaly Detection Model
Following the granulation phase, each segmented subseries is transformed into a triplet of the form , which captures the local distribution characteristics of the raw signal. To facilitate nonlinear modeling using KPCA, each granule is further encoded into a compact numerical representation. Specifically, for each segment k, two quantities are extracted: the interval range , and the representative value . These jointly describe the span and central tendency of the segment in its local context.
Each segment is thus represented by a feature vector
. By aggregating all such vectors, the entire sequence is transformed into a two-dimensional feature matrix
, where
K denotes the total number of granulated segments. To ensure consistency in scale, each row of
Z is standardized to zero mean and unit variance, yielding the normalized matrix
, which serves as the input to the KPCA model [
24].
Although each segment is compactly represented by two parameters, the joint distribution of granulated features across different segments is often nonlinear due to process complexity, multivariate interactions, and varying operating conditions. Linear PCA tends to capture only global variance directions, which may overlook subtle but important nonlinear dependencies among these features. By contrast, KPCA equipped with the RBF kernel can project the compact feature set into a high-dimensional space, where nonlinear relations among granulated segments are more effectively separated and represented. This allows the model to preserve both the interpretability of information granulation and the nonlinear structural information essential for accurate anomaly detection.
These quantities are stacked column-wise to construct the matrix, as follows:
The standardization is then performed as
where
and
are the row-wise mean and standard deviation vectors, respectively. Each element in
is obtained by subtracting the corresponding row mean and dividing by the row standard deviation, ensuring normalized feature scales.
Based on the standardized input, the RBF kernel matrix is computed as
where
is the
i-th column of
, and
is the kernel bandwidth parameter. The kernel matrix
is then centered via
where
is a matrix with all entries equal to
.
Subsequently, KPCA is performed on the centered kernel matrix , and the top c components that explain at least a cumulative variance ratio are retained. Let denote the selected eigenvectors, with corresponding eigenvalues . The projection scores for each segment form a matrix , where .
In the proposed KPCA-based monitoring scheme, two statistics are computed for each segment
k: Hotelling’s
statistic and the squared prediction error (SPE). The
statistic characterizes the variation of a segment within the retained kernel principal component subspace, while the SPE quantifies the residual reconstruction error outside this subspace in the kernel feature space. They are defined as
where
denotes the score of the
k-th segment on the
j-th principal component,
is the associated eigenvalue, and
represents the kernel mapping of the segment. The reconstruction
using the first
c kernel principal components is obtained as
To establish the decision boundaries, control limits for both statistics are determined from the training set, as follows:
where
and
are the sample means,
and
are the corresponding standard deviations, and
and
are user-defined scaling factors. Alternatively, quantile-based or distribution-based strategies may be adopted for threshold estimation. A test segment is flagged as anomalous if it satisfies
or
.
To ensure reproducibility, the thresholds in this study are determined by the classical
principle [
25,
26]. Specifically, both
and
are set to 3 so that the control limits are defined as the sample mean plus three times the standard deviation, which covers approximately 99.7% of normal samples under Gaussian assumptions. The anomaly detection procedure based on KPCA is illustrated in
Figure 3.
4. Conclusions
The analysis clarifies the scope and quality of the proposed method. The tested cases cover step-type faults (TE-4, TE-7, CSTR), random disturbances (TE-8), and more complex unknown faults (TE-12, TE-17). The method successfully detects all these failure modes with stable performance. In terms of sensitivity, it achieves extremely low missed detection rates, for instance, only in TE-8 and even in TE-12, demonstrating high responsiveness while suppressing false alarms. Regarding data applicability, the framework generalizes well across heterogeneous process datasets, confirming its suitability for diverse industrial scenarios. These results collectively indicate that the method balances sensitivity and reliability, ensuring its effectiveness in real-time monitoring applications. Nevertheless, while the method shows strong adaptability and sensitivity across the tested scenarios, its performance under extremely noisy or highly imbalanced datasets remains to be further explored in future work.
By integrating information granulation with KPCA, the proposed method effectively captures both local structural features and global nonlinear correlations in industrial time series. Experimental results on TE and CSTR datasets show consistently low false alarm and missed detection rates across various fault types. The approach demonstrates strong robustness, accuracy, and generalization, offering a practical solution for anomaly detection in complex process systems.
Nevertheless, it should be acknowledged that, when applied to extremely large-scale and high-frequency time series, the computational burden of KPCA and PSO can become significant, and the framework may be sensitive to parameters such as granulation window length and kernel bandwidth. These challenges may restrict its efficiency in real-time big-data scenarios. To mitigate this, future work will explore scalable extensions, including incremental KPCA for dynamic updates, adaptive parameter selection to enhance stability, and parallel optimization strategies to accelerate computation. Such improvements are expected to strengthen both robustness and practicality in broader industrial applications.