Next Article in Journal
Data-Driven Solutions and Parameters Discovery of the Chiral Nonlinear Schrödinger Equation via Deep Learning
Previous Article in Journal
An Editorial to the Journal Mathematics Special Issue “Recent Trends in Convex Analysis and Mathematical Inequalities”
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph-Regularized Orthogonal Non-Negative Matrix Factorization with Itakura–Saito (IS) Divergence for Fault Detection

1
College of Electronic and Information Engineering, Southwest University, Chongqing 400715, China
2
School of Mechanical Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
3
School of Computing and Information Science, Faculty of Science and Engineering, Anglia Ruskin University, Cambridge CB1 1PT, UK
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(15), 2343; https://doi.org/10.3390/math13152343
Submission received: 14 May 2025 / Revised: 15 June 2025 / Accepted: 22 July 2025 / Published: 23 July 2025

Abstract

In modern industrial environments, quickly and accurately identifying faults is crucial for ensuring the smooth operation of production processes. Non-negative Matrix Factorization (NMF)-based fault detection technology has garnered attention due to its wide application in industrial process monitoring and machinery fault diagnosis. As an effective dimensionality reduction tool, NMF can decompose complex datasets into non-negative matrices with practical and physical significance, thereby extracting key features of the process. This paper presents a novel approach to fault detection in industrial processes, called Graph-Regularized Orthogonal Non-negative Matrix Factorization with Itakura–Saito Divergence (GONMF-IS). The proposed method addresses the challenges of fault detection in complex, non-Gaussian industrial environments. By using Itakura–Saito divergence, GONMF-IS effectively handles data with probabilistic distribution characteristics, improving the model’s ability to process non-Gaussian data. Additionally, graph regularization leverages the structural relationships among data points to refine the matrix factorization process, enhancing the robustness and adaptability of the algorithm. The incorporation of orthogonality constraints further enhances the independence and interpretability of the resulting factors. Through extensive experiments, the GONMF-IS method demonstrates superior performance in fault detection tasks, providing an effective and reliable tool for industrial applications. The results suggest that GONMF-IS offers significant improvements over traditional methods, offering a more robust and accurate solution for fault diagnosis in complex industrial settings.

1. Introduction

In the current era dominated by vast amounts of data, data-centric fault detection technology has become an indispensable part of modern industrial systems. The rapid and precise identification of potential issues is crucial for minimizing production disruptions, extending equipment lifespan, and safeguarding employee safety. Fault detection methods can generally be categorized into three main types based on their technical approaches: model-based methods, signal-based methods, and data-driven methods. Model-based methods rely on accurate mathematical models of the system; they perform well when the models are precise but face limitations when dealing with complex or unknown systems. Signal-based methods require high-quality input signals and are susceptible to external noise, which affects their stability and reliability. In contrast, data-driven methods do not depend on specific system models. Instead, they leverage large volumes of historical data to train machine learning models that can identify both normal operating patterns and fault modes. This approach offers significant adaptability and flexibility, enabling effective fault detection by comparing real-time data with training data. Traditional data-driven methods include Principal Component Analysis (PCA) and Independent Component Analysis (ICA), both of which assist in identifying fault patterns by extracting key features from the data. PCA transforms the data into a new coordinate system through linear transformation, where the first principal component corresponds to the direction of maximum variance in the data projection, the second principal component represents the direction of the second largest variance, and so on. On the other hand, ICA is a computational method used to decompose multivariate signals or datasets into statistically independent non-Gaussian signal sources. A promising technique known as Non-negative Matrix Factorization (NMF) has gained significant attention in the field of data-driven fault detection in recent years. The innovation of this method lies in its ability to decompose observed industrial process data under non-negativity constraints, not only effectively uncovering latent features in system operations but also yielding decomposition results that have clear physical interpretations [1,2]. Traditional Non-negative Matrix Factorization (NMF) methods still exhibit notable limitations when applied to complex industrial scenarios. On one hand, they struggle to effectively handle processes characterized by non-Gaussian distributions. On the other hand, these methods often fail to adequately capture the similarity and structural relationships among data points, which hinders further improvement in feature extraction performance. This is especially so when dealing with complex operating condition data that exhibit probabilistic distribution characteristics; the current NMF algorithms require further enhancement in terms of accuracy and robustness.
In 1999, Lee and Seung pioneered the theoretical framework of Non-negative Matrix Factorization (NMF) and proposed an associated optimization algorithm for matrix decomposition [3]. As research progressed, numerous improved variants of NMF were developed [4,5,6]. Among these advancements, Cai’s research team introduced a regularization strategy for NMF within a theoretical graph framework, leading to the development of the Graph-regularized NMF (GNMF) method [7]. Meanwhile, Berahmand et al. (2025) comprehensively reviewed the application of Spectral Clustering in high-dimensional data clustering in the article, particularly emphasizing the key role of Graph Structure Learning (GSL) in improving clustering performance [8]. Choi proposed the Orthogonal NMF (ONMF) model by incorporating orthogonality constraints into the original formulation [9,10]. Saberi-Movahed et al. (2024) together provided a comprehensive review of NMF in the field of dimensionality reduction, including its applications in feature extraction and feature selection [11]. Chen et al.(2022) provides a comprehensive overview of Deep Non-negative Matrix Factorization (Deep NMF) [12]. These enhanced NMF algorithms have demonstrated significant advantages in various engineering applications [13,14,15,16]. In the field of industrial fault diagnosis, Li was among the first to introduce NMF into fault detection. The innovation lies in the use of two statistical indicators—Hotelling’s T 2 and squared prediction error (SPE)—for fault feature extraction, combined with kernel-density estimation to determine control thresholds [17]. In the field of fault detection based on graph neural networks, Xiao proposed a bearing fault detection method based on graph neural networks, which uses graph autoencoders for unsupervised anomaly detection, significantly improving the accuracy of fault detection [18]. The effectiveness of these approaches was validated on the benchmark Tennessee Eastman Process (TEP) platform. Building upon this foundation, subsequent studies have proposed several improved models, including the Deep NMF (DNMF), Structured Joint Sparsity NMF (SJSNMF), and Structured Joint Sparsity Orthogonal NMF (SJSONMF) frameworks. These developments have achieved notable success in practical fault-diagnosis tasks [19,20,21,22].
The GONMF-IS framework significantly enhances fault detection performance by integrating advanced techniques such as Non-negative Matrix Factorization (NMF), graph regularization, orthogonal constraints, and the Itakura–Saito divergence (IS divergence). Within this framework, graph regularization delves into the internal similarity relationships within the data, providing the model with rich structural information. This precise capture of intrinsic data correlations enables the model to better understand the overall data distribution, laying a solid foundation for subsequent fault detection. At the same time, the introduction of orthogonal constraints further optimizes the feature extraction process. By ensuring the independence of extracted features, these constraints effectively eliminate redundant information among features and notably improve the model’s interpretability. This independence allows each feature to clearly reflect specific characteristics of the data, offering more accurate and reliable insights for fault detection. Furthermore, the GONMF-IS framework innovatively incorporates the Itakura–Saito divergence (IS divergence) as a core optimization tool [23,24,25]. The IS divergence is particularly effective in measuring and adjusting the discrepancies between different data distributions, excelling at capturing subtle variations within complex datasets. Its sensitivity to distributional differences enhances the model’s ability to distinguish between normal and faulty conditions. As a result, the GONMF-IS framework not only improves the accuracy of fault detection but also strengthens the robustness and stability of the detection outcomes.
Future research will focus on exploring the synergistic optimization mechanisms between Non-negative Matrix Factorization (NMF) and multi-dimensional divergence measures. This includes the development of novel divergence metrics, the design of adaptive algorithms for optimal measure selection, and the enhancement of model interpretability and visualization capabilities. Additionally, efforts will be directed toward improving the model’s robustness against noise and interference. Such investigations aim to advance the theoretical foundation of divergence-based NMF methods while also offering innovative solutions for industrial fault diagnosis. By leveraging cutting-edge methodologies, this line of research is expected to provide reliable technical support for ensuring the stable operation and secure control of complex industrial systems.

2. Preliminaries

Definition 1.
The Frobenius norm of a matrix, denoted as X F , is defined as the square root of the sum of the squares of all its elements. In mathematical terms, it is expressed as follows:
X F = i = 1 m j = 1 n x i j 2
where x i j is the element in the i-th row and j-th column of matrix X .
Definition 2.
For matrix X R m n , the 2,1-norm is computed as the sum of the Euclidean norms (i.e., 2-norms) of each individual row in the matrix. In other words, it computes the Euclidean norm across the columns for every row and then sums these values together. Mathematically, it is expressed as follows:
X 2 , 1 = i = 1 m j = 1 n x i j 2 1 / 2
Definition 3.
The Kullback–Leibler divergence (KL divergence), also known as relative entropy, is a measure of how one probability distribution P is different from a second, reference probability distribution Q . It is defined as an asymmetric measure of the difference between the two distributions. The formula for KL divergence is as follows:
D K L ( P Q ) = x P ( x ) log P ( x ) Q ( x )
Definition 4.
The Jensen–Shannon divergence (JS divergence) is a method for measuring the similarity between two probability distributions. It is defined based on the Kullback–Leibler divergence (KL divergence), but unlike KL divergence, JS divergence is symmetric and finite, even when the two distributions do not intersect. The definition of JS divergence is as follows:
D J S ( P Q ) = 1 2 D K L ( P M ) + 1 2 D K L ( Q M )
where P and Q are two probability distributions. M is the average distribution of P and Q , defined as follows: M = 1 2 ( P + Q ) .
Definition 5.
The Itakura–Saito (IS) Divergence is a non-symmetric measure used to quantify the difference between two spectra or probability distributions. It is defined as follows:
D I S ( P Q ) = P Q log P Q 1 .

3. Related Works

3.1. Non-Negative Matrix Factorization

Given a non-negative matrix V R m × N where V 0 , the goal of Non-negative Matrix Factorization is to decompose V into the product of two non-negative matrices A R m × r and S R r × N , it can be described as the following optimization problem:
min A , S V A S F 2 s . t . A 0 ,   S 0 .
Here, A and S are non-negative matrices, and r is a positive integer representing the rank of the decomposition.

3.2. Variants of Non-Negative Matrix Factorization

The objective function for NMF with KL divergence can be formulated as follows:
min A , S D K L ( V A S ) s . t . A 0 ,   S 0 .
where V is the original non-negative matrix, A and S are the non-negative matrices to be found, and denotes the Kullback–Leibler divergence [26,27]. The KL divergence between two matrices A and B is defined element-wise as follows:
D K L ( A B ) = i , j A i j log A i j B i j A i j B i j
In the context of NMF, this can be written as follows:
D K L ( V A S ) = i , j V i j log V i j ( AS ) i j ( V i j ( AS ) i j )
The optimization problem is to find A and S such that the above divergence is minimized, subject to the constraints that all elements of A and S are non-negative. GNMF incorporates graph regularization into the classical Non-negative Matrix Factorization (NMF) to better capture the local structure of the data, and its objective function can be expressed as follows [28,29]:
min A , S V AS F 2 + λ tr ( S L S T ) s . t . A 0 ,   S 0 .
where L is the graph Laplacian matrix, and λ is the regularization parameter. The graph Laplacian matrix is L = D W , where D is the degree matrix, and W is the adjacency matrix constructed based on the distance or similarity between data points. The Itakura–Saito (IS) divergence is a metric commonly used in signal processing and information theory, especially in applications like linear predictive coding. The definition of IS divergence is as follows: d ( x | y ) = x y log x y 1 . The nmf function incorporating the IS Divergence is as follows [23]:
min A , S V AS log V AS 1 . s . t . A > 0 ,   S > 0 .
Unlike the classical nmf model, here, the decomposed matrix A , S is required to be greater than 0 and not equal to 0.

3.3. Model Construction

In this section, the paper presents an innovative model that integrates IS divergence into the framework of Non-negative Matrix Factorization (NMF), further incorporating a graph regularization term. The specific form of the model is as follows:
min A , S V AS log V AS + λ tr ( S L S T ) . s . t . A T A = I , A > 0 , S > 0 .
KL divergence and Itakura–Saito divergence are both measures used to quantify the difference between probability distributions, but they differ in their mathematical forms, properties, and applications. KL divergence is widely used in machine learning and information theory, being sensitive to small changes in the tail of distributions and asymmetric when comparing different distributions. Itakura–Saito divergence, on the other hand, is particularly suited for handling non-negative data and is commonly used in speech recognition and audio processing. It is more robust to extreme values and is also asymmetric. Both are non-negative, but KL divergence is more sensitive to tail and outlier values, while Itakura–Saito divergence is better suited for signal-processing tasks. The model schematic is shown in Figure 1.

4. Optimization Process

Alternating Direction Method of Multipliers (ADMM) is an optimization method designed to minimize functions with multiple sets of variables, especially when these variables can be divided into distinct groups [30,31]. The algorithm works by iteratively optimizing each group of variables while holding the others constant. This technique is particularly effective for large-scale optimization tasks, as it reduces the computational complexity at each step. For Equation (8), the Alternating Direction Method of Multipliers (ADMM) algorithm is used to update variables one by one [32,33]. The initial formula for the objective function is as follows:
min A , S = V A S log V A S + λ 1 tr S L S T s . t . Z = AS ,   A T A = I ,   A > 0 ,   S > 0 .
The associated augmented Lagrangian is as follows:
L V , A , S , A + , S + , N , P , Q , Y = V Z log V Z + λ 1 tr S L S T + N , Z A S + β 2 Z A S F 2 + P , A A + + β 2 A A + F 2 + Q , S S + + β 2 S S + F 2 + Y , A T A I + β 2 A T A I F 2 .

4.1. Update Z k + 1 , Fixed A k , S k

If other variables are fixed, the variable Z subproblem can be written as follows:
min A , S V Z log V Z + N , Z A S + β 2 Z A S F 2
The derivation of Z yields the following equation:
β Z 3 β Z 2 A S + Z V = 0
for this cubic equation with one variable, the Newton method is used to solve it in this paper.

4.2. Update A k + 1 , Fixed Z k + 1 , S k

For variable A the subproblem can be written as follows:
min W , H N , Z A S + β 2 Z A S F 2 + P , A A + + β 2 A A + F 2 + Y , A T A I + β 2 A T A I F 2 .
The derivation of A yields the following equation:
A = N S T + β Z S T + β A + P β S ( S T ) + β I + Y T + Y 1

4.3. Update S k + 1 , Fixed Z k + 1 , A k + 1

For variable S , the subproblem can be written as follows:
min W , H λ 1 tr S L S T + N , Z A S + β 2 Z A S F 2 + Q , S S + + β 2 S S + F 2
The derivation of S yields the following equation:
β A T A + I S + S · 2 λ 1 L = β A T Z + A T N + β S + Q
The above is a Sylvester equation that can be solved directly with a function in matlab.

4.4. Complexity Analysis

The computational complexity of the model mainly consists of two parts: the objective function and the constraints. In the objective function, the matrix multiplication AS has a complexity of O ( m n k ) , while computing both requires O ( m n ) operations. The term λ 1 t r ( SLS T ) involves a complexity of O ( k n 2 ) . Therefore, the total complexity of the objective function is O ( m n k + m n + k n 2 ) . Regarding the constraints, computing Z = AS also takes O ( m n k ) , and evaluating A T A = I involves a complexity of O ( m k 2 + k 2 ) . Assuming the algorithm runs for t iterations, the overall computational complexity of the model is O ( t ( m n k + m n + k n 2 + m k 2 + k 2 ) ) , indicating that the computational cost scales with the matrix dimensions m , n , k , the number of iterations t, and the structure of the graph Laplacian matrix L .

5. Simulation Experiments

5.1. Fault Detection Process

This section describes the fault detection process based on (ISGONMF). By optimizing Algorithm 1, the decomposed matrices A and S are obtained, and the coefficient matrix S serves as a low-rank approximation of the original data V , reflecting the state of the industrial process represented by V . When given a test set V t e s t , according to [17], the reconstruction S ^ of S is expressed as follows:
S ^ = A T A 1 A T V test = A T V test .
Subsequently, the monitoring model based on Non-negative Matrix Factorization (NMF) is formulated as follows:
V ^ = A S ^ .
Furthermore, there are two monitoring metrics: the T 2 statistic and the Squared Prediction Error ( S P E ) statistic, which are utilized to monitor the industrial process. These are defined as follows:
T 2 = S ^ T S ^ , S P E = ( V test V ^ ) T ( V test V ^ ) .
Use Kernel Density Estimation (KDE) to estimate the probability density function and calculate control limits. Assuming x i represents a sample set where each x i is independently and identically distributed as x according to the probability density function p ( x ) , the estimation of which can be expressed as follows:
p ^ ( x ) = 1 n h i = 1 n K x x i h
h represents the bandwidth, and K ( · ) is the kernel function that meets +   K ( x ) d x = 1 and K ( x ) 0 . The Gaussian kernel function is selected for K ( · ) . Given the significant impact of bandwidth choice on p ( x ) estimation, the optimal bandwidth h is determined by minimizing the mean integrated square error (MISE), which is expressed as follows:
h opt = 1.06 σ n 1 5
As the training phase is completed offline, the online detection process is notably swift, ensuring the efficiency of the detection method. The NMF-based fault detection process is shown in Figure 2.
Algorithm 1. Algorithm for solving ISGONMF
Input: Sample matrix V , parameter λ 1 , β , the rank of the decomposition k, iterations.
Initialize: A init , S init .
1. Calculated Diagonal matrix D .
2. Calculated Laplace matrix L .
3. For i = 1 to k
  calculated gradient matrix Z k + 1 by Equation (10);
  calculated gradient matrix A k + 1 by Equation (12);
  calculated gradient matrix S k + 1 by Equation (14);
4. end for
Output  A k + 1 ,   S k + 1 .

5.2. Application on the TEP

The TEP, developed by Eastman Chemical Company, aims to assess process control and monitoring techniques through a realistic industrial simulation [34,35,36]. It features a simulation of an authentic industrial process, adjusted for proprietary concerns. Comprising five key components—reactor, condenser, compressor, separator, and distillation tower—and eight substances: A, B, C, D, E, F, G, and H. The process simulation for the Tennessee Eastman problem includes 21 predefined faults. Among these, 16 are known, and 5 are unknown; each fault contains 52 variables. In these faulty datasets, a fault is introduced at the 161st sampling point, indicating the initial 160 samples are without faults, and the subsequent 800 samples contain faults. All experiments in this article were conducted on Windows 10, Intel(R) Core(TM) i7-8750H, CPU of 2.21 GHz, RAM of 16.0 GB, using Matlab R2020b.

5.2.1. Data Preparation

The TE dataset consists of samples with 52 observed variables each. The training set is defined by files from d00.dat to d21.dat, while the test set includes samples from d00_te.dat to d21_te.dat. Both d00.dat and d00_te.dat represent normal operating conditions. The training samples in d00.dat were collected during a 25-h simulation, resulting in 500 data points, while the test samples in d00_te.dat were gathered over 48 h, totaling 960 data points. Faulty samples in the training set were generated through a 25-h simulation, starting with no faults and introducing them after one hour. Data collection began after the fault was introduced, capturing 480 data points. The test set with faults comes from a 48 h simulation, where the fault was introduced at hour eight, resulting in 960 data points, including the first 160 data points representing normal operation before the fault. The fault-free set d00 is used for training, which is a matrix of 500 × 52 , and d01_te.dat to d21_te.dat are the fault sets, all of which are matrices of 52 × 960 .
The FD rate (FDR) and false alarm rate (FAR) serve as metrics to assess the detection capability. FDR is defined as the proportion of false detections among all detections. It is calculated as follows:
number of samples T 2 J th , T 2 f 0 total of samples ( f 0 ) × 100 %
and FAR is defined as follows:
number of samples T 2 J th , T 2 f = 0 total of samples ( f = 0 ) × 100 %
The FDR and FAR rate of T 2 and S P E statistic are all calculated as above. A higher FDR value or lower FAR value indicates better FD performance.
For PCA, the principal components are retained. Its contribution rate exceeds 85%. In SNMF, the sparsity level parameters should be selected starting from a lower value and gradually increased until the best FD performance is obtained. For the proposed algorithm, the confidence limit is set to 0.99, the parameters λ 1 and β are optimized on the fault-free data in the candidate set through five-fold cross-validation { 10 5 , 10 4 , 10 3 , 10 2 , 10 1 , 1 , 10 1 , 10 2 }, the rank of the decomposition matrix is set to 20, and the maximum number of iterations is set to 60.

5.2.2. Results of the Experiment

From Table 1, we can see that the average value of the SPE index based on NMF is better than NMF_KL and NMF_IS. However, as can be seen from Table 2, the average value of the FAR based on NMF is not as good as NMF_KL and NMF_IS, while the FDR based on NMF_IS is better than NMF_KL. This is the reason why we chose NMF based on IS divergence. And from Table 1, we can see that the average FDR of ISGONMF is superior to that of NMF_IS and most other traditional models based on NMF. Meanwhile, the analysis of the data indicates that ISGONMF performs exceptionally well in terms of FAR (false alarm rate). Across all tested Industrial Data Vectors (IDVs), ISGONMF achieves an average FAR value of T 2 being only 1.59%, as shown in Table 2, significantly lower than other methods. Notably, for IDVs such as IDV1, IDV8, and IDV19, the FAR of ISGONMF is close to zero, indicating virtually no false alarms. Moreover, compared to conventional methods, ISGONMF consistently maintains the lowest or near-lowest FAR across all IDVs, demonstrating its significant advantage in reducing false alarms. This makes ISGONMF highly effective in minimizing erroneous alerts and significantly enhancing system reliability in practical applications. Figure 3 shows the visualization of the eight compared algorithms in IDV7. Figure 4 shows the visualization of the model proposed in this paper on two datasets.

5.3. XJTU

The XJTU-SY dataset from Xi’an Jiaotong University offers valuable data on rolling bearings’ vibration and lifespan under various conditions [37]. It is mainly used for bearing-lifespan prediction and health monitoring research, supporting the development and assessment of machine learning and deep learning models [38,39]. The experimental setup, depicted in Figure 5, features an AC motor, speed controller, shaft, support bearing, hydraulic loading system, and test bearings. Data is gathered at 25.6 kHz using a portable dynamic signal acquisition unit, lasting 1 min and 28 s per sample. The CSV dataset includes vibration signals for analyzing different bearing fault types and characteristics.

5.3.1. Data Preparation

The study defined three distinct operational settings, each involving five bearings. Accelerometers attached to the bearings’ exterior captured vibration data at a rate of 25.6 kHz, with each session lasting 1.28 s. The data, including horizontal and vertical vibration measurements, were stored in CSV files. Testing continued until the peak vibration amplitude exceeded ten times that of normal operation, providing insight into the bearing’s deterioration from normal to critical failure. The dataset includes detailed information on the number of CSV files, operational lifespan, and failure types, such as inner race wear, cage rupture, outer race wear, and fracture. This dataset serves as valuable data for evaluating predictive maintenance algorithms for rolling element bearings, with the analysis focusing on the horizontal vibration data in the first column. Table 3 of this study presents 15 different faults evaluated under three operational conditions. For each fault, the training dataset contains 1000 normal instances, while the test dataset consists of 200 normal instances and 800 fault instances.

5.3.2. Results of the Experiment

The ISGONMF algorithm exhibits significant strengths in the XJTU datasets. Regarding FDR, in terms of mean values, the ISGONMF algorithm has mean values of 77.55 and 84.37 for both T2 and SPE metrics, respectively, which indicates that the ISGONMF algorithm outperforms most of the other algorithms in terms of overall fault detection performance, in Table 3, Bearings 1-2, 1-4, 1-5, 2-1, 2-2, 2-5, and 3-3 showed better results. In terms of FAR, the ISGONMF algorithm achieved the lower average SPE value at just 0.96%, as shown in Table 4, indicating its effectiveness. On Bearings 2-2, 3-1, and 3-2, both T2 and SPE statistics are optimal, confirming that the algorithm maintains a considerably low false alarm rate across the bearing dataset. Figure 6 shows the visualization of the eight compared algorithms in Bearing2–5.

5.4. Ablation Study

Orthogonality constraints are essential to improve the performance of the model in fault detection tasks. As shown in the Table 5, we analyze the model performance on both TEP and XJTU-SY datasets in detail. In terms of fault detection rate (FDR), the orthogonality constraint enables the ISGONMF model to exhibit slightly higher performance than the ISGNMF model without the orthogonality constraint on both datasets, especially on the XJTU-SY dataset, where the ISGONMF model achieves an FDR T 2 of 73.03%, which is significantly higher than that of the ISGNMF model of 72.11%. This indicates that the orthogonality constraint helps the model to identify the faulty situations more effectively. In terms of false alarm rate (FAR), the orthogonality constraint also plays a significant role in improving it. On the TEP dataset, the FAR T 2 of the ISGONMF model is 1.59%, which is lower than the 1.71% of the ISGNMF model.
Summarizing the above analysis, we can conclude that the orthogonality constraint plays a crucial role in the ISGONMF model. It not only improves the fault detection rate of the model and reduces the false alarm rate, but also enhances the model’s ability to adapt to complex datasets. The orthogonality constraint helps the model to better capture the data features, thus improving the accuracy and reliability of fault detection.

5.5. Experimental Convergence Results

Figure 7 demonstrates the convergence results of the algorithm on the datasets TEP and XJTU, the residuals Z A S F and A A + F decrease rapidly within 20 iterations, indicating numerical stability. The algorithm terminates when Z k A k S k F < ϵ ϵ = 10 5 or reaches the maximum iterations (60). The convergence results of the TEP simulation experiments are shown in terms of IDV7, and the convergence results of the XJTU bearing dataset are shown in terms of bearings 2–5.

6. Conclusions

This study proposes a novel fault detection approach named Graph-Regularized Orthogonal Non-negative Matrix Factorization with Itakura–Saito (IS) Divergence (ISGONMF), specifically designed to address the challenges in complex industrial processes. By integrating graph regularization, orthogonal constraints, and the IS divergence, the method effectively captures the probabilistic characteristics of data and leverages the intrinsic relationships among data points to enhance the quality of matrix factorization. Experimental results demonstrate that ISGONMF achieves superior performance in reducing false alarm rates (FAR) and improving fault detection rates (FDR), particularly when applied to the Tennessee Eastman Process (TEP) dataset and the XJTU-SY dataset commonly used for bearing fault diagnosis. To further enhance the model’s adaptability and robustness, future work may explore the incorporation of additional divergence measures and the development of mechanisms for automatically selecting the most suitable divergence based on data characteristics. Moreover, efforts to improve model interpretability, visualization capabilities, and resilience against external disturbances will be key directions for further research.

Author Contributions

Conceptualization, Y.L. and J.Z.; Methodology, Y.L. and J.Z.; Software, Y.L. and J.W.; Validation, Y.L.; Formal analysis, J.W.; Resources, M.-F.L.; Data curation, J.Z.; Writing—original draft, Y.L.; Writing—review & editing, Y.L. and J.W.; Funding acquisition, M.-F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, Y.X.; Zhang, Y.J. Nonnegative matrix factorization: A comprehensive review. IEEE Trans. Knowl. Data Eng. 2012, 25, 1336–1353. [Google Scholar] [CrossRef]
  2. Kim, H.; Park, H. Sparse non-negative matrix factorizations via alternating non-negativity-constrained least squares for microarray data analysis. Bioinformatics 2007, 23, 1495–1502. [Google Scholar] [CrossRef] [PubMed]
  3. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef]
  4. Belachew, M.T. Efficient algorithm for sparse symmetric nonnegative matrix factorization. Pattern Recognit. Lett. 2019, 125, 735–741. [Google Scholar] [CrossRef]
  5. Huang, S.; Wang, H.; Li, T.; Li, T.; Xu, Z. Robust graph regularized nonnegative matrix factorization for clustering. Data Min. Knowl. Discov. 2018, 32, 483–503. [Google Scholar] [CrossRef]
  6. Yoo, J.; Choi, S. Orthogonal nonnegative matrix factorization: Multiplicative updates on Stiefel manifolds. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Daejeon, Republic of Korea, 2–5 November 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 140–147. [Google Scholar]
  7. Cai, D.; He, X.; Han, J.; Huang, T.S. Graph regularized nonnegative matrix factorization for data representation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 1548–1560. [Google Scholar] [CrossRef]
  8. Berahmand, K.; Saberi-Movahed, F.; Sheikhpour, R.; Li, Y.; Jalili, M. A Comprehensive Survey on Spectral Clustering with Graph Structure Learning. arXiv 2025, arXiv:2501.13597. [Google Scholar]
  9. Choi, S. Algorithms for orthogonal nonnegative matrix factorization. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1828–1832. [Google Scholar]
  10. Ding, C.; Li, T.; Peng, W.; Park, H. Orthogonal nonnegative matrix t-factorizations for clustering. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, 20–23 August 2006; pp. 126–135. [Google Scholar]
  11. Saberi-Movahed, F.; Berahman, K.; Sheikhpour, R.; Li, Y.; Pan, S. Nonnegative matrix factorization in dimensionality reduction: A survey. arXiv 2024, arXiv:2405.03615. [Google Scholar] [CrossRef]
  12. Chen, W.S.; Zeng, Q.; Pan, B. A survey of deep nonnegative matrix factorization. Neurocomputing 2022, 491, 305–320. [Google Scholar] [CrossRef]
  13. Li, C.; Che, H.; Leung, M.F.; Liu, C.; Yan, Z. Robust multi-view non-negative matrix factorization with adaptive graph and diversity constraints. Inf. Sci. 2023, 634, 587–607. [Google Scholar] [CrossRef]
  14. Zhang, W.; Yu, S.; Wang, L.; Guo, W.; Leung, M.F. Constrained symmetric non-negative matrix factorization with deep autoencoders for community detection. Mathematics 2024, 12, 1554. [Google Scholar] [CrossRef]
  15. Dong, Y.; Che, H.; Leung, M.F.; Liu, C.; Yan, Z. Centric graph regularized log-norm sparse non-negative matrix factorization for multi-view clustering. Signal Process. 2024, 217, 109341. [Google Scholar] [CrossRef]
  16. Yang, X.; Che, H.; Leung, M.F.; Liu, C. Adaptive graph nonnegative matrix factorization with the self-paced regularization. Appl. Intell. 2023, 53, 15818–15835. [Google Scholar] [CrossRef]
  17. Li, X.b.; Yang, Y.p.; Zhang, W.d. Fault detection method for non-Gaussian processes based on non-negative matrix factorization. Asia-Pac. J. Chem. Eng. 2013, 8, 362–370. [Google Scholar] [CrossRef]
  18. Xiao, L.; Yang, X.; Yang, X. A graph neural network-based bearing fault detection method. Sci. Rep. 2023, 13, 5286. [Google Scholar] [CrossRef]
  19. Ren, Z.; Zhang, W.; Zhang, Z. A deep nonnegative matrix factorization approach via autoencoder for nonlinear fault detection. IEEE Trans. Ind. Inform. 2019, 16, 5042–5052. [Google Scholar] [CrossRef]
  20. Xiu, X.; Fan, J.; Yang, Y.; Liu, W. Fault detection using structured joint sparse nonnegative matrix factorization. IEEE Trans. Instrum. Meas. 2021, 70, 3513011. [Google Scholar] [CrossRef]
  21. Zhang, X.; Xiu, X.; Zhang, C. Structured joint sparse orthogonal nonnegative matrix factorization for fault detection. IEEE Trans. Instrum. Meas. 2023, 72, 2506015. [Google Scholar] [CrossRef]
  22. Ahmadian, S.; Berahmand, K.; Rostami, M.; Forouzandeh, S.; Moradi, P.; Jalili, M. Recommender Systems based on Non-negative Matrix Factorization: A Survey. IEEE Trans. Artif. Intell. 2025, 1–21. [Google Scholar] [CrossRef]
  23. Févotte, C.; Bertin, N.; Durrieu, J.L. Nonnegative matrix factorization with the Itakura-Saito divergence: With application to music analysis. Neural Comput. 2009, 21, 793–830. [Google Scholar] [CrossRef]
  24. Cichocki, A.; Lee, H.; Kim, Y.D.; Choi, S. Non-negative matrix factorization with α-divergence. Pattern Recognit. Lett. 2008, 29, 1433–1440. [Google Scholar] [CrossRef]
  25. Févotte, C.; Idier, J. Algorithms for nonnegative matrix factorization with the β-divergence. Neural Comput. 2011, 23, 2421–2456. [Google Scholar] [CrossRef]
  26. Yang, Z.; Zhang, H.; Yuan, Z.; Oja, E. Kullback-Leibler divergence for nonnegative matrix factorization. In Proceedings of the International Conference on Artificial Neural Networks, Espoo, Finland, 14–17 June 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 250–257. [Google Scholar]
  27. Hien, L.T.K.; Gillis, N. Algorithms for nonnegative matrix factorization with the Kullback–Leibler divergence. J. Sci. Comput. 2021, 87, 93. [Google Scholar] [CrossRef]
  28. Zhang, L.; Liu, Z.; Pu, J.; Song, B. Adaptive graph regularized nonnegative matrix factorization for data representation. Appl. Intell. 2020, 50, 438–447. [Google Scholar] [CrossRef]
  29. Chen, K.; Che, H.; Li, X.; Leung, M.F. Graph non-negative matrix factorization with alternative smoothed L 0 regularizations. Neural Comput. Appl. 2023, 35, 9995–10009. [Google Scholar] [CrossRef]
  30. Zhang, G.; Wang, Y.; Lessard, L.; Grosse, R.B. Near-optimal local convergence of alternating gradient descent-ascent for minimax optimization. In Proceedings of the International Conference on Artificial Intelligence and Statistics—Proceedings of Machine Learning Research, Virtual, 28–30 March 2022; pp. 7659–7679. [Google Scholar]
  31. Harper, J.; Wells, D. Recent advances and future developments in PGD. Prenat. Diagn. 1999, 19, 1193–1199. [Google Scholar] [CrossRef]
  32. Hajinezhad, D.; Chang, T.H.; Wang, X.; Shi, Q.; Hong, M. Nonnegative matrix factorization using ADMM: Algorithm and convergence analysis. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 4742–4746. [Google Scholar]
  33. Lin, C.J. Projected gradient methods for nonnegative matrix factorization. Neural Comput. 2007, 19, 2756–2779. [Google Scholar] [CrossRef]
  34. Yin, S.; Ding, S.X.; Haghani, A.; Hao, H.; Zhang, P. A comparison study of basic data-driven fault diagnosis and process monitoring methods on the benchmark Tennessee Eastman process. J. Process Control 2012, 22, 1567–1581. [Google Scholar] [CrossRef]
  35. Lee, G.; Han, C.; Yoon, E.S. Multiple-fault diagnosis of the Tennessee Eastman process based on system decomposition and dynamic PLS. Ind. Eng. Chem. Res. 2004, 43, 8037–8048. [Google Scholar] [CrossRef]
  36. Zhang, C.; Guo, Q.; Li, Y. Fault detection in the Tennessee Eastman benchmark process using principal component difference based on k-nearest neighbors. IEEE Access 2020, 8, 49999–50009. [Google Scholar] [CrossRef]
  37. Wang, B.; Lei, Y.; Li, N. XJTU-SY bearing datasets. Github Github Repos. 2018. [Google Scholar]
  38. Wang, B.; Lei, Y.; Li, N.; Li, N. A hybrid prognostics approach for estimating remaining useful life of rolling element bearings. IEEE Trans. Reliab. 2018, 69, 401–412. [Google Scholar] [CrossRef]
  39. Xue, Y.; Yang, R.; Chen, X.; Tian, Z.; Wang, Z. A novel local binary temporal convolutional neural network for bearing fault diagnosis. IEEE Trans. Instrum. Meas. 2023, 72, 3525013. [Google Scholar] [CrossRef]
Figure 1. The framework diagram of Graph-Regularized Orthogonal Non-negative Matrix Factorization with Itakura–Saito (IS) Divergence. The top left of the image shows the classical NMF model, the bottom left shows the added regularization and constraint terms, and the right side of the image shows the fault detection process, which detects whether a fault has occurred based on the fault collection samples.
Figure 1. The framework diagram of Graph-Regularized Orthogonal Non-negative Matrix Factorization with Itakura–Saito (IS) Divergence. The top left of the image shows the classical NMF model, the bottom left shows the added regularization and constraint terms, and the right side of the image shows the fault detection process, which detects whether a fault has occurred based on the fault collection samples.
Mathematics 13 02343 g001
Figure 2. NMF-based fault detection process.
Figure 2. NMF-based fault detection process.
Mathematics 13 02343 g002
Figure 3. Detection results of fault IDV7 for TEP datasets.
Figure 3. Detection results of fault IDV7 for TEP datasets.
Mathematics 13 02343 g003
Figure 4. Detectionresults for ISGONMF on two datasets.
Figure 4. Detectionresults for ISGONMF on two datasets.
Mathematics 13 02343 g004
Figure 5. XJTU-SY bearing dataset experiment platform.
Figure 5. XJTU-SY bearing dataset experiment platform.
Mathematics 13 02343 g005
Figure 6. Detection results of Bearing 2–5 for XJTU datasets.
Figure 6. Detection results of Bearing 2–5 for XJTU datasets.
Mathematics 13 02343 g006
Figure 7. Convergence results.
Figure 7. Convergence results.
Mathematics 13 02343 g007
Table 1. FDR values for TEP.
Table 1. FDR values for TEP.
PCANMFNMF_KLNMF_ISGNMFONMFSNMFNMF_JSISGONMF
 T2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPE
IDV159.3099.7599.5099.6399.1399.0098.7598.8898.3899.1397.6399.0099.0099.3899.3899.6399.3899.75
IDV230.3898.6397.7597.6197.7598.0097.1398.5097.5098.0096.7597.3897.5098.7597.7598.1398.0096.63
IDV310.633.752.384.132.001.633.004.008.758.3810.7514.380.631.504.886.502.003.25
IDV46.0093.8811.3852.132.252.631.386.633.135.501.255.380.755.2512.8832.3815.7547.00
IDV530.7528.6322.3827.0019.3822.3820.6324.0016.0025.2518.2527.1318.6320.3826.3819.3819.8819.50
IDV699.3899.3899.8899.88100.00100.00100.00100.0099.25100.0098.50100.00100.0099.8899.5098.5099.6399.63
IDV750.12100.00100.0099.8899.5099.1392.5099.75100.00100.0062.2583.8871.6399.8897.2588.38100.00100.00
IDV889.0097.5093.6397.3388.5091.1390.7592.3870.5092.2568.7590.1372.1392.8887.3864.8894.7587.00
IDV98.134.383.753.881.381.003.002.637.137.384.3810.380.251.502.382.131.252.88
IDV1052.7542.7531.1338.6326.6330.8828.8829.3831.3843.8837.1347.8812.6518.3315.5013.1313.5023.25
IDV1113.8868.6323.0044.255.3813.003.7520.5012.7520.6313.5019.130.5014.7533.7520.1314.3838.00
IDV1292.1398.6388.6395.3881.0082.2581.7584.3868.3888.3869.1386.2577.8389.3875.8856.3879.0090.38
IDV1391.3891.6392.3891.3890.5091.3888.8891.0087.6392.0088.6392.3882.8892.1389.7588.8392.7593.13
IDV1419.5088.3856.6381.886.8848.383.6357.135.6329.753.0011.755.1362.6362.0052.0034.3897.63
IDV1514.885.503.505.755.635.003.137.5013.6312.2512.1316.250.383.504.135.001.753.00
IDV1638.7527.1351.7561.6329.8830.3822.6320.5033.2547.3843.7555.6325.1353.3857.1320.6360.5063.75
IDV1754.8823.3839.2557.2535.2526.007.8827.1320.5030.7512.3822.001.0035.3853.1355.7535.2552.13
IDV1888.3890.0088.7398.1388.3887.8888.5089.2583.1387.6381.3886.6384.8888.8887.6395.6381.1392.25
IDV192.1313.752.385.751.881.132.131.504.884.383.885.250.252.384.7514.635.004.63
IDV2021.3828.8821.8842.1312.8820.5011.0022.1323.2529.2518.3835.6313.2520.3863.7564.3840.8847.75
IDV2128.3844.8834.1347.3532.1332.0013.8817.5017.1241.3844.3846.5011.6335.3323.8822.3820.6339.88
avreage42.9659.5050.6659.5644.1046.8441.1047.3642.9650.6442.1950.1336.9549.3252.3348.5148.0857.21
Table 2. FAR values for TEP.
Table 2. FAR values for TEP.
PCANMFNMF_KLNMF_ISGNMFONMFSNMFNMF_JSISGONMF
 T2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPE
IDV10.002.500.000.000.000.001.250.000.000.000.000.000.630.001.880.000.002.50
IDV20.001.880.630.630.000.000.000.000.000.000.000.000.001.251.881.250.631.25
IDV36.381.252.502.500.000.000.630.630.630.000.001.250.000.633.755.630.631.88
IDV40.002.500.631.250.000.001.251.253.131.251.251.250.000.000.001.251.881.88
IDV50.002.500.631.250.000.001.251.253.131.251.251.250.000.004.211.881.250.63
IDV60.000.630.001.250.000.001.252.500.631.881.251.880.000.631.251.881.881.88
IDV70.630.002.504.380.000.000.631.251.250.636.8811.880.000.630.001.254.381.25
IDV81.880.001.881.250.000.000.003.130.007.505.6316.250.001.250.632.501.880.00
IDV924.381.888.787.630.000.002.505.6315.6321.2517.3818.731.250.635.631.880.633.13
IDV100.001.250.631.250.000.000.000.630.000.630.630.000.000.632.500.000.634.38
IDV110.003.130.633.130.000.000.001.886.880.002.502.500.001.251.255.001.251.88
IDV1213.131.251.881.250.250.251.253.138.755.631.2510.631.882.382.501.881.882.50
IDV130.000.000.000.630.000.000.001.250.000.000.000.000.000.000.002.501.250.63
IDV140.882.750.001.880.000.000.000.636.253.135.633.750.630.631.631.381.250.63
IDV150.001.250.630.630.000.001.250.000.630.000.001.250.631.252.501.881.253.13
IDV1632.503.756.385.250.000.006.2511.880.6318.1311.8817.251.255.631.251.882.503.75
IDV170.001.881.254.380.000.001.882.5010.137.2513.1316.251.251.251.252.504.381.88
IDV180.004.385.6375.640.000.004.383.130.000.000.001.250.000.630.0056.250.6310.00
IDV190.000.000.631.880.000.001.251.883.752.500.633.130.001.250.630.631.250.00
IDV200.000.630.630.630.000.000.000.000.630.630.000.000.000.001.880.331.251.25
IDV212.503.130.630.630.000.001.881.880.000.630.000.630.251.381.880.631.251.25
avreage3.911.731.735.580.010.011.282.112.953.603.295.190.361.011.744.401.592.17
Table 3. FDR values for XJTU.
Table 3. FDR values for XJTU.
PCANMFNMF_KLNMF_ISGNMFONMFSNMFNMF_JSISGONMF
 T2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPE
Bearing 1-195.6095.4895.1195.4795.2495.6095.1195.4895.2495.6095.3695.3694.8895.4895.2495.4895.3695.24
Bearing 1-295.3695.2495.2495.3695.2495.3695.2495.3695.3695.3695.2495.3695.2495.2495.2495.2495.3695.36
Bearing 1-395.2495.2495.2495.4895.2495.2495.3695.4895.2495.3695.2495.3695.3695.2495.2495.3695.2495.36
Bearing 1-421.4578.245.2481.795.8379.8814.8864.8812.8644.6613.2476.903.0572.2921.8677.814.1782.14
Bearing 1-55.0013.464.4123.134.5119.895.8314.294.3816.483.2416.763.338.145.0027.197.864.76
Bearing 2-195.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.24
Bearing 2-295.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.2495.36
Bearing 2-395.2495.2495.2495.2495.2495.2495.2495.3695.2495.2495.3695.3695.2495.2495.4895.2495.3695.36
Bearing 2-43.3615.4814.1233.367.7132.6210.7117.625.9828.953.5225.522.1935.8112.9616.092.866.07
Bearing 2-595.3695.2495.2495.3695.2495.4895.3695.3695.2495.3695.4895.2495.2495.2495.3695.2495.2495.95
Bearing 3-195.2495.2495.2495.4895.2495.2495.2495.2495.2495.2495.3695.2495.3695.3695.2495.3695.2495.24
Bearing 3-295.2495.3695.2495.2495.2495.3695.2495.2495.2495.2495.3695.3695.2495.2495.2495.3695.3695.24
Bearing 3-395.4895.2495.4795.3695.4895.3695.4895.2495.4895.4895.2495.6095.3695.3695.2495.4895.3695.48
Bearing 3-495.2495.3695.3695.4895.3695.3695.3695.2495.7195.8195.2495.4895.3695.3695.3695.4895.3695.36
Bearing 3-529.0595.2431.6795.4828.3495.1271.7995.2428.5795.4877.3895.3811.6795.8175.3395.8132.2295.56
avreage73.8283.3773.5585.5172.9585.0776.7682.7173.3582.3176.3884.2271.2084.0277.5584.3773.0382.51
Table 4. FAR values for XJTU.
Table 4. FAR values for XJTU.
PCANMFNMF_KLNMF_ISGNMFONMFSNMFNMF_JSISGONMF
 T2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPET2SPE
Bearing 1-15.001.751.880.631.881.252.501.251.881.251.880.632.501.881.880.631.881.25
Bearing 1-21.881.250.002.500.003.130.631.880.002.501.253.750.001.880.633.131.881.25
Bearing 1-30.000.631.250.001.250.631.250.630.001.880.631.250.630.630.001.882.501.25
Bearing 1-40.620.621.871.871.882.501.252.501.251.251.251.251.883.131.880.631.253.13
Bearing 1-51.250.001.251.250.000.631.880.000.621.230.001.250.630.000.630.631.251.88
Bearing 2-10.000.630.630.630.630.000.630.630.631.250.630.630.631.251.251.880.001.88
Bearing 2-20.631.250.001.250.001.880.000.000.002.500.003.120.001.250.000.630.000.00
Bearing 2-30.003.131.880.001.880.000.630.630.000.000.630.001.860.001.251.251.250.63
Bearing 2-40.620.000.000.630.620.620.630.630.620.620.000.000.000.000.620.621.250.63
Bearing 2-50.630.001.250.001.250.000.630.631.251.250.630.631.250.000.631.880.630.63
Bearing 3-10.630.001.250.631.880.002.500.632.500.000.001.251.860.001.250.001.250.00
Bearing 3-21.251.253.132.502.501.252.500.631.881.880.000.631.860.632.502.501.880.63
Bearing 3-30.001.880.001.250.000.000.630.630.000.001.250.630.000.000.000.000.000.63
Bearing 3-41.250.000.620.620.630.630.630.630.630.631.250.630.631.250.630.000.000.63
Bearing 3-50.630.633.751.251.251.250.000.003.752.501.252.503.753.752.500.003.160.00
avreage0.960.871.251.001.040.921.080.751.001.250.711.211.161.041.041.041.210.96
Table 5. Ablation study results on TEP and XJTU-SY datasets.
Table 5. Ablation study results on TEP and XJTU-SY datasets.
DatasetModelFDRFAR
T 2 (%)SPE (%) T 2 (%)SPE (%)
TEPISGONMF48.0857.211.592.17
ISGNMF46.6352.131.712.35
XJTU-SYISGONMF73.0382.511.210.96
ISGNMF72.1181.950.891.12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Wu, J.; Zhang, J.; Leung, M.-F. Graph-Regularized Orthogonal Non-Negative Matrix Factorization with Itakura–Saito (IS) Divergence for Fault Detection. Mathematics 2025, 13, 2343. https://doi.org/10.3390/math13152343

AMA Style

Liu Y, Wu J, Zhang J, Leung M-F. Graph-Regularized Orthogonal Non-Negative Matrix Factorization with Itakura–Saito (IS) Divergence for Fault Detection. Mathematics. 2025; 13(15):2343. https://doi.org/10.3390/math13152343

Chicago/Turabian Style

Liu, Yabing, Juncheng Wu, Jin Zhang, and Man-Fai Leung. 2025. "Graph-Regularized Orthogonal Non-Negative Matrix Factorization with Itakura–Saito (IS) Divergence for Fault Detection" Mathematics 13, no. 15: 2343. https://doi.org/10.3390/math13152343

APA Style

Liu, Y., Wu, J., Zhang, J., & Leung, M.-F. (2025). Graph-Regularized Orthogonal Non-Negative Matrix Factorization with Itakura–Saito (IS) Divergence for Fault Detection. Mathematics, 13(15), 2343. https://doi.org/10.3390/math13152343

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop