Abstract
The deterministic discrete-time method is a dominant approach for analyzing neural network algorithms. To address the issue where conventional convergence conditions impose stringent restrictions on the range of learning factors, this paper proposes a principal component estimation algorithm with an adaptive learning factor, which guarantees global convergence. The convergence of the algorithm is analyzed using the deterministic discrete-time method, and conditions ensuring convergence are established. Unlike convergence conditions for other algorithms, the proposed algorithm’s convergence conditions eliminate restrictions on the learning factor, thereby extending its feasible range. Simulation results demonstrate that the proposed algorithm effectively resolves ill-conditioned matrix problems. When compared with existing algorithms, the proposed algorithm exhibits significantly faster convergence speed than several current methods.
1. Introduction
The principal component in signal processing and deep learning is identified by the direction of a signal’s greatest variability []. Principal component analysis (PCA) is widely adopted for data preprocessing and feature extraction, with the resulting features serving as inputs for downstream tasks. For instance, feature vectors from the signal subspace can be utilized with a flow matrix in weighted subspace fitting to achieve high-accuracy estimation of signal source directions []. The utility of PCA further encompasses domains such as image compression [], fault diagnosis [], and various other disciplines [,,,,,,,]. Consequently, the problem of estimating a signal’s principal component could be transformed into a matrix eigenvalue decomposition problem, which could then be solved using eigenvalue decomposition algorithms. However, in real-time signal processing scenarios [], sensors could only acquire the signal value at the current time instant, not the entire autocorrelation matrix. In such cases, batch processing algorithms become unsuitable for the online estimation of principal components. To enable real-time estimation of signal principal components, researchers have proposed Hebbian neural network-based algorithms [,,,]. This class of algorithms could directly estimate the principal component from the input signal stream, thus avoiding the computation of the autocorrelation matrix.
Based on distinct update rules, numerous neural network-based algorithms have been successively proposed [,,,]. Convergence analysis of neural network-based algorithms constitutes a critically important research topic. In the field of convergence analysis, three primary methodological frameworks have gained prominence []: Stochastic Discrete-Time (SDT), Deterministic Continuous-Time (DCT), and Deterministic Discrete-Time (DDT). Typically, algorithms are formulated as SDT systems, for which direct convergence analysis proves highly challenging []. Among indirect analysis methods, the DCT approach imposes excessively restrictive conditions that are often impractical in real-world applications []. In contrast, the DDT method imposes fewer constraints and has consequently become the dominant approach for convergence analysis in neural network-based algorithms [].
In neural networks, the learning rate stands as one of the most critical hyperparameters, directly governing the step size for parameter updates during training. The assumption of a constant learning rate constitutes a standard premise in traditional DDT-based convergence analysis []. To ensure algorithm convergence, restrictive constraints are usually imposed, confining the learning rate to a very narrow range. However, in practical applications, a larger learning rate can be appropriately selected during the initial iterations when far from the stationary point to accelerate convergence. Conversely, during later stages of operation, as the solution approaches the stationary point, a smaller learning rate should be chosen to produce more precise update steps []. Since restricting the learning rate inherently limits algorithm performance, exploring adaptive algorithms constitutes a meaningful research direction.
In this paper, an adaptive algorithm is proposed, which can directly extract the principal components of the input signal without the need to calculate the autocorrelation matrix of the algorithm. Compared with batch processing, the proposed algorithm has higher real-time performance and lower computational complexity. Consequently, the proposed algorithm is highly effective for online signal processing and time-varying systems, particularly in applications like real-time signal tracking and adaptive filtering. The paper is organized as follows: Section 2 details the novel adaptive learning rate algorithm. Section 3 examines its convergence properties via a DDT-based analysis. Section 4 validates the algorithm’s performance through numerical simulations and practical case studies. Finally, Section 5 summarizes the main conclusions.
2. Adaptive Algorithm Design
The theoretical framework of this study is structured around a standard Hebbian neural network, characterized by the input-output relationship below,
where is the -dimensional input signal, the evolving weight vector, and the corresponding output. The development of this Hebbian algorithm is directed towards establishing an update rule that ensures convergence of to the first generalized eigenvector.
The principal component estimation algorithm proposed by Chen []. is derived from this neural network model.
By employing the DDT method, a convergence proof for the modified algorithm was established by Kong et al. [], contingent on the learning rate satisfying . Evidently, the permissible range for the learning rate is severely restricted. To develop an algorithm with an adaptive learning rate, the following method is proposed:
By comparing the two equations, it appears that Equation (3) replaces the learning factor in Equation (2) with variable . However, there is a distinct difference between them. In Equation (2), the learning factor must satisfy , whereas Equation (3) imposes no such requirement. When the norm of the weight vector reaches , we have , at which point it already violates the requirement . The learning rate is a parameter of paramount importance for algorithmic convergence. An inappropriately low value leads to prohibitively slow progress, whereas an excessively high value adversely induces oscillatory or divergent behavior. The subsequent analysis shows that the norm of the weight vector monotonically rises throughout the iterative process. This implies that, during iteration process, the learning factor in this algorithm is a monotonically decreasing sequence. Unlike a fixed learning rate, our algorithm employs a dynamically self-adjusting learning factor that optimizes the trade-off between convergence speed and stability throughout the iterative process, thereby ensuring robust performance and enhanced practicality.
This section examines the overall computational complexity of the online algorithm, with a detailed cost breakdown provided in Table 1 for clarity. The resulting requirement is 5n + 3 multiplications per update, placing the overall complexity on the same order as that of the algorithms in [,]. Relying solely on elementary vector addition and multiplication, the proposed algorithm is naturally suited for an efficient systolic array implementation.
Table 1.
Computational complexity.
For a stationary stochastic input signal , the dynamics of (3) can be analyzed using the DDT approach. The DDT system is derived by applying the conditional expectation to (3) and defining the resultant expectation as the subsequent iterate.
where is the autocorrelation matrix of the input signal.
The norm of the weight vector at time could be calculated according to Equation (4).
where . Comparing the magnitudes of the weight vectors at two consecutive iteration steps yields:
From Equation (6), it could be concluded that the norm of the weight vector exhibits monotonic increase during the algorithm’s iteration. Consequently, the adaptive learning factor decreases monotonically, which is consistent with the algorithm’s requirement for the variation in the learning factor.
3. Convergence of the Algorithm
3.1. Preliminaries
Assume that scalar and vectors are the eigenvalues and mutually orthogonal eigenvectors, respectively, of the signal autocorrelation matrix . For convenience of use, the eigenvalues are arranged here in descending order, namely:
The order of the eigenvectors is consistent with that of their respective eigenvalues. In accordance with signal processing theory [], the eigenvector associated with the largest eigenvalue of a matrix represents the first principal component of the signal. The properties of linear independence and mutual orthogonality allow the eigenvectors to serve as an orthogonal basis for the space . Any vector in the space could be represented as a linear combination of this basis.
where represents the coefficient for each basis. Utilizing the properties of matrix eigendecomposition, we have:
Equation (8) is substituted into Equation (4) to obtain
By the property of the Rayleigh quotient, it follows that:
where is a symmetric positive definite matrix and . Equation (11) is substituted into Equation (10) to obtain
The above relation indicates that forms a monotonically increasing sequence. That is, the norm of the projection of the weight vector onto the principal component exhibits monotonic increase throughout the algorithm’s iteration process.
The constant is defined. Given that is monotonically increasing, the convergence analysis of the algorithm proceeds by considering the following two cases:
Case 1. For all , holds.
Case 2. There exists a constant such that , where is a positive integer.
3.2. Convergence Analysis of the Algorithm in Case 1
The convergence of in this scenario is governed by the following theorem.
Theorem 1.
Under Case 1 with the initialization
, it follows that .
Proof.
could be obtained from . has an upper bound in Case 1, so also has an upper bound and the monotonically increasing sequence will converge to a constant value.
From , it follows that . Under Case 1, is bounded above. Consequently, is also bounded above. Therefore, the monotonically increasing sequence must converge to a constant value. From Equations (10)–(12), it could be deduced that during the iteration process, the sign of remains unchanged. That is, if , then holds for all ; conversely, if holds for all , then . Hence, when converges, must also converge. Supposing
where is a constant. According to
we get
From and Equation (8), we can deduce:
That is, when , there is
It could be obtained by combining Equations (13) and (17).
Therefore, . □
3.3. Convergence Analysis of the Algorithm in Case 2
In this case, the convergence of will still be established by analyzing the convergence of . The following three lemmas will be proved for later use.
Lemma 1.
Under the premises of Case 2 with , the two inequalities below hold for all .
Proof.
Under Case 2, when , the condition is satisfied, we obtain
When , according to Equations (10) and (8) have
From the above formula, for all have: , and then we can get
□
Lemma 2.
In Case 2, if the initialization condition satisfies
, then .
Proof.
According to Equation (10), for all , the following holds.
This equality indicates that is a monotonically decreasing sequence. Since is a monotonically increasing sequence, it follows under Case 2 that:
Let
According to Equations (10) and (8), we can get
According to Equations (23) and (26), we can get
Since the sequence is divergent, we have , and, thus,
□
Lemma 3.
Under Case 2, if the initial condition is met, then we have , where .
Proof.
Since , it necessarily follows that . Given that is a monotonically increasing sequence, we necessarily have . We now turn to prove Lemma 3 using mathematical induction.
According to Lemma 2, we have .
Assume that when , holds. Then there necessarily exists a positive integer such that for all , the following two inequalities hold:
According to Equations (10) and (29), for all , there are
then
The above equation shows that the sequence is monotonically decreasing. According to Lemma 2 and Equations (29)–(31), for all , we have
Let
According to Equations (12) and (32), for all there are
where . Since the series is divergent, there is
Then
□
Theorem 2.
In Case 2, if the initialization condition satisfies
, then we can get .
Proof.
could be obtained from . Since is a monotonically increasing sequence, we can get . According to Equation (8) to get
The following equation is obtained according to Lemma 2 and Lemma 3.
Substituting into Equation (37) and according to Equation (38), we obtained
Since does not change the symbol during the iteration, there is , and then
□
3.4. Convergence Analysis of the Algorithm
Synthesizing the results of Theorems 1 and 2, we derive the following convergence guarantee:
Theorem 3.
For the proposed algorithm, convergence is ensured provided that the initial weight vector satisfies
, then there is .
Theorem 3 outlines the convergence conditions for our algorithm. This algorithm is closely related to the one in [], which, in contrast, has a convergence condition of and , while the convergence conditions of the proposed algorithm do not restrict the learning factor. Theorem 3 stipulates the initial condition . This condition is straightforward to satisfy in practice, for instance, through random initialization of the weight vector. In the extreme case, if , then we can directly calculate all the eigenvectors by constructing an orthogonal space, which has a very small amount of computation. Due to the wide range of changes in the weight vector modulus during the iteration process, the adaptive learning factor has a large value space.
4. Simulation Experiment
Three simulation experiments are conducted in this section to validate the performance of the proposed algorithm. Experiment 1 verifies the algorithm’s capability to directly estimate principal components from input signals. Experiment 2 examines the conclusions drawn from the convergence analysis. Experiment 3 assesses the algorithm’s performance under extreme conditions.
4.1. Signal Principal Component Estimation Experiment
For this experimental setup, the input signal is synthesized using the following first-order sliding regression process.
where is a Gaussian random process with zero mean and unit variance. Eight consecutive, non-overlapping data points form the input signal vector .The first principal component of this signal is then extracted using Oja’s algorithm [], the PASTd algorithm [], and the proposed algorithm, respectively. To ensure operational stability, both the Oja and PASTd algorithms normalize the weight vector, a feature fundamental to their design. The initialization parameters of the algorithm are set as follows: (Oja’s algorithm); , (the PASTd algorithm). The proposed algorithm does not need to set parameters. For fair comparison, the three algorithms use the same initialization weight vector.
In the iterative process, the direction cosine between the weight vector and the principal component is calculated continuously.
Equation (42) shows that the weight vector aligns with the principal component direction as converges to 1. That is to say, the represents the algorithm’s ability to extract principal components. In adaptive signal processing, where neural network algorithms are predominantly applied, the convergence speed emerges as the primary performance metric once the algorithm’s curve approaches 1.
Figure 1 displays the directional cosine curves computed by the three algorithms, representing the average of 30 independent trials. Simulation results illustrate the convergence trajectory of the weight vector . It asymptotically approaches the principal component direction, reaching it after around 200 iterations, which confirms the algorithm’s principal component extraction capability.
Figure 1.
Direction cosine of three algorithms.
A comparative evaluation confirmed that the proposed algorithm was observed to converge more rapidly than the two alternative methods by a significant margin.
4.2. Verification Experiment on Convergence Analysis
The following signal autocorrelation matrix is generated by using the method in [].
The maximum and minimum eigenvalues of the matrix are and , respectively, and its principal component is . The largest principal component of this autocorrelation matrix is estimated using the proposed algorithm (4). During the iteration process, the absolute value of the basis coefficients (i.e., ), the norm of the weight vector and the direction cosine are calculated.
Figure 2 and Figure 3 present the simulation results obtained when the initial vector satisfies . Since holds, this condition fulfills the convergence requirements specified in Theorem 3. As observed in Figure 2, increases monotonically and converges to a constant value, while ultimately converges to zero. These findings are consistent with the conclusions derived from the convergence analysis. Figure 3 demonstrates that the directional cosine curve converges to 1, indicating that Algorithm (4) accurately estimates the direction of the principal component. Concurrently, the magnitude of the weight vector also exhibits monotonic increase, aligning with the analytical conclusions presented in Section 3. This further validates the correctness of Theorem 3.
Figure 2.
The absolute value curve of the basis coefficient.
Figure 3.
Modulus and direction cosine curve.
4.3. Algorithm Performance Experiment Under Pathological Conditions
The condition number serves as a critical metric for assessing whether a matrix is ill-conditioned. Typically, the larger the condition number, the more ill-conditioned the matrix becomes, imposing more stringent requirements on algorithms. To assess the performance of the proposed algorithm on ill-conditioned matrices, we employ the following symmetric positive definite matrix:
Computations reveal that the condition number of this matrix is . When employing other algorithms to estimate the principal components of ill-conditioned matrices, a very small learning factor is typically chosen to ensure algorithm convergence; however, this significantly impacts the convergence speed. Here, the proposed algorithm is employed for principal component estimation of this matrix. Figure 4 illustrates the trajectory of the basis coefficient’s absolute value under the initial condition . Similar to Figure 2, the proposed algorithm readily achieves convergence even when handling the ill-conditioned matrix.
Figure 4.
Absolute value curve of basis coefficient under ill-conditioned.
5. Conclusions
The choice of learning rate significantly impacts the convergence speed of algorithms. To extend the feasible range of learning rates, a principal component estimation algorithm with an adaptive learning rate is proposed. The convergence analysis of the proposed algorithm is rigorously conducted by considering two distinct cases. The derived convergence conditions impose requirements solely on the initialization of the weight vector, with no restrictions applied to the learning rate. Simulation results demonstrate the algorithm’s functional capability for principal component extraction, while also corroborating the validity of the theoretical convergence analysis. In the future, the proposed algorithm can be used to solve image compression, fault diagnosis and other problems.
Author Contributions
Conceptualization, Y.G.; methodology, Y.G. and H.D.; software, H.D. and Y.Z.; validation, Y.G.; formal analysis, Z.X.; investigation, H.D. and Y.Z.; resources, J.L.; data curation, Y.G.; writing—original draft preparation, Y.G.; writing—review and editing, Y.G.; visualization, K.J.; supervision, Z.X.; project administration, Y.G. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported in part by the National Natural Science Foundation of China (62101579, 62106242).
Data Availability Statement
The data that support the findings of this study are available within the article.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Du, K.-L.; Swamy, M. Principal component analysis. In Neural Networks and Statistical Learning; Springer: Berlin/Heidelberg, Germany, 2019; pp. 373–425. [Google Scholar]
- Zhou, C.; Gu, Y.; Fan, X.; Shi, Z.; Mao, G.; Zhang, Y.D. Direction-of-arrival estimation for coprime array via virtual array interpolation. IEEE Trans. Signal Process. 2018, 66, 5956–5971. [Google Scholar] [CrossRef]
- Bartecki, K. Classical vs. neural network-based PCA approaches for lossy image compression: Similarities and differences. Appl. Soft Comput. 2024, 161, 111721. [Google Scholar] [CrossRef]
- Zhang, Q.; Song, C.; Yuan, Y. Fault diagnosis of vehicle gearboxes based on adaptive wavelet threshold and LT-PCA-NGO-SVM. Appl. Sci. 2024, 14, 1212. [Google Scholar] [CrossRef]
- Chen, P.; Wu, L.; Peng, D.; Gong, X.; De Albuquerque, V.H.C. An improved minor component analysis algorithm based on convergence analysis of 5G multi-dimensional signals. IEEE Access 2019, 7, 91860–91871. [Google Scholar] [CrossRef]
- Qian, G.; Lin, B.; Mei, J.; Qian, J.; Wang, S. Adaptive Learning Algorithm and Its Convergence Analysis with Complex-Valued Error Loss Network. Neural Netw. 2025, 190, 107677. [Google Scholar] [CrossRef]
- Cha, G.-W.; Choi, S.-H.; Hong, W.-H.; Park, C.-W. Developing a prediction model of demolition-waste generation-rate via principal component analysis. Int. J. Environ. Res. Public Health 2023, 20, 3159. [Google Scholar] [CrossRef]
- Gao, Y.; Dong, H.; Xu, Z.; Li, H.; Li, J.; Yuan, S. Multiple Minor Components Extraction in Parallel Based on Möller Algorithm. Electronics 2025, 14, 4073. [Google Scholar] [CrossRef]
- Lei, D.; Zhang, Y.; Lu, Z.; Lin, H.; Fang, B.; Jiang, Z. Slope stability prediction using principal component analysis and hybrid machine learning approaches. Appl. Sci. 2024, 14, 6526. [Google Scholar] [CrossRef]
- Liu, Y.; Ruan, X. Parameter-Optimal-Gain-Arguable Iterative Learning Control for Linear Time-Invariant Systems with Quantized Error. Appl. Sci. 2023, 13, 9551. [Google Scholar] [CrossRef]
- Zhang, F.; Chen, Z. A Novel Reinforcement Learning-Based Particle Swarm Optimization Algorithm for Better Symmetry between Convergence Speed and Diversity. Symmetry 2024, 16, 1290. [Google Scholar] [CrossRef]
- Zhang, Y.; Wu, W.; He, W.; Zhao, N. Algorithm design and convergence analysis for coexistence of cognitive radio networks in unlicensed spectrum. Sensors 2023, 23, 9705. [Google Scholar] [CrossRef]
- Naik, G.R.; Selvan, S.E.; Gobbo, M.; Acharyya, A.; Nguyen, H.T. Principal component analysis applied to surface electromyography: A comprehensive review. IEEE Access 2016, 4, 4025–4037. [Google Scholar] [CrossRef]
- Kalla, U.K. Minor component analysis based anti-hebbian neural network scheme of decoupled voltage and frequency controller (DVFC) for nanohydro system. In Proceedings of the 2016 IEEE 7th Power India International Conference (PIICON), Bikaner, India, 25–27 November 2016; pp. 1–6. [Google Scholar]
- Nguyen, T.D.; Yamada, I. A unified convergence analysis of normalized PAST algorithms for estimating principal and minor components. Signal Process. 2013, 93, 176–184. [Google Scholar] [CrossRef]
- Nguyen, V.-D.; Abed-Meraim, K.; Linh-Trung, N.; Weber, R. Generalized minimum noise subspace for array processing. IEEE Trans. Signal Process. 2017, 65, 3789–3802. [Google Scholar] [CrossRef]
- Pehlevan, C.; Hu, T.; Chklovskii, D.B. A hebbian/anti-hebbian neural network for linear subspace learning: A derivation from multidimensional scaling of streaming data. Neural Comput. 2015, 27, 1461–1495. [Google Scholar] [CrossRef]
- Bartelmaos, S.; Abed-Meraim, K. Fast adaptive algorithms for minor component analysis using Householder transformation. Digit. Signal Process. 2011, 21, 667–678. [Google Scholar] [CrossRef]
- Feng, X.; Kong, X.; Duan, Z.; Ma, H. Adaptive generalized eigen-pairs extraction algorithms and their convergence analysis. IEEE Trans. Signal Process. 2016, 64, 2976–2989. [Google Scholar] [CrossRef]
- Feng, X.; Kong, X.; Ma, H.; Liu, H. Unified and coupled self-stabilizing algorithms for minor and principal eigen-pairs extraction. Neural Process. Lett. 2017, 45, 197–222. [Google Scholar] [CrossRef]
- Zhang, W.-T.; Lou, S.-T.; Feng, D.-Z. Adaptive quasi-Newton algorithm for source extraction via CCA approach. IEEE Trans. Neural Netw. Learn. Syst. 2013, 25, 677–689. [Google Scholar] [CrossRef] [PubMed]
- Kong, X.; Hu, C.; Duan, Z. Principal Component Analysis Networks and Algorithms; Springer: Durham, NC, USA, 2017. [Google Scholar]
- Tan, K.K.; Lv, J.C.; Yi, Z.; Huang, S. Adaptive multiple minor directions extraction in parallel using a PCA neural network. Theor. Comput. Sci. 2010, 411, 4200–4215. [Google Scholar] [CrossRef]
- Nguyen, T.D.; Yamada, I. Necessary and sufficient conditions for convergence of the DDT systems of the normalized PAST algorithms. Signal Process. 2014, 94, 288–299. [Google Scholar] [CrossRef]
- Kong, X.; An, Q.; Ma, H.; Han, C.; Zhang, Q. Convergence analysis of deterministic discrete time system of a unified self-stabilizing algorithm for PCA and MCA. Neural Netw. 2012, 36, 64–72. [Google Scholar] [CrossRef] [PubMed]
- Miao, Y.; Hua, Y. Fast subspace tracking and neural network learning by a novel information criterion. IEEE Trans. Signal Process. 2002, 46, 1967–1979. [Google Scholar] [CrossRef]
- Chen, T.; Amari, S.-I. Amari. Unified stabilization approach to principal and minor components extraction algorithms. Neural Netw. 2001, 14, 1377–1387. [Google Scholar] [CrossRef] [PubMed]
- Antoniou, A. Digital Signal Processing; McGraw-Hill: New York, NY, USA, 2016. [Google Scholar]
- Wang, R.; Yao, M.; Zhang, D.; Zou, H. Stable and orthonormal OJA algorithm with low complexity. IEEE Signal Process. Lett. 2011, 18, 211–214. [Google Scholar] [CrossRef]
- Chan, S.-C.; Tan, H.-J.; Lin, J.-Q.; Liao, B. A new local polynomial modeling based variable forgetting factor and variable regularized PAST algorithm for subspace tracking. IEEE Trans. Aerosp. Electron. Syst. 2018, 54, 1530–1544. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).