Next Article in Journal
H Preview Tracking Control of Time-Delay Discrete Systems and Its Application in Nuclear Reactor Problems
Previous Article in Journal
m-Isometric Operators with Null Symbol and Elementary Operator Entries
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

LDC-GAT: A Lyapunov-Stable Graph Attention Network with Dynamic Filtering and Constraint-Aware Optimization

1
School of Science, Zhejiang Sci-Tech University, Hangzhou 310018, China
2
School of Computer Science, Zhejiang Sci-Tech University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(7), 504; https://doi.org/10.3390/axioms14070504 (registering DOI)
Submission received: 6 June 2025 / Revised: 25 June 2025 / Accepted: 26 June 2025 / Published: 27 June 2025
(This article belongs to the Section Mathematical Analysis)

Abstract

Graph attention networks are pivotal for modeling non-Euclidean data, yet they face dual challenges: training oscillations induced by projection-based high-dimensional constraints and gradient anomalies due to poor adaptation to heterophilic structure. To address these issues, we propose LDC-GAT (Lyapunov-Stable Graph Attention Network with Dynamic Filtering and Constraint-Aware Optimization), which jointly optimizes both forward and backward propagation processes. In the forward path, we introduce Dynamic Residual Graph Filtering, which integrates a tunable self-loop coefficient to balance neighborhood aggregation and self-feature retention. This filtering mechanism, constrained by a lower bound on Dirichlet energy, improves multi-head attention via multi-scale fusion and mitigates overfitting. In the backward path, we design the Fro-FWNAdam, a gradient descent algorithm guided by a learning-rate-aware perceptron. An explicit Frobenius norm bound on weights is derived from Lyapunov theory to form the basis of the perceptron. This stability-aware optimizer is embedded within a Frank–Wolfe framework with Nesterov acceleration, yielding a projection-free constrained optimization strategy that stabilizes training dynamics. Experiments on six benchmark datasets show that LDC-GAT outperforms GAT by 10.54% in classification accuracy, which demonstrates strong robustness on heterophilic graphs.
Keywords: LDC-GAT; DRG-Filtering; Fro-FWNAdam; multi-head weight threshold LDC-GAT; DRG-Filtering; Fro-FWNAdam; multi-head weight threshold

Share and Cite

MDPI and ACS Style

Chen, L.; Zhu, H.; Han, S. LDC-GAT: A Lyapunov-Stable Graph Attention Network with Dynamic Filtering and Constraint-Aware Optimization. Axioms 2025, 14, 504. https://doi.org/10.3390/axioms14070504

AMA Style

Chen L, Zhu H, Han S. LDC-GAT: A Lyapunov-Stable Graph Attention Network with Dynamic Filtering and Constraint-Aware Optimization. Axioms. 2025; 14(7):504. https://doi.org/10.3390/axioms14070504

Chicago/Turabian Style

Chen, Liping, Hongji Zhu, and Shuguang Han. 2025. "LDC-GAT: A Lyapunov-Stable Graph Attention Network with Dynamic Filtering and Constraint-Aware Optimization" Axioms 14, no. 7: 504. https://doi.org/10.3390/axioms14070504

APA Style

Chen, L., Zhu, H., & Han, S. (2025). LDC-GAT: A Lyapunov-Stable Graph Attention Network with Dynamic Filtering and Constraint-Aware Optimization. Axioms, 14(7), 504. https://doi.org/10.3390/axioms14070504

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop