1. Introduction
Knowledge graphs (KGs) have been widely used to represent structured knowledge and play an important role in tasks such as recommendation systems and semantic search [
1,
2,
3]. However, the static characteristics of traditional knowledge graphs limit their ability to model dynamic relationships and cannot adapt to time-sensitive tasks (such as event prediction and supply chain risk management). Temporal Knowledge Graphs (TKGs) introduce the time dimension and model facts that evolve over time in the form of quadruples
, which can effectively capture the dynamic changes of entity relationships [
4,
5]
At present, Temporal Knowledge Graph Reasoning is mainly divided into two types of methods: interpolation and extrapolation. Interpolation models (e.g., TA-DistMult [
6], TTransE [
7], HyTE [
8]) are used to fill in missing facts within the observed time range. Extrapolation models (e.g., Know-Evolve [
2], DyRep [
9], RE-NET [
10]) predict unknown facts in the future by analyzing time series patterns.
Although extrapolation models are crucial in applications such as financial forecasting and supply chain risk assessment [
3,
11], existing methods still face the following three core challenges:
Noisy Confounder Interference: In actual TKG data, causal features (such as Tesla’s R&D progress driving product releases) and confounding factors (such as supply chain disruptions causing Model Q delays) are often intertwined, making the model sensitive to noise. For example, RE-GCN’s MRR dropped by 12.3% on the GDELT dataset, indicating that its generalization ability is weak in high-noise scenarios [
12].
Temporal–Semantic Misalignment: Existing methods either focus on temporal consistency (e.g., RE-NET [
10] processes temporal patterns through RNN) or optimize semantic alignment (e.g., CyGNet [
13] focuses on entity–relation similarity), but few methods can optimize both at the same time, resulting in unstable performance in multi-task reasoning (e.g., predicting entities and relations simultaneously).
Static–Dynamic Feature Imbalance: Static knowledge graphs provide structured prior information (as shown in
Figure 1, where Tesla’s long-term product roadmap reflects static strategy), while dynamic features reflect temporal evolution patterns (such as short-term supply chain fluctuations). This figure illustrates the interplay between consistent long-term strategies and unpredictable short-term variations—emphasizing the need for TKG models that can balance static structure with temporal dynamics to make robust future predictions. However, TGformer [
14] only uses static information and fails to effectively model the multi-scale dependencies between long-term trends and short-term dynamics, limiting its reasoning capabilities.
Despite recent progress, existing TKG reasoning models suffer from three critical limitations that hinder their performance in real-world, noisy, and temporally complex environments: they lack explicit mechanisms for disentangling causal signals from noise; they treat temporal consistency and semantic alignment as separate optimization objectives; and they fail to effectively fuse static and dynamic knowledge across different temporal scales.
To the best of our knowledge, no prior work has addressed all these limitations in a unified framework. This gap motivates the development of TCCGN—a novel model that integrates causal–noise decoupling, semantic–temporal alignment, and static–dynamic fusion into a cohesive architecture for robust temporal reasoning.
4. Experiments
4.7. Analysis of Module Contributions and Synergistic Effects
To verify the contribution of each module in the model to the performance of TKG reasoning, we conducted ablation experiments on the ICEWS14, ICEWS05-15, ICEWS18, YAGO, and GDELT datasets. The results are summarized in
Table 9. From the table, it is clear that the Causal and Confounding Representation Learning (CD) module provides a strong baseline. For instance, MRR scores with CD alone reach 41.66% on ICEWS14 and 46.33% on ICEWS05-15, and 30.97% on ICEWS18, forming the foundation for causal-aware modeling.
With the addition of the Dynamic Dual Contrastive Learning (DDCL) module, performance improves across the board. For example, MRR increases to 41.93% and 47.10% on ICEWS14 and ICEWS05-15, and to 31.08% on ICEWS18. These results validate DDCL’s utility in enhancing temporal smoothness and suppressing noisy supervision.
Similarly, introducing the Global Static–Dynamic Fusion (GSDF) module independently also improves performance, with the MRR on the ICEWS18 dataset rising to 30.90%. When GSDF is combined with the CD module, the MRR further increases to 31.10%, while the GSDF + DDCL combination shows a slight drop among these variants, with the MRR reaching 30.88%. These results highlight the advantage of fusing static and dynamic contexts under temporal contrastive modeling.
Notably, while the margin between the CD+DDCL combination (31.00%) and the full TCCGN model (31.23%) on ICEWS18 appears small (+0.23%), this trend is consistent and reproducible across datasets and multiple runs. The incremental gain reflects a saturation point common in modular designs—where the final component (e.g., GSDF) builds on an already strong backbone. Similar incremental behaviors have been reported in robust temporal KG frameworks [
20]. It is important to note that all results in
Table 9 are averaged over three repeated runs to reduce noise and account for variance. While standard deviations are not explicitly reported in the table, our internal analysis confirmed that the variation across runs was small (typically < 0.10 MRR), and trends remained consistent.
Finally, TCCGN achieves the highest overall performance by integrating CD, DDCL, and GSDF, reaching 42.46%, 47.33%, and 31.23% on ICEWS14, ICEWS05-15, and ICEWS18, respectively. This confirms the design’s synergistic effect across causal modeling, contrastive learning, and global fusion.
In summary, although individual module gains may appear numerically modest, their combined effect leads to statistically meaningful improvements, enhanced robustness, and better generalization—crucial in dynamic and noisy real-world TKG environments.
5. Conclusions
In this paper, we propose TCCGN, a novel and unified reasoning framework for temporal knowledge graphs (TKGs) that robustly captures temporal dynamics while suppressing noise. TCCGN integrates three complementary modules—Causal and Confounding Representation Learning (CD), Dynamic Dual-domain Contrastive Learning (DDCL), and Global Static–Dynamic Fusion (GSDF)—to address the challenges of temporal irregularity, information sparsity, and noisy supervision. By disentangling causal signals from spurious correlations, aligning representations across structural and temporal views, and fusing evolving and static contexts, TCCGN enhances the expressiveness and generalization of temporal reasoning models.
Comprehensive experiments conducted on five standard TKG benchmarks—ICEWS14, ICEWS05-15, ICEWS18, YAGO, and GDELT—demonstrate that TCCGN consistently outperforms state-of-the-art methods in both entity prediction and relation prediction tasks. Furthermore, we perform ablation studies and hyperparameter analyses to assess the individual contributions of each module. These experiments are repeated under consistent settings to ensure statistical validity, and the results confirm that even small gains across modules are robust and reproducible.
Despite its strong performance, TCCGN also has limitations. The current design assumes an exponentially decaying influence of historical noise, which may fail to model structured, periodic, or cyclic interference patterns often seen in real-world data. This limitation can affect the accuracy of causal-confounding separation in time-dependent contexts with regularities, such as seasonality or weekly cycles.
Moreover, while our experiments focus on clean and structured benchmarks, many real-world TKGs—such as those in healthcare, finance, or scientific discovery—exhibit higher noise levels, incomplete schema, and complex event semantics. TCCGN’s modular design shows promising robustness under such conditions, but further validation is needed. Future work will extend our evaluation to real-world noisy TKGs and explore improvements such as noise-aware training objectives, interpretable causal modules, and schema-adaptive components that dynamically adjust to heterogeneous or evolving ontology structures.
We also plan to enhance TCCGN’s temporal modeling capacity by incorporating periodic basis functions or neural temporal kernels to better represent recurring and long-range temporal patterns. Furthermore, we aim to extend TCCGN to support multi-hop reasoning and multitask learning across related temporal tasks, such as temporal question answering, forecasting, and anomaly detection.
Finally, several open challenges remain: improving performance on temporally sparse or irregular datasets; integrating multimodal knowledge to enrich temporal understanding; and enhancing the interpretability of learned causal pathways to support trustworthy and explainable decision-making. Addressing these challenges will be critical for deploying temporal KG reasoning systems in real-world dynamic environments.
In summary, this work presents a principled, modular, and empirically validated framework for robust temporal reasoning in dynamic and noisy environments. By explicitly addressing the entanglement of causal and confounding signals, aligning multi-view temporal semantics, and integrating static–dynamic entity representations, TCCGN provides both theoretical grounding and practical effectiveness for advancing temporal knowledge graph applications. We hope this work will inform future developments in robust and interpretable temporal reasoning across diverse domains.
Author Contributions
Conceptualization, S.F. and H.L.; Methodology, S.F. and H.L.; Software, S.F., H.L., Q.L. and Y.Z.; Validation, H.L., Q.L., P.X. and B.C.; Formal analysis, H.L., Q.L. and P.X.; Resources, H.L. and Y.Z.; Writing—original draft preparation, S.F. and H.L.; Writing—review and editing, Q.L., M.H. and B.C.; Project administration, S.F. and M.H.; Funding acquisition, S.F. and M.H. All authors have read and agreed to the published version of the manuscript.
Funding
This research received funding from the National Natural Science Foundation of China (Grants 62466016 and 62241202), the National Key Research and Development Program of China (Grant 2021ZD0111000), and the Key Research and Development Plan of the Ministry of Science and Technology (Grant 2021ZD0111002).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data are contained within the article.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Gao, H.; Wu, L.; Hu, P.; Wei, Z.; Xu, F.; Long, B. Graph-augmented Learning to Rank for Querying Large-scale Knowledge Graph. In Proceedings of the AACL-IJCNLP, Online, 20–23 November 2022; pp. 82–92. [Google Scholar]
- Trivedi, R.; Dai, H.; Wang, Y.; Chang, L. Know-evolve: Deep temporal reasoning for dynamic knowledge graphs. In Proceedings of the 34th International Conference on Machine Learning (ICML), Sydney, Australia, 6–11 August 2017. [Google Scholar]
- Chen, Z.; Zhang, Y.; Yu, S.; Wang, Y.; Shen, H. Temporal knowledge graph question answering via subgraph reasoning. Knowl.-Based Syst. 2022, 251, 109134. [Google Scholar] [CrossRef]
- Chen, W.; Liang, X.; Zhang, M.; He, F.; Wang, Y.; Yang, T. Building and exploiting spatial–temporal knowledge graph for next POI recommendation. Knowl.-Based Syst. 2022, 258, 109951. [Google Scholar] [CrossRef]
- Kazemi, S.M.; Goel, R.; Jain, K.; Kobyzev, I.; Sethi, A.; Forsyth, P.; Poupart, P. Representation learning for dynamic graphs: A survey. J. Mach. Learn. Res. 2020, 21, 1–73. [Google Scholar]
- García-Durán, A.; Dumančić, S.; Niepert, M. Learning sequence encoders for temporal knowledge graph completion. arXiv 2018, arXiv:1809.03202. [Google Scholar] [CrossRef]
- Leblay, J.; Chekol, M.W. Deriving validity time in knowledge graph. In Proceedings of the Companion Proceedings of The Web Conference 2018, Lyon, France, 23–27 April 2018. [Google Scholar]
- Dasgupta, S.S.; Ray, S.N.; Talukdar, P. HyTE: Hyperplane-based temporally aware knowledge graph embedding. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP), Brussels, Belgium, 31 October–4 November 2018. [Google Scholar]
- Trivedi, R.; Farajtabar, M.; Biswal, P.; Zha, H. DyRep: Learning representations over dynamic graphs. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Jin, W.; Yang, Y.; Liu, X.; Sun, Z.; Huang, H. Recurrent event network: Autoregressive structure inference over temporal knowledge graphs. arXiv 2019, arXiv:1904.05530. [Google Scholar]
- Xu, C.; Kou, B.; Zhang, L.; Li, P.; Liu, Y.; Wu, B. Temporal knowledge graph completion using a linear temporal regularizer and multivector embeddings. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021. [Google Scholar]
- Li, Z.; Sun, Z.; Yu, J.; Zhang, W.; Ji, H. Temporal knowledge graph reasoning based on evolutional representation learning. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual Event, 11–15 July 2021. [Google Scholar]
- Zhu, C.; Li, C.; Cao, J.; Xiong, F.; Zhang, L. Learning from history: Modeling temporal knowledge graphs with sequential copy-generation networks. Proc. AAAI Conf. Artif. Intell. 2021, 35, 4733–4741. [Google Scholar] [CrossRef]
- Shi, F.; Li, D.; Wang, X.; Li, B.; Wu, X. TGformer: A Graph Transformer Framework for Knowledge Graph Embedding. IEEE Trans. Knowl. Data Eng. 2024, 37, 526–541. [Google Scholar] [CrossRef]
- Ma, Q.; Zhang, X.; Ding, Z.; Gao, C.; Shang, W.; Nong, Q.; Ma, Y.; Jin, Z. Temporal knowledge graph reasoning based on evolutional representation and contrastive learning. Appl. Intell. 2024, 54, 10929–10947. [Google Scholar] [CrossRef]
- Fang, Z.; Lei, S.L.; Zhu, X.; Yang, C.; Zhang, S.X.; Yin, X.C.; Qin, J. Transformer-based Reasoning for Learning Evolutionary Chain of Events on Temporal Knowledge Graph. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, Washington, DC, USA, 14–18 July 2024. [Google Scholar]
- Sun, H.; Geng, S.; Zhong, J.; Hu, H.; He, K. Graph Hawkes Transformer for Extrapolated Reasoning on Temporal Knowledge Graphs. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022. [Google Scholar]
- Liu, K.; Zhao, F.; Chen, H.; Li, Y.; Xu, G.; Jin, H. Da-net: Distributed Attention Network for Temporal Knowledge Graph Reasoning. In Proceedings of the 31st ACM International Conference on Information & Knowledge Management (CIKM), Atlanta, GA, USA, 17–21 October 2022. [Google Scholar]
- Zhang, D.; Rong, Z.; Xue, C.; Li, G. Simre: Simple Contrastive Learning with Soft Logical Rule for Knowledge Graph Embedding. Inf. Sci. 2024, 661, 120069. [Google Scholar] [CrossRef]
- Liu, Z.; Tan, L.; Li, M.; Zhang, W. Simfy: A Simple yet Effective Approach for Temporal Knowledge Graph Reasoning. In Findings of the Association for Computational Linguistics: EMNLP 2023; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 3825–3836. [Google Scholar]
- Xu, Y.; Li, P.; Zhang, Y.; Zhao, M.; Liu, X. Temporal knowledge graph reasoning with historical contrastive learning. Proc. AAAI Conf. Artif. Intell. 2023, 37, 4521–4528. [Google Scholar] [CrossRef]
- Yang, J.; Wang, X.; Wang, Y.; Wang, J.; Wang, F.Y. AMCEN: An Attention Masking-based Contrastive Event Network for Two-stage Temporal Knowledge Graph Reasoning. arXiv 2024, arXiv:2405.10346. [Google Scholar]
- Xu, Y.; Shi, B.; Ma, T.; Dong, B.; Zhou, H.; Zheng, Q. CLDG: Contrastive Learning on Dynamic Graphs. In Proceedings of the 2023 IEEE 39th International Conference on Data Engineering (ICDE), Anaheim, CA, USA, 3–7 April 2023. [Google Scholar]
- Peng, M.; Liu, B.; Xu, W.; Jiang, Z.; Zhu, J.; Peng, M. Deja Vu: Contrastive Historical Modeling with Prefix-tuning for Temporal Knowledge Graph Reasoning. arXiv 2024, arXiv:2404.00051. [Google Scholar] [CrossRef]
- Xu, W.; Liu, B.; Peng, M.; Jia, X.; Peng, M. Pre-trained language model with prompts for temporal knowledge graph completion. arXiv 2023, arXiv:2305.07912. [Google Scholar]
- Chen, W.; Wan, H.; Wu, Y.; Zhao, S.; Cheng, J.; Li, Y.; Lin, Y. Local-Global History-Aware Contrastive Learning for Temporal Knowledge Graph Reasoning. In Proceedings of the 2024 IEEE 40th International Conference on Data Engineering (ICDE), Utrecht, The Netherlands, 13–16 May 2024. [Google Scholar]
- Shao, P.; Wang, Y.; Zhang, Y.; Wang, H.; Liu, Y. Tucker decomposition-based temporal knowledge graph completion. Knowl.-Based Syst. 2022, 238, 107841. [Google Scholar] [CrossRef]
- Zhang, J.; Zhu, Z.; Li, H.; Li, J. Spatial-temporal attention network for temporal knowledge graph completion. In Proceedings of the Database Systems for Advanced Applications 26th nternational Conference DASFAA 2021, Taipei, Taiwan, 11 April 2021; Proceedings, Part I. Springer International Publishing: Cham, Switzerland, 2021; Volume 12681. [Google Scholar]
- Sui, Y.; Xie, J.; Hou, X.; Chen, Z.; Tian, C. Causal attention for interpretable and generalizable graph classification. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022. [Google Scholar]
- Han, Z.; Ye, H.; Sun, Z.; Lin, Y.; Han, X.; Zhou, J.; Liu, Z.; Li, P.; Sun, M.; Zhou, J. Dyernie: Dynamic evolution of riemannian manifold embeddings for temporal knowledge graph completion. arXiv 2020, arXiv:2011.03984. [Google Scholar] [CrossRef]
- Han, Z.; Sun, Z.; Lin, Y.; Ye, H.; Liu, Z.; Li, P.; Zhou, J. Learning neural ordinary equations for forecasting future links on temporal knowledge graphs. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP), Virtual Event, 7–11 November 2021. [Google Scholar]
- Wu, J.; Feng, Y.; Wang, W.; Chen, M.; Zhao, Y. Temp: Temporal message passing for temporal knowledge graph completion. arXiv 2020, arXiv:2010.03526. [Google Scholar] [CrossRef]
- Sun, X.; Zhang, J.; Wu, X.; Cheng, H.; Xiong, Y.; Li, J. Graph Prompt Learning: A Comprehensive Survey and Beyond. arXiv 2023, arXiv:2311.16534. [Google Scholar] [CrossRef]
- Korkmaz, G.; Cadena, J.; Kuhlman, C.J.; Marathe, A.; Vullikanti, A.; Ramakrishnan, N. Combining heterogeneous data sources for civil unrest forecasting. In Proceedings of the 2015 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Paris, France, 25–28 August 2015; pp. 258–265. [Google Scholar]
- Bousmalis, K.; Trigeorgis, G.; Silberman, N.; Krishnan, D.; Erhan, D. Domain separation networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Barcelona, Spain, 5–10 December 2016; Volume 29. [Google Scholar]
- Locatello, F.; Bauer, S.; Lucic, M.; Raetsch, G.; Gelly, S.; Schölkopf, B.; Bachem, O. Challenging common assumptions in the unsupervised learning of disentangled representations. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 4114–4124. [Google Scholar]
- Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
- Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning (ICML), Virtual Event, 13–18 July 2020. [Google Scholar]
- O’Brien, S.P. Crisis early warning and decision support: Contemporary approaches and thoughts on future research. Int. Stud. Rev. 2010, 12, 87–104. [Google Scholar] [CrossRef]
- Han, Z.; Chen, P.; Ma, Y.; Tresp, V. Explainable Subgraph Reasoning for Forecasting on Temporal Knowledge Graphs. In Proceedings of the International Conference on Learning Representations, Online, 3–7 May 2021. [Google Scholar]
- Hoffart, J.; Suchanek, F.M.; Berberich, K.; Weikum, G. YAGO2: A spatially and temporally enhanced knowledge base from Wikipedia. Artif. Intell. 2013, 194, 28–61. [Google Scholar] [CrossRef]
- Leetaru, K.; Schrodt, P.A. GDELT: Global data on events, location, and tone, 1979–2012. ISA Annu. Conv. 2013, 2, 1–49. [Google Scholar]
- Trouillon, T.; Welbl, J.; Riedel, S.; Gaussier, É.; Bouchard, G. Complex embeddings for simple link prediction. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; pp. 2071–2080. [Google Scholar]
- Yang, B.; Yih, W.T.; He, X.; Gao, J.; Deng, L. Embedding entities and relations for learning and reasoning in knowledge bases. In Proceedings of the International Conference on Machine Learning (ICLR) (Poster), Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
- Sun, Z.; Deng, Z.H.; Nie, J.Y.; Tang, X. Rotate: Knowledge graph embedding by relational rotation in complex space. arXiv 2019, arXiv:1902.10197. [Google Scholar] [CrossRef]
- Dettmers, T.; Minervini, P.; Stenetorp, P.; Riedel, S. Convolutional 2D knowledge graph embeddings. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. Art. 221. [Google Scholar]
- Shang, C.; Tang, Y.; Huang, J.; Bi, J.; He, X.; Zhou, B. End-to-end structure-aware convolutional networks for knowledge base completion. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019. Art. 3060. [Google Scholar]
- Goel, R.; Kazemi, S.M.; Brubaker, M.A.; Poole, D. Diachronic embedding for temporal knowledge graph completion. Proc. AAAI Conf. Artif. Intell. 2020, 34, 3988–3995. [Google Scholar] [CrossRef]
- Kazemi, S.M.; Poole, D.L. Simple Embedding for Link Prediction in Knowledge Graphs. In Proceedings of the Advances in Neural Information Processing Systems, Montr’eal, QC, Canada, 3–8 December 2018; pp. 428–438. [Google Scholar]
- Lacroix, T.; Obozinski, G.; Usunier, N. Tensor decompositions for temporal knowledge base completion. arXiv 2020, arXiv:2004.04926. [Google Scholar] [CrossRef]
- Gao, Y.; Feng, L.; Kan, Z.; Han, Y.; Qiao, L.; Li, D. Modeling Precursors for Temporal Knowledge Graph Reasoning via Auto-encoder Structure. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Vienna, Austria, 23–29 July 2022. [Google Scholar]
- Li, Z.; Feng, S.; Shi, J.; Zhou, Y.; Liao, Y.; Yang, Y.; Li, Y.; Yu, N.; Shao, X. Future Event Prediction Based on Temporal Knowledge Graph Embedding. Comput. Syst. Sci. Eng. 2023, 44, 2411–2423. [Google Scholar] [CrossRef]
Figure 1.
Tesla’s product timeline as a temporal knowledge graph, showing the interplay of static trends and temporal dynamics for future reasoning.
Figure 2.
An illustrative diagram of the proposed TCCGN model. The CD component represents the causal decoupling module. The DDCL component represents the dual-domain contrastive learning module. The GSDF represents the gated static–dynamic fusion module.
Figure 3.
The variations in losses during training on five datasets.
Figure 4.
Performance (%) using different history length settings of ICEWS14.
Figure 5.
Performance (%) of different dilate length settings using ICEWS14.
Figure 6.
Performance (%) of various embedding dimensions on ICEWS14. Values shown are from the final training epoch (500).
Table 1.
Performance comparison on representative benchmarks (higher better for MRR/Hits@1; lower better for YAGO epoch time).
Model | GDELT-MRR | ICEWS14 Hits@1 | YAGO Epoch Time (Relative to RE-GCN) |
---|
CyGNet [13] | 0.1805 | 0.2535 | — |
RE-GCN [12] | 0.1931 | 0.3046 | 1.17× |
TCCGN (ours) | 0.1963 | 0.3163 | 1.00× |
Table 2.
Key symbols and their descriptions.
Symbol | Description |
---|
| Set of triples at time t. |
| Subject, relation, object in a quadruple. |
| Weight matrix for relation r at layer . |
| GCN output at time t. |
| Update gate: . |
| Updated state: . |
| Causal and noise embeddings at time t. |
| Combined embedding: . |
| Projection matrices for causal and noise components. |
| Causal–noise decomposition loss. |
| Gradient reversal layer. |
| Adversarial loss on noise embedding. |
| GRU modules for causal and noise sequences. |
| Fusion gate: . |
| Dynamic memory decay coefficient. |
| Time-step contrastive loss. |
| Entity–relation alignment loss. |
| Margin hyperparameters for contrastive losses. |
| Static embedding of entity i and dynamic embedding at time t. |
| Static–dynamic fusion gate: . |
| Static–dynamic fusion contrastive loss. |
| Total loss of static–dynamic module. |
| Weights and regularization coefficients. |
| Convolutional TransE decoder. |
Table 3.
Statistics of the datasets.
Dataset | Entities | Relations | Training | Validation | Test |
---|
ICEWS18 | 23,033 | 256 | 373,018 | 45,995 | 49,545 |
ICEWS14 | 12,498 | 260 | 323,895 | 8514 | 341,409 |
ICEWS05-15 | 10,094 | 251 | 368,868 | 46,302 | 46,159 |
GDELT | 7691 | 240 | 1,734,399 | 238,765 | 305,241 |
YAGO | 10,623 | 10 | 161,540 | 19,523 | 20,026 |
Table 4.
Paired t-test results comparing TCCGN and baselines on MRR and Hits@1 (5 runs).
Model | Dataset | MRR (±std) | Hits@1 (±std) | p-Value (MRR) | p-Value (Hits@1) |
---|
RE-GCN | ICEWS14 | 0.413 ± 0.003 | 0.303 ± 0.002 | 1.1 × 10−5 | 2.6 × 10−5 |
CyGNet | GDELT | 0.1805 ± 0.0002 | 0.1113 ± 0.0001 | <1 × 10−6 | <1 × 10−6 |
TCCGN | ICEWS14 | 0.428 ± 0.003 | 0.315 ± 0.002 | — | — |
TCCGN | GDELT | 0.1963 ± 0.0002 | 0.1223 ± 0.0002 | — | — |
Table 5.
Performance (in percentage) of the entity prediction task using ICEWS14, ICEWS05-15, and ICEWS18. The best result is highlighted in bold.
Model | ICEWS14 | ICEWS05-15 | ICEWS18 |
---|
MRR | H@1 | H@3 | H@10 | MRR | H@1 | H@3 | H@10 | MRR | H@1 | H@3 | H@10 |
---|
DisMult [2015] | 20.32 | 6.13 | 27.59 | 46.61 | 19.91 | 5.63 | 27.22 | 47.33 | 13.86 | 5.61 | 15.22 | 31.26 |
ComplEx [2014] | 22.61 | 9.88 | 28.93 | 47.57 | 20.26 | 6.66 | 26.43 | 47.31 | 15.45 | 8.04 | 17.19 | 30.73 |
R-GCN [2022] | 28.03 | 19.42 | 31.95 | 44.83 | 27.13 | 18.83 | 30.04 | 43.16 | 15.05 | 8.13 | 16.49 | 29.00 |
ConvE [2018] | 30.30 | 21.30 | 34.42 | 47.89 | 31.40 | 21.56 | 35.70 | 50.96 | 22.81 | 13.63 | 25.83 | 41.43 |
ConvTransE [2018] | 31.50 | 22.46 | 34.98 | 50.03 | 30.28 | 20.79 | 33.80 | 49.95 | 23.22 | 14.26 | 26.13 | 41.34 |
RotatE [2019] | 25.71 | 16.41 | 29.01 | 45.16 | 19.01 | 10.42 | 21.35 | 36.92 | 14.53 | 6.47 | 15.78 | 31.86 |
HyTE [2018] | 16.78 | 2.13 | 24.84 | 43.94 | 16.05 | 6.53 | 20.20 | 34.72 | 7.41 | 3.10 | 7.33 | 16.01 |
TTransE [2018] | 12.86 | 3.14 | 15.72 | 33.65 | 16.53 | 5.51 | 20.77 | 39.26 | 8.44 | 1.85 | 8.95 | 22.38 |
TA-DistMult [2018] | 26.22 | 16.83 | 29.72 | 45.23 | 27.51 | 17.57 | 31.46 | 47.32 | 16.42 | 8.60 | 18.13 | 34.80 |
DE-Simple [2020] | 32.67 | 24.43 | 35.69 | 49.11 | 35.02 | 25.91 | 38.99 | 52.75 | 19.30 | 11.53 | 21.86 | 36.91 |
TNF-ComplEx [2020] | 32.12 | 23.35 | 36.03 | 49.13 | 27.54 | 9.52 | 30.80 | 42.86 | 21.23 | 13.28 | 24.02 | 36.91 |
CyGNet [2021] | 34.68 | 25.35 | 38.88 | 53.16 | 35.46 | 25.44 | 40.20 | 54.47 | 24.98 | 15.54 | 28.58 | 43.54 |
RE-NET [2020] | 35.77 | 25.99 | 40.10 | 54.87 | 36.86 | 26.24 | 41.85 | 57.60 | 26.17 | 16.43 | 29.89 | 44.37 |
TANGO-DistMult [2021] | 22.87 | 14.22 | 25.43 | 40.32 | 40.23 | 30.53 | 44.95 | 59.05 | 26.21 | 16.92 | 29.77 | 44.41 |
TANGO-Tucker [2021] | 24.36 | 15.12 | 27.15 | 43.07 | 41.82 | 31.10 | 47.55 | 62.19 | 24.36 | 15.12 | 27.15 | 43.07 |
RE-GCN [2021] | 41.25 | 30.46 | 46.26 | 62.05 | 45.61 | 34.43 | 51.85 | 66.64 | 30.55 | 20.00 | 34.73 | 51.46 |
xERTE [2021] | 32.23 | 24.29 | 24.29 | 24.29 | 38.07 | 28.45 | 43.92 | 57.62 | 27.98 | 19.26 | 32.43 | 46.00 |
GHT [2022] | 37.40 | 27.77 | 41.66 | 56.19 | 41.50 | 30.79 | 46.85 | 62.73 | 27.40 | 18.08 | 30.76 | 45.76 |
rGalT [2022] | 38.33 | 28.57 | 42.86 | 58.13 | 38.89 | 27.58 | 44.19 | 59.10 | 27.88 | 18.01 | 31.59 | 47.02 |
PPT [2023] | 38.42 | 28.94 | 42.50 | 57.01 | 38.85 | 28.57 | 43.35 | 58.63 | 26.63 | 16.94 | 30.64 | 45.43 |
CENET [2023] | 39.02 | 29.62 | 43.23 | 57.49 | 41.95 | 32.17 | 46.93 | 60.43 | 27.85 | 18.15 | 31.63 | 46.98 |
ERSP [2024] | 42.65 | 31.88 | 47.99 | 63.64 | 47.10 | 35.68 | 53.42 | 68.70 | 31.17 | 20.45 | 35.39 | 52.39 |
TCCGN | 42.46 | 31.63 | 47.90 | 63.51 | 47.33 | 35.89 | 53.83 | 68.79 | 31.23 | 20.63 | 35.48 | 52.05 |
Table 6.
Performance (in percentage) of the entity prediction task with YAGO and GDELT. The best results are highlighted in bold.
Model | GDELT | YAGO |
---|
MRR
| Hits@1
| Hits@3
| Hits@10
| MRR
| Hits@3
| Hits@10
|
---|
DisMult [2015] | 8.61 | 3.91 | 8.27 | 17.04 | 44.05 | 49.70 | 59.94 |
ComplEx [2014] | 9.84 | 5.17 | 9.58 | 18.23 | 44.09 | 49.57 | 59.64 |
R-GCN [2022] | 12.17 | 7.40 | 12.37 | 20.63 | 24.25 | 24.01 | 37.30 |
ConvE [2018] | 18.37 | 11.29 | 19.36 | 32.13 | 41.22 | 47.03 | 59.90 |
Conv-TransE [2018] | 19.07 | 11.85 | 20.32 | 34.13 | 46.67 | 52.22 | 65.52 |
RotatE [2019] | 3.62 | 0.52 | 2.26 | 8.37 | 42.08 | 46.77 | 59.39 |
HyTE [2018] | 6.69 | 0.01 | 7.57 | 19.06 | 14.42 | 39.73 | 46.98 |
TTransE [2018] | 5.53 | 0.46 | 4.97 | 15.37 | 26.10 | 36.28 | 47.73 |
TA-DistMult [2018] | 10.34 | 4.44 | 10.44 | 21.63 | 44.98 | 50.64 | 61.11 |
RGCRN [2018] | 18.63 | 11.53 | 19.80 | 32.42 | 43.71 | 48.53 | 56.98 |
CyGNet [2021] | 18.05 | 11.13 | 19.11 | 31.50 | 46.72 | 52.48 | 61.52 |
RE-NET [2020] | 19.60 | 12.03 | 20.56 | 33.89 | 46.81 | 52.71 | 61.93 |
TANGO-DistMult [2021] | — | — | — | — | 49.49 | 55.42 | 63.74 |
TANGO-Tucker [2021] | — | — | — | — | 49.31 | 55.12 | 63.73 |
RE-GCN [2021] | 19.31 | 11.99 | 20.61 | 33.59 | 62.50 | 70.24 | 81.55 |
rGalT [2022] | 19.56 | 12.11 | 20.89 | 34.15 | 51.45 | 57.76 | 68.31 |
RE-GAT [2023] | 19.11 | 11.80 | 20.44 | 33.34 | — | — | — |
TCCGN | 19.63 | 12.23 | 20.95 | 34.07 | 63.61 | 72.11 | 83.53 |
Table 7.
Performance (in percentage) of the relation-prediction task with ICEWS18, ICEWS14, ICEWS05-15, YAGO, and GDELT. The best results are highlighted in bold.
Model | ICE18 | ICE14 | ICE05-15 | YAGO | GDELT |
---|
ConvE 1 | 37.73 | 38.80 | 37.89 | 91.33 | 18.84 |
ConvTransE 1 | 38.00 | 38.40 | 38.26 | 90.98 | 18.97 |
RGCRN 1 | 37.14 | 38.04 | 38.37 | 90.18 | 18.58 |
RE-GCN 1 | 40.53 | 41.06 | 40.63 | 93.85 | 19.22 |
TCCGN * | 41.24 | 41.63 | 41.14 | 93.93 | 19.44 |
Table 8.
Average per-epoch training time (GPU-hours) on five datasets.
Model | ICEWS14 | ICEWS05-15 | ICEWS18 | GDELT | YAGO |
---|
RE-GCN | 0.0133 | 0.1850 | 0.0169 | 0.1910 | 0.0028 |
TiRGN | 0.0283 | 0.5670 | 0.1890 | 0.4340 | 0.0325 |
TCCGN | 0.0089 | 0.1500 | 0.0156 | 0.1180 | 0.0022 |
Table 9.
Performance comparison of different models on various datasets.
Model | ICE14 | ICE05-15 | ICE18 | YAGO | GDELT |
---|
CD | 41.66 | 46.33 | 30.97 | 63.12 | 19.23 |
DDCL | 41.93 | 47.10 | 31.08 | 63.35 | 19.36 |
GSDF | 41.90 | 46.95 | 30.90 | 63.25 | 19.29 |
CD + GSDF | 42.15 | 47.36 | 31.10 | 63.40 | 19.38 |
GSDF + DDCL | 42.10 | 46.91 | 30.88 | 63.42 | 19.41 |
CD + DDCL | 42.30 | 47.20 | 31.00 | 63.50 | 19.45 |
TCCGN | 42.46 | 47.33 | 31.23 | 63.61 | 19.63 |
Table 10.
Estimated contribution ratio of static and dynamic features across datasets (%).
Dataset | Static Contribution | Dynamic Contribution |
---|
ICEWS14 | 46.0% | 54.0% |
ICEWS05-15 | 45.2% | 54.8% |
ICEWS18 | 45.7% | 54.3% |
GDELT | 31.0% | 69.0% |
YAGO | 59.4% | 40.6% |
Table 11.
Ablation results on five datasets. We compare the full model (A) with its ablated variants. Metrics include MRR(%) and Hits@K(%) (K = 1, 3, 10).
ICEWS14 | MRR | Hits@1 | Hits@3 | Hits@10 |
A: Full TCCGN | 42.46 | 31.63 | 47.90 | 63.51 |
B: w/o causal | 41.85 | 31.02 | 47.09 | 62.77 |
C: w/o adv. loss | 41.25 | 30.45 | 46.22 | 62.47 |
D: w/o both | 41.13 | 30.36 | 46.44 | 61.86 |
ICEWS05-15 | MRR | Hits@1 | Hits@3 | Hits@10 |
A: Full TCCGN | 47.33 | 35.89 | 53.83 | 68.79 |
B: w/o causal | 46.82 | 35.42 | 53.33 | 68.36 |
C: w/o adv. loss | 46.67 | 35.32 | 53.08 | 68.14 |
D: w/o both | 46.43 | 35.05 | 52.95 | 67.83 |
ICEWS18 | MRR | Hits@1 | Hits@3 | Hits@10 |
A: Full TCCGN | 31.23 | 20.63 | 35.48 | 52.05 |
B: w/o causal | 31.16 | 20.43 | 35.64 | 52.22 |
C: w/o adv. loss | 30.86 | 20.18 | 35.12 | 51.94 |
D: w/o both | 30.66 | 19.96 | 35.07 | 51.67 |
GDELT | MRR | Hits@1 | Hits@3 | Hits@10 |
A: Full TCCGN | 19.63 | 12.23 | 20.95 | 34.07 |
B: w/o causal | 19.46 | 12.06 | 20.82 | 33.89 |
C: w/o adv. loss | 19.40 | 12.03 | 20.76 | 33.77 |
D: w/o both | 19.35 | 12.01 | 20.70 | 33.72 |
YAGO | MRR | Hits@1 | Hits@3 | Hits@10 |
A: Full TCCGN | 63.61 | 52.08 | 72.11 | 83.53 |
B: w/o causal | 63.22 | 51.82 | 71.51 | 82.83 |
C: w/o adv. loss | 63.17 | 51.75 | 71.46 | 82.76 |
D: w/o both | 62.93 | 51.86 | 70.80 | 82.13 |
Table 12.
MRR (%) on five datasets under different and settings. Best results per dataset are in bold.
| | ICEWS14 | ICEWS05-15 | ICEWS18 | GDELT | YAGO |
---|
0.5 | 0.05 | 40.25 | 44.17 | 28.51 | 18.44 | 61.17 |
0.5 | 0.10 | 41.38 | 44.95 | 29.26 | 18.73 | 62.13 |
0.5 | 0.15 | 40.57 | 43.88 | 28.95 | 18.36 | 61.88 |
0.6 | 0.05 | 41.81 | 45.77 | 29.95 | 18.90 | 62.84 |
0.6 | 0.10 | 42.10 | 46.21 | 30.45 | 19.31 | 63.09 |
0.6 | 0.15 | 41.63 | 45.36 | 30.17 | 18.94 | 62.52 |
0.7 | 0.05 | 42.12 | 46.90 | 30.98 | 19.41 | 63.29 |
0.7 | 0.10 | 42.46 | 47.33 | 31.23 | 19.63 | 63.61 |
0.7 | 0.15 | 41.92 | 46.41 | 30.81 | 19.29 | 63.02 |
0.8 | 0.05 | 41.66 | 46.48 | 30.59 | 19.18 | 63.10 |
0.8 | 0.10 | 41.84 | 46.52 | 30.71 | 19.26 | 62.94 |
0.8 | 0.15 | 41.42 | 46.01 | 30.45 | 18.95 | 62.41 |
Table 13.
Examples of representative failure cases in temporal reasoning.
Failure Type | Input Event Context | Incorrect Prediction |
---|
Ambiguous Roles | (US, threatens, Iran) at t; (Iran, protests, US) at | Predicted: Iran threatens US |
Rare Entity–Relation Pair | (Kenya, signs agreement, Denmark)—only 1 prior occurrence | Predicted: Kenya protests Denmark |
Noisy History | Input history includes conflicting event: (France, supports, Mali) vs. (France, sanctions, Mali) | Prediction oscillates between “supports” and “sanctions” |
Time Shifted Effect | (Russia, annexes, Crimea) appears at ; no immediate follow-up | Model fails to propagate effect to t |
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).