Next Article in Journal
Dynamic MAML with Efficient Multi-Scale Attention for Cross-Load Few-Shot Bearing Fault Diagnosis
Previous Article in Journal
Cross-Domain OTFS Detection via Delay–Doppler Decoupling: Reduced-Complexity Design and Performance Analysis
Previous Article in Special Issue
Dual-Gated Graph Convolutional Recurrent Unit with Integrated Graph Learning (DG3L): A Novel Recurrent Network Architecture with Dynamic Graph Learning for Spatio-Temporal Predictions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CasDacGCN: A Dynamic Attention-Calibrated Graph Convolutional Network for Information Popularity Prediction

1
School of Computer Science and Technology, Kashi University, Kashi 844000, China
2
School of Computer and Information Engineering, Institute for Artificial Intelligence, Shanghai Polytechnic University, Shanghai 201209, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(10), 1064; https://doi.org/10.3390/e27101064
Submission received: 21 July 2025 / Revised: 8 October 2025 / Accepted: 8 October 2025 / Published: 14 October 2025

Abstract

Information popularity prediction is a critical problem in social network analysis. With the increasing prevalence of social platforms, accurate prediction of the diffusion process has become increasingly important. Existing methods mainly rely on graph neural networks to model structural relationships, but they are often insufficient in capturing the complex interplay between temporal evolution and local cascade structures, especially in real-world scenarios involving sparse or rapidly changing cascades. To address this issue, we propose the Cascading Dynamic attention-calibrated Graph Convolutional Network, named CasDacGCN. It enhances prediction performance through spatiotemporal feature fusion and adaptive representation learning. The model integrates snapshot-level local encoding, global temporal modeling, cross-attention mechanisms, and a hypernetwork-based sample-wise calibration strategy, enabling flexible modeling of multi-scale diffusion patterns. Results from experiments demonstrate that the proposed model consistently surpasses existing approaches on two real-world datasets, validating its effectiveness in popularity prediction tasks.

1. Introduction

In recent years, with the rising impact of social media platforms like Twitter, Weibo, and Facebook, people increasingly use social media to express opinions and share daily experiences. As a result, online public opinion has become a critical factor in social governance and stability. Understanding the diffusion patterns of online content and predicting its future popularity has aroused considerable attention since its wide applications in public management, commercial decisions, and public safety [1,2,3,4,5].
Nevertheless, owing to the public nature of social networks and the massive scale of user interactions, the complex relationships and noisy content present in social media make popularity prediction highly challenging. An information cascade represents the spread of content among users via follow and repost behaviors. A single user’s action may significantly influence the final popularity of a post, making cascade prediction inherently difficult. The goal of cascade popularity prediction is to estimate the future trend derived from observed diffusion history.
One approach involves modeling the dynamic process of information diffusion over networks. If the diffusion pattern can be summarized by a simple model with a few parameters and the underlying network topology is known, the model can be fitted using observed data to predict future trends. However, the complex link structures involved in real-world propagation processes are often beyond the capacity of such simplified models to capture. Generative models, such as Hawkes processes and their variants, have been widely used to model self-exciting diffusion dynamics [6,7,8,9], but they usually assume homogeneous triggering effects and often underperform when cascades exhibit heterogeneous temporal bursts. Another approach adopts traditional machine learning techniques, which predict future trends by feeding selected key features into a predictor [10]. Nevertheless, this method suffers from significant uncertainty and subjectivity in both feature selection and quantification [11].
With the advancement of deep learning, end-to-end models have emerged to bypass manual feature engineering. Many recent works utilize network structures and temporal dynamics as input, as these modalities are platform-independent and generalizable. A growing number of studies combine graph convolutional networks (GCN) for structural encoding [12,13,14,15,16] with RNN for temporal modeling. Despite these advances, most existing studies evaluate on limited observation windows (e.g., 1–3 h on Weibo), rely on fixed GCN aggregation that ignores directional asymmetry in cascades, and seldom address the long-tail phenomenon where the majority of cascades are small and sparse. GCN often employ entropy-based loss functions, such as cross-entropy, to minimize prediction errors during training by measuring the divergence between predicted and true node labels. Models such as CasCN [17] and AECasN [18,19] further explore spatiotemporal fusion through subgraph sequences and autoencoders [20].
Although the end-to-end models show strong performance, they still face several key challenges: First, there is a lack of deep integration between dynamic temporal sequences and local structural evolution. Second, cross-level interaction is insufficient. Third, generalization to sparse, long-tail cascades is limited. Fixed-parameter models are prone to noise and fail to adapt to high variance across samples.
To overcome these limitations, we propose the Cascading Dynamic attention-calibrated Graph Convolutional Network, named CasDacGCN. The model incorporates a bidirectional cross-attention module to align local structural cues with global temporal evolution in latent space, effectively mitigating the decoupling of spatial and temporal signals. Furthermore, we introduce a hypernetwork-based calibration mechanism that dynamically adjusts feature representations at the sample level, enhancing robustness against sparsity and long-range degradation during long-term prediction.
  • We propose CasDacGCN, a dynamic attention-calibrated graph convolutional network for information popularity prediction. Our model introduces the Bidirectional Cross-Attention Fusion (BCAF) module, which dynamically integrates snapshot-level local structural features with global temporal context, mitigating the decoupling of spatial and temporal signals, and thereby better capturing the evolution of information cascades.
  • We propose a Hypernetwork Calibration Module to address the challenge of sparse and long-tail cascades. It employs a lightweight feedforward network to generate sample-specific scaling coefficients, and when combined with the GCN–GRU backbone, this adaptive calibration ensures robustness across diverse cascade densities, enabling the model to generalize better than existing spatiotemporal approaches.
  • Experimental results on the Weibo and DBLP datasets validate the effectiveness of our approach, demonstrating that the CasDacGCN model outperforms existing baseline methods in information propagation prediction tasks. Ablation studies further confirm the critical roles of the Bidirectional Cross-Attention Fusion (BCAF) and Hypernetwork Calibration (HFCM) modules in improving model performance.
The remainder of this paper is structured as follows. We provide a review of related works in Section 2, introduce our proposed method in Section 3, describe datasets and experimental settings in Section 4, report results in Section 5, and conclude in Section 6.

2. Related Work

2.1. Traditional Methods

Traditional approaches to popularity prediction fall into the category of feature-based methods and generative models. Feature-based methods extract signals from historical data—such as content attributes, user profiles, structural topology, and temporal dynamics—and feed them into machine learning models like regression or classification [21,22,23,24]. However, because these methods rely heavily on handcrafted feature design, they often fail to capture the complexity of large-scale and dynamic networks, which limits their generalization ability.
For content features, studies have shown that text length, hashtag patterns, and semantic relevance significantly influence repost behavior. Non-textual signals like hyperlinks, emojis, and visual features also affect content popularity. User features, such as follower count, user activeness, and social influence, have been widely utilized. Structural features describe the cascade’s network topology—depth, breadth, and early-stage shape—all of which correlate with final popularity. Temporal features capture diffusion dynamics, with many works identifying temporal patterns, burst intensity, and time-aware embeddings as key factors.
Generative models treat cascade diffusion as a temporal point process. Poisson processes assume independent events with fixed rates, while Hawkes processes incorporate self-excitation, modeling how past events influence future ones. Applications of Poisson-based models include early-stage cascade modeling and decay-aware variants. Hawkes-based models extend to time-slice prediction, brand popularity, hybrid epidemiological frameworks, and even neural implementations like Hawkesformer [25]. Despite their mathematical elegance, these models often rely on strong assumptions and struggle in complex real-world environments.

2.2. Deep Learning-Based Prediction Methods

The rise of deep learning has transformed popularity prediction by enabling end-to-end spatiotemporal modeling. Recurrent models such as RNNs and GRUs excel at capturing temporal dependencies, while GNNs automatically learn structural patterns from cascade graphs, overcoming the limitations of manual feature design. Entropy can quantify the uncertainty in GCN outputs, aiding in tasks like confidence estimation or structural analysis of graphs for improved decision-making. Notable models include DeepCas [26], which generates random walk sequences and encodes them via Bi-GRU with attention, and DeepHawkes [27], which integrates RNNs with Hawkes processes to jointly model event triggers.
To address incomplete topologies and long-range dependencies, hybrid architectures have been proposed. MINDS [28] optimizes multi-scale diffusion prediction through serialized hypergraphs and adversarial learning. Casformer [29] uses adaptive cascade sampling and graph transformers to predict the popularity of information propagation in social networks. MUCas [30] integrates multi-scale capsule networks and influence-aware attention, while HeDAN [31] constructs heterogeneous diffusion graphs to capture complex multi-relation influence. Other approaches incorporate dynamic routing [32] for efficient temporal encoding.
Recent advancements have also leveraged Transformer-enhanced Hawkes processes [25], autoencoder-based structural learning [22], temporal convolutional networks [33], and degree-distribution-aware deep neural networks to further enhance cascade prediction. These models demonstrate state-of-the-art performance, though challenges remain in handling sparse or incomplete network structures. We also note a broader trend across learning systems: multi-scale feature fusion and adaptive alignment between local and global signals can enhance robustness in disparate domains (e.g., AI-generated image detection) [34], echoing our use of BCAF for local–global alignment and HFCM for sample-level calibration.
In summary, existing methods are limited by handcrafted feature dependence, oversimplified generative assumptions, or insufficient structural–temporal integration. In contrast, CasDacGCN overcomes these challenges by integrating a direction-aware GCN with BiGRU for joint structural and temporal modeling, introducing the BCAF module to align local and global signals, and employing HFCM to adaptively calibrate features at the sample level, thereby achieving improved robustness and generalization.

3. Preliminaries

The task of modeling information cascade prediction is relatively challenging. To address this, cascade graphs are commonly adopted, as they retain adequate topological characteristics required for prediction. For clarity, this section begins with definitions and explanations of relevant concepts before introducing the proposed model. The relevant symbols are shown in Table 1.
Definition 1
(Cascade Graph).  Let G = ( V , E ) represent a static social network, where V denotes the set of all nodes (users) and E denotes the set of edges. For each message i disseminated within the network, an information cascade is formed. The corresponding cascade graph is defined as the subgraph C i = ( V i , E i ) , where V i refers to the set of nodes that the message passes through, and E i denotes the edges connecting these nodes. Since message diffusion is time-dependent, we further define C i T = ( V i T , E i T ) as the cascade graph of message i observed within a time window T from the beginning of its propagation.
Definition 2
(Activation State). A node is considered activated once it retweets a message, which is equivalent to stating that the message has reached this node. Conversely, a node is inactive if it has not been activated. At time t, if a message passes through a node, its state is assigned a value of 1; otherwise, it is 0. The state of a node is denoted as  X i T ( t ) , representing its status within the cascade subgraph  C i T  at time t, where T refers to the observation window.
Definition 3
(Cascade Snapshot). As illustrated in Figure 1, only one node is marked in orange, indicating that the message has just been broadcast. All other nodes have state vectors  X i T ( t )  = 0. In the middle diagram, three additional nodes are marked in orange besides the initial sender, showing that three users have forwarded the message. The cascade subgraph   C i T  at this moment consists of these four nodes, each with a state of 1, while the remaining nodes are 0. In the bottom diagram, all nodes are involved in forwarding, so the cascade subgraph becomes the entire graph and the state vectors of all nodes are 1. By combining the cascade subgraph  C i T  with the node states  X i T , we obtain the snapshot  S i T = V i T , E i T , X i T ( t )  , which represents both the structure of the cascade graph and the node states at time t.
Definition 4
(Popularity Prediction). Our focus lies in the cascade scale, which is characterized by the count of retweets or, equivalently, by the number of nodes (users) that a message reaches. In particular, we estimate the expansion size Δ R i T p = R i T + T p R i T of a cascade after an observation window T, as illustrated in Figure 2.

4. The Model

We propose CasDacGCN, as an end-to-end framework for popularity prediction. As illustrated in Figure 3, the model consists of four interdependent modules. (a) A direction-aware GCN encoder extracts snapshot-level structural features by aggregating incoming and outgoing node information separately, capturing role asymmetry in diffusion. (b) A bidirectional GRU encodes the sequence of snapshots to model long-term temporal dependencies. (c) A cross-attention fusion module adaptively combines structural and temporal representations using a learned gating mechanism. (d) A hypernetwork-based calibrator generates sample-specific scaling factors to enhance robustness under sparse conditions. (e) Prediction layer with an MLP for log-scale popularity prediction.

4.1. Snapshot-Level Local Feature Extraction

The dynamic characteristics of information diffusion require the model to accurately capture the local structural evolution within each time window. To achieve this, this paper employs directed graph convolutional networks (GCNs) and global pooling operations to extract snapshot-level local features from two dimensions: directional propagation and multi-order neighborhood aggregation.
First, the snapshot partitioning and node state construction process divides the information propagation process into T consecutive snapshots { G 1 , G 2 , , G T } over time, where each snapshot G t characterizes the propagation state within time window [ t Δ t , t ] . Specifically, a node state matrix X R N × 1 is first constructed, where N is a predefined fixed node count covering all potentially participating nodes. If node v i is activated within the window, its state is set to X i , 1 = 1 ; otherwise X i , 1 = 0 . To unify node counts across snapshots, zero-padding is applied to newly added nodes, ensuring matrix dimensional consistency. For example, if a snapshot has M < N actually activated nodes, the remaining N M rows are filled with zero vectors.
Next, we construct the directed adjacency matrices A o u t and A i n to represent the outgoing and incoming edges, respectively. Both of these matrices are N × N square matrices, representing the directional information flow between nodes. Specifically, the out-edge adjacency matrix A o u t [ i , j ] = 1 if there is an edge v i v j , otherwise 0, and the in-edge adjacency matrix A i n [ i , j ] = 1 if there is an edge v j v i , otherwise 0. Assuming snapshot G t contains a propagation path v 1 v 2 v 3 , the matrices are expressed as
A o u t = 0 1 0 0 0 1 0 0 0 , A i n = 0 0 0 1 0 0 0 1 0
Traditional GCNs aggregate neighboring information symmetrically and fail to distinguish between propagation directions, which can lead to confusion regarding node roles. To address this, we employ a two-layer directed GCN that separately processes the out-edge and in-edge features. The first-layer GCN extracts low-order structural features, and its out-edge feature aggregation is given by
H out = σ A ˜ out X W out ( 0 )
where W out ( 0 ) R d × 1 is the out-edge weight matrix, A ˜ out is the normalized adjacency matrix and σ represents the ReLU activation function. The in-edge feature aggregation is expressed as
H in = σ A ˜ in X W in ( 0 ) .
Here, W in ( 0 ) R d × 1 is the in-edge weight matrix, and A ˜ in is the normalized in-edge adjacency matrix. The resulting feature matrices H out and H in both have dimensions N × d , where d is the feature dimension of each node.
To preserve directional information and avoid overfitting, the out-edge and in-edge features are fused by simple addition to form the fused feature representation H 1 :
H 1 = H o u t + H i n
where H 1 has dimensions N × d . The second-layer GCN further extracts high-order structural features by aggregating multi-hop neighbor information based on the first-layer features. Specifically, the second-layer GCN processes H 1 , resulting in high-order structural features H 2 :
H 2 = σ W · ( A o u t + A i n , H 1 )
where W R d × d is the second-layer GCN weight matrix, and A out + A in represents the fused out-edge and in-edge adjacency matrices. The aggregation captures the influence of a node’s multi-hop neighbors. Finally, a global mean pooling operation is applied to transform the node-level features H 2 R N × d into a snapshot-level representation H t :
H t = 1 N i = 1 N H 2 [ i , : ] R d
At this stage, H t R d , which is the global feature vector for the snapshot. By compressing the spatial dimension, it preserves the statistical characteristics of node connection patterns within the snapshot while suppressing interference from noisy nodes.

4.2. Global Temporal Modeling

To comprehensively capture the dynamic characteristics of information propagation, we employ a global temporal modeling module over snapshot features. Specifically, we adopt a standard GRU [35] to model the hidden state sequence { s 1 , s 2 , , s T } from inputs { h 1 , h 2 , , h T } , where h t is the snapshot-level representation from Section 4.1. The GRU updates are given by
z t = σ W z h t + U z s t 1 + b z ,
r t = σ W r h t + U r s t 1 + b r ,
s ˜ t = tanh W h t + U r t s t 1 + b ,
s t = ( 1 z t ) s t 1 + z t s ˜ t .
where s t R d is the hidden state at time step t, σ denotes the sigmoid function, ⊙ represents element-wise multiplication, W z , U z , W r , U r , W , U are trainable weight matrices, and b z , b r , b denote bias vectors. After processing the entire sequence through the GRU, the hidden state s t at the final time step is taken as the global temporal representation of the cascade propagation trajectory. To align with the dimensionality of local features, a linear projection layer is added:
T g l o b a l = W p r o j s T + b p r o j
where W p r o j and b p r o j are the learnable parameters of the projection layer. Finally, T g l o b a l R d represents the context vector integrating global temporal dynamics.

4.3. Bidirectional Cross-Attention Fusion

The snapshot-level local feature h t emphasizes the propagation structure within the current time window, while the global temporal feature T global reflects the overall evolution trend of the cascade. Direct concatenation or weighted summation may lead to insufficient sensitivity to critical time points due to fixed weights. Moreover, the importance of different snapshots varies significantly across the cascade’s lifecycle, necessitating dynamic adjustment of fusion weights between local and global features. To address this, this paper proposes a gated bidirectional cross-attention fusion mechanism, which achieves adaptive calibration of local features and global temporal context through a dynamic gating strategy.
The local snapshot feature h t and the global temporal context T global are concatenated on the feature axis to generate a joint representation vector u t R 2 d :
u t = [ h t | | T g l o b a l ]
The concatenation operation preserves the independent distribution information of local and global features, preventing feature confusion, while also providing richer input for gating coefficient computation and eliminating the impact of scale differences on weight learning.
A dynamic gating coefficient g t , quantifying the relative importance of local features in the current snapshot, is generated through a fully connected layer and a sigmoid function:
g t = σ ( w T u t + b )
where w R 2 d is the learnable weight vector, b is the bias term, and σ is the sigmoid function.
Based on the gating coefficient g t , a weighted fusion of the local feature h t and the global context T global is performed to generate the calibrated snapshot representation z t :
z t = g t · h t + ( 1 g t ) · T g l o b a l
Through end-to-end training, the global context T global can be indirectly refined by the local features { h t } t = 1 T , enabling bidirectional information interaction. Finally, the fused representations of all snapshots { z 1 , z 2 , , z T } are aggregated along the temporal dimension to generate the cascade’s global representation:
z = 1 T t = 1 T z t

4.4. Hypernetwork Calibration Module

Different samples exhibit significant variations in propagation patterns, making fixed-parameter models difficult to generalize. Directly using uncalibrated global feature representations may render the model sensitive to sparse data or noise. Motivated by this challenge, this paper presents a Hypernetwork-based Feature Calibration Module (HFCM), which dynamically generates sample-level calibration coefficients through a lightweight network to achieve adaptive adjustment at the feature dimension. The HFCM generates sample-specific calibration coefficients via a lightweight hypernetwork and applies them to rescale the fused representation. Intuitively, in sparse cascades with very few reposts, the calibration coefficients tend to emphasize global temporal information to mitigate the lack of structural signals; in dense cascades with many reposts, they place more weight on structural cues, allowing the model to fully exploit rich local interactions. For clarity of positioning, our hypernetwork is conditioned on the post-fusion cascade representation z (from the bidirectional cross-attention module), rather than on raw inputs or a single branch, and it outputs a dimension-wise bounded vector γ ( 0 , 1 ) d (instead of a single scalar gate), enabling fine-grained calibration under long-tail and sparse settings. The scaling is applied at the global representation stage—after fusion and before the MLP predictor—via Equation (17), without altering layer-internal normalizations; the module is trained end-to-end under the same MSLE objective as the predictor, adding no auxiliary losses. Instead of fixed parameters, the hypernetwork generates input-dependent calibration coefficients. Taking the cascade’s global representation z R d as input, it employs a single-layer fully connected network with a sigmoid activation function to produce dimension-wise calibration coefficients γ ( 0 , 1 ) d :
γ = σ W calib z + b calib
where W calib R d × d is the weight matrix, b calib R d is the bias term, and σ denotes the element-wise sigmoid function, ensuring γ i ( 0 , 1 ) for all elements in γ .
After obtaining the calibration coefficient vector, the module performs element-wise adjustment on the input holistic representation z , computing the calibrated representation z ca as
z cal = γ z
where ⊙ denotes element-wise multiplication. Through this operation, each dimensional feature value is scaled according to its corresponding calibration coefficient, enabling the model to adaptively amplify or suppress information across specific dimensions. This enhances both the robustness of overall feature representation and the predictive capability of the model.

4.5. Prediction Layer

After completing local snapshot feature extraction, global temporal modeling, bidirectional cross-attention fusion, and hypernetwork calibration, the model obtains the calibrated global cascade representation z cal R d , which integrates local structural patterns, temporal dynamics, and sample-specific calibration. To map this d-dimensional vector to the prediction target, we employ a multilayer perceptron (MLP) head f θ : R d R that outputs a scalar:
y ^ = f θ ( z cal ) R .
The cascade growth scale is defined as the increment of reposts Δ R within the prediction window T p , with the log-transformed target
y = log 2 Δ R + 1 , Δ R = | R T + T p R T | ,
where R t denotes the cumulative repost count within the observation interval ( 0 , t ] .
During training, model parameters are optimized by minimizing the Mean Squared Logarithmic Error (MSLE) over mini-batches. For a mini-batch , with y ^ i = f θ ( z cal , i ) and y i = log 2 ( Δ R i + 1 ) , the objective is
MSLE = 1 | | i y i y ^ i 2 ,
which is also used as the primary evaluation metric at test time.

5. Experiment

5.1. Experimental Setup and Datasets

The experimental environment of this model consists of a server running the Windows Server 2019 operating system, equipped with 128 GB RAM (manufactured by Kingston Technology, Fountain Valley, CA, USA) and an Intel Xeon Processor (Icelake) 2.59 GHz CPU (manufactured by Intel Corporation, Santa Clara, CA, USA).
In this study, we employ two datasets: the Weibo dataset for predicting repost cascades of Sina Weibo posts and the paper citation dataset for predicting academic paper citation cascades. Weibo: The dataset provided by Zhang et al. [36] was collected from the Sina Weibo platform and contains approximately 300,000 information cascade records, fully documenting the follow relationships between users and the repost chains of messages. Each repost behavior is defined as a directed edge, effectively characterizing the information propagation paths in the social network. DBLP: This dataset describes the structure of academic paper citation networks [37], constructing a directed network by parsing citation relationships between papers. Its propagation mechanism shares structural similarities with Weibo reposts.
For each sub-dataset, we adopt a fixed split ratio of 70% training, 10% validation, and 20% testing to ensure robust evaluation. Specifically, for the Weibo dataset (≈29,000 cascades) and the DBLP dataset (≈30,000 cascades), the corresponding splits are approximately 20,500 training, 3000 validation, and 5800 testing cascades each. As shown in Figure 4, the distributions of cascade popularity across train/validation/test splits are highly consistent, with overlapping curves and no significant distributional bias. The x-axis intervals [ 0 , 1 ) , [ 1 , 2 ) , represent discrete bins of cascade popularity, where popularity is defined as the final number of reposts (or activated nodes). For instance, [ 0 , 1 ) corresponds to cascades with fewer than one repost, [ 1 , 2 ) to those with one repost, and so forth. This histogram-style binning provides a clear view of how cascades are distributed across different popularity levels.
Table 2 summarizes the statistical properties of the datasets. The table data statistics display the total number of nodes and edges in both the Sina Weibo dataset and the DBLP paper citation dataset. Additionally, this study provides statistics on the number of cascades, average nodes, average edges, and average popularity under different observation windows across the entire dataset. These statistical metrics visually reflect the data distribution and further provide data support for model performance evaluation. For the Weibo dataset, the observation window T is set to 1/2/3 h, corresponding to prediction time windows T p of 23, 22, and 21 h, respectively. For the DBLP dataset, three temporal subsets are constructed, with observation window T set to 3/5/7 years, corresponding to prediction time windows T p of 7/5/3 years.

5.2. Parameter Settings and Baselines

The model employs a two-layer GCN with hidden dimension 64, which is consistent with the output feature size of snapshots. This design balances expressive capacity and computational efficiency. In addition, preliminary experiments explored multiple learning rates { 0.0001 , 0.0005 , 0.001 , 0.005 , 0.01 } , and the results confirm that 0.001 yields the best convergence speed and accuracy, which aligns with prior works adopting similar settings. The GRU consists of two layers, and the hidden state dimension is configured to match the output feature dimension of the GCN. The gated fusion linear layer in the bidirectional cross-attention fusion module maintains a hidden dimension of 64, consistent with the snapshot feature dimension. The hypernetwork calibration module adopts a single-layer lightweight fully connected network to reduce computational complexity. The learning rate of the model is set to 0.001.
This paper selects several classical baseline models for comparative analysis with the proposed model. The selected baseline methods are described in detail as follows:
  • Feature-linear [38]: A linear regression model is used to fit the cascade growth, with the learning rate set to 0.01.
  • Feature-deep [39]: A two-layer fully connected neural network is employed to capture complex features, with the number of hidden layers set to 3 and the learning rate set to 0.001.
  • DeepCas [26]: The first end-to-end deep learning model for cascade prediction.
  • DeepHawkes [27]: Converts the cascade graph into forwarding paths and integrates RNN with the self-exciting mechanism of Hawkes processes for cascade prediction.
  • CasCN [17]: Divides the cascade graph into subgraphs, learns subgraph representations using GCN, and models the structural evolution with LSTM.
  • AECasN [18]: Employs an autoencoder to learn deep representations and outputs the predicted cascade growth.
  • CasSeqGCN [32]: Generates subgraph sequences from cascade snapshots and learns structural and temporal features for prediction.
  • CasDO [40] integrates probabilistic diffusion models with temporal neural ODEs to model uncertainties and irregular dynamics in cascade evolution.
  • Casformer [29] introduces an adaptive cascade sampling strategy and a graph-based Transformer that jointly capture structural and temporal features of cascade graphs.
Based on existing research findings, This paper adopts the Mean Squared Logarithmic Error (MSLE) as the primary evaluation metric. This choice is motivated by the long-tail distribution and exponential growth of information cascades: while MSE and MAE are dominated by large cascades, MSLE reduces this bias through logarithmic scaling, thus enabling fairer evaluation across different cascade sizes. Moreover, this metric has been consistently adopted in prior cascade prediction studies such as DeepCas, CasFlow, CasSeqGCN, and AECasN, which ensures comparability of our results. It is formally defined as follows:
M S L E = 1 N j = 1 N l o g 2 Δ R ˜ i T p l o g 2 Δ R i T p 2
Among them, Δ R i T p represents the actual repost increment of the information within the time window, Δ R ˜ i T p denotes the predicted repost increment by the model, and d indicates the total number of information cascades in the dataset.

5.3. Performance Comparison

We first analyzed the model’s convergence speed, error variation process, and performance differences under varying observation time windows, further revealing its advantages in capturing cascading propagation characteristics. The CasDacGCN model demonstrated a robust convergence process across training, validation, and test sets. Under the conditions of a 1-h observation window for the Weibo dataset and a 7-year observation window for the DBLP dataset as shown in Figure 5, the model exhibited continuously declining error curves with significant advantages in convergence speed.
The experimental results of the CasDacGCN model and baseline methods on the Weibo and DBLP datasets are presented in Table 3 and Table 4. The experimental results demonstrate that CasDacGCN consistently outperforms all baseline methods across both datasets. On the Weibo social network, CasDacGCN achieves prediction errors of 1.985 (1 h), 1.894 (2 h), and 1.627 (3 h), yielding clear improvements over strong baselines such as CasSeqGCN (1.704 at 3 h) and Casformer (2.118 at 1 h). CasDacGCN reduces MSLE by 6.28% on Weibo-1h and 4.52% on Weibo-3h. On the DBLP academic network, CasDacGCN further demonstrates strong advantages in long-term prediction, reaching 0.517 at 7 years, which represents a 27.4% improvement over AECasN (0.712). These results verify that the bidirectional cross-attention fusion module (BCAF) effectively captures local propagation fluctuations, while the hypernetwork calibration module (HFCM) enhances the modeling of long-range sparse collaboration patterns. Moreover, when the observation time window is expanded, the predictive performance of all models improves accordingly. This improvement can be attributed to two factors: first, a larger window provides richer cascade information, enabling the directed GCN to capture second-order structural dependencies and the bidirectional GRU to extract long-range temporal patterns; second, longer propagation processes tend to stabilize, which helps establish more reliable relationships between popularity and propagation features. Through the integration of these mechanisms, CasDacGCN fully leverages additional cascade information under extended observation windows, thereby achieving more precise and robust predictions.

5.4. Ablation Study

To further validate the necessity and effectiveness of each module in the CasDacGCN model, comparative ablation experiments were conducted by removing different modules from CasDacGCN. Specifically, the bidirectional cross-attention fusion module (BCAF) and the hypernetwork feature calibration module (HFCM) were removed, resulting in CasDacGCN-BCAF and CasDacGCN-HFCM, respectively. The experimental setup for these comparative tests remained identical to that of the complete CasDacGCN model, using the same datasets. The experimental results are summarized in Table 5.
The experimental results demonstrate that both CasDacGCN-BCAF and CasDacGCN-HFCM models underperform the complete CasDacGCN model across both datasets, unequivocally confirming the core role of BCAF and HFCM modules. The experiments further reveal that the BCAF module is crucial for effectively capturing the interaction between local structural features and global temporal context. When removed, the model’s performance decreases significantly, particularly in long-term prediction scenarios. This highlights the importance of the dynamic alignment of local and global signals to enhance prediction accuracy. Further analysis indicates that the HFCM is designed to address the challenge of sparse and long-tail cascades by generating adaptive scaling factors for sample-specific calibration. The removal of this module leads to less stable predictions, especially when handling sparse cascades where the available information is limited. The HFCM improves model robustness and generalization by dynamically adjusting the feature representation based on the characteristics of each sample, particularly for cases with low-density cascades.
Furthermore, to conduct an in-depth analysis of the specific performance of each comparative model in the ablation experiments, this study performed a visual analysis of the training process under a 1-h observation window on the Weibo dataset. The results are shown in Figure 6. It can be observed that the two comparative models, CasDacGCN-BCAF and CasDacGCN-HFCM, exhibit similar performance during training, with each showing relative advantages at different training epochs. The figure displays the first 100 training epochs, clearly indicating that the complete CasDacGCN model not only significantly outperforms the others in terms of error convergence speed but also achieves the optimal final error value.

5.5. Parameter Analysis

During the training process, this study investigated the performance of the CasDacGCN model on the Weibo dataset under different learning rates. The learning rate was selected from the set [0.0001, 0.0005, 0.001, 0.005, 0.01]. Experimental results, as illustrated in Figure 7, demonstrate that the model achieves optimal performance when the learning rate is set to 0.001, exhibiting the fastest convergence speed and the lowest error level. In contrast, lower learning rates (e.g., 0.0001 or 0.0005), while providing better convergence stability, result in slower gradient descent, requiring more iterations to reach the desired error level. Conversely, higher learning rates (e.g., 0.005 or 0.01) initially show rapid error reduction but are prone to error rebound in later stages, indicating potential overshooting of the global optimum.

6. Conclusions

This paper proposes a Dynamic Cascade Attention Calibration Network (CasDacGCN), which significantly improves the accuracy and robustness of information popularity prediction through direction-aware graph modeling, dynamic spatiotemporal interaction, and sparse data calibration mechanisms. Experiments validate the model’s effectiveness in social network and academic citation scenarios, demonstrating superior performance over mainstream baseline methods and providing a novel solution for dynamic network modeling. In future work, we plan to explore multimodal data sources to enrich cascade representations. These include textual content (such as post text and comments), visual features (images accompanying posts), and user profile information. We also intend to investigate causal inference approaches, including counterfactual analysis and causality-regularized learning. The goal is to better disentangle true causal effects of user interactions from spurious correlations. These directions are expected to further enhance the interpretability and robustness of popularity prediction models.

Author Contributions

Conceptualization, B.Z. and B.L.; methodology, K.L. and S.N.; software implementation, Y.Z. and Z.Z.; validation and experimental setup, Y.Z., Z.Z. and H.L.; formal analysis and model evaluation, H.L.; investigation and experiment design, S.N.; resources and data acquisition, B.Z.; data curation, K.L.; writing—original draft preparation, Y.Z.; writing—review and editing, S.N. and Y.Z.; visualization, K.L.; supervision, B.L. and B.Z.; project administration, S.N.; funding acquisition, H.L. and B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Fundamental Research Funds for the Xinjing University under Grant No. XJEDU2024Z004, the Natural Science Foundation of Xinjiang Uygur Autonomous Region under Grant No. 2022D01A236, and the Shanghai Central Guide Local Science and Technology Development Fund Projects under Grant No. YDZX20253100004004005.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to express their sincere gratitude to colleagues and collaborators for their valuable discussions and constructive suggestions during the course of this research. The authors also acknowledge the Institute of Artificial Intelligence at Shanghai Polytechnic University for providing computational support, and they are deeply grateful to the anonymous reviewers for their insightful comments and suggestions, which have significantly improved the quality of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ta, T.A.N.; Li, K.; Yang, Y.; Jiao, F.; Tang, Z.; Li, G. Evaluating public anxiety for topic-based communities in social networks. IEEE Trans. Knowl. Data Eng. 2020, 34, 1191–1205. [Google Scholar] [CrossRef]
  2. Szabo, G.; Huberman, B.A. Predicting the popularity of online content. Commun. ACM 2010, 53, 80–88. [Google Scholar] [CrossRef]
  3. Centola, D. The spread of behavior in an online social network experiment. Science 2010, 329, 1194–1197. [Google Scholar] [CrossRef] [PubMed]
  4. Barabási, A. The origin of bursts and heavy tails in human dynamics. Nature 2005, 435, 207–211. [Google Scholar] [CrossRef] [PubMed]
  5. Yang, M.; Seklouli, A.S.; Zhang, H.; Ren, L.; Yu, X.; Ouzrout, Y. Popularity Prediction of Micro-Blog Hot Topics Based on Time-Series Data. In Proceedings of the 2023 15th International Conference on Software, Knowledge, Information Management and Applications (SKIMA), Kuala Lumpur, Malaysia, 8–10 December 2023; pp. 199–204. [Google Scholar]
  6. Bao, P.; Shen, H.W.; Jin, X.; Cheng, X.Q. Modeling and predicting popularity dynamics of microblogs using self-excited Hawkes processes. In Proceedings of the 24th International Conference on World Wide Web (WWW), Florence, Italy, 18–22 May 2015; pp. 9–10. [Google Scholar]
  7. Rizoiu, M.A.; Mishra, S.; Kong, Q.; Carman, M.; Xie, L. SIR-Hawkes: Linking epidemic models and Hawkes processes to model diffusions in finite populations. In Proceedings of the 2018 World Wide Web Conference (WWW), Lyon, France, 23–27 April 2018; pp. 419–428. [Google Scholar]
  8. Kong, Q.; Rizoiu, M.A.; Xie, L. Modeling information cascades with self-exciting processes via generalized epidemic models. In Proceedings of the 13th International Conference on Web Search and Data Mining (WSDM), Houston, TX, USA, 3–7 February 2020; pp. 286–294. [Google Scholar]
  9. Okawa, M.; Iwata, T.; Tanaka, Y.; Toda, H.; Kurashima, T.; Kashima, H. Dynamic Hawkes processes for discovering time-evolving communities’ states behind diffusion processes. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual, 14–18 August 2021; pp. 1276–1286. [Google Scholar]
  10. Cui, P.; Jin, S.; Yu, L.; Wang, F.; Zhu, W.; Yang, S. Cascading outbreak prediction in networks: A data-driven approach. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 901–909. [Google Scholar]
  11. Cai, Y.; Hu, W.; Liu, X.; Wang, H.; Zhang, X.; Wang, J. User Dual Intents Graph Modeling for Information Diffusion Prediction. In Proceedings of the 2024 IEEE 10th Conference on Big Data Security on Cloud (BigDataSecurity), New York, NY, USA, 10–12 May 2024; pp. 35–40. [Google Scholar]
  12. Jiao, P.; Song, W.; Wang, Y.; Zhang, W.; Chen, H.; Zhao, Z.; Wu, J. Cascade Popularity Prediction: A Multi-View Learning Approach with Socialized Modeling. IEEE Trans. Netw. Sci. Eng. 2025, 12, 1198–1209. [Google Scholar] [CrossRef]
  13. Jiao, P.; Yan, P.; Zhang, J.; Wang, B.; Zhang, W.; Zhao, N. Combine the Growth of Cascades and Impact of Users for Diffusion Prediction. IEEE Trans. Big Data 2025, 11, 887–895. [Google Scholar] [CrossRef]
  14. Xu, X.; Zhou, F.; Zhang, K.; Liu, S.; Trajcevski, G. CasFlow: Exploring hierarchical structures and propagation uncertainty for cascade prediction. IEEE Trans. Knowl. Data Eng. 2023, 35, 3484–3499. [Google Scholar] [CrossRef]
  15. Ye, J.; Bao, Q.; Xu, M.; Xu, J.; Qiu, H.; Jiao, P. RD-GCN: A Role-Based Dynamic Graph Convolutional Network for Information Diffusion Prediction. IEEE Trans. Netw. Sci. Eng. 2024, 11, 4923–4937. [Google Scholar] [CrossRef]
  16. Huang, S.; Zhang, G.; Liu, H.; Zhang, C.; Liu, J. CaSTGCN: Deep Learning Method for Information Cascade Prediction. In Proceedings of the 2024 China Automation Congress (CAC), Qingdao, China, 1–3 November 2024; pp. 2174–2179. [Google Scholar]
  17. Chen, X.; Zhao, Y.; Liu, G.; Sun, R.; Zhou, X.; Zheng, K. Efficient similarity-aware influence maximization in geo-social network. IEEE Trans. Knowl. Data Eng. 2020, 34, 4767–4780. [Google Scholar] [CrossRef]
  18. Feng, X.; Zhao, Q.; Li, Y. AECasN: An information cascade predictor by learning the structural representation of the whole cascade network with autoencoder. Expert Syst. Appl. 2022, 191, 116260. [Google Scholar] [CrossRef]
  19. Zhang, Z.; Xie, X.; Zhang, Y.; Zhang, L.; Jiang, Y. HierCas: Hierarchical Temporal Graph Attention Networks for Popularity Prediction in Information Cascades. In Proceedings of the 2024 International Joint Conference on Neural Networks (IJCNN), Yokohama, Japan, 30 June–5 July 2024; pp. 1–8. [Google Scholar]
  20. Wang, R.; Xu, X.; Zhang, Y. Multiscale Information Diffusion Prediction With Minimal Substitution Neural Network. IEEE Trans. Neural Netw. Learn. Syst. 2025, 36, 1069–1080. [Google Scholar] [CrossRef] [PubMed]
  21. Yang, J.; Leskovec, J. Patterns of temporal variation in online media. In Proceedings of the 4th ACM International Conference on Web Search and Data Mining (WSDM), Hong Kong, China, 9–12 February 2011; pp. 177–186. [Google Scholar]
  22. Gao, S.; Ma, J.; Chen, Z. Popularity prediction in microblogging network. In Proceedings of the Asia-Pacific Web Conference, Changsha, China, 5–7 September 2014; pp. 379–390. [Google Scholar]
  23. Kong, S.; Mei, Q.; Feng, L.; Ye, F.; Zhao, Z. Predicting bursts and popularity of hashtags in real-time. In Proceedings of the 37th ACM SIGIR International Conference on Research and Development in Information Retrieval, Gold Coast, Australia, 6–11 July 2014; pp. 927–930. [Google Scholar]
  24. Weng, L.; Menczer, F.; Ahn, Y.Y. Predicting successful memes using network and community structure. In Proceedings of the International AAAI Conference on Web and Social Media, Ann Arbor, MI, USA, 1–4 June 2014; Volume 8, pp. 535–544. [Google Scholar]
  25. Yu, L.; Xu, X.; Trajcevski, G.; Zhou, F. Transformer-enhanced Hawkes process with decoupling training for information cascade prediction. Knowl.-Based Syst. 2022, 255, 109740. [Google Scholar] [CrossRef]
  26. Li, C.; Ma, J.; Guo, X.; Mei, Q. DeepCas: An end-to-end predictor of information cascades. In Proceedings of the 26th International Conference on World Wide Web (WWW), Perth, Australia, 3–7 April 2017; pp. 577–586. [Google Scholar]
  27. Cao, Q.; Shen, H.; Cen, K.; Ouyang, J.; Cheng, X. DeepHawkes: Bridging the gap between prediction and understanding of information cascades. In Proceedings of the 2017 ACM Conference on Information and Knowledge Management (CIKM), Singapore, 6–10 November 2017; pp. 1149–1158. [Google Scholar]
  28. Jiao, P.; Chen, H.; Bao, Q.; Zhang, W.; Wu, H. Enhancing multi-scale diffusion prediction via sequential hypergraphs and adversarial learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 8571–8581. [Google Scholar]
  29. Wang, B.; Li, Z.; Xu, Z.; Zhang, J. Casformer: Information Popularity Prediction With Adaptive Cascade Sampling and Graph Transformer in Social Networks. IEEE Trans. Big Data 2025, 11, 1652–1663. [Google Scholar] [CrossRef]
  30. Feng, X.; Zhao, Q.; Zhu, R.J. Towards popularity prediction of information cascades via degree distribution and deep neural networks. J. Inf. 2023, 17, 101413. [Google Scholar] [CrossRef]
  31. Jia, X.; Shang, J.; Liu, D.; Zhang, H.; Ni, W. HeDAN: Heterogeneous diffusion attention network for popularity prediction of online content. Knowl.-Based Syst. 2022, 254, 109659. [Google Scholar] [CrossRef]
  32. Wang, Y.; Wang, X.; Ran, Y.; Michalski, R.; Jia, T. CasSeqGCN: Combining network structure and temporal sequence to predict information cascades. Expert Syst. Appl. 2022, 206, 117693. [Google Scholar] [CrossRef]
  33. Zhao, Q.; Zhang, Y.; Feng, X. Predicting information diffusion via deep temporal convolutional networks. Inf. Syst. 2022, 108, 102045. [Google Scholar] [CrossRef]
  34. Yao, L.; Niu, S.; Liao, K.; Zou, G.; Li, K.; Zhao, T. MSAFNet: Multi-Scale Self-Adaptive Feature Fusion Network for AI-Generated Image Detection. Meas. Sci. Technol. 2025, 36, 095402. [Google Scholar] [CrossRef]
  35. Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar]
  36. Zhang, J.; Liu, B.; Tang, J.; Chen, T.; Li, J. Social Influence Locality for Modeling Retweeting Behaviors. In Proceedings of the 23rd International Joint Conference on Artificial Intelligence (IJCAI), Macao, China, 19–25 August 2013; pp. 2761–2767. [Google Scholar]
  37. Tang, J.; Zhang, J.; Yao, L.; Li, J.; Zhang, L.; Su, Z. Arnetminer: Extraction and mining of academic social networks. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 990–998. [Google Scholar]
  38. Cheng, J.; Adamic, L.; Dow, P.A.; Kleinberg, J.M.; Leskovec, J. Can cascades be predicted? In Proceedings of the 23rd International Conference on World Wide Web (WWW), Seoul, Republic of Korea, 7–11 April 2014; pp. 925–936. [Google Scholar]
  39. Chen, X.; Zhou, F.; Zhang, K.; Trajcevski, G.; Zhong, T.; Zhang, F. Information diffusion prediction via recurrent cascades convolution. In Proceedings of the 2019 IEEE 35th International Conference on Data Engineering (ICDE), Macao, China, 8–11 April 2019; pp. 770–781. [Google Scholar]
  40. Cheng, Z.; Zhou, F.; Xu, X.; Zhang, K.; Trajcevski, G.; Zhong, T.; Yu, P.S. Information Cascade Popularity Prediction via Probabilistic Diffusion. IEEE Trans. Knowl. Data Eng. 2024, 36, 8541–8555. [Google Scholar] [CrossRef]
Figure 1. Illustration of snapshot diagrams at different time steps. Activated nodes are marked in orange, and non-activated nodes are in white.
Figure 1. Illustration of snapshot diagrams at different time steps. Activated nodes are marked in orange, and non-activated nodes are in white.
Entropy 27 01064 g001
Figure 2. A schematic representation of the information cascade.
Figure 2. A schematic representation of the information cascade.
Entropy 27 01064 g002
Figure 3. The framework of CasDacGCN.
Figure 3. The framework of CasDacGCN.
Entropy 27 01064 g003
Figure 4. This figure shows the popularity distribution of cascades in train/validation/test splits.
Figure 4. This figure shows the popularity distribution of cascades in train/validation/test splits.
Entropy 27 01064 g004
Figure 5. Change curve of MSLE values during the training process of the CasDacGCN model.
Figure 5. Change curve of MSLE values during the training process of the CasDacGCN model.
Entropy 27 01064 g005
Figure 6. Ablation study of CasDacGCN on the 1-h Weibo dataset with different module configurations. MSLE is used as the evaluation metric.
Figure 6. Ablation study of CasDacGCN on the 1-h Weibo dataset with different module configurations. MSLE is used as the evaluation metric.
Entropy 27 01064 g006
Figure 7. MSLE performance of CasDacGCN under different learning rates.
Figure 7. MSLE performance of CasDacGCN under different learning rates.
Entropy 27 01064 g007
Table 1. Summary of notations.
Table 1. Summary of notations.
SymbolMeaning
G = ( V , E ) Static social network graph
C i = ( V i , E i ) Cascade subgraph of message i
C i T = ( V i T , E i T ) Cascade graph observed within window T
X i T ( t ) State of node i at time t in window T
A o u t , A i n Out-edge and in-edge adjacency matrices
W , H Weight matrix; feature matrix
h t Local snapshot feature at time t
s t GRU hidden state at time t
T g l o b a l Global temporal representation (from GRU)
σ Sigmoid activation function
Element-wise multiplication
z Snapshot–temporal fused representation
γ Calibration coefficient (from hypernetwork)
z c a l Calibrated global representation
Δ R Cascade growth size within prediction window
M S L E Mean Squared Logarithmic Error
Table 2. Statistics of the datasets.
Table 2. Statistics of the datasets.
DatasetWeiboDBLP
Obs (T) 1 h 2 h 3 h 3 y 5 y 7 y
Cascades29.3k30.1k29.6k30.1k30.0k30.0k
Ave Nodes36.638.439.927.631.334.3
Ave Edges30.732.933.433.444.550.5
Ave Popularity184.6128.6101.630.016.98.5
Table 3. Experimental comparison between CasDacGCN and baseline approaches on Weibo datasets across different observation times evaluated by MSLE.
Table 3. Experimental comparison between CasDacGCN and baseline approaches on Weibo datasets across different observation times evaluated by MSLE.
MethodWeibo
1 h 2 h 3 h
FeatureLinear4.4684.1873.697
FeatureDeep4.4314.1973.733
DeepCas2.8742.6732.256
DeepHawkes2.9142.7862.313
CasCN2.7222.5742.203
AECasN2.8972.6462.241
CasSeqGCN2.1842.0751.704
CasDo2.1781.9261.782
Casformer2.1181.8641.767
CasDacGCN (Ours)1.9851.8941.627
Table 4. Experimental comparison between CasDacGCN and baseline approaches on DBLP datasets across different observation times evaluated by MSLE.
Table 4. Experimental comparison between CasDacGCN and baseline approaches on DBLP datasets across different observation times evaluated by MSLE.
MethodDBLP
3 Years 5 Years 7 Years
FeatureLinear3.4512.8252.035
FeatureDeep3.2732.4611.463
DeepCas1.8491.3380.948
DeepHawkes1.9561.5580.990
CasCN1.2371.1040.722
AECasN1.0420.8950.712
CasSeqGCN0.8840.7350.527
CasDacGCN (Ours)0.8060.6950.517
Table 5. Experimental comparison between CasDacGCN and its variants on two datasets across multiple observation times evaluated by MSLE.
Table 5. Experimental comparison between CasDacGCN and its variants on two datasets across multiple observation times evaluated by MSLE.
MethodWeiboDBLP
1 h 2 h 3 h 3 y 5 y 7 y
CasDacGCN-HFCM2.5952.4702.3151.6841.6681.515
CasDacGCN-BCAF2.7352.5752.3951.2351.7131.555
CasDacGCN1.9851.8941.6270.8060.6950.517
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, B.; Zhu, Y.; Zhang, Z.; Liao, K.; Niu, S.; Li, B.; Li, H. CasDacGCN: A Dynamic Attention-Calibrated Graph Convolutional Network for Information Popularity Prediction. Entropy 2025, 27, 1064. https://doi.org/10.3390/e27101064

AMA Style

Zhang B, Zhu Y, Zhang Z, Liao K, Niu S, Li B, Li H. CasDacGCN: A Dynamic Attention-Calibrated Graph Convolutional Network for Information Popularity Prediction. Entropy. 2025; 27(10):1064. https://doi.org/10.3390/e27101064

Chicago/Turabian Style

Zhang, Bofeng, Yanlin Zhu, Zhirong Zhang, Kaili Liao, Sen Niu, Bingchun Li, and Haiyan Li. 2025. "CasDacGCN: A Dynamic Attention-Calibrated Graph Convolutional Network for Information Popularity Prediction" Entropy 27, no. 10: 1064. https://doi.org/10.3390/e27101064

APA Style

Zhang, B., Zhu, Y., Zhang, Z., Liao, K., Niu, S., Li, B., & Li, H. (2025). CasDacGCN: A Dynamic Attention-Calibrated Graph Convolutional Network for Information Popularity Prediction. Entropy, 27(10), 1064. https://doi.org/10.3390/e27101064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop