Next Article in Journal
ICC-VulKG-TAER: Industrial Control Component Vulnerability Knowledge Graph-Based Target Attack Entity Reasoning
Previous Article in Journal
Intelligent Control and Optimization of Cooperative Transportation Between a Single Drone and an Autonomous Vehicle Under Dynamic Weather Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

USTGCN: A Unified Spatio-Temporal Graph Convolutional Network for Stock-Ranking Prediction

1
College of Information Science and Technology, Jinan University, Guangzhou 510632, China
2
Big Data Decision Institute, Jinan University, Guangzhou 510632, China
3
School of Medicine, Jinan University, Guangzhou 510632, China
4
School of Management, Jinan University, Guangzhou 510632, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2026, 15(6), 1317; https://doi.org/10.3390/electronics15061317
Submission received: 2 March 2026 / Revised: 17 March 2026 / Accepted: 20 March 2026 / Published: 21 March 2026

Abstract

Stock-ranking prediction is an important task in quantitative finance because it directly influences portfolio construction and alpha generation. Recent Graph Neural Network (GNN) models provide a promising way to describe inter-stock dependencies, but many existing methods still have difficulty balancing rapidly changing market interactions with relatively stable structural relationships. They are also easily affected by financial micro-structure noise. To address these issues, this paper proposes USTGCN, a Unified Spatio-Temporal Graph Convolutional Network for stock-ranking prediction. USTGCN adopts a dual-stream temporal encoder based on ALSTM and GRU to capture short-term dynamic patterns and longer-horizon structural information, respectively. We further introduce a rolling-window correlation smoothing strategy to build a more stable dynamic graph, and then integrate the dynamic and structural graph views through a shared fusion layer. Skip connections are used to preserve original temporal information during spatial aggregation. Experiments on the CSI100 and CSI300 benchmark datasets show that USTGCN achieves IC values of 0.141 and 0.154, respectively, and exhibits improved drawdown control during stressed market periods, indicating its practical value for quantitative trading.

1. Introduction

Stock-ranking prediction is a core task in quantitative finance because it supports data-driven portfolio construction and alpha discovery [1,2,3]. Rooted in the Efficient Market Hypothesis [4] and Modern Portfolio Theory [5], early studies often modeled financial assets independently and relied only on individual historical price series [6,7]. In real markets, however, stocks rarely evolve in isolation. Sector membership, supply-chain relations, and investor sentiment can all induce clear spatio-temporal dependencies among stocks [8,9,10]. Ignoring these links may therefore discard useful relational information.
This task remains difficult for two main reasons. First, traditional factor-based models usually rely on fixed-weight linear combinations of handcrafted alpha features, which adapt poorly to changing market regimes [11]. Second, financial time-series data are noisy and strongly non-stationary. Without suitable structural inductive biases, deep models can easily overfit transient market fluctuations [12]. Similar stability concerns have also been discussed in deep learning for other complex systems [13]. This motivates a framework that can capture persistent stock relations while remaining responsive to short-term changes.
The development of Graph Neural Networks (GNNs) [14,15] has provided a natural way to model stocks as nodes in a relational graph [16]. Spatio-temporal graph models have also shown strong performance in related domains [17,18], which motivates their use in financial prediction. Representative stock-oriented graph models include RSR [1], HIST [10], SDGNN [2], and DTSRN [3]. Among them, SDGNN jointly models static and dynamic graphs, but its static branch depends on stock ID embeddings, and its dynamic branch uses single-step correlations that are easily disturbed by temporary price noise. DTSRN avoids the ID-embedding issue, but discards static structural relations altogether. In addition, many existing approaches still rely on a single temporal encoder and use separate fusion mechanisms for different graph views, which increases architectural complexity and may weaken feature integration.
To address these limitations, we propose USTGCN, a Unified Spatio-Temporal Graph Convolutional Network for stock-ranking prediction. As shown in Figure 1, stock trajectories contain both rapidly changing co-movements and more persistent structural patterns. USTGCN models these temporal characteristics with complementary streams before graph construction and then fuses the resulting graph information through a unified fusion module. The main contribution is a task-oriented integration of temporal modeling, graph construction, and graph-view fusion for stock-ranking under noisy and evolving inter-stock relations.
Our primary contributions are summarized as follows:
  • We design a dual-stream temporal encoder that uses ALSTM to capture short-term volatility and GRU to summarize longer-term structural patterns.
  • We introduce a rolling-window correlation smoothing strategy to stabilize the dynamic graph and reduce the effect of transient micro-structure noise.
  • We build a shared fusion layer to aggregate dynamic and structural relations within one downstream prediction pipeline, while skip connections retain the original temporal features.

2. Related Work

2.1. Deep Temporal Models for Financial Time-Series

Early stock prediction studies relied heavily on classical econometric models such as ARIMA [19]. With the growth of deep learning, recurrent networks such as LSTM [20] and GRU [21] became common tools for modeling non-linear temporal dependence. These models were later improved by feature-fusion strategies [22], attention-based variants such as ALSTM [23], and more recent Transformer-based variants [24,25,26]. For a broader background on deep learning for financial time-series forecasting and recent graph-based stock forecasting literature, see [27,28]. Together, these studies form the temporal modeling background for this work.

2.2. Spatio-Temporal Graph Neural Networks

The success of GCNs [14] and GATs [29] also promoted the development of spatiotemporal GNNs, which first achieved strong results in applications such as traffic forecasting [17,30,31]. Models such as Graph WaveNet [18] and AGCRN [32] further showed that useful graph structures can be learned directly from data. In finance, treating stocks as interconnected nodes has improved predictive modeling [9,28]. Early graphbased approaches mainly relied on predefined relations, such as industry information [10] or explicit company links [16], while RSR [1] modeled relations through temporal response strength. However, purely static graphs are often not flexible enough for rapidly changing markets.

2.3. Dynamic Graph Learning and Temporal Fusion

Dynamic graph learning has therefore attracted increasing attention in evolving environments [33,34]. In finance, TRAN [35] introduces time-aware relational attention, while FinGAT [36] and HyperStockGAT [37] explore more expressive graph structures. Hybrid approaches have also attempted to combine static and dynamic views [38]. SDGNN [2] jointly learns both graphs but relies on noisy ID embeddings, whereas DTSRN [3] removes static relations altogether. Matsunaga et al. [9] further emphasized the usefulness of rolling windows in financial graph construction. These observations motivate the design choices adopted in USTGCN.

3. Preliminary

3.1. Problem Formulation

Let the investable stock universe be defined as a set of N stocks S = { s 1 , s 2 , , s N } . At any given trading day t, each stock s i is associated with a historical multidimensional time-series data X i t R T × F , where T = 60 denotes the defined lookback window, and F represents the number of observed daily trading features (e.g., open, close, high, low, volume, VWAP).
The fundamental objective of stock-ranking prediction is to accurately estimate the relative order of stocks based on their future profitability to facilitate algorithmic portfolio construction. Specifically, we define the target variable as the actual return of stock i at time t + 1 :
y i t + 1 = Close i t + 1 Close i t Close i t .
The core learning task aims to optimize a neural mapping function f that translates historical multidimensional data into predicted future returns:
y ^ t + 1 = f ( X t ) .
Ultimately, the equities are sorted in descending order based on y ^ t + 1 . Portfolios are formed by executing long positions on the top-ranked assets.

3.2. Graph Construction Paradigm

To mathematically represent inter-stock dependencies, we formalize the financial market as a set of nodes within two complementary graph topologies:
  • Dynamic Graph: Designed to encapsulate transient and highly reactive inter-stock correlations. It is dynamically generated via the cosine similarity of short-horizon latent representations extracted from recent temporal patterns.
  • Static Graph: Designed to capture relatively stable structural dependencies. It is derived from the cosine similarity of longer-horizon latent representations encoded via a separate recurrent network and is refreshed at each trading day, but typically evolves more smoothly than the dynamic graph.

4. Methodology

As illustrated in Figure 2, the operational pipeline of USTGCN comprises four synergistic modules: dual-stream temporal encoding, robust graph construction, unified spatio-temporal fusion, and ranking prediction.

4.1. Dual-Stream Temporal Encoding

Given the raw input tensor X t R N × T × F , where X t contains observed trading features rather than latent node embeddings, a monolithic sequential encoder inherently struggles to isolate high-frequency volatility from low-frequency structural trends. We therefore map the raw inputs into two learned latent representations with embedding dimension d = 128 :
  • Dynamic Features ( H dyn ): To capture transient market shocks and short-term sequential patterns, we deploy an Attention-augmented LSTM (ALSTM). The attention mechanism allows the network to assign varying importance weights to informative trading days, and attention-based LSTM models have been used in stock forecasting to highlight important time steps [23,39], which makes this branch well suited to emphasize abrupt short-horizon fluctuations:
    H dyn = ALSTM ( X t ) R N × d .
  • Static Features ( H stat ): Conversely, to extract robust, longer-horizon structural characteristics that are less sensitive to daily noise, we utilize a standard Gated Recurrent Unit (GRU). GRU-based recurrent modeling provides a compact gated summary over sequence histories [21], and GRU has also been adopted for relatively longer prediction windows in stock trading applications [40]. By extracting the final hidden state over the whole lookback window, we obtain a smoother historical summary for each asset:
    H stat = GRU ( X t ) [ : , 1 , : ] R N × d .
This ALSTM/GRU assignment is motivated by their different inductive biases. Attention-based LSTM models can emphasize informative time steps in stock sequences [39], whereas GRU-based modeling has also been used for relatively longer prediction windows in trading applications [40]. In our setting, ALSTM is therefore used for the dynamic stream, while GRU provides a comparatively smoother summary over the whole lookback window for the structural stream. Section 5 summarizes an encoder-assignment experiment showing that the default ALSTM/GRU pairing attains the strongest overall performance, whereas homogeneous encoder choices are consistently weaker, supporting the complementarity of the two streams.

4.2. Robust Graph Construction

The raw, instantaneous dynamic adjacency matrix at time t is computed via pairwise cosine similarity of the dynamic features:
A i j dyn , t = ( h i dyn ) h j dyn h i dyn h j dyn .
In high-frequency financial data, relying solely on A dyn , t introduces severe micro-structure noise, leading to erratic graph topologies. To mathematically enforce temporal consistency and stabilize the manifold, we apply a sliding rolling-window average over the preceding W = 20 days:
A ¯ dyn , t = 1 W τ = t W + 1 t A dyn , τ .
We adopt a simple moving average because it suppresses transient correlation spikes without introducing an additional decay hyperparameter. In parallel, the structural graph A stat , t is recomputed at each trading day from H stat . Because H stat summarizes longer-horizon behavior, A stat , t typically evolves more slowly than the dynamic graph and serves as a relatively stable relational anchor without relying on stock ID embeddings:
A i j stat , t = ( h i stat ) h j stat h i stat h j stat .
During both training and inference, the two graph views are refreshed for each trading day using only information available up to day t; the distinction lies in temporal sensitivity rather than whether the graph is updated. To ensure numerical stability during message passing, both adjacency matrices are subjected to row-wise SoftMax normalization, yielding the stochastic matrices Λ dyn , t and Λ stat , t .

4.3. Unified Spatio-Temporal Fusion

USTGCN uses a unified fusion module to combine information from the dynamic and structural graph views. We first project the temporally specialized embeddings across their corresponding graph topologies:
H ˜ dyn = Λ dyn , t H dyn ,
H ˜ stat = Λ stat , t H stat .
The propagated dynamic and structural representations are then integrated within a shared fusion stage through a linear transformation matrix W R 2 d × d :
H fused = [ H ˜ dyn H ˜ stat ] · W ,
where denotes the matrix concatenation operation.

4.4. Ranking Prediction

Deep graph convolutions inevitably risk over-smoothing, potentially erasing idiosyncratic stock signatures critical for precise ranking. To explicitly preserve the fidelity of original temporal signals, we establish a dense skip connection encompassing both initial feature spaces:
H skip = [ H dyn H stat ] .
The final alpha score predictions are generated by routing the concatenated fused and skip tensors through a deep 4-layer Multi-Layer Perceptron (MLP):
y ^ t + 1 = MLP [ H fused H skip ] ,
configured with an aggressive dimensionality reduction cascade ( 3 d d d / 2 d / 4 1 ) utilizing non-linear ReLU activations.
To specifically optimize for the relative ordering of assets rather than absolute scalar regression, we deploy the scale-invariant RankIC Loss:
L RankIC = i ( y ^ i y ^ ¯ ) ( y i y ¯ ) N · σ y ^ · σ y ,
where y ^ ¯ , y ¯ , and σ y ^ , σ y represent the sample mean and standard deviation of the predicted alpha scores and the ground truth returns, respectively. Minimizing this objective directly maximizes the Pearson correlation between predictions and true market outcomes in rank space.

5. Experiments

5.1. Datasets and Experimental Setup

We evaluate USTGCN on two Chinese A-share benchmark datasets, CSI100 and CSI300. Consistent with SDGNN and DTSRN, we use 2007–2014 for training, 2015–2016 for validation, and 2017–2020 for out-of-sample testing. Each stock is represented by six raw trading features, namely open, close, high, low, volume, and VWAP, over a lookback window of T = 60 .

5.2. Baseline Methods

We compare USTGCN with 11 representative baselines across three categories:
  • Traditional ML: MLP, SFM.
  • Deep Sequence Models: GRU, LSTM, ALSTM, Transformer, ALSTM+TRA.
  • Advanced Graph Models: GATs, HIST, SDGNN, DTSRN.

5.3. Evaluation Metrics

To rigorously quantify both the ranking accuracy and the practical trading viability of the models, we evaluate performance from two dimensions: predictive metrics and portfolio backtesting metrics.

5.3.1. Predictive Metrics

To assess the cross-sectional ranking capability of the models at each trading day t, we employ three standard predictive indicators. Let N denote the total number of valid stocks in the daily cross-section, while y i and y ^ i represent the true return and the model’s predicted alpha score for stock i, respectively (the time superscript t + 1 is omitted for brevity).
  • Information Coefficient (IC): Calculates the cross-sectional Pearson correlation coefficient between the predicted scores and the actual true returns. A higher IC indicates stronger linear predictive power:
    IC = i = 1 N ( y ^ i y ^ ¯ ) ( y i y ¯ ) i = 1 N ( y ^ i y ^ ¯ ) 2 i = 1 N ( y i y ¯ ) 2 ,
    where y ^ ¯ and y ¯ are the cross-sectional means of the predictions and true returns.
  • Rank Information Coefficient (Rank IC): Computes the Spearman’s rank correlation coefficient. It measures the monotonic relationship by replacing continuous values with their respective ordinal ranks r ( y ^ i ) and r ( y i ) :
    Rank IC = i = 1 N ( r ( y ^ i ) r ¯ ) ( r ( y i ) r ¯ ) i = 1 N ( r ( y ^ i ) r ¯ ) 2 i = 1 N ( r ( y i ) r ¯ ) 2 ,
    where r ¯ = ( N + 1 ) / 2 is the mean of the ranks. Compared to a standard IC, Rank IC is significantly more robust to financial extreme outliers and directly evaluates the effectiveness of the portfolio sorting order.
  • Precision@N (P@N): Measures the accuracy of the top algorithmic recommendations. It is defined as the proportion of the actual true top-N most profitable stocks successfully identified within the model’s top-N predictions:
    P @ N = | Top N ( y ^ ) Top N ( y ) | N .

5.3.2. Portfolio Backtesting Metrics

For algorithmic trading evaluation, we establish a daily rebalanced Long-Top-K portfolio ( K = 30 ). Let R s , t and R b , t denote the daily return of the strategy and the benchmark index at day t, respectively. Assuming 252 trading days per year and a total of D trading days in the backtest period, the quantitative metrics are defined as follows:
  • Annualized Return (Ann. Ret.): The geometric average of annualized wealth accumulation. Given the cumulative return R c u m = t = 1 D ( 1 + R s , t ) 1 , the annualized return is calculated as:
    Ann . Ret . = ( 1 + R c u m ) 252 D 1 .
  • Maximum Drawdown (Max DD): Measures profound downside risk by computing the maximum observed loss from a historical peak to a trough of the portfolio’s net asset value (NAV):
    Max DD = min t ( 1 , D ) NAV t max τ t NAV τ max τ t NAV τ ,
    where NAV t = τ = 1 t ( 1 + R s , τ ) .
  • Sharpe Ratio (Sharpe): Evaluates risk-adjusted performance. Setting the risk-free rate to zero for strict daily evaluation, it is computed by scaling the daily ratio by 252 :
    Sharpe = E [ R s ] σ ( R s ) × 252 ,
    where E [ R s ] and σ ( R s ) are the sample mean and standard deviation of the daily strategy returns.
  • Information Ratio (IR): Measures the active return of the strategy relative to its tracking error against the benchmark:
    IR = E [ R s R b ] σ ( R s R b ) × 252 .
  • Annualized Alpha ( α ): Gauges excess profitability strictly independent of systemic market movements (Beta). We compute the daily alpha ( α d a i l y ) via Ordinary Least Squares (OLS) regression ( R s , t = α d a i l y + β R b , t + ϵ t ), and project the annualized alpha as:
    Alpha = α d a i l y × 252 .

5.4. Implementation Details and Hyperparameter Configuration

The proposed USTGCN framework is implemented using PyTorch (v2.4.1+cu121). To ensure reproducibility and provide transparent structural details, the key hyperparameters governing the data inputs, model architecture, and optimization process are summarized in Table 1.
The uniform hidden size (d) is configured to 128 based on empirical sensitivity analysis. In the dynamic graph generation module, the window size is set to W = 20 based on the same analysis, which provides a favorable balance between noise suppression and responsiveness to market changes. During optimization, model parameters are updated using the Adam optimizer with a designated learning rate of 2 × 10 4 . To prevent overfitting on highly stochastic financial time-series data, training spans a maximum of 100 epochs with early stopping patience set to 30 epochs based on the validation set. Furthermore, batch processing is executed on a daily cross-sectional basis to preserve holistic structural integrity.

5.5. Results and Quantitative Analysis

Predictive Performance: As detailed in Table 2, USTGCN achieves the best overall results across the reported evaluation metrics (IC, Rank IC, and Top-N Precisions). The performance gains are also somewhat larger on the CSI300 dataset, suggesting that the proposed design remains effective when the stock universe becomes larger and relation patterns become more heterogeneous.

5.6. Ablation Studies

To quantify the contribution of individual design choices, we conducted component ablations (Table 3) and an encoder-assignment study (Table 4). Removing the rolling window (USTGCN-RW) noticeably degrades IC metrics, indicating the interference caused by micro-structure noise. Replacing the unified fusion module (USTGCN-UC) or removing the dual-stream design (USTGCN-DS) also leads to clear performance drops, indicating the contribution of temporally specialized encoders and shared dual-graph fusion. The drop from USTGCN to USTGCN-UC is larger on CSI300 than on CSI100, especially in IC and Rank IC, which indicates that the shared fusion stage plays a larger role when the stock universe is larger and the cross-stock relation patterns are more heterogeneous. Removing residual skip connections (USTGCN-SK) causes a further decline, indicating that preserving stock-specific temporal signals is important for ranking quality.
Table 4 further examines the ALSTM/GRU allocation. The default configuration attains the best overall IC/Rank IC on both datasets. Swap-1 remains competitive on CSI300, which indicates that the two recurrent encoders can both contribute useful information. However, the homogeneous settings (GRU/GRU and ALSTM/ALSTM) consistently underperform, indicating that complementary encoder choices are preferable to identical dual branches.

5.7. Hyperparameter Sensitivity

To examine the sensitivity of USTGCN to several key hyperparameters, we monitored the variations of Test IC and Rank IC on both CSI100 and CSI300 while varying the hidden size, the learning rate, and the window size (Figure 3).
Impact of Hidden Size: Figure 3a shows that performance on both datasets peaks around d = 128 . A smaller hidden size (e.g., 64) appears to limit representation capacity, while larger sizes such as 256 or 512 are associated with weaker out-of-sample performance, which is consistent with overfitting to noisy market fluctuations.
Impact of Learning Rate: Figure 3b shows that performance is strongest around 2 × 10 4 . Lower learning rates lead to slower convergence, while a larger rate such as 10 3 is associated with noticeably weaker ranking performance.
Impact of Window Size: Figure 3c shows that very short windows retain more day-to-day noise, while overly long windows can oversmooth time-varying relations and reduce responsiveness to market changes. Both datasets achieve their best or near-best results around W = 20 , which is therefore adopted in all experiments. We use a simple moving average because it stabilizes correlation estimates without introducing an additional decay parameter; exploring exponentially weighted smoothing is left for future work. The similar hyperparameter preferences observed on CSI100 and CSI300 suggest that the model behavior is reasonably stable across the two datasets.

5.8. Practical Backtest Performance and Risk Management

The practical value of a financial model is also reflected in trading performance. As visualized in Figure 4 and Figure 5, USTGCN achieves competitive cumulative returns in a daily rebalanced Long-Top-30 portfolio scheme spanning from 2017 to 2020. The drawdown analyses in Figure 6 and Figure 7 also indicate comparatively stable risk behavior. During severe systemic market downturns (e.g., Q1 2020), USTGCN shows shallower drawdown trajectories than several baselines.
To assess quantitative trading viability, Table 5 reports five backtesting metrics. USTGCN performs competitively across these dimensions:
1. Alpha Generation: USTGCN yields Annualized Returns of 20.75% (CSI100) and 31.16% (CSI300). The corresponding Annualized Alpha values of 7.09% and 16.36% indicate that the model captures excess returns beyond passive market exposure.
2. Risk-Adjusted Performance: While achieving strong returns, USTGCN also obtains the highest Sharpe Ratios (1.117 and 1.396) and Information Ratios (0.632 and 1.538). The Maximum Drawdowns are −17.48% and −20.65%, indicating competitive risk-adjusted profitability.

6. Conclusions

In this work, we proposed USTGCN, a Unified Spatio-Temporal Graph Convolutional Network for stock-ranking prediction. The model combines a dual-stream temporal encoder, rolling-window dynamic graph smoothing, and a unified dual-graph fusion layer with skip connections to capture both short-term market variation and longer-term structural information. Experimental results on the CSI100 and CSI300 benchmarks show that USTGCN improves ranking performance and delivers competitive backtest results with favorable drawdown behavior. The results also show that a compact design tailored to stock-ranking can achieve strong performance under noisy and evolving inter-stock relations. In future work, we plan to incorporate broader external information, such as macroeconomic indicators and news signals, to further enrich graph construction.

Author Contributions

Conceptualization, W.Y., L.G. and X.Z.; methodology, W.Y. and L.G.; software, W.Y.; validation, W.Y., L.G. and H.C.; formal analysis, W.Y.; writing—original draft preparation, W.Y. and L.G.; writing—review and editing, X.Z., M.L. and Y.H.; supervision, X.Z., M.L. and Y.H.; funding acquisition, Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant No. 72371116) and the Guangdong Engineering Technology Research Center for Big Data Precision Healthcare (Grant No. 603141789047).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data are available from Qlib (https://github.com/microsoft/qlib) (accessed on 15 January 2026). The preprocessed data and code used in this study are not publicly available due to confidentiality restrictions. Requests regarding research materials should be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Feng, F.; He, X.; Wang, X.; Luo, C.; Liu, Y.; Chua, T. Temporal Relational Ranking for Stock Prediction. ACM Trans. Inf. Syst. 2019, 37, 27:1–27:30. [Google Scholar] [CrossRef]
  2. He, Y.; Li, Q.; Wu, F.; Gao, J. Static-dynamic graph neural network for stock recommendation. In 34th International Conference on Scientific and Statistical Database Management; Association for Computing Machinery: New York, NY, USA, 2022; pp. 1–4. [Google Scholar] [CrossRef]
  3. Zhong, Y.; Chen, J.; Gao, J.; Wang, J.; Wan, Q. DTSRN: Dynamic Temporal Spatial Relation Network for Stock Ranking Recommendation. In International Conference on Neural Information Processing; Springer: Cham, Switzerland, 2023; pp. 346–359. [Google Scholar] [CrossRef]
  4. Fama, E.F. Efficient capital markets: A review of theory and empirical work. J. Financ. 1970, 25, 383–417. [Google Scholar] [CrossRef]
  5. Markowitz, H.M. Portfolio Selection. J. Financ. 1952, 7, 77–91. [Google Scholar] [CrossRef]
  6. Baek, Y.; Kim, H.Y. ModAugNet: A new forecasting framework for stock market index value with an overfitting prevention LSTM module and a prediction LSTM module. Expert Syst. Appl. 2018, 113, 457–480. [Google Scholar] [CrossRef]
  7. Patel, J.; Shah, S.; Thakkar, P.; Kotecha, K. Predicting stock market index using fusion of machine learning techniques. Expert Syst. Appl. 2015, 42, 2162–2172. [Google Scholar] [CrossRef]
  8. Bathla, G. Stock Price prediction using LSTM and SVR. In 2020 Sixth International Conference on Parallel, Distributed and Grid Computing (PDGC); IEEE: Piscataway, NJ, USA, 2020; pp. 211–214. [Google Scholar] [CrossRef]
  9. Matsunaga, D.; Suzumura, T.; Takahashi, T. Exploring Graph Neural Networks for Stock Market Predictions with Rolling Window Analysis. arXiv 2019, arXiv:1909.10660. [Google Scholar] [CrossRef]
  10. Xu, W.; Liu, W.; Wang, L.; Xia, Y.; Bian, J.; Yin, J.; Liu, T.Y. HIST: A graph-based framework for stock trend forecasting via mining concept-oriented shared information. arXiv 2021, arXiv:2110.13716. [Google Scholar] [CrossRef]
  11. Shi, H.; Song, W.; Zhang, X.; Shi, J.; Luo, C.; Ao, X.; Arian, H.; Seco, L.A. Alphaforge: A framework to mine and dynamically combine formulaic alpha factors. Proc. AAAI Conf. Artif. Intell. 2025, 39, 12524–12532. [Google Scholar] [CrossRef]
  12. Duan, Y.; Wang, W.; Li, J. FactorGCL: A Hypergraph-Based Factor Model with Temporal Residual Contrastive Learning for Stock Returns Prediction. Proc. AAAI Conf. Artif. Intell. 2025, 39, 173–181. [Google Scholar] [CrossRef]
  13. Noorizadegan, A.; Cavoretto, R.; Young, D.L.; Chen, C.S. Stable weight updating: A key to reliable PDE solutions using deep learning. Eng. Anal. Bound. Elem. 2024, 168, 105933. [Google Scholar] [CrossRef]
  14. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar] [CrossRef]
  15. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  16. Chen, Y.; Wei, Z.; Huang, X. Incorporating corporation relationship via graph convolutional neural networks for stock price prediction. In 27th ACM International Conference on Information and Knowledge Management; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1655–1658. [Google Scholar] [CrossRef]
  17. Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar] [CrossRef]
  18. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Zhang, C. Graph wavenet for deep spatial-temporal graph modeling. arXiv 2019, arXiv:1906.00121. [Google Scholar] [CrossRef]
  19. Ariyo, A.A.; Adewumi, A.O.; Ayo, C.K. Stock price prediction using the ARIMA model. In 2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation; IEEE: Piscataway, NJ, USA, 2014; pp. 106–112. [Google Scholar] [CrossRef]
  20. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  21. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar] [CrossRef]
  22. Kim, T.; Kim, H.Y. Forecasting stock prices with a feature fusion LSTM-CNN model using different representations of the same data. PLoS ONE 2019, 14, e0212320. [Google Scholar] [CrossRef]
  23. Feng, F.; Chen, H.; He, X.; Ding, J.; Sun, M.; Chua, T.S. Enhancing stock movement prediction with adversarial training. arXiv 2018, arXiv:1810.09936. [Google Scholar] [CrossRef]
  24. Nie, Y.; Nguyen, N.H.; Sinthong, P.; Kalagnanam, J. A time series is worth 64 words: Long-term forecasting with transformers. arXiv 2022, arXiv:2211.14730. [Google Scholar] [CrossRef]
  25. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. iTransformer: Inverted transformers are effective for time series forecasting. arXiv 2023, arXiv:2310.06625. [Google Scholar] [CrossRef]
  26. Wen, Q.; Zhou, T.; Zhang, C.; Chen, W.; Ma, Z.; Yan, J.; Sun, L. Transformers in time series: A survey. arXiv 2022, arXiv:2202.07125. [Google Scholar] [CrossRef]
  27. Sezer, O.B.; Gudelek, M.U.; Ozbayoglu, A.M. Financial time series forecasting with deep learning: A systematic literature review: 2005–2019. Appl. Soft Comput. 2020, 90, 106181. [Google Scholar] [CrossRef]
  28. Patel, M.; Jariwala, K.; Chattopadhyay, C. A Systematic Review on Graph Neural Network-based Methods for Stock Market Forecasting. ACM Comput. Surv. 2024, 57, 34. [Google Scholar] [CrossRef]
  29. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, P.; Liò, P.; Bengio, Y. Graph attention networks. arXiv 2018, arXiv:1710.10903. [Google Scholar] [CrossRef]
  30. Li, Y.; Yu, R.; Shahabi, C.; Liu, Y. Diffusion convolutional recurrent neural network: Data-driven traffic forecasting. arXiv 2017, arXiv:1707.01926. [Google Scholar] [CrossRef]
  31. Jiang, W.; Luo, J. Graph neural network for traffic forecasting: A survey. Expert Syst. Appl. 2022, 207, 117921. [Google Scholar] [CrossRef]
  32. Bai, L.; Yao, L.; Li, C.; Wang, X.; Wang, C. Adaptive graph convolutional recurrent network for traffic forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17804–17815. [Google Scholar] [CrossRef]
  33. Pareja, A.; Domeniconi, G.; Chen, J.; Ma, T.; Suzumura, T.; Kanezashi, H.; Kaler, T.; Schardl, T.; Leiserson, C. EvolveGCN: Evolving graph convolutional networks for dynamic graphs. Proc. AAAI Conf. Artif. Intell. 2020, 34, 5363–5370. [Google Scholar] [CrossRef]
  34. Rossi, E.; Chamberlain, B.; Frasca, F.; Eynard, D.; Monti, F.; Bronstein, M. Temporal graph networks for deep learning on dynamic graphs. arXiv 2020, arXiv:2006.10637. [Google Scholar] [CrossRef]
  35. Gao, J.; Ying, X.; Xu, C.; Wang, J.; Zhang, S.; Li, Z. Graph-Based Stock Recommendation by Time-Aware Relational Attention Network. ACM Trans. Knowl. Discov. Data 2022, 16, 4:1–4:21. [Google Scholar] [CrossRef]
  36. Hsu, Y.-L.; Tsai, Y.-C.; Li, C.-T. FinGAT: Financial Graph Attention Networks for Recommending Top-K Profitable Stocks. IEEE Trans. Knowl. Data Eng. 2023, 35, 469–481. [Google Scholar] [CrossRef]
  37. Sawhney, R.; Agarwal, S.; Wadhwa, A.; Shah, R.R. Exploring the Scale-Free Nature of Stock Markets: Hyperbolic Graph Learning for Algorithmic Trading. In Proceedings of the Web Conference 2021, ACM/IW3C2, Ljubljana, Slovenia, 19–23 April 2021; pp. 11–22. [Google Scholar] [CrossRef]
  38. Sawhney, R.; Agarwal, S.; Wadhwa, A.; Shah, R. Deep attentive learning for stock movement prediction from social media text and company correlations. In 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP); Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 8415–8426. [Google Scholar] [CrossRef]
  39. Yu, Y.; Kim, Y.J. Two-dimensional attention-based LSTM model for stock index prediction. J. Inf. Process. Syst. 2019, 15, 1231–1242. [Google Scholar] [CrossRef]
  40. Touzani, Y.; Douzi, K. An LSTM and GRU based trading strategy adapted to the Moroccan market. J. Big Data 2021, 8, 126. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Illustration of spatio-temporal co-movements among representative stocks. The y-axis shows standardized stock prices, and the x-axis shows trading days. The alternating divergence and convergence across stocks suggest the coexistence of time-varying and relatively stable cross-stock dependencies.
Figure 1. Illustration of spatio-temporal co-movements among representative stocks. The y-axis shows standardized stock prices, and the x-axis shows trading days. The alternating divergence and convergence across stocks suggest the coexistence of time-varying and relatively stable cross-stock dependencies.
Electronics 15 01317 g001
Figure 2. Overall architecture of USTGCN. The GRU branch corresponds to the structural/static stream, while the ALSTM branch corresponds to the dynamic stream. The framework contains four stages: dual-stream temporal encoding, robust graph construction, unified spatio-temporal fusion, and ranking prediction.
Figure 2. Overall architecture of USTGCN. The GRU branch corresponds to the structural/static stream, while the ALSTM branch corresponds to the dynamic stream. The framework contains four stages: dual-stream temporal encoding, robust graph construction, unified spatio-temporal fusion, and ranking prediction.
Electronics 15 01317 g002
Figure 3. Hyperparameter sensitivity analysis of USTGCN. The y-axis is the performance metric (Test IC and Rank IC); CSI100 curves are shown in blue and CSI300 in orange, with solid lines/markers for IC and dashed lines for Rank IC (see legends). (a) Hidden size d: performance peaks at d = 128 and weakens toward 512. (b) Learning rate: best results around 2 × 10 4 , with a sharp drop at larger rates; learning-rate ticks use × 10 n notation. (c) Rolling-window length W: both markets peak near W = 20 , consistent with Table 1.
Figure 3. Hyperparameter sensitivity analysis of USTGCN. The y-axis is the performance metric (Test IC and Rank IC); CSI100 curves are shown in blue and CSI300 in orange, with solid lines/markers for IC and dashed lines for Rank IC (see legends). (a) Hidden size d: performance peaks at d = 128 and weakens toward 512. (b) Learning rate: best results around 2 × 10 4 , with a sharp drop at larger rates; learning-rate ticks use × 10 n notation. (c) Rolling-window length W: both markets peak near W = 20 , consistent with Table 1.
Electronics 15 01317 g003
Figure 4. Empirical cumulative returns on the CSI100 dataset (2017–2020). The curves illustrate the out-of-sample wealth accumulation of daily rebalanced, equally weighted portfolios consisting of the top 30 consistently ranked stocks. The USTGCN portfolio (solid red line) stays above several baselines and the passive market index (dashed black line) for much of the test period, indicating competitive alpha-generation capability.
Figure 4. Empirical cumulative returns on the CSI100 dataset (2017–2020). The curves illustrate the out-of-sample wealth accumulation of daily rebalanced, equally weighted portfolios consisting of the top 30 consistently ranked stocks. The USTGCN portfolio (solid red line) stays above several baselines and the passive market index (dashed black line) for much of the test period, indicating competitive alpha-generation capability.
Electronics 15 01317 g004
Figure 5. Empirical cumulative returns on the broader CSI300 dataset (2017–2020). Evaluated under identical top-30 long-only conditions, USTGCN maintains competitive portfolio performance on the larger asset universe, indicating that the method remains effective beyond the smaller CSI100 setting.
Figure 5. Empirical cumulative returns on the broader CSI300 dataset (2017–2020). Evaluated under identical top-30 long-only conditions, USTGCN maintains competitive portfolio performance on the larger asset universe, indicating that the method remains effective beyond the smaller CSI100 setting.
Electronics 15 01317 g005
Figure 6. Dynamic drawdown analysis of the top-30 portfolios on the CSI100 dataset (2017–2020). The graph tracks the percentage decline from historical equity peaks. USTGCN (solid red line) shows comparatively smaller drawdowns during periods of severe market turbulence (e.g., the 2018 bear market and the early 2020 systemic shock), indicating better downside-risk control than several baselines.
Figure 6. Dynamic drawdown analysis of the top-30 portfolios on the CSI100 dataset (2017–2020). The graph tracks the percentage decline from historical equity peaks. USTGCN (solid red line) shows comparatively smaller drawdowns during periods of severe market turbulence (e.g., the 2018 bear market and the early 2020 systemic shock), indicating better downside-risk control than several baselines.
Electronics 15 01317 g006
Figure 7. Dynamic drawdown analysis of the top-30 portfolios on the CSI300 dataset (2017–2020). USTGCN maintains relatively shallow drawdown trajectories compared with several evaluated models, which is consistent with its competitive portfolio performance under systemic market stress.
Figure 7. Dynamic drawdown analysis of the top-30 portfolios on the CSI300 dataset (2017–2020). USTGCN maintains relatively shallow drawdown trajectories compared with several evaluated models, which is consistent with its competitive portfolio performance under systemic market stress.
Electronics 15 01317 g007
Table 1. Summary of detailed hyperparameters, configurations, and their specific architectural purposes in USTGCN.
Table 1. Summary of detailed hyperparameters, configurations, and their specific architectural purposes in USTGCN.
CategoryHyperparameterValueRemark/Functional Purpose
Data and InputFeature Dimension (F)6Raw daily trading indicators (e.g., OHLCV, VWAP).
Temporal Lookback Window (T)60Length of the historical trading day sequence.
Batch SizeDaily Cross-SectionPreserves holistic market spatial topology per day.
Model ArchitectureDual-Stream Hidden Dim. (d)128Latent representation capacity for sequence encoders.
Recurrent Layers2Depth of ALSTM and GRU sequential extractors.
Graph Rolling Window (W)20Days aggregated to smooth transient market noise.
Attention Reduction Factor (v)4Reduces parameter overhead in self-attention projection.
Predictor MLP Cascade 512 128 1 Progressive feature-fusion mapping to final alpha scores.
Dropout Rate0.0Network regularization threshold (empirically disabled).
Training and Opt.OptimizerAdamAdaptive gradient descent algorithm.
Learning Rate ( η ) 2 × 10 4 Optimal step size determined via sensitivity analysis.
Maximum Epochs100Upper bound for training iterations.
Early Stopping Patience30Epochs without validation gain before early termination.
Table 2. Predictive performance comparison on CSI100 and CSI300 test sets. Bold formatting indicates the superior result in each column.
Table 2. Predictive performance comparison on CSI100 and CSI300 test sets. Bold formatting indicates the superior result in each column.
ModelCSI100CSI300
IC Rank IC Precision@N (%) IC Rank IC Precision@N (%)
(↑) (↑) 3 (↑) 5 (↑) 10 (↑) 30 (↑) (↑) (↑) 3 (↑) 5 (↑) 10 (↑) 30 (↑)
MLP0.0710.06756.5356.1755.4953.550.0820.07957.2157.1056.7555.56
SFM0.0810.07457.7956.9655.9253.880.1020.09659.8458.2857.8956.82
GATs0.0960.09059.1758.7157.4854.590.1110.10560.4959.9659.0257.41
LSTM0.0970.09160.1259.4959.0454.770.1040.09859.5159.2758.4056.98
ALSTM0.1020.09760.7959.7658.1355.000.1150.10959.5159.3358.9257.47
GRU0.1030.09759.9758.9958.3755.090.1130.10859.9559.2858.5957.43
Transformer0.0890.09059.6259.2057.9454.800.1060.10460.7660.0659.4857.71
ALSTM+TRA0.1070.10260.2759.0957.6655.160.1190.11260.4559.5259.1658.24
HIST0.1200.11561.8760.8259.3856.040.1310.12661.6061.0860.5158.79
SDGNN0.1260.12062.4961.4159.8156.390.1370.13262.2361.7661.1859.56
DTSRN0.1370.13262.8561.7960.6856.840.1460.14162.7262.0361.3759.74
USTGCN0.1410.13763.5662.6061.3357.420.1540.14863.6262.9262.2160.50
Table 3. Ablation study evaluating the specific impact of USTGCN components. Precision@N metrics are scaled by 100.
Table 3. Ablation study evaluating the specific impact of USTGCN components. Precision@N metrics are scaled by 100.
VariantCSI100CSI300
IC Rank IC Precision@N (%) IC Rank IC Precision@N (%)
(↑) (↑) 3 (↑) 5 (↑) 10 (↑) 30 (↑) (↑) (↑) 3 (↑) 5 (↑) 10 (↑) 30 (↑)
USTGCN0.1410.13763.5662.6061.3357.420.1540.14863.6262.9262.2160.50
USTGCN-RW0.1390.13564.1762.5861.2357.220.1490.14463.9662.5561.4860.48
USTGCN-UC0.1330.12963.3062.7160.8856.950.1350.12962.1561.4860.8959.46
USTGCN-DS0.1260.12063.8862.2360.2656.300.1380.13262.0161.2761.2459.54
USTGCN-SK0.1170.11362.1560.8960.1456.170.1260.12361.0960.7259.9659.25
Table 4. Effect of encoder assignment in the dual-stream temporal module.
Table 4. Effect of encoder assignment in the dual-stream temporal module.
DatasetMethodDynamicStaticICRank IC
CSI100USTGCN (Default)ALSTMGRU0.1410.137
Swap-1GRUALSTM0.1380.134
Swap-2GRUGRU0.1380.132
Swap-3ALSTMALSTM0.1280.125
CSI300USTGCN (Default)ALSTMGRU0.1540.148
Swap-1GRUALSTM0.1520.148
Swap-2GRUGRU0.1460.140
Swap-3ALSTMALSTM0.1480.143
Table 5. Comprehensive quantitative backtest performance metrics (2017–2020). Bold formatting indicates the best result in each column.
Table 5. Comprehensive quantitative backtest performance metrics (2017–2020). Bold formatting indicates the best result in each column.
ModelCSI100CSI300
Ann. Ret. Sharpe Max DD IR Alpha Ann. Ret. Sharpe Max DD IR Alpha
(% ↑) (↑) (% ↑) (↑) (% ↑) (% ↑) (↑) (% ↑) (↑) (% ↑)
MLP9.740.606−24.65−0.602−2.4021.861.051−23.670.8588.88
SFM8.750.551−25.63−0.670−3.2626.971.249−19.331.30312.95
GATs16.680.924−20.360.2113.5425.691.182−22.681.09612.07
LSTM15.430.870−25.240.0752.4624.891.094−26.661.00910.83
ALSTM16.370.905−25.850.1783.2425.721.197−21.041.14212.08
GRU18.100.974−20.810.3604.6327.091.227−24.021.21513.06
Transformer14.150.812−20.08−0.0691.4322.281.027−23.710.8638.85
ALSTM+TRA14.720.839−21.98−0.0062.0321.741.021−28.220.8658.41
HIST14.660.828−20.25−0.0081.7526.381.180−27.111.11612.42
SDGNN18.751.002−18.110.4145.2426.851.273−19.271.23013.30
DTSRN15.400.852−17.390.0732.5228.131.265−24.721.25513.99
USTGCN20.751.117−17.480.6327.0931.161.396−20.651.53816.36
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yao, W.; Gao, L.; Zhang, X.; Chen, H.; Liu, M.; Hu, Y. USTGCN: A Unified Spatio-Temporal Graph Convolutional Network for Stock-Ranking Prediction. Electronics 2026, 15, 1317. https://doi.org/10.3390/electronics15061317

AMA Style

Yao W, Gao L, Zhang X, Chen H, Liu M, Hu Y. USTGCN: A Unified Spatio-Temporal Graph Convolutional Network for Stock-Ranking Prediction. Electronics. 2026; 15(6):1317. https://doi.org/10.3390/electronics15061317

Chicago/Turabian Style

Yao, Wenjie, Lele Gao, Xiangzhou Zhang, Haotao Chen, Mingzhe Liu, and Yong Hu. 2026. "USTGCN: A Unified Spatio-Temporal Graph Convolutional Network for Stock-Ranking Prediction" Electronics 15, no. 6: 1317. https://doi.org/10.3390/electronics15061317

APA Style

Yao, W., Gao, L., Zhang, X., Chen, H., Liu, M., & Hu, Y. (2026). USTGCN: A Unified Spatio-Temporal Graph Convolutional Network for Stock-Ranking Prediction. Electronics, 15(6), 1317. https://doi.org/10.3390/electronics15061317

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop