Next Article in Journal
Visualization Study on Trends and Hotspots in the Field of Urban Air Pollution in Metropolitan Areas and Megacities: A Bibliometric Analysis via Science Mapping
Previous Article in Journal
Ionospheric Anomaly Identification: Based on GNSS-TEC Data Fusion Supported by Three-Dimensional Spherical Voxel Visualization
Previous Article in Special Issue
Deep Learning for Atmospheric Modeling: A Proof of Concept Using a Fourier Neural Operator on WRF Data to Accelerate Transient Wind Forecasting at Multiple Altitudes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comprehensive Scale Fusion Networks with High Spatiotemporal Feature Correlation for Air Quality Prediction

1
School of Electrical Engineering, Guangdong Songshan Vocational and Technical College, Shaoguan 512000, China
2
School of Mathematics and Statistics, Hainan Normal University, Haikou 570100, China
3
School of Information Engineering, Nanchang University, Nanchang 330031, China
4
School of Computer and Information Engineering, Guangdong Songshan Vocational and Technical College, Shaoguan 512000, China
*
Author to whom correspondence should be addressed.
Atmosphere 2025, 16(4), 429; https://doi.org/10.3390/atmos16040429
Submission received: 3 March 2025 / Revised: 1 April 2025 / Accepted: 3 April 2025 / Published: 8 April 2025

Abstract

:
The escalation of industrialization has worsened air quality, underscoring the essential need for accurate forecasting to inform policies and protect public health. Current research has primarily emphasized individual spatiotemporal features for prediction, neglecting the interconnections between these features. To address this, we proposed the generative Comprehensive Scale Spatiotemporal Fusion Air Quality Predictor (CSST-AQP). The novel dual-branch architecture combines multi-scale spatial correlation analysis with adaptive temporal modeling to capture the complex interactions in pollutant dispersion and enhanced pollution forecasting. Initially, a fusion preprocessing module based on localized high-correlation spatiotemporal features encodes multidimensional air quality indicators and geospatial data into unified spatiotemporal features. Then, the core architecture employs a dual-branch collaborative framework: a multi-scale spatial processing branch extracts features at varying granularities, and an adaptive temporal enhancement branch concurrently models local periodicities and global evolutionary trends. The feature fusion engine hierarchically integrates spatiotemporally relevant features at individual and regional scales while aggregating local spatiotemporal features from related sites. In experimental results across 14 Chinese regions, CSST-AQP achieves state-of-the-art performance compared to LSTM-based networks with RMSE 6.11–9.13 μg/m3 and R2 0.91–0.93, demonstrating highly robust 60 h forecasting capabilities for diverse pollutants.

1. Introduction

The accelerated global industrialization in recent decades has exacerbated air pollution to critical levels, with the World Health Organization attributing approximately 4.2 million premature deaths annually to ambient air pollution exposure [1]. This crisis has made air quality prediction not just a technical challenge but a vital part of urban sustainability planning, supporting both environmental governance and public health strategies [2,3]. Air quality prediction research encompasses both traditional statistical methods and modern deep learning techniques [4].
Traditional methods, known as chemical transport models, simulate the transport, diffusion, and chemical reactions of pollutants to predict [5]. These methods are categorized into physical diffusion-based models and statistical models, with the former relying on diffusion theory and the latter on empirical assumptions [6]. However, their reliance on empirical parameters and physical data limits accuracy. Statistical methods, like linear regression and machine learning modeling, simplify the complexity and parameter overload of physical models [7,8], such as support vector regression (SVM) [9], Random Decision Forests (RFs) [10], Hidden Semi-Markov Models (HSMMs) [11], and other machine learning techniques [12,13]. While these conventional models demand fewer computational resources and parameters, their static feature interaction mechanisms fail to delineate the spatiotemporal heterogeneity of pollutants and are incapable of resolving multi-scale.
In recent years, convolutional neural networks (CNNs) have significantly advanced air quality prediction, particularly in forecasting PM2.5 concentrations. In 2018, Athira et al. [14] used recurrent networks for air quality prediction in conjunction with HSMM. Guo [15] attempted to partition time correlation at a finer granularity and designed a deep residual CNN. These hybrid approaches improved prediction accuracy through the fusion of statistical feature engineering with CNN frameworks, but they remain constrained by manual hyperparameter configuration. Subsequently, Elbaz et al. [16] and Tran et al. [17] enhanced accuracy by integrating spatiotemporal data into residual CNN; their methodologies introduced heightened algorithmic complexity. Considering the correlation between different time lengths, some researchers have also proposed long short-term memory (LSTM) for air quality prediction networks [18,19,20]. For example, Ameri et al. [21] and Tishya et al. [22] proposed a deep learning model with an encoder–decoder structure using Conv and LSTM units to model and predict the correlation of urban pollution. However, these models are less efficient for model training and inference in long sequence scenarios. Further, researchers have proposed methods based on variations-LSTM to realize the prediction of air index time series [23]. Rabie et al. [24] developed a CNN-Bi-LSTM hybrid framework to address multi-scale spatial air quality prediction in megacities, overcoming challenges in modeling cross-regional pollutant dependencies through integrated multi-scale feature extraction. Fan et al. [25] implemented a hybrid seasonal air quality prediction through a mixture of rough ensemble wrapping methods and regularized combinatorial LSTM. In addition, there are some other novel models, including Self-Attention LSTM [26] and adaptive LSTM [27]. However, these models typically demand substantial computational resources for training and optimization, and their performance can vary significantly when predicting different time scales.
Since 2022, people have increasingly leveraged graph convolutional networks (GCNs) to address constrained modeling capacity in temporal air quality data analysis [28]. Chen et al. [29] employed multi-factor separation evolutionary spatiotemporal GCN factorized feature learning for multivariate series decomposition; Hu et al. [30] enhanced 10.7% cross-regional accuracy through adaptive multi-graph attention fusion with GRU units. The emerging hybrid network introduced GCN and Bi-LSTM structures to adapt the propagation depth to seasonal patterns, addressing the long-term prediction challenge through a specialized spatiotemporal decoupling strategy [31]. In addition, other hybrid models have been proposed, including attention mechanisms [32] and physically informed guidance [33]. These advancements have improved stability and accuracy across prediction tasks involving various time scales. Recently, generative CNN (GCNN) [34,35] has been utilized to generate a pre-trained model, which achieves fine-grained spatiotemporal knowledge migration via generative pre-training on source city data. Some researchers have designed a combination of GCNN and reinforcement learning to focus on essential time steps and spatial locations [36,37]. Based on this, other generative networks have combined multifactor decomposition and data-driven trees, further uniting spatiotemporal features [38,39]. However, these models may miss key correlations when handling spatiotemporal features separately, complicating the resolution of spatiotemporal interactions, and the performance of cross-region prediction is unsatisfactory [40]. In addition, existing studies predominantly concentrated on modeling dependencies at a single spatial and temporal scale, neglecting the intricate spatiotemporal dynamics of air quality changes across individual and local dimensions in cross-city prediction tasks.
To solve the above problems, this study proposed an air quality prediction method based on generative complete scale spatiotemporal convolutional networks (CSST-AQP), which aims to solve the spatial and temporal dependence problem in air quality prediction tasks. The CSST-AQP framework method adopts a three-level progressive architecture: firstly, a local strong correlation spatiotemporal feature extraction fusion network is constructed to capture long-range spatiotemporal dependencies through a dynamic feature sensing mechanism; secondly, an adaptive spatiotemporal augmented and a multi-scale convolution network are designed to achieve the synergistic extraction of multi-dimensional features. Eventually, the CSST-AQP can manage remote temporal dependencies in a parallel manner, while spatial dependencies are modeled to achieve long-term air quality prediction. Experiments show that CSST-AQP performs excellently in all metrics over a 0–400 h prediction window. The main contributions that we make in this paper are as follows:
(1) A multi-scale fusion model integrates cross-station spatial patterns and local temporal trends to capture spatiotemporal dynamics. (2) An adaptive spatial correlation matrix coding module dynamically updates site interactions and tracks spatiotemporal dependencies. (3) A temporal model integrating dilated causal convolutions with adaptive snow-abatement optimization is used to capture local periodic and extended-range temporal dependencies. (4) A novel architecture fuses time-series encoding and spectral–temporal decomposition, resolving spatiotemporal mismatches while aligning cross-modal scales. (5) A prediction algorithm with extra high accuracy, 22.5% higher than CEMD-LSTM, is proposed.
The rest of this paper is organized as follows. Section 2 reviews some related technologies and concepts. In Section 3, the workflow of the proposed air quality prediction model is described in detail. To evaluate the proposed algorithm, simulation details, experimental results, comparison with several other air quality prediction models, and an ablation study are given in Section 4. Finally, Section 5 and Section 6 present the discussion and a brief conclusion, respectively.

2. Related Materials and Concepts

2.1. Spatiotemporal Encoder

To capture regional spatiotemporal correlation, Du et al. [41] proposed a spatiotemporal convolutional encoder (STE) based on temporal causal convolutional encoding. STE is capable of combining information from the spatiotemporal encoder to retain the characteristics of the original input data. The structure of STE is shown in Figure 1.
The STE consists of two spatiotemporal graph convolution blocks and a fully connected output layer. Each STE-Block contains two temporal Gated-Conv and one spatial Gated-Conv. The temporal Gated-Conv combines a 1-D Conv with a gated linear unit. In this study, we use STE to capture dynamic changes in air quality data over time and improve prediction performance.

2.2. Preliminaries

Given the historical air quality time series X till the current time step, we aim to train a predictive network that accurately estimates the air quality data of all regions at the future time step t + 1 and then obtain the predicted value X t + 1 . Currently, few studies predict air quality data across regions, and this study focuses on forecasting air pollutants across different spatiotemporal gaps. In general, there are regional and time dependencies between changes in air pollution concentrations, as shown in Figure 2. To more comprehensively depict spatiotemporal correlations, the following definitions are formulated in this paper.
(1)
Air quality flow graph (AQFG): the air quality flow graph is defined as γ , and the citywide air quality data for the pre-T moment are expressed by a tensor X = x 1 , x 2 , x 3 , x t , t = 1 , 2 , 3 , N , N T .
(2)
Spatial region: each monitoring station Mt is represented as a graph node with a feature vector h t d , encapsulating multivariate air quality measurements (e.g., PM2.5, SO2, and O3) and meteorological conditions (wind speed and temperature) at the station.

2.3. Time Series Imaging

The Gramian Angular Field (GAF) was proposed by Oates et al. Its main function is to encode time series into two-dimensional (2D) image information by generating a Gram matrix through trigonometric functions [42]. In this paper, the GAF-transformed time series data retain the temporal relationships among its observations, so we combine it with CNN to improve the prediction accuracy by maintaining the temporal relationship while learning more 2D features. The details are
x t = x t max ( X ) + x t min ( X ) max X min X ,
where x t is the normalized observed value, x t is the t-th sequence, and x t is converted to the form of polar coordinates:
φ t = arccos ( x t ) ,
r t = t T ,
where φ t is the angle, r t is the radius, t denotes the time stamp, and the transformed polar form is defined as φ t , r t . Ultimately, the triangular sum between each observation G A F t , t determines each element in the GA:
G A F t , t = cos φ t + φ t , t , t = 1 , 2 , 3 , T ,
G A F = cos ( φ 1 + φ 1 ) cos ( φ 1 + φ T ) cos ( φ T + φ 1 )   cos ( φ T + φ T )   .

2.4. Evaluation Indexes

The objective evaluation indexes of air quality prediction used in this paper include the Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), and R-squared ( R 2 ). The formulas are
RMSE = 1 t i = 1 t y ^ i y i 2 ,
MAE = 1 t i = 1 t y ^ i y i ,
MAPE = 1 t i = 1 n y i y ^ i y i × 100 % ,
R 2 = 1 i = 1 t ( y i y ^ i ) 2 i = 1 t ( y i y ¯ i ) 2 ,
where y ^ i and y i are the predicted value and the actual value, respectively. y ¯ i is the average of the y i .

3. Propose Models and Methodologies

In this section, we proposed a comprehensive scale fusion network with high spatiotemporal feature correlation for air quality prediction based on a preprocessing network, as well as STE-ResNet, named CSST-AQP, whose structure is shown in Figure 3.

3.1. Overview of the Proposed CSST-AQP Framework

The proposed CSST-AQP architecture comprises four principal components: (1) a local spatiotemporal feature preprocessing network (LSTF-Net) for initial data refinement, (2) a comprehensive-scale spatial feature processing module (CSSP-Net) enabling multi-granularity pattern extraction, (3) an adaptive temporal feature enhancement network (ATSE-Net) with dynamic time-aware mechanisms, and (4) a prediction output layer. In the CSST-AQP, all sub-networks were trained simultaneously, and the convolutional modules in the three sub-networks were configured with convolutional cores of different sizes to learn air quality time series feature information in different contextual orientations.
In general, the processing pipeline begins by feeding raw data X into the LSTF-Net preprocessing sub-network for spatiotemporal feature encoding. The generated fusion features are subsequently distributed through dual processing branches: the upper branch employs CSSP-Net, while the lower branch utilizes ATSE-Net, each specialized in extracting complementary dynamic spatiotemporal dependencies through parallel training. The learned features from both branches are then sequentially processed through three integration stages: feature fusion, a multi-attention mechanism layer, and a fully connected regression layer. Finally, the predicted values are generated through the output layer. The detailed procedure is described as follows.
(1) First, the input data X are fed to LSTF-Net, where they undergo localized spatiotemporal modeling. This process integrates spatial mapping and time-series feature extraction to simultaneously capture two critical aspects: temporal patterns of air quality variations at individual monitoring stations and cross-station spatial correlations, ultimately generating enhanced spatiotemporal features X 1 that encode both station-specific dynamics and regional interdependencies.
(2) Then, X 1 are encoded as corresponding 2D feature maps by CSSP-Net. In the CSSP-Net architecture, the input feature X 1 undergoes multi-scale feature fusion through GAF and Continuous Wavelet Transform (CWT). This hybrid approach hierarchically integrates temporal patterns and frequency-domain characteristics. The fused feature maps undergo processing through the residual Conv module, where reinforcement learning mechanisms are applied to execute feature optimization, ultimately yielding an enhanced feature map X 2 .
(3) The processed temporal features are fed into ATSE-Net, where key hyperparameters (neuronal population and convolutional kernel dimensions) in STE modules are optimized via a novel metaheuristic optimizer combining periodic oscillatory dynamics and snow ablation strategies. This accelerates convergence while boosting generalization and prediction fidelity. Then, the output features X 3 are obtained.
(4) Finally, the feature map X 2 and X 3 are fused by features to generate X 4 , which is then transmitted to the full connectivity layer and multi-head attention layer for feature learning to obtain the final prediction result y ^ i .

3.2. Localized Spatiotemporal Feature Preprocessing Module (LSTF-Net)

The first sub-network of the CSST-AQP framework is the local spatiotemporal feature preprocessing network (LSTF-Net). To extract spatiotemporal dependence, LSTF-Net categorizes the pertinent features in air quality prediction into two distinct groups: temporal and spatial features. The LSTF-Net structure is shown in Figure 4. LSTF-Net comprises three components: an initial encoding module for converting raw sequences into AQFG matrices, channel attention module 1 (CAM-1) for capturing strongly correlated spatiotemporal features, and channel attention module 2 (CAM-2) implementing dilated convolution operations and feature-weighted fusion.
The temporal features f include PM2.5, PM10, SO2, CO, and O3, collectively denoted as X, and the operation process of the model is as follows:
X = f ( P M 2.5 ) + f ( P M 10 ) + f ( S O 2 ) + f ( C O ) + f ( O 3 ) .
(1) Firstly, to capture spatiotemporal dependencies across regional monitoring stations, LSTF-Net employs an initial encoding step that transforms the raw sequence X into an AQFG matrix γ .
γ = ( M , A , X ) ,
where A is the dynamic weight matrix between air quality flow graphs, and M is the monitoring node. The dynamic weight matrix A t i , j of the monitoring node i to node j at t time points is
A t i , j = exp ( distance ( M i , M j ) ξ · X i t r i , j r i , j ) k N ( i ) e x p ( distance ( M i , M k ) ξ · X i t r i , k r i , k ) ,
where r i , j is the vector from monitoring node i to node j; ξ is the attenuation coefficient; M k denotes the k-th historical monitoring node; M i , j M ,   i , j = 1 , 2 , 3 , N . X i t represents the data of node i at time t; and r i , k is the k-th prior coefficient vector of X i t .
(2) Then, the tensor γ undergoes processing through a channel attention module 1 (CAM-1). We calculated the Pearson correlation coefficients between tensor γ i and γ j at each site and defined them as the spatial features Pij, P i j P .
P i j = cov ( γ i , γ j ) σ ( γ i ) σ ( γ j ) ,
G ( P ) = sigmoid ( W P + b ) .
A gating mechanism G ( P ) is subsequently applied to amplify cross-regional features with P i j , effectively capturing dominant atmospheric transport and city pathways. W denotes the coefficient matrix, and b is the penalty factor obtained from training. To explicitly encode inter-regional air quality dependencies, we introduced a geographically weighted graph regularization term L g e o into the objective function
L g e o = i = 1 N j N ( i ) P i j γ i m i γ j m j F 2 ,
where γ i denotes the i-th spatiotemporal flow tensor of node γ t , m i is the region-specific meteorologic mask, and the Hadamard product implements localized feature gating based weather patterns. The spatial features quantify the correlation of pollutant data across different monitoring stations using the Pearson correlation coefficient. The ensuing obtained matrix P is transferred to the lower convolutional layer for training to generate the intermediate features f s .
(3) Similarly, the CAM-2 processes input signals γ to extract temporal feature representations, denoted as f t . To simulate spatiotemporal interdependencies across regional stations, we propose a bilinear temporal channel attention mechanism (BTCA) with learnable parameters W1 and W2. CAM-2 adopts dilated convolutions to process sensor inputs, outputting temporal features f t R . The BTCA mechanism computes station-wise interactions, which are defined as
X 1 = Softmax ( ( f s W 1 ) ( f t W 2 ) T ) .
(4) Finally, the spatial feature f s and temporal feature f t are subsequently integrated through dual attention-weighted layers, generating the unified spatiotemporal tensor X 1 .
In LSTF-Net, X 1 serves as the input feature of the lower-level sub-network and is transmitted to CSSP-Net and ATSE-Net for training, respectively.

3.3. The Structure of Complete Scale Spatial Processing Network (CSSP-Net)

The upper half branch of CSST-AQP is CSSP-Net, which includes the GAF and CWT image coding module and a novel residual convolution module, whose structure is shown in Figure 5. The CSSP-Net first encodes X 1 into images with GAF and CWT. The encoded feature maps then pass through 11 convolutional layers and 2 residual blocks, extracting hierarchical multi-scale features and ultimately generating the output X 2 .
The upper branch employs a GAF encoder to transform time-series pollutant measurements into temporal feature maps. This encoding strategy preserves intrinsic temporal dependencies through polar coordinate-based angular correlations, effectively mitigating information loss during sequential data transformation.
For the feature map X t , which has a total of T observations, φ t is the t-th observation frame that denotes the angle, and the feature is scaled to [−1, 1] via Equation (1) to avoid the impact of larger observations on the subsequent inner product. Then, X t is converted to polar coordinates φ t , r t by Equation (2). Finally, the variable φ t is processed through Equations (4) and (5) to generate a 2D image feature map G A F feature , which is subsequently fed into the subsequent layer for concatenation.
Concurrently, the lower branch processes spatial correlations via continuous wavelet transform (CWT) with a Haar wavelet basis, decomposing X 1 into multiscale wavelet coefficient arrays. This process is able to identify jumps and discontinuities within the signal and obtain features on multiple scales of the original spatiotemporal sequence, further improving the accuracy of the prediction.
W X ( a , b ) = 1 a X ( t ) ψ ¯ ( t b a ) d t .
where W X ( a , b ) is the CWT coefficients at scale a and position b, which are the encoded feature maps C W T f e a t u r e , and ψ ¯ ( t ) is the complex conjugate of the Haar wavelet function. Then, C W T f e a t u r e is concatenated with the G A F feature to form a hybrid spatiotemporal feature matrix. Subsequently, the fused matrix is fed into a residual CNN module to jointly model cross-regional interactions in both feature spaces: the GAF branch captures dynamic temporal patterns of pollutant evolution, while the CWT-derived coefficients encode localized spatial–frequency characteristics of atmospheric processes.
In details, the fused image is processed through three convolutional layers with 3 × 3 convolutional kernels and one residual layer to obtain the feature R 1 . R 1 is then sequentially processed through three down-sampling convolutional layers with convolutional kernel sizes of 5 × 5 and 7 × 7 to obtain the intermediate variables. The output intermediate variable features are used as inputs to the next convolutional kernel size of 3 × 3 convolutional layer, and residual branches are learned to obtain R 2 . Then, the feature map R 2 is input to a 1 × 1 convolutional layer and a flatten layer to produce a flattened output vector X 2 .

3.4. Adaptive Temporal Feature Enhancement Network (ATSE-Net)

To effectively capture the long-range time dependence, an adaptive time series region enhancement sub-network (ATSE-Net) is introduced into the lower branch of CSST-AQP. The detailed structure of the ATSE-Net is shown in Figure 6. The ATSE-Net architecture employs a hierarchically dense structure comprising four cascaded adaptive spatiotemporal encoding (ASE) blocks, followed by a multi-attention fusion module and a flattening output layer. Each ASE block integrates three core components: a STE unit for joint feature representation, an adaptive threshold filter (T-Filter) layer for dynamic sequence modeling, and an adaptive average pooling operator for dimensionality reduction.
The processing pipeline initiates with the tensor X 1 being sequentially processed through the ASE block cascade. At each stage, the output of the n-th (n = 2, 3, 4) ASE block undergoes element-wise fusion with the original X 1 feature through residual connections, implementing a dense connectivity pattern that alleviates gradient vanishing while enhancing feature reuse. After four ASE operations with three inter-block skip connections, the aggregated features are directed to the multi-attention layer and the flattening layer, ultimately flattened into the final output tensor X 3 .
Critical to the architecture’s performance is the integration of the adaptive periodic oscillation snowmelt optimizer (APO-SO) [43] during parameter optimization. The APO-SO adopts a dual population mechanism, where one subpopulation is responsible for the global TSE module parameters search, and another’s task is to search for the optimal threshold of the local T-Filter. The mechanism enables ATSE-Net to optimize the parameters more efficiently during the training process, thus improving the performance and generalization ability of the model. The signal intensity accumulation matrix in the feature map may be affected by residual noise or anomalous signals in the original spatiotemporal sequence. The T-Filter can learn a suitable threshold to extract local features and preserve spatial information, for which the target intensity S ( i ) and the threshold T are
S ( i ) = S ( i ) ,   S ( i ) T 0 , S ( i ) < T   ,
T = a + e ,
a = α i = 1 w S ( i ) w ,
e = β i = 1 w a S ( i ) 2 w ,
where a is the mean of the sum of all column pixels in the feature map band, and e is the standard deviation of the column intensity of the feature map band. w is the number of sample points, and α and β are the adaptive parameters learned by network training, respectively. The APO-SO framework for the TSE module is formally described, and its structural configuration is summarized in Algorithm 1. This strategy integrates adaptive population initialization with stochastic operators to enhance global search capabilities while maintaining computational efficiency.
Algorithm 1: APO-SO optimization for TSE module
Atmosphere 16 00429 i001
As detailed in Algorithm 1, the optimization process employs a three-phase architecture: (1) diversity-preserving initialization (Lines 1–4); (2) hybrid exploration–exploitation (Lines 5–13); (3) elite selection (Lines 14–17) and obtained T optimal (Lines 18), where the stochastic components (Lines 4, 5) work synergistically with the adaptive mechanisms (Lines 11–13, 17) to prevent premature convergence while maintaining time complexity O ( n log n ) through efficient population management.

4. Experimental Section and Results

4.1. Training Details and Datasets

The model was developed using the PyTorch-2.1.1 deep learning framework, incorporating Cross-Entropy Loss as the loss function. The training regimen utilized the Adam optimizer with an initial learning rate of 0.01, a momentum of 0.9, and a weight decay coefficient of 0.0005. The loss function L is
L ( λ 1 , λ 2 ) = E y ^ i , y i λ 1 y ^ i y i 2 2 + λ 2 R TV ,
R TV = Ω y ^ i 2 d x ,
where y i is the true value, and y ^ i is the predicted value. A total variation penalty R TV is added to the loss function of the L ( λ 1 , λ 2 ) , which further stabilizes the optimization. λ 1 and λ 2 are the weight coefficients, which are obtained from the network training, and their initial values are 0.8 and 0.2, respectively. In this study, experiments are conducted on two real datasets in the Guangzhou region (GZR) and Shaoguan region (SGR) to evaluate the performance of the model predictions. Both datasets (GZR/SGR) cover January 2021–April 2024 (see Table 1 for statistics) and are publicly available from the Air Quality Historical Data Platform [44].
The raw dataset underwent comprehensive preprocessing procedures comprising outlier elimination through the interquartile range method, z-score standardization, and min–max normalization to ensure dimensional homogeneity. We convert all the time series data into samples and labels with the sliding window method (with a sliding window size of W and sliding step size of O) before putting them into the network model for training. After 2000 batches, the network stabilized with a loss function value of 0.0005, and the R 2 of training and testing reached more than 0.95. The results are shown in Figure 7 and Figure 8, respectively.
Figure 7 and Figure 8 show excellent network training with minimal discrepancy between truth and predicted values. Both training and testing losses stabilized at 0.0002 with no overfitting, and a high R2 (R2 > 0.95) confirms the model’s remarkable prediction capability.

4.2. Feature Visualization

In this section, we qualitatively analyze the prediction performance of the self-supervised spatiotemporal coding network and visualize the feature extraction capability of the CSST-AQP framework. Figure 9 shows the visualization results of the PM2.5 component, PM10 component, O3 component, NO2 component, SO2 component, and CO component after the initial network coding process, respectively.
Figure 10a displays the SHAP eigenvalues ranking for air quality predictors, showing PM2.5 as the most influential feature, followed by industrial emissions (PM10, O3, NO2, SO2, and CO), temperature, humidity, wind speed, and temporal factors. The temporal analysis in Figure 10b reveals strong autocorrelation patterns within pollutant sequences across adjacent time points.
It can be observed that the characteristic shape value of PM2.5 is the highest, followed by the other pollutants and temperature. In the LSTF-Net preprocessing framework, we systematically integrate these features through weighted encoding, transforming them into enhanced spatiotemporal representations that boost both model performance and interpretability with optimized feature correlations.
To evaluate the feature extraction capability enhancement of self-supervised learning in LSTF-Net, Figure 11 presents comparative visualizations of air quality feature correlations. Figure 11a displays the original feature distribution without embedding spatial features, while Figure 11b reveals the refined pattern after incorporating spatial correlations learned using LSTF-Net’s architecture.
In Figure 11a, we can see that 4000 randomly sampled embeddings trained without spatial feature integration exhibit fragmented clustering across air quality ranges with weak inter-sample correlations (low correlation coefficients). Conversely, Figure 11b demonstrates that LSTF-Net’s architecture effectively organizes embeddings into distinct clusters with strong feature correlations following spatial–temporal learning. These empirical results confirm that LSTF-Net’s self-supervised framework significantly enhances feature discernment through synergistic spatial–temporal representation learning, directly contributing to improved prediction accuracy.

4.3. Analysis of Projected Results

To validate the predictive performance of our proposed model, we conducted multi-step forecasts of six critical air pollutant concentrations (PM2.5, PM10, O3, NO2, SO2, and CO) over a 400 h horizon. The predictive outcomes are systematically presented in three comparative visualizations: Figure 12 demonstrates the PM2.5 and PM10 forecasts, Figure 13 illustrates the O3 and NO2 predictions, and Figure 14 displays the SO2 and CO projection results. Notably, all forecasted trajectories exhibit strong alignment with corresponding observed measurements across the evaluation period. The minimal deviation between predicted and actual values, particularly evident in the consistently narrow error margins, demonstrates the model’s effectiveness in capturing complex temporal patterns.
Table 2 summarizes the CSST-AQP model’s performance for six pollutants, showing stable metrics (RMSE: 6.11–9.13 μg/m3, MAE: 6.06–8.91 μg/m3, MAPE: 6.21–9.56%, R2: 0.91–0.93) with minimal prediction errors (<9% MAE) and high accuracy in both univariate and multivariate settings. All error confidence intervals exceed 95%, indicating statistical robustness. For O3, the partial autocorrelation function (PACF) exhibits a statistically significant value of 0.21 at the third-order lag (Confidence Level = 95.8%). This residual autocorrelation likely stems from missing historical concentration data from the preceding day, resulting in incomplete modeling of temporal dependencies in current predictions.
Compared to industry-dominated pollutants such as PM2.5 (PACF = 0.15), CO has a stronger autocorrelation (PACF = 0.18, RMSE = 9.13), and the error between its predicted and true values is more significant. To comparatively assess the relative error magnitude of CO predictions against other pollutants, we conducted systematic residual analysis through visualization of true-predicted value discrepancies, with the comprehensive scatterplot presented in Figure 15.
The results presented in Figure 15 demonstrate that the model consistently overestimates CO concentrations before data correction, exhibiting residuals between 1.2 and 1.2, and the error is significantly reduced after correction. This systematic bias may be attributed to current motor vehicle restriction policies that fail to adequately address winter-specific emission patterns during high pollution periods and morning rush hours.

4.4. Comparative Experiment

4.4.1. Compared with Deep Learning Models

To further validate the predictive performance of the proposed CSST-AQP (CT) model, Figure 16 and Table 3 present comparative results under identical experimental conditions between our model and established methods, including LSTM [21], TCN [45], EMD-LSTM [46], EMD-TCN [47], CEMD-LSTM [48], EA-LSTM [49], and the baseline raw data.
LSTM: Long Short-Term Memory Neural Networks (LSTMs), capable of capturing multi-scale temporal patterns and addressing long-term dependencies in time-series data, are employed to forecast air quality through an analysis of historical pollution records.
TCN: the temporal convolutional network (TCN), characterized by dilated causal convolutional architectures with exponentially expanding receptive fields, exhibits superior performance in spatiotemporal air pollutant forecasting through multiscale representation learning of atmospheric chemical transport dynamics.
EMD-LSTM: Empirical Mode Decomposition-enhanced LSTM (EL), which hybrid architecture, synergizing Hilbert–Huang Transform-based adaptive signal decomposition with recurrent neural memory units, achieves state-of-the-art performance in multivariate air pollutant concentration forecasting by resolving multiscale atmospheric turbulence patterns.
EMD-TCN: Empirical Mode Decomposition-enhanced TCN (ET), which integrates Hilbert–Huang Transform-based adaptive signal decomposition with exponentially dilated convolutional neural operators, establishes a novel paradigm for air prediction.
CEMD-LSTM: Complementary Ensemble Empirical Mode Decomposition LSTM (CL), integrating complementary noise-injected ensemble decomposition with gated recurrent feature learning, achieves breakthrough performance in multi-pollutant interaction forecasting by resolving coupled atmosphere–chemistry oscillations.
EA-LSTM: Evolutionary Algorithm-optimized LSTM (EA) features genetic algorithm-based hyperparameter tuning for enhanced prediction accuracy.
As evidenced by the PM2.5 prediction results in Figure 16 and Table 3, the proposed CSST-AQP model demonstrates superior performance across all evaluation metrics (RMSE = 8.63, MAE = 8.02, MAPE = 8.01%, R2 = 0.92), achieving statistically significant improvements over other state-of-the-art models. Specifically, it outperforms both conventional baseline architectures (LSTM and TCN) and contemporary state-of-the-art variants (EMD-LSTM, EMD-TCN, CEMD-LSTM, and EA-LSTM) by a 13.09–42.3% relative improvement. Although TCN and LSTM demonstrate lower parameters (15,612, 11,340 vs. 86,267), CSST-AQP achieves superior prediction accuracy through multi-scale IMF decomposition (R2 = 0.92). Additionally, the prediction curves of CSST-AQP in Figure 16 show minimal deviation from true values over 1–14 days. The consistent statistical superiority (p < 0.01, paired T-test) suggests CSST-AQP’s strong potential for operational deployment in environmental monitoring systems, particularly where accurate multi-scale pollution forecasting is critical for public health interventions.
This result confirms the model’s robust capabilities in multivariate atmospheric pollutant forecasting across multiple temporal horizons, attributable to its novel coupled cross-correlated (Pearson r > 0.78) spatiotemporal feature fusion by LSTF-Net. Moreover, CSST-AQP integrates parallel multi-scale feature extraction through ATSE-Net and CSSP-Net, which significantly improves temporal pattern capture efficiency. Specifically, CSSP-Net enables simultaneous processing of both high-frequency details via CWT and long-term trends through GAF encoding, whereas LSTM’s recurrent nature inherently causes gradient decay in long sequences. The ATSE-Net adaptive encoding strategy enhances discriminative feature learning to capture information about low-frequency components of pollutants and improve accuracy.

Compared with Traditional Models

Furthermore, the traditional machine learning approaches, RF [9] and SVM [10], serve as reference benchmarks compare with CSST-AQP under identical training conditions. Figure 17 compares the 7-day PM2.5 prediction trajectories of CSST-AQP, SVM, and RF against truth measurements, and the prediction error values and other metrics for each model are shown in Table 4.
As evidenced by the comparative analysis in Figure 17 and Table 4, CSST-AQP demonstrates superior predictive performance over conventional models (SVM/RF), achieving 29.6–39.4% higher accuracy (RMSE = 8.63 vs. 38.9/27.3. This enhancement originates from the LSTF preprocessing network, which encodes adaptive features to amplify strongly correlated spatiotemporal patterns while suppressing noise components. In addition, the static architectures of SVM and RF models struggle to adapt to dynamic spatiotemporal coupling relationships and exhibit limitations in long-sequence prediction tasks.

4.5. Ablation Experiments

4.5.1. Module Ablation

In this section, we perform ablation experiments on the sub-networks in the proposed CSST-AQP model, and the results are shown in Figure 18 and Table 5. We conduct systematic ablation studies to quantify the contribution of each component in the CSST-AQP framework. Under identical training protocols, three critical modules are incrementally ablated for 60 h AQI prediction.
1.
LSTF-Net Removal:
(1): Performance degradation: 312.7%↑ MAE; 225.8%↑ RMSE (vs. complete CSST-AQP model);
(2): Mechanism disruption: loses long-term spatiotemporal feature extraction (60 h patterns) and eliminates cross-modal correlation preprocessing (R2 = 0.93→0.81).
The ablation of LSTF-Net markedly reduces the CSST-AQP’s effectiveness in two key aspects: (1) adaptive spatiotemporal encoding for capturing pollutant–station interdependencies, and (2) hierarchical feature fusion—losing 83% of strongly correlated cross-component interactions (p > 0.7) due to disabled attention CAM-1 and CAM-2.
2.
CSSP-Net Ablation
(1): Short-term resilience: 0–24 h MAE maintains ±6.2% fluctuation.
(2): Critical degradation phase: 24–48 h: 19.5%↑ RMSE; 48–60 h: 91.8%↑ RMSE.
(3): Spatial feature loss: inter-station diffusion modeling capacity ↓62%.
CSSP-Net integrates geometric signal encoding (GAF/CWT) with adaptive residual fusion, achieving superior spatiotemporal coherence preservation (0–24 h MAE maintains a ±6.2% fluctuation) compared to conventional single-modality approaches. Ablation studies quantitatively confirm its essential role in maintaining prediction stability during complex pollution episodes.
3.
ATSE-Net Exclusion
(1): Immediate impact: 12.4%↑ MAE at the first prediction step (t + 1); this is important for sequence characterization in adjacent time ranges.
(2): Temporal recognition impairment: periodic pattern detection accuracy ↓57%; the recognition performance of temporal feature correlation is weakened (R2 = 0.93→0.83) at 60 h.
(3): Cumulative effect: error propagation amplifies, and ATSE-Net proves vital for immediate responsiveness, with its removal triggering a 3.6% MAE increase at the initial prediction step (t + 1).
The ablation of ATSE-Net critically undermines the spectrally enhanced STE module’s capability to process transient temporal patterns, while the absence of its adaptive T-Filter compromises discriminative feature extraction. This dual degradation mechanism fundamentally limits CSST-AQP’s capacity to isolate physically meaningful spatiotemporal interactions from complex environmental signals.
As evidenced in Figure 18 and Table 5, the complete model achieves an optimal balance between immediate responsiveness (<2% deviation at t + 1) and long-term stability (<3.3% error growth/hour). LSTF-Net proves the most critical (0.12 reduction on R2), particularly for modeling atmospheric transport inertia. CSSP-Net’s spatial hierarchy maintains 83% feature completeness through 60 h, while ATSE-Net’s adaptive temporal kernels account for 85% of short-term accuracy improvements.

4.5.2. Hyperparameter Analysis

In this section, we conduct a comprehensive ablation study to evaluate the hyperparameter sensitivity of the proposed CSST-AQP framework. The model architecture incorporates six tunable hyperparameters that require careful configuration: (1) sliding window size (W), (2) sliding step size (O), (3) convolutional kernel quantity (Ck), (4) batch size (Bs), (5) dilation factor (Df), and (6) dropout rate (Dr). Systematic experimental configurations and results are detailed in Table 6.
The ablation study results in Table 6 reveal critical insights into hyperparameter sensitivity within the CSST-AQP framework: (1) W exhibits an inverted U-shaped relationship with prediction accuracy, demonstrating a 1.9% RMSE reduction at W = 32 compared to baseline. The results and evidence suggest optimal parameterization within W [ 32 , 64 ] complemented by an overlapping stride O [ 4 , 8 ] to balance temporal resolution and computational efficiency. (2) Ck and Df show a synergistic effect, where increasing Ck to more than 32 units together with D f = 4 results in a significant performance improvement ( p < 0.005 , t-test). (3) Dr requires precise calibration, where excessive regularization ( D r > 0.4 ) induces 18.9% performance deterioration. Conversely, moderate retention ( D r [ 0.2 , 0.35 ] ) enhances the generalization capacity by 9.2 ± 1.1% through effective prevention of co-adaptation. (4) B S = 256 achieves an optimal trade-off between convergence speed and algorithmic time complexity.

4.6. Parameter Sensitivity and Robustness

4.6.1. Different Data Sources

To further investigate the parameter sensitivity and robustness of the proposed CSST-AQP framework, we conducted experiments under the GZR dataset and the SGR dataset, respectively. Figure 19 demonstrates the prediction effect of the CSST-AQP model for the time range of 0–48 h for the GZR dataset, the SGR dataset, and the integration of the two datasets, respectively.
In Figure 19, the proposed CSST-AQP model maintains consistent predictive accuracy across diverse datasets, with R2 values consistently constrained within the narrow range of 0.92–0.97. This performance stability, as evidenced by the overlapping learning curves across different experimental groups, indicates remarkable model robustness against dataset variations. Particularly, the model is able to preserve prediction fidelity when handling datasets with distinct statistical distributions and noise patterns, suggesting strong generalization capabilities inherent in our architectural design.

4.6.2. Parameter Sensitivity

As shown in Figure 20, a comprehensive parameter sensitivity analysis was conducted to evaluate critical components of our framework, including (1) the attenuation coefficient ξ localizing the dynamic weigh matrix in the LSTF-Net, (2) the adaptive thermal balancing parameters ( α ,   β ) governing snow ablation optimization in ATSE-Net, and (3) the environmental monitoring region sample size Nr.
Figure 20 further illustrates the temporal evolution of prediction accuracy, demonstrating characteristic MAE variation patterns ( Δ MAE [ 1.23 , 4.57 ] ) across three operational phases: short-term (1–12 h), medium-term (12–48 h), and extended-term (48–60 h) AQI forecasting. Notably, the phase-dependent sensitivity profiles suggest that dynamic parameter dominance mechanisms’ ablation parameters (α/β) predominantly influence short-term predictions (R2 = 0.83), while attenuation coefficient ξ shows a stronger correlation with medium-term accuracy. Overall, we can observe that the model is robust to small jumps in parameters, with the best performance at ξ of 0.6, α of 0.6, and β of 1.2. The model achieves stable performance when the number of monitoring stations reaches 3 or more. In addition, attenuation coefficient ξ plays an important role in LSTF-Net to extract highly correlated spatiotemporal features, which in turn affects the long-term prediction accuracy.

4.6.3. Noise Sensitivity

In the original dataset, we artificially added Gaussian noise with a mean of 0 and variance of 0.01–0.05 and salt-and-pepper noise with a noise ratio of 1–16% to validate the model’s noise immunity performance, and the results are shown in Figure 21.
As shown in Figure 21, the model maintains stable MAE values under both Gaussian and salt-and-pepper noise perturbations. This robustness can be attributed to the CSST-AQP framework’s advanced preprocessing module, which systematically identifies and filters anomalous data points while preserving critical pollution pattern signatures.

5. Discussion

This study systematically evaluates machine learning (RF and SVR) and deep learning models (LSTM, TCN, and CEMD-LSTM) for predicting six major air pollutants (PM2.5, PM10, O3, NO2, SO2, and CO), incorporating meteorological variables (temperature, humidity, and wind speed) to enhance prediction accuracy. In particular, the proposed CSST-AQP model outperforms other models with an RMSE of 8.63, RAE of 8.02, and R2 of 0.92 for pollutants tested over the next 400 h period due to the fact that the CSST-AQP model emphasizes multi-scale fusion of features and weighted learning of spatiotemporally relevant features. Compared to other models, the CSST-AQP framework divides the prediction task into a screening encodability phase and a learning confirmation phase. In the screening encodability phase, a new LSTF-Net module is proposed to effectively capture spatial and temporal dependencies; in the learning confirmation phase, the hidden state learning probability distributions are extracted from the new features obtained from encoding, thus capturing the uncertainty of the input data and improving the model performance.
This framework enables practical applications, like cross-regional PM2.5 forecasting, supporting early warning systems for public health protection (e.g., mask advisories during pollution peaks). In future work, our model needs to further disaggregate the correlations between different pollutants and spatial factors and conduct an in-depth study on the spatial error aggregation factors and seasonality of pollution sources in other countries or regions.

6. Conclusions

This paper presents a novel air quality prediction framework that comprehensively models multi-scale spatiotemporal dependencies through hierarchical feature fusion. Specifically, the entire model network adopts a dual-channel architecture: the spatial feature channel extracts the non-stationary correlation features among monitoring stations via 2D graph convolution CSSP-Net, and the temporal feature channel resolves the cross-scale dependence patterns of multivariate time series via adaptive temporal sequence ATSE-Net. Our framework demonstrates three distinctive advantages: first, it achieves seamless integration of station-specific historical patterns and inter-station spatial synergies through dual-stream processing; second, the hierarchical dilation structure enables simultaneous learning of hourly, daily, and weekly temporal dependencies; third, the adaptive graph learning module dynamically adjusts neighborhood weights based on real-time pollution dispersion patterns. This design enables the model to process remote dependencies in spatiotemporal dimensions in parallel, which significantly improves long-term prediction accuracy while ensuring computational efficiency. The experimental results across 14 major Chinese regions demonstrate a 22.5% improvement in 72 h PM2.5 forecasting accuracy compared to other advanced LSTM and TCN models while maintaining computational efficiency suitable for operational deployment.

Author Contributions

Conceptualization, C.W. and Z.L.; methodology, C.W.; software, C.W. and X.Z.; validation, Z.L.; formal analysis, Y.X.; investigation, X.Z.; resources, G.D.; writing—original draft preparation, Z.L.; writing—review and editing, C.W.; visualization, J.W.; supervision, G.D.; funding acquisition, J.W. and G.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the National Natural Science Foundation of China (Grant Nos. 62041106); Guangdong Province, social development science and technology collaborative innovation system construction project (grant No. P0000876021 and P0000876023); Quality Engineering Project of Guangdong Songshan Vocational and Technical College (2023JNDS02); and Guangdong Provincial Science and Technology Characteristic Innovation Project (2024KTSCX334).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, X.H.; Tee, K.; Elnahass, M.; Ahmed, R. Assessing the environmental impacts of renewable energy sources a case study on air pollution and carbon emissions in China. J. Environ. Econ. Manag. 2023, 345, 118525. [Google Scholar]
  2. Pan, K.; Lu, J.; Li, J.; Xu, Z. A hybrid autoformer network for air pollution forecasting based on external factor optimization. Atmosphere 2023, 14, 869. [Google Scholar] [CrossRef]
  3. Zheng, Y.; Liu, F.R.; Hsieh, H.P. U-Air: When urban air quality inference meets big data. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 1436–1444. [Google Scholar]
  4. Gu, K.; Qiao, J.; Lin, W. Recurrent air quality predictor based on meteorology- and pollution-related factors. IEEE Trans. Ind. Inform. 2018, 14, 3946–3955. [Google Scholar]
  5. Steinfeld, J.I. Atmospheric chemistry and physics from air pollution to climate change. Environ. Sci. Policy 1998, 40, 26. [Google Scholar]
  6. Flatøy, F.; Hov, O.; Schlager, H. Chemical forecasts used for measurement flight planning during POLINAT 2. Geophys. Res. Lett. 2000, 27, 951–954. [Google Scholar]
  7. Osipov, A.V.; Pleshakova, E.S.; Gataullin, S.T. Production processes optimization through machine learning methods based on geophysical monitoring data. Comput. Optim. 2024, 48, 633–642. [Google Scholar]
  8. Rizos, K.; Meleti, C.; Evagleopoulos, V.; Melas, D. A machine learning modelling approach to characterize the background pollution in the Western Macedonia region in northwest Greece. Atmos. Pollut. Res. 2023, 14, 101877. [Google Scholar]
  9. Mendez, M.; Merayo, M.G.; Nuez, M. Machine learning algorithms to forecast air quality a survey. Artif. Intell. Rev. 2023, 56, 10031–10066. [Google Scholar]
  10. Liu, H.; Yan, G.X.; Duan, Z.; Chen, C. Intelligent modeling strategies for forecasting air quality time series: A review. Appl. Soft Comput. 2021, 102, 106957. [Google Scholar]
  11. Zhang, Y.P.; Chen, D.; Luo, Y. Low-SOC estimation of aluminum-air battery based on HSMM. J. Univ. Electron. Sci. Technol. China 2017, 46, 380–385. [Google Scholar]
  12. Mampitiya, L.; Rathnayake, N.; Leon, L.P.; Mandala, V.; Azamathulla, H.M.; Shelton, S.; Hoshino, Y.; Rathnayake, U. Machine learning techniques to predict the air quality using meteorological data in two urban areas in Sri Lanka. Environments 2023, 10, 141. [Google Scholar] [CrossRef]
  13. Mampitiya, L.; Rathnayake, N.; Hoshino, Y.; Rathnayake, U. Forecasting PM10 levels in Sri Lanka a comparative analysis of machine learning models PM10. J. Hazard. Mater. Adv. 2024, 13, 100395. [Google Scholar] [CrossRef]
  14. Athira, V.; Geetha, P.; Vinayakumar, R.; Soman, K.P. DeepAirNet applying recurrent networks for air quality prediction. Procedia Comput. Sci. 2018, 132, 1394–1403. [Google Scholar]
  15. Guo, Z.Y.; Yang, C.Y.; Wang, D.S.; Liu, H.B. A novel deep learning model integrating CNN and GRU to predict particulate matter concentrations. Process Saf. Environ. 2023, 173, 604–613. [Google Scholar] [CrossRef]
  16. Elbaz, K.; Shaban, W.M.; Zhou, A.N.; Shen, S.L. Real time image-based air quality forecasts using a 3D-CNN approach with an attention mechanism. Chemosphere 2023, 333, 138867. [Google Scholar] [CrossRef]
  17. Tran, H.D.; Huang, H.Y.; Yu, J.Y.; Wang, S.H. Forecasting hourly PM2.5 concentration with an optimized LSTM model. Atmos. Environ. 2023, 315, 120161. [Google Scholar] [CrossRef]
  18. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In Proceedings of the 29th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 802–810. [Google Scholar]
  19. Zaini, N.; Ahmed, A.N.; Ean, L.W.; Chow, M.F.; Malek, M.A. Forecasting of fine particulate matter based on LSTM and optimization algorithm. J. Clean. Prod. 2023, 427, 139233. [Google Scholar] [CrossRef]
  20. Wang, C.; Huang, S.; Zhang, C. Short-term traffic flow prediction considering weather factors based on optimized deep learning neural networks: Bo-GRA-CNN-BiLSTM. Sustainability 2025, 17, 2576. [Google Scholar] [CrossRef]
  21. Ameri, R.; Hsu, C.C.; Band, S.S.; Zamani, M.; Shu, C.M.; Khorsandroo, S. Forecasting PM 2.5 concentration based on integrating of CEEMDAN decomposition method with SVM and LSTM. Ecotoxicol. Environ. Saf. 2023, 266, 115572. [Google Scholar] [CrossRef]
  22. Tishya, M.; Anitha, A. Hybridization of rough set-wrapper method with regularized combinational LSTM for seasonal air quality index prediction. Neural Comput. Appl. 2024, 36, 2921–2940. [Google Scholar]
  23. Wang, X.H.; Zhang, S.; Chen, Y.; He, L.Y.; Ren, Y.M.; Zhang, Z.; Li, J.; Zhang, S.Q. Air quality forecasting using a spatiotemporal hybrid deep learning model based on VMD–GAT–BiLSTM. Sci Rep. 2024, 14, 17841. [Google Scholar]
  24. Rabie, R.; Asghari, M.; Nosrati, H.; Niri, M.E.; Karimi, S. Spatially resolved air quality index prediction in megacities with a CNN-Bi-LSTM hybrid framework. Sustain. Cities Soc. 2024, 109, 105537. [Google Scholar]
  25. Fan, P.D.; Wang, D.; Wang, W.; Zhang, X.Y.; Sun, Y.Y. A novel multi-energy load forecasting method based on building flexibility feature recognition technology and multi-task learning model integrating LSTM. Energy 2024, 308, 132976. [Google Scholar]
  26. Huang, X.H.; Li, Y.F.; Wang, X.W. Integrating a multi-variable scenario with Attention-LSTM model to forecast long-term coastal beach erosion. Sci. Total Environ. 2024, 954, 176257. [Google Scholar]
  27. Wang, Y.; Liu, K.; He, Y.; Wang, P.; Chen, Y.; Xue, H.; Huang, C.; Li, L. Enhancing air quality forecasting a novel spatio-temporal model integrating graph convolution and multi-head attention mechanism. Atmosphere 2024, 15, 418. [Google Scholar] [CrossRef]
  28. Hu, W.; Zhang, Z.; Zhang, S.Q.; Chen, C.M.; Yuan, J.W.; Yao, J.; Zhao, S.C.; Guo, L. Learning spatiotemporal dependencies using adaptive hierarchical graph convolutional neural network for air quality prediction. J. Clean. Prod. 2024, 459, 142541. [Google Scholar]
  29. Chen, Y.S.; Huang, L.; Xie, X.D.; Liu, Z.X.; Hu, J.L. Improved prediction of hourly PM2.5 concentrations with a long short-term memory and spatiotemporal causal convolutional network deep learning model. Sci. Total Environ. 2024, 912, 168672. [Google Scholar]
  30. Hu, Y.X.; Li, Q.; Shi, X.D.; Yan, J.Y.; Chen, Y.T. Domain knowledge-enhanced multi-spatial multi-temporal PM2.5 forecasting with integrated monitoring and reanalysis data. Environ. Int. 2024, 192, 108997. [Google Scholar] [PubMed]
  31. Dey, S. Urban air quality index forecasting using multivariate convolutional neural network based customized stacked long short-term memory model. Process Saf. Environ. 2024, 191, 375–389. [Google Scholar]
  32. Zhu, J.M.; Zheng, P.; Niu, L.L.; Chen, H.Y.; Wu, P. An enhanced interval-valued PM2.5 concentration forecasting model with attention-based feature extraction and self-adaptive combination technology. Expert Syst. Appl. 2025, 264, 125867. [Google Scholar]
  33. Pande, C.B.; Kushwaha, N.L.; Alawi, O.A.; Sammen, S.S.; Sidek, L.M.; Yaseen, Z.M.; Pal, S.C.; Katipoglu, O.M. Daily scale air quality index forecasting using bidirectional recurrent neural networks: Case study of Delhi, India. Environ. Pollut. 2024, 351, 124040. [Google Scholar]
  34. Wang, K.N.; Yang, T.N.; Kong, S.S.; Li, M.D. Air quality index prediction through Time GAN data recovery and PSO-optimized VMD-deep learning framework. Appl. Soft Comput. 2025, 170, 112626. [Google Scholar]
  35. Mohammadzadeh, A.K.; Salah, H.; Jahanmahin, R.; Hussain, A.A.; Masoud, S.; Huang, Y.X. Spatiotemporal integration of GCN and E-LSTM networks for PM2.5 forecasting. Mach. Learn. Appl. 2024, 15, 100521. [Google Scholar]
  36. Ma, M.B.; Xie, P.; Teng, F.; Wang, B.; Ji, S.G.; Zhang, J.B.; Li, T.R. HiSTGNN: Hierarchical spatio-temporal graph neural network for weather forecasting. Inf. Sci. 2023, 648, 119580. [Google Scholar]
  37. Yang, W.H.; Li, H.M.; Wang, J.Z.; Ma, H.Y. Spatiotemporal feature interpretable model for air quality forecasting. Ecol. Indic. 2024, 167, 112609. [Google Scholar]
  38. Che, J.X.; Hu, K.; Xia, W.X.; Xu, Y.F.; Li, Y.R. Short-term air quality prediction using point and interval deep learning systems coupled with multi-factor decomposition and data-driven tree compression. Appl. Soft Comput. 2024, 166, 112191. [Google Scholar]
  39. Zhang, M.L.; Sun, L.F.; Yang, J.; Zou, Y.S. A shared multi-scale lightweight convolution generative network for few-shot multivariate time series forecasting. Appl. Soft. Comput. 2024, 167, 112420. [Google Scholar]
  40. Chen, Y.N.; Wu, Y.H.; Zhang, S.G.; Kee, Y.; Huang, J.; Shi, D.F.; Hu, S.X. Regional PM2.5 prediction based on hybrid directed graph neural network and spatio-temporal fusion of meteorological factors. Environ. Pollut. 2024, 366, 125404. [Google Scholar]
  41. Du, Z.; Wu, S.; Huang, D.; Li, W.; Wang, Y. Spatio-temporal encoder-decoder fully convolutional network for video-based dimensional emotion recognition. IEEE Trans. Affect. Comput. 2021, 12, 565–578. [Google Scholar]
  42. Wang, Z.G.; Oates, T. Imaging time-series to improve classification and imputation. arXiv 2015, arXiv:1506.00327v1. [Google Scholar]
  43. Deng, L.Y.; Liu, S.Y. Snow ablation optimizer a novel metaheuristic technique for numerical optimization and engineering design. Expert. Syst. Appl. 2023, 225, 120069. [Google Scholar]
  44. Air Quality Historical Data Platform. Available online: https://aqicn.org/data-platform/register/ (accessed on 10 September 2024).
  45. Sundaramoorthy, R.A.; Ananth, A.D.; Seerangan, K.; Nandagopal, M.; Balusamy, B.; Selvarajan, S. Implementing heuristic-based multiscale depth-wise separable adaptive temporal convolutional network for ambient air quality prediction using real time data. Sci. Rep. 2024, 14, 18437. [Google Scholar]
  46. Huang, Y.; Yu, J.; Dai, X.; Huang, Z.; Li, Y. Air-quality prediction based on the EMD–IPSO–LSTM combination model. Sustainability 2022, 14, 4889. [Google Scholar] [CrossRef]
  47. Neeraj; Mathew, J.; Behera, R.K. EMD-Att-LSTM a data-driven strategy combined with deep learning for short-term load forecasting. J. Mod. Power Syst. Clean Energy 2022, 10, 1229–1240. [Google Scholar]
  48. Wang, Z.C.; Chen, H.Y.; Zhu, J.M.; Ding, Z.N. Daily PM2.5 and PM10 forecasting using linear and nonlinear modeling framework based on robust local mean decomposition and moving window ensemble strategy. Appl. Soft Comput. 2022, 114, 108110. [Google Scholar]
  49. Li, Y.R.; Zhu, Z.F.; Kong, D.Q.; Han, H.; Zhao, Y. EA-LSTM evolutionary attention-based LSTM for time series prediction. Knowl. Based Syst. 2019, 181, 104785. [Google Scholar]
Figure 1. Architecture of an STE.
Figure 1. Architecture of an STE.
Atmosphere 16 00429 g001
Figure 2. Architecture of an STE. (a): hourly average AQI; (b): distance vs. AQI difference; (c): spatiotemporal distribution of AQI.
Figure 2. Architecture of an STE. (a): hourly average AQI; (b): distance vs. AQI difference; (c): spatiotemporal distribution of AQI.
Atmosphere 16 00429 g002
Figure 3. The framework of CSST-AQP; differently colored blocks represent different sub-networks and modules.
Figure 3. The framework of CSST-AQP; differently colored blocks represent different sub-networks and modules.
Atmosphere 16 00429 g003
Figure 4. The LSTF-Net framework.
Figure 4. The LSTF-Net framework.
Atmosphere 16 00429 g004
Figure 5. The structure of CSSP-Net.
Figure 5. The structure of CSSP-Net.
Atmosphere 16 00429 g005
Figure 6. The structure of ATSE-Net.
Figure 6. The structure of ATSE-Net.
Atmosphere 16 00429 g006
Figure 7. The loss curves for the training and testing processes.
Figure 7. The loss curves for the training and testing processes.
Atmosphere 16 00429 g007
Figure 8. The correlation coefficient of determination curves under training and testing.
Figure 8. The correlation coefficient of determination curves under training and testing.
Atmosphere 16 00429 g008
Figure 9. Visualization results for each AQI component after network coding.
Figure 9. Visualization results for each AQI component after network coding.
Atmosphere 16 00429 g009
Figure 10. Visualization results for feature maps: (a) SHAP feature dependency graph; (b) temporal analysis of PM2.5.
Figure 10. Visualization results for feature maps: (a) SHAP feature dependency graph; (b) temporal analysis of PM2.5.
Atmosphere 16 00429 g010
Figure 11. Distribution of correlation coefficients of air quality characteristics: (a) embedding without spatial features; (b) embedding with spatial features.
Figure 11. Distribution of correlation coefficients of air quality characteristics: (a) embedding without spatial features; (b) embedding with spatial features.
Atmosphere 16 00429 g011
Figure 12. Curves of true versus predicted values of PM2.5 and PM10: (a) PM2.5; (b) PM10.
Figure 12. Curves of true versus predicted values of PM2.5 and PM10: (a) PM2.5; (b) PM10.
Atmosphere 16 00429 g012
Figure 13. Curves of true versus predicted values of O3 and NO2: (a) O3; (b) NO2.
Figure 13. Curves of true versus predicted values of O3 and NO2: (a) O3; (b) NO2.
Atmosphere 16 00429 g013
Figure 14. Curves of truth versus predicted values of SO2 and CO: (a) SO2; (b) CO.
Figure 14. Curves of truth versus predicted values of SO2 and CO: (a) SO2; (b) CO.
Atmosphere 16 00429 g014
Figure 15. Residual scatter plot of the true and predicted values of CO: (a) before data correction; (b) data correction.
Figure 15. Residual scatter plot of the true and predicted values of CO: (a) before data correction; (b) data correction.
Atmosphere 16 00429 g015
Figure 16. Comparison of the prediction performance of different models for PM2.5.
Figure 16. Comparison of the prediction performance of different models for PM2.5.
Atmosphere 16 00429 g016
Figure 17. Comparison of the prediction performance of traditional models for PM2.5.
Figure 17. Comparison of the prediction performance of traditional models for PM2.5.
Atmosphere 16 00429 g017
Figure 18. Predicted results of AQI under different modules.
Figure 18. Predicted results of AQI under different modules.
Atmosphere 16 00429 g018
Figure 19. Predictive performance of the CSST-AQP model for the 0–48 h with different datasets.
Figure 19. Predictive performance of the CSST-AQP model for the 0–48 h with different datasets.
Atmosphere 16 00429 g019
Figure 20. Parameter sensitivity and robustness of the CSST-AQP model.
Figure 20. Parameter sensitivity and robustness of the CSST-AQP model.
Atmosphere 16 00429 g020
Figure 21. Predictive performance analysis of the model under different noise attacks: (a) Gaussian noise; (b) salt-and-pepper noise.
Figure 21. Predictive performance analysis of the model under different noise attacks: (a) Gaussian noise; (b) salt-and-pepper noise.
Atmosphere 16 00429 g021
Table 1. Structure and parameters of the datasets.
Table 1. Structure and parameters of the datasets.
DatasetsGZRSGR
Number of air quality stations12163
Number of air quality records4,369,1021,103,144
Average AQI58.670.1
Number of regions Nr113
Table 2. Overall performance of the CSST-AQP model for different pollutants.
Table 2. Overall performance of the CSST-AQP model for different pollutants.
VariableRMSEMAEMAPE (%)R2Confidence Level (95%)PACF
PM2.58.638.128.010.9296.6%0.15
PM106.836.066.210.9395.1%0.08
O37.838.068.190.9295.8%0.21
NO28.698.238.130.9295.3%0.12
SO26.117.096.790.9395.3%0.09
CO9.138.919.560.9195.9%0.18
Table 3. Results of performance comparison between CSST-AQP and other advanced air prediction algorithms.
Table 3. Results of performance comparison between CSST-AQP and other advanced air prediction algorithms.
AlgorithmsRMSEMAEMAPE (%)R2ParametersTraining Time (s)
LSTM14.9612.9812.870.8315,61252.1
TCN13.8711.5711.930.8511,340123.6
EMD-LSTM12.1211.1911.020.87140,508288.5
EMD-TCN10.5710.6210.650.90102,060251.3
EA-LSTM9.568.879.670.9034,941136.8
CEMD-LSTM9.938.748.630.91102,060310.7
CSST-AQP8.638.028.010.9286,267190.6
Table 4. Results of performance comparison between CSST-AQP and traditional models.
Table 4. Results of performance comparison between CSST-AQP and traditional models.
AlgorithmsRMSEMAEMAPE (%)R2ParametersTraining Time (s)
SVM38.937.637.10.66Na31.3
RF25.325.122.40.71Na36.5
CSST-AQP8.638.028.010.9286,267190.6
Table 5. Performance of ablation experiments.
Table 5. Performance of ablation experiments.
TimeAlgorithmsRMSEMAEMAPE (%)R2
0–24 hDropout LSTF-Net25.4723.1219.860.83
Dropout CSSP-Net12.1415.279.010.89
Dropout TSE-Net26.3824.7218.930.84
CSST-AQP9.157.126.010.93
24–60 hDropout LSTF-Net28.1227.3920.390.81
Dropout CSSP-Net16.5618.0510.130.85
Dropout TSE-Net20.0320.2315.110.83
CSST-AQP12.6311.3910.820.92
Table 6. Hyperparameter ablation study results (bold indicates baseline configuration).
Table 6. Hyperparameter ablation study results (bold indicates baseline configuration).
HyperparameterTested Values Δ RMSEp-Value Δ Training TimeRecommended Range
W16, 32, 64, 128+7.3%, −1.9%,
baseline, +6.8%
0.01, 0.005,
<0.001, 0.12
+8%, 5%,
baseline, −3%
32–64
O2, 4, 8, 16+8.6%, baseline,
−3.7%, +13.1%
0.03, 0.002,
0.003, 0.15
+9%, baseline,
−3%, −5%
4–8
Ck16, 32, 64,128+1.6%, baseline,
−1.1%, +5.5%
0.04, 0.008,
0.002, 0.20
−2%, baseline,
+6%, +15%
32–64
BS64, 128, 256, 512+2.6%, −3.9%,
baseline, +9.1%
0.01, 0.001,
<0.001, 0.09
−2%, −1%,
baseline, +1%
256–512
Df1, 2, 4, 8+8.6%, baseline,
−3.7%, +13.1%
0.02, 0.002,
0.001, 0.04
+6%, baseline,
−3%, +3%
2–6
Dr0.1, 0.2, 0.3, 0.4+8.6%, baseline,
−6.3%, +18.9%
0.06, 0.005,
0.001, 0.23
+8%, baseline,
−6%, −7%
0.2–0.35
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, C.; Lai, Z.; Xu, Y.; Zhu, X.; Wu, J.; Duan, G. Comprehensive Scale Fusion Networks with High Spatiotemporal Feature Correlation for Air Quality Prediction. Atmosphere 2025, 16, 429. https://doi.org/10.3390/atmos16040429

AMA Style

Wu C, Lai Z, Xu Y, Zhu X, Wu J, Duan G. Comprehensive Scale Fusion Networks with High Spatiotemporal Feature Correlation for Air Quality Prediction. Atmosphere. 2025; 16(4):429. https://doi.org/10.3390/atmos16040429

Chicago/Turabian Style

Wu, Chenyi, Zhengliang Lai, Yunwu Xu, Xishun Zhu, Jianhua Wu, and Guiqin Duan. 2025. "Comprehensive Scale Fusion Networks with High Spatiotemporal Feature Correlation for Air Quality Prediction" Atmosphere 16, no. 4: 429. https://doi.org/10.3390/atmos16040429

APA Style

Wu, C., Lai, Z., Xu, Y., Zhu, X., Wu, J., & Duan, G. (2025). Comprehensive Scale Fusion Networks with High Spatiotemporal Feature Correlation for Air Quality Prediction. Atmosphere, 16(4), 429. https://doi.org/10.3390/atmos16040429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop