Next Article in Journal
Research on an ID-PCA Early Fault Detection Method for Rolling Bearings
Next Article in Special Issue
Performance Evaluation of Planetary Boundary Layer Schemes in Simulating Structures of Wintertime Lower Troposphere in Seoul Using One-Hour Interval Radiosonde Observation
Previous Article in Journal
A Multidisciplinary Methodology for Technological Knowledge, Characterization and Diagnostics: Sandstone Facades in Florentine Architectural Heritage
Previous Article in Special Issue
Validating Aerosol Optical Depth Estimation Methods Using the National Institute of Environmental Research Operational Numerical Forecast Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting Fine-Grained Air Quality for Locations without Monitoring Stations Based on a Hybrid Predictor with Spatial-Temporal Attention Based Network

1
Department of Electrical Engineering, National Cheng Kung University, Tainan 70101, Taiwan
2
Institute of Computer, Communication Engineering, National Cheng Kung University, Tainan 70101, Taiwan
3
Department of Medical Imaging, Chi Mei Medical Center, Tainan 710402, Taiwan
4
Department of Health and Nutrition, Chia Nan University of Pharmacy and Science, Tainan 71710, Taiwan
5
Institute of Biomedical Sciences, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
6
College of Arts and Humanities, Swansea University, Swansea SA2 8PP, UK
7
Research Center for Information Technology Innovation, Academia Sinica, Taipei 115201, Taiwan
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(9), 4268; https://doi.org/10.3390/app12094268
Submission received: 23 March 2022 / Revised: 17 April 2022 / Accepted: 19 April 2022 / Published: 23 April 2022
(This article belongs to the Special Issue Advances in Monitoring and Modeling of Urban Air Quality)

Abstract

:
Air pollution in cities is a severe and worrying problem because it causes threats to economic development and health. Furthermore, with the development of industry and technology, rapid population growth, and the massive expansion of cities, the total amount of pollution emissions continue to increase. Hence, observing and predicting the air quality index (AQI), which measures fatal pollutants to humans, has become more and more critical in recent years. However, there are insufficient air quality monitoring stations for AQI observation because the construction and maintenance costs are too high. In addition, finding an available and suitable place for monitoring stations in cities with high population density is difficult. This study proposes a spatial-temporal model to predict the long-term AQI in a city without monitoring stations. Our model calculates the spatial-temporal correlation between station and region using an attention mechanism and leverages the distance information between all existing monitoring stations and target regions to enhance the effectiveness of the attention structure. Furthermore, we design a hybrid predictor that can effectively combine the time-dependent and time-independent predictors using the dynamic weighted sum. Finally, the experimental results show that the proposed model outperforms all the baseline models. In addition, the ablation study confirms the effectiveness of the proposed structures.

1. Introduction

In recent years, with the increasing development of industry and technology, the rapid growth of the population, and the massive expansion of cities, air pollution has become significantly worse and threatens both health and economy. For example, PM2.5, which consists of particles less than 2.5 μm, is fatal and leads to cardiovascular disease or respiratory illness [1]. Hence, an air quality index (AQI) is used by government agencies (https://www.airnow.gov/aqi/aqi-basics/) (accessed on 17 April 2022) to communicate to the public how polluted the air currently is or how polluted it is forecast to become. By monitoring AQI levels, the government and citizens can forecast air quality and thus take significant actions in advance to prevent being affected by air pollution. Even though the demand to forecast the air quality has increased, forecasting air quality is still challenging for several reasons. Air quality values vary significantly over time and across locations because of various complicated factors, such as meteorological effects, surrounding land usages, and the daily behaviors of people. Moreover, sudden changes and accidents might affect the air quality as well and thus make it difficult for a general air quality prediction model to accurately forecast the future. For example, an industrial accident such as a fire in a factory might cause the concentration of pollutants to increase rapidly. Conversely, the outbreak of widespread diseases like COVID-19 [2] might decrease pollution emissions because of substantially reduced human activities.
In general, air quality prediction in cities has focused on two conditions. One is to predict future air pollution in areas with monitoring stations, known as the air pollution forecasting problem. This problem has been widely discussed and has had many high-performance solutions in recent research [3,4]. The other condition is to estimate air pollution in areas without monitoring stations in the current hour, also known as fine-grained air pollution estimation [5,6]. The second condition is also more challenging because there is no sufficient labeled data to train an effective module [5]. In the real world, monitoring stations are usually insufficient and sparsely distributed in a large city because of the high construction and maintenance costs. In addition, suitable places in densely populated cities are rare and unavailable. As a result, people have to analyze and predict the current fine-grained AQI with limited historical data.
Traditionally, researchers [7,8,9] used historical air pollution data crawled from established monitoring stations to predict air quality. However, with the increasing amount of open data and sensors, people can collect more heterogeneous and dynamic data highly correlated with air pollutants. The studies [10,11] utilized these data and developed deep-learning-based models that outperform the traditional methods in air quality modeling. However, deep learning-based methods were still unable to accurately predict fine-grained air quality throughout a city because of the scarcity of air pollution monitoring stations and the incompleteness of historical data. Therefore, recently, it has become popular to discuss how existing monitoring stations can predict future air pollution levels and assess the current air pollution levels in minuscule urban areas. For example, the works [12,13] used past air quality records as a reference to calculate the spatial quality relationship between “areas with monitoring stations” and “areas without monitoring stations” and applied this relationship to predictions. Most studies have attempted to achieve the two goals stated above separately or in one model [14,15,16].
In this study, we focus on forecasting fine-grained air quality for every region, including the one without a monitoring station. The proposed model has three major components: feature extractor, inversed distance attention layer, and hybrid predictor. We use the attention layer to learn the relationship between the embedding results of the light feature extractor and the general feature extractor. Then, we apply the inversed distance as an amended coefficient to strengthen the influence of high correlation monitoring stations. Finally, the hybrid predictor generates long-term air quality predictions. To summarize, the main contributions of this study are as follows:
  • To the best of our knowledge, we are the first to combine the air quality forecasting of monitoring stations, and predict the air quality of sites without monitoring stations, forecast the long-term fine-grained air quality without monitoring stations and realize the goal of long-term prediction for locations without monitoring stations.
  • This study proposes a hybrid predictor with two complementary data flows that can perform well in long-term prediction for the locations without monitoring stations.
  • The experimental results show that the proposed model outperforms the baseline methods. In addition, the ablation study confirms the effectiveness of the proposed structure.
The rest of this paper is organized as follows. In Section 2, we introduce the related work. We formulate the problem and introduce the data in Section 3. In Section 4, we introduce our proposed model in detail. In Section 5, we experimentally evaluate the effectiveness of the proposed solution by comparing different methods. Finally, we conclude our work with a discussion about future work in Section 6.

2. Related Work

Air pollution prediction has become a popular research topic in recent years. Researchers have proposed several methods to fulfill the objectives. In this section, we introduce the previous works of air quality forecasting, fine-grained air quality estimation, and the combination of air quality forecasting and fine-grained air quality estimation.
  • Air Quality Forecasting: Previous studies have proposed many data-driven approaches for air quality modeling. For example, the work [7] introduced autoregressive integrated moving average (ARIMA), which is a famous and widely adopted data-driven model for air pollution forecast at monitoring stations. In [8], the authors proposed a traditional physical-based numerical model that used historical air pollution data [9] to formulate the urban pollution process and pollutant dispersion. Moreover, studies [3,4,17] used the combination of spatial and temporal features to model air quality forecasting. Recently, big data and neural network techniques have improved the traditional methods and achieved state-of-the-art performance in urban air quality forecasting [10,11]. Graph convolution neural network (GCN) has become particularly popular in many works. The GCN-based air quality prediction takes monitoring stations as nodes of the graph and uses static features to calculate the node embedding to forecast the predicted values [13].
  • Fine-Grained Estimation: Several studies have proposed to use basic interpolation methods such as spatial averaging, nearest neighbor, inversed distance weighting (IDW), and kriging to solve the fine-grained estimation of air quality for locations without monitoring stations [18,19]. In addition, the work [5] proposed an affinity graph framework to infer real-time air quality of a location without a monitoring station, and recommend the best locations for new monitoring stations. As air quality forecasting, several studies have investigated interpolation methods using the combination of spatial and temporal features [3,6,20]. In addition, a neural network with an attention mechanism [12] has become popular for estimating the air quality of a region without monitoring stations.
  • Joint Air Quality Forecasting and Fine-Grained Estimation: Few studies have attempted to achieve both air quality forecasting and fine-grained estimation. The work [15] used the combination of GCN and LSTM networks to learn the spatial–temporal correlation and then predict these two targets. The work in [16] proposed a hybrid model to perform feature selection, air quality forecasting, and fine-grained estimation jointly in one single model. Moreover, the studies in [14,21] show that using air quality forecasting and fine-grained estimating as multi-task learning can improve the generalization performance by leveraging the commonalities and differences between these two tasks.
Although previous studies [14,16] have proposed methods to jointly model air quality forecasting and fine-grained estimation and enhance the performance, there is still much for improvement. For example, these methods do not solve long-term prediction challenges in regions without sufficient historical data or monitoring stations. To solve this problem, we propose a model that can capture the complex spatial and temporal interaction between air pollution and geographical data, and then predict the air quality in places without sufficient historical data.

3. Preliminaries

3.1. Problem Statement

We divide cities into disjoint grids (e.g., 2.2 km × 2.2 km). Let U and S represent the lack and the existence of air quality monitoring stations separately. Given a grid G U and a set of grids G S = { G S n } ( n = 1 ) K , where K is the number of monitoring stations, our goal is to infer future AQI of G U (say, next 24 h) based on historical and current dynamic and static features of G U and G S .

3.2. Definition of Input Features

In this study, we aim to infer spatially fine-grained city air quality value based on existing air quality monitoring stations and achieve a long-term prediction on areas without monitoring stations. The input features of the proposed model are categorized into dynamic features (i.e., AQI and meteorological data) and static features (i.e., point of interest (POIs), road network, and land use). The details of input features are as follows:
  • Grid: We use longitude and latitude to divide the target area into disjoint square grids as the basic unit. Notably, we adjust the size of the grid to ensure that each grid contains only one air quality monitoring station. A grid region with no monitoring station is denoted as unlabeled grids G U n , and a grid region with a monitoring station is denoted as labeled grids G S n .
  • AQI: In this study, we use the dataset collected by [22]. We choose PM2.5 values observed by air-quality monitoring stations every hour as the predictive features because PM2.5 is the main pollutant that influences the air quality level and also the most widely used indicators of an air quality report.
  • Meteorological Data: The meteorological data are critical factors to air quality level. The grid-level meteorological data in [22] includes temperature, pressure, humidity, wind direction, and wind speed.
  • POIs: The amounts of POIs represent activities and the density of people in a region, which might have a huge influence on the air quality. In this study, we collect POIs data from Gaode Map and select 14 types of POIs.
  • Road Network: It is known that air quality is strongly affected by traffic condition. The type and accumulated length of a road might be strongly correlated to the real-time traffic condition. In this study, the road network data are collected using OpenStreetMap and four features are identified for each grid: (1) total length of highway, (2) total length of primary road, (3) total length of secondary road, and (4) total length of the pedestrian walk.
  • Land Use: The air quality of a region is highly affected by its surrounding land. For example, “Industrial Area” could be the source of air pollution; “Residential” is the place that attracts people and traffic, which worsens the air quality. Conversely, the region with a larger area of open “Green” or “Water” land might improve the air quality. In this study, we use OpenStreetMap to collect four types of land use data including (1) total area of green land, (2) total area of industrial area, (3) total area of residential area and (4) total area of waters for each grid.

4. Methodology

In this section, we introduce our proposed model to forecast the PM2.5 values in 24 h at a target location without a monitoring station. The data preprocessing and representations of input features were described in Section 4.1. Figure 1 presents the overview of our proposed framework. We divide our model into three hierarchical layers including (a) feature extractor, (b) inversed distance attention, and (c) hybrid predictor. The feature extractor has two independent structures that calculate the embedding vectors of unlabeled target grids without monitoring stations and labeled grids with monitoring stations. The inversed distance attention calculates the influences of different labeled grids on unlabeled target grids. Finally, the hybrid predictor makes a long-term prediction of air quality by using the parallel structure of two independent predictors.

4.1. Data Preprocessing and Symbols

The AQI values X A collected from monitoring sensors sometimes have missing values that might decrease the performance of the model predictions. To solve this problem, a previous study [13] proposed replacing the missing values by computing the mean value in a six-hour sliding window. For example, after imputation, the missing value “NaN” in a time series of AQIs [ …, 34, 37, 46, “NaN”, 40, 31, 33, …] will be replaced by 37. Note that samples with at least two consecutive missing values are excluded from the training and testing because the artificial data might affect the effectiveness of prediction. Finally, we apply standard normalization to the AQI data.
The time series of meteorological features X M in this study have five categories: (1) temperature, (2) pressure, (3) humidity, (4) wind direction, and (5) wind speed. Among these features, only the wind direction is a categorical feature with categories, while the others are numerical. We apply one-hot encoding to the wind direction data and standard normalization to the values of numerical data.
For POI features POIs X P , 14 POI categories include Company, Culture, Daily Life, Entertainment, Institution, Transportation Spots, Medical, Vehicle Service, Shopping Mall, Sports, Food and Beverage, Hotels, Education, and Financial. The POI features are the total numbers of each POI categories within the grid. After computing the POI features of all grids, the POI features are then normalized by the sum of each category among all grids so that all values are in the range of [0, 1]. For Road Networks X R and Land Use s X L features, the values are normalized to the range of [0, 1] by the total length and area of each type, respectively.
In addition, we calculate the distance between the unlabeled grid and all labeled grids by using their central latitudes and longitudes. For instance, IDW and K-NN are the common baselines for inferring fine-grained AQI from monitoring stations, and the most important key point for prediction is the distance between target regions and each monitoring station. In Equation (1), we normalize the distance between G U and G S as the Distance feature X D and yield the Inversed Distance feature X I D simultaneously. Normalization on the Inversed Distance feature is not performed, since the minimum Distance feature is 1.
X D = N o r m a l i z e ( D i s t a n c e ( G U , S ) ) , f o r S i n G S X I D = 1 / D i s t a n c e ( G U , S ) , f o r S i n G S
After data preprocessing, we have static features and sequential features in each grid. However, there are a few differences between input features between each grid due to the existence of monitoring stations. We extract the feature set F S n , which includes Meteorological X U M , POIs X U P , Road Net X U R , Land Use X U L , Distance X U D , and Inversed Distance X U I D for unlabeled girds G U . For a set of grids with a monitoring station, we denote them as labeled grids and use AQI X S n A , Meteorological X S n M , POIs X S n P , Road Net X S n R , Land Use X S n L , Distance X S n D , and Inversed Distance X S n I D as labeled grids features F S n .

4.2. Feature Extractor

We first introduce the feature extractor, which is the input layer of our model. The input layer consists of two independent feature extractors that process the unlabeled grid G U and labeled grids G S separately due to the dimensional difference between the input features. Moreover, the input features are composed of non-sequential data and sequential data. Thus, we extract their embedding vectors separately in different ways.
Since the input features, air quality index X A and meteorological features X M are long-term time-related, we can use recurrent neural network (RNN) to encode each of the sequential features. Although RNN is designed to deal with time series data, this technique suffers from a vanishing or exploding gradient problem. To deal with the flaws, long short-term memory (LSTM) is proposed in [23]. In Equation (2), the first two gates control the proportion of previous state C ( t 1 ) and generate the current state C t . Then, the output gate is used to control the amount of cell states to generate the current output value h t .
i t = σ ( W i [ X t , h t 1 ] + b i ) , f t = σ ( W f [ X t , h t 1 ] + b f ) , o t = σ ( W o [ X t , h t 1 ] + b o ) , c t = f t c t 1 + i t t a n h ( W c [ X t , h t 1 ] + b c ) , h t = o t t a n h ( c t ) ,
where W i , W f , W o , and W c are the weights for concatenated input vectors at time step t for each gate, i t is the input gate’s activation vector, f t is the forget gate’s activation vector, o t is the output gate’s activation vector, and b i , b f , b o , and b c are the biases vectors. To make the model nonlinear, σ and t a n h are applied as activation functions and ⊙ is the element-wise multiplication of the matrix.
Other static features such as X P , X R , X L and X D are non-dynamic related. We employ a single dense layer (the first dense layer in Figure 1a) to extract and learn the interaction between static features.
Then we calculate the static embedding in Equation (3):
e U 1 = R e L U ( W U 1 ( X U P X U R X U L ) + b U 1 ) , e S n 1 = R e L U ( W S 1 ( X S n P X S n R X S n L X S n D ) + b S 1 ) , for n = 1 , , k
where e U 1 and e S n 1 are the output of the first dense layer of unlabeled grids and labeled grids, ⊕ denotes the operation of feature concatenation, and rectified linear unit (ReLU) is our activation function. After extracting both static features and sequential features, we aim to capture the correlation between them. We further construct another dense layer on top of LSTM and the first dense layer, in order to obtain the feature interaction and calculate the embedding feature of each grid. We can have an interactive dense layer in Equation (4):
e U 2 = R e L U ( W U 2 ( e U 1 h U t ) + b U 2 ) , e S n 2 = R e L U ( W S 2 ( e S n 1 h S n t ) + b S 2 ) , for n = 1 , , k
in which e U 2 and e S n 2 are the outputs of the feature extractor for unlabeled grids and labeled grids, and h t is the last hidden state of LSTM. In Equation (4), we use the same weight and bias in the feature extractor for all k labeled grids. Since the labeled grids are the same, learning shared weight is a method to reduce duplicated computation and parameter numbers.

4.3. Inversed Distance Attention

The proposed inversed distance attention is shown in Figure 1b. We can acquire the spatio-temporal embedding vectors of unlabeled grids and all labeled grids through the first layer feature extractor. However, not all labeled grids have the same contribution to each unlabeled grid. The attention mechanism [24] is a beneficial structure that computes their different contributions. Hence, we apply the attention technique in our model to learn the importance between unlabeled grids and all labeled grids. There are a total of k embedding vectors of labeled grids. Generally, attention networks calculate the attention score of two vectors: e U 2 and e S n 2 for k times. Then we exploit each attention score as the pooling coefficient to multiply e S n 2 and sum up the results of all grids as one embedding vector.
Inner product, cosine similarity, and multilayer perceptron (MLP) are common methods to calculate the attention score. In our model, we make use of MLP as the base attention network because its performance is the best of common methods. In Equation (5), we first concatenate e U 2 and e S n 2 and use two dense layers as MLP attention mechanism to obtain the correlation values.
e A T T n = R e L U ( W A T T 2 ( R e L U ( W A T T 1 ( e U 2 e S n 2 ) + b A T T 1 ) ) + b A T T 2 ) , for n = 1 , , k
a S n = e x p ( e A T T n X S n I D ) n = 1 k e x p ( e A T T n X S n I D ) ) , for n = 1 , , k
e S 3 = n = 1 k e S n 2 × a S n
Generally, the attention score can be obtained by directly encoding all correlation values by softmax layer. However, the distance between G U and G S n might have a high influence when calculating the correlation. It is conceivable that the closer to G U in G S n , the more impact there will be on G U , so in Equation (6) we multiply the inversed distance feature ( X S n I D ) and correlation values before the softmax layer. We can have an attention score for each labeled grid after the softmax layer, and use them as a weighted coefficient to obtain the final output e S 3 of inversed distance attention in Equation (7).

4.4. Hybrid Predictor

In this section, we detail our proposed structure for prediction. The main purpose of the prediction layer in Figure 1c is to obtain the predicted AQI values from the final embedding vector. First, the final embedding vector is obtained by concatenating the output of the inversed distance attention ( e S 3 ) and the embedding vector of unlabeled grids ( e U 2 ). Although our predicted target can be taken as a regression problem, we can use MLP to process embedding vectors and predict the results. Compared to the common regression problem, directly using traditional MLP does not yield good accuracy, due to the lack of historical AQI data from unlabeled grids. Therefore, we propose a hybrid predictor to solve the problem and achieve a better long-term prediction by splitting the predicted values into two data flows. We assume that predicted values can be separated into two independent fractions: time-independent and time-dependent. We design two divided predictor structures that have two independent data flows to predict the AQI, respectively. In the end of the predictor, we sum the predicted values up with dynamic weights to combine these two data flows on each prediction timestamp.
We apply the first dense layer behind the inversed distance attention layer to learn the cross features of unlabeled grid embedding and attention output. In Equation (8), we can have a cross embedding vector ( e Z ) for the prediction of two data flows after the dense layer.
e Z = R e L U ( W Z ( e U 2 e S 3 ) + b Z )
We then directly use MLP, a common method for a regression problem, to predict one data flow. As Equation (9), we can have time-independent predicted values for future time t by using an additional dense layer.
X m = W x e Z + b x , for m = 0 , 1 , , t
Another predicted data flow, time-dependent flow, means the prediction for the next timestamp is based on the current prediction and so on for further timestamp prediction. To achieve this goal, we propose an LSTM predictor that is based on the structure of LSTM AutoEncoder [25]. AutoEncoder is a common unsupervised learning method that usually trains a hidden embedding vector for the image, or compresses the dimension of vectors. AutoEncoder is a structure that minimizes the error between input and output vectors. We use the concept of the decoder part for time-dependent data flow in our model.
LSTM Enc. refers to the structure of LSTM Encoder in Figure 2. Because the input of LSTM should be a sequence, we repeat the cross embedding vector ( e Z ) several times as a sequence input. Through the processing of the encoder, we can have output ( e E n c . ) and hidden state ( h t A E , c t A E ) of the encoder for the next step in Equations (10) and (11).
e L S T M , ( h t A E , c t A E ) = L S T M E n c . ( e Z )
e E n c . = R e L U ( e L S T M )
LSTM Dec. in Figure 2 is composed of one LSTM and one dense layer. We have a total of t LSTM decoders which have the same hidden sizes as the t number of the encoder’s LSTM. In Equation (12), we take the output of the encoder that usually represents the last layer of h t A E as the initial input of the decoder. The hidden state of the encoder is also the initial hidden state for all decoder units. In other words, we set the same hidden state for both the encoder and all decoders in back-propagation. Meanwhile, in Equation (13), the output of the first decoder would be the input of the second decoder, and so on until t time periods. With our proposed LSTM predictor, the data flow of time-dependent prediction can be dealt with.
Y 0 = L S T M D e c . ( e E n c , ( h t A E , c t A E ) )
Y m = L S T M D e c . ( Y m 1 , ( h t A E , c t A E ) ) , for m = 1 , 2 , , t
From two separate predicting data flows, we acquire two complementary predicted values that are time-independent, as well as time-dependent values for the future t time periods. Then we aggregate the results of the two data flows. Although it is intuitive to simply sum the two components to obtain final predicted values, we should also take into account that the proportion of two components is not equal in all timestamps. Hence, we exploit the dynamic weights sum to yield results y ^ in Equation (14):
y ^ = m = 0 t ( α m X m + β m Y m )
in which α t is the weight of time-independent predictor and β t is the weight of time dependent on predicted time t. We do not define α t and β t as fixed values but use them as dynamic weights to learn the proportions automatically in all timestamps during the training processing. A better performance can be achieved by our proposed dynamic weights summation method. In Figure 2, we illustrate our technical data flows and the components of the proposed framework.

5. Experiments

In this section, we first introduce the Beijing dataset used in this study. Then, we include the details of the experimental setting. Finally, we present the experimental results.

5.1. Dataset

In this study, we set 115.9 E∼ 117.2 E and 39.38 N∼ 40.6 N as our target region. We then divided the target region into a total of 3172 grids with a resolution of approximately 2.2 km × 2.2 km ( 0.025 longitude and 0.02 latitude.) In addition, we made sure there was no more than one station in a grid. Figure 3 shows the target region, where each green mark represents the location of a certain monitoring station. In total, there are 35 monitoring stations in the target region.
The input features of the models include the AQI, meteorological data, POIs, road networks, and land use. These features were collected from 1 January 2017 to 31 December 2017 hourly. Notably, we have the historical PM2.5 data of all monitoring stations, and fine-grained meteorological records of the target location for all experiments.

5.2. Experimental Settings and Evaluation Metrics

Because the fine-grained air quality data were unavailable, we selected some monitoring stations as unlabeled locations and the predicted targets. Specifically, in the total 35 grids with monitoring stations, we randomly selected 15 grids as labeled grids G S n and the rest of grids as unlabeled grids G U n . We used all 8760 timestamps in a year to collect valid input data and predicted values for training and testing. All available data are partitioned into a training set, validation set, and testing set at an 8:1:1 ratio. Figure 4 is an example with k monitoring stations selected as labeled grids, and the current time is Tc. According to the problem definition in Section 3, we take Tc − 23 PM2.5 data of labeled grids as input features and Tc + 24 PM2.5 data of remaining unlabeled grids as a single piece of available data. In other words, the total number of all remaining unlabeled grids is the number of valid instances in time Tc (monitoring sensors sometimes have missing values). To obtain sufficient training and testing data, it is impossible to ensure that all timestamps in all k selected stations have complete historical data. Therefore, we made sure that at least one unlabeled grid will not have missing data in the future. We looked over all timestamps to collect available data and set k as 15 to calculate the amount of all available data in our experiment.
The mean absolute error (MAE) and the root mean square error (RMSE) are used as the evaluation metrics to evaluate the predicted and the ground truth air quality value. The definitions of MAE and RMSE are as follows:
M A E = 1 n i = 1 n | y ^ y |
R M S E = 1 n i = 1 n ( y ^ y ) 2
where y ^ and y are the predicted and observed ground truth values, respectively. We calculated average MAE and RMSE for the current time (0 h) and the future times (1–6, 7–12, 13–18, and 19–24 h).
In order to obtain robust evaluation results, we ran the experiment 10 times and used the average MAE and RMSE scores as representative. In each run, we randomly selected labeled and unlabeled grids and made sure that the combinations of monitoring stations selections were different from the previous runs.

5.3. Experimental Results

First, we show the performance comparison between our model and some baseline methods. Then, we discuss how different numbers of labeled grids affect performance. Finally, we conduct the ablation studies to show the advantages of the proposed inversed distance attention and hybrid predictor.

5.3.1. Comparison with Baseline Models

Baseline models can be divided into two categories, non-NN-based methods and NN-based methods. The non-NN-based methods include K-nearest neighbor (K-NN), weighted K-NN, GBDT regression, and linear regression. The NN-based methods include MLP, LSTM, and LSTM-IDW. The details of the baseline models are as follows.
  • K-nearest neighbor (K-NN) regression: The idea of the K-NN regressor is to predict the value by using the nearest monitoring stations as references. In this study, we set the nearest neighbors to three (i.e., k = 3). That is, the predicted value was calculated by using the average scores of the three nearest stations.
  • Weighted KNN regression: The weighted K-NN predicts the target values by using the inverse-distance-weighted [26] average scores of the nearest stations. That is, a closer station will have a greater influence on the predicted value of the target station. In this study, we used k = 3 as the setting of the K-NN regressor.
  • Gradient Boosting Decision Tree (GBDT) regression: GBDT is a state-of-art tree-based model that captures complex cross features [27].
  • Linear Regression: A linear model that assumes a linear relationship between the input variables and the output variable.
  • Multilayer Perceptron (MLP): For the MLP model, we concatenate all features on all timestamps as input. The MLP model has two layers.
  • LSTM: For the LSTM model, we concatenated all static features with every time step of time-series data as input features. The LSTM model contains two LSTM layers and one dense output layer.
  • LSTM-IDW: The LSTM-IDW model is a combination of the LSTM model and the IDW method. We use LSTM to make an AQI prediction of the labeled grids using their historical data first, and then use IDW interpolation to infer the future predictive values of unlabeled grids based on the prediction values of LSTM.
Table 1 is the overall comparison between baseline models and our proposed model. Our model outperforms baselines particularly in long-term prediction. The results of LSTM-based models are worse than other methods, for the reason that is also the challenge of our target: the lack of historical air quality data of unlabeled grids makes it difficult for LSTM to make predictions.
In addition, the error of LSTM-IDW is slightly smaller than IDW and K-NN, and LSTM-IDW performs much better than LSTM. LSTM-IDW is mostly affected by the accuracy of the first stage predicted method before interpolation, and the limitation of IDW interpolation’s calculating spatial correlation still influences the overall performance. From the results, our proposed model is able to calculate more critical spatial-temporal correlations and performs better than comparative baselines with the same input features.

5.3.2. Ablation Study of Inversed Distance Attention

We used the attention mechanism to calculate the correlation between target prediction regions and grids with monitoring stations. By multiplying the inversed distance feature as a modification term before the softmax layer, the performance of our model will be improved. In this section, we show an ablation analysis with the attention mechanism and the influence of the inversed distance feature. Here we introduce three models for comparison:
  • Model without Attention mechanism: We disabled all the attention structures in our model and used summation to aggregate the embedding vectors of all labeled grids.
  • Model without Inversed Distance Attention: In Figure 1b, we only disregard the inversed distance feature X S n I D to calculate the attention weight, while the process and parameter settings are the same as our proposed method.
  • Model with Inversed Distance Attention: This is our proposed model with inversed distance attention structure. We compare our prediction result with the two above models to show the advantage of the proposed inversed distance attention layer.
The overall results for inversed distance attention are shown in Table 2. In the first and second rows, a simple attention layer can improve the performance by giving the high influential grids critical weights. Moreover, in the second and third rows, we prove that multiplying the inversed distance before the softmax layer can further decrease errors. Thus, the distances between unlabeled grids and labeled grids can significantly impact the finding of related reference grids with monitoring stations.

5.3.3. Ablation Study of Hybrid Predictor

We study the advantages of our hybrid predictor in this section. Hence, we test our proposed hybrid predictor with the ablation analysis and verify whether our overall framework can achieve the best performance. First, we compare our performance with different individual predictors. The first one is MLP as the time-independent predictor, and the second one is the LSTM AutoEncoder as the time-dependent predictor. Namely, we divide our framework into two data flows to predict the result separately in this experiment. Second, we combine these two predictors into a hybrid predictor as a testing model with a simple sum. We sum up the results of two predictors with equal weight. Finally, we use the proposed dynamic weights sum to aggregate the results and compare them to the results of the summation that has no dynamic weights sum. In this part, we show our proposed idea step by step with the results.
However, we expect a prediction model might have better performance during the current time than the more extended periods. That is to say, the trend of the error from T 0 to T 19 24 should be incremental, meeting our expectations for real life. From Table 3 and Table 4, we only show the results with MAE to compare each predictor, since the trends of MAE and RMSE are similar. We list all 10 experiment results simultaneously for a more detailed comparison between two different predictors.
The results with the MLP predictor (Table 3) are obviously superior in that the overall MAE is better than the results with the LSTM AutoEncoder predictor (Table 4), whether in every experiment or the average results. However, the accuracy trend of the LSTM AutoEncoder predictor is more realistic than the MLP predictor. We mark the values in T 0 that are worse than T 1 6 in Table 3 and Table 4 and denote them as unexpected T 0 , and find that the number of unexpected T 0 in Table 3 is more than in Table 4. Thus, we can have a hypothesis that each predictor structure has its advantages and disadvantages, and their combined effectiveness proves that their results are complementary.
Hence, the aggregation of two data flows shows that the two predictors are complementary and beneficial to our proposed method. Table 5 shows the result from integrating two predictors by simply adding up. Through simple summation, the overall MAE is better than with each predictor individually, and the number of unexpected T 0 is less than the results in Table 3. Thus, we easily confirm that a simple sum hybrid predictor can enhance the performance and that separating the predicted values into two independent data flows is effective.
Last but not least, we assume that the proportions of the MLP predictor and LSTM AutoEncoder predictor might not be the same for all cases. Hence, we propose to sum up the dynamic weights in our model and show the results in Table 6. We exploit the MAE of 10 groups data and the overall MAE and RMSE to compare the difference between simple sum in Table 5 and dynamic weights sum in Table 6. Even though the number of unexpected T 0 is the same, the overall MAE and RMSE have both improved with the dynamic weights sum. In conclusion, our proposed hybrid predictor with dynamic weights sum efficiently reduces overall errors and can meet the demand in real life as well.

5.3.4. Affection of Number of Labeled Grids

Generally, errors are fewer when more labeled grids are selected. Missing observed values in a real case is usual. Hence, in our experimental settings in Section 5.2, the number of labeled grids might influence available data for training and testing. In other words, it is more difficult to have complete historical PM2.5 data with at least 20 labeled grids than with at least 15 labeled grids. In this part, we experiment with the different numbers of labeled grids (k) to discuss the effect of k.
To set up the experiment, we select the same amount (15,210) of data with different k. Namely, k as the only variation means that the result will not be influenced by the total amount of data or predicted target. We experiment with k at 10, 15, and 20 as appropriate values in our dataset. First, we randomly select 20 labeled grids from 35 grids with monitoring stations and collect all available data. Next, we randomly select 15 grids from the previously selected 20 as the labeled grids and with k = 15, and randomly select 10 from the selected 15 grids. We then filter the selected number of k stations in previously collected data. The predicted unlabeled grids, predicted values and the number of all collected available data are the same with a different k. The only difference is the set of labeled grids used for reference. We then repeat the steps to select grids and collect available data 10 times and also evaluate the effect of k with average MAE and RMSE in Table 7. As we expected, our model performs better with a larger k because there is more referable monitoring stations. However, a larger k is not feasible in real life due to the completeness of observed data stated in the previous section. Thus, we choose an appropriate k = 15 as our setting in this dataset.

6. Conclusions

In this study, we propose a new topic in air quality prediction that forecasts the fine-grained air quality in regions without monitoring stations. The challenge to this objective is the lack of historical air quality data, which is crucial to precise long-term prediction. Hence, we propose a hybrid predictor structure that consists of two data flows to predict time-independent and time-dependent values respectively. We then combine the two complementary results with the dynamic weights sum, which effectively increases the accuracy of long-term prediction. Additionally, we propose inversed distance attention based on the MLP attention mechanism that further improves the overall performance. Our attention networks can provide crucial correlation information of all monitoring stations, as well as any region in the city for an accurate fine-grained air quality prediction. According to the experiment results, our model outperforms all baselines significantly for the next 19–24 h in MAE and RMSE. As a result, our fine-grained air quality prediction not only performs well in the current time but satisfies the demand for long-term prediction. In our future plan, we aim to devise a score specific to the air quality prediction problem that quantifies the human health improvement induced by the model to measure the model prediction performance.

Author Contributions

Supervision, H.-P.H.; methodology, H.-P.H., S.W., Y.-W.C. and Z.-T.Y.; validation, C.-C.K. and Z.-T.Y.; investigation, H.-P.H., S.W., and Z.-T.Y.; writing—original draft preparation, Y.-W.C., S.W. and Z.-T.Y.; writing—review and editing, H.-P.H., C.-C.K. and C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology (MOST) of Taiwan under grants MOST 109-2636-E-006-025 and MOST 110-2636-E-006-011.

Data Availability Statement

Not applicable.

Acknowledgments

This work was partially supported by Ministry of Science and Technology (MOST) of Taiwan under grants MOST 109-2636-E-006-025 and MOST 110-2636-E-006-011(MOST Young Scholar Fellowship).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xing, Y.F.; Xu, Y.H.; Shi, M.H.; Lian, Y.X. The impact of PM2.5 on the human respiratory system. J. Thorac. Dis. 2016, 8, E69. [Google Scholar] [PubMed]
  2. Menut, L.; Bessagnet, B.; Siour, G.; Mailler, S.; Pennel, R.; Cholakian, A. Impact of lockdown measures to combat Covid-19 on air quality over western Europe. Sci. Total Environ. 2020, 741, 140426. [Google Scholar] [CrossRef] [PubMed]
  3. Qin, S.; Liu, F.; Wang, C.; Song, Y.; Qu, J. Spatial-temporal analysis and projection of extreme particulate matter (PM10 and PM2.5) levels using association rules: A case study of the Jing-Jin-Ji region, China. Atmos. Environ. 2015, 120, 339–350. [Google Scholar] [CrossRef]
  4. Zheng, Y.; Yi, X.; Li, M.; Li, R.; Shan, Z.; Chang, E.; Li, T. Forecasting fine-grained air quality based on big data. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 2267–2276. [Google Scholar]
  5. Hsieh, H.P.; Lin, S.D.; Zheng, Y. Inferring air quality for station location recommendation based on urban big data. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 437–446. [Google Scholar]
  6. Le, V.D.; Bui, T.C.; Cha, S.K. Spatiotemporal deep learning model for citywide air pollution interpolation and prediction. In Proceedings of the 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), Busan, Korea, 19–22 February 2020; IEEE: Piscataway, NJ, USA; pp. 55–62. [Google Scholar]
  7. Kumar, K.; Yadav, A.; Singh, M.; Hassan, H.; Jain, V. Forecasting daily maximum surface ozone concentrations in Brunei Darussalam—An ARIMA modeling approach. J. Air Waste Manag. Assoc. 2004, 54, 809–814. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Lateb, M.; Meroney, R.N.; Yataghene, M.; Fellouah, H.; Saleh, F.; Boufadel, M. On the use of numerical modelling for near-field pollutant dispersion in urban environments—A review. Environ. Pollut. 2016, 208, 271–283. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Rybarczyk, Y.; Zalakeviciute, R. Machine learning approaches for outdoor air quality modelling: A systematic review. Appl. Sci. 2018, 8, 2570. [Google Scholar] [CrossRef] [Green Version]
  10. Xayasouk, T.; Lee, H.; Lee, G. Air pollution prediction using long short-term memory (LSTM) and deep autoencoder (DAE) models. Sustainability 2020, 12, 2570. [Google Scholar] [CrossRef] [Green Version]
  11. Zheng, Y.; Liu, F.; Hsieh, H.P. U-air: When urban air quality inference meets big data. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 1436–1444. [Google Scholar]
  12. Cheng, W.; Shen, Y.; Zhu, Y.; Huang, L. A neural attention model for urban air quality inference: Learning the weights of monitoring stations. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  13. Lin, Y.; Mago, N.; Gao, Y.; Li, Y.; Chiang, Y.Y.; Shahabi, C.; Ambite, J.L. Exploiting spatiotemporal patterns for accurate air quality forecasting using deep learning. In Proceedings of the 26th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Seattle, WA, USA, 6–9 November 2018; pp. 359–368. [Google Scholar]
  14. Chen, L.; Ding, Y.; Lyu, D.; Liu, X.; Long, H. Deep multi-task learning based urban air quality index modelling. Proc. ACM Interact. Mobile Wearable Ubiquitous Technol. 2019, 3, 1–17. [Google Scholar] [CrossRef]
  15. Qi, Y.; Li, Q.; Karimian, H.; Liu, D. A hybrid model for spatiotemporal forecasting of PM2.5 based on graph convolutional neural network and long short-term memory. Sci. Total Environ. 2019, 664, 1–10. [Google Scholar] [CrossRef] [PubMed]
  16. Qi, Z.; Wang, T.; Song, G.; Hu, W.; Li, X.; Zhang, Z. Deep air learning: Interpolation, prediction, and feature analysis of fine-grained air quality. IEEE Trans. Knowl. Data Eng. 2018, 30, 2285–2297. [Google Scholar] [CrossRef] [Green Version]
  17. Soh, P.W.; Chang, J.W.; Huang, J.W. Adaptive deep learning-based air quality prediction model using the most relevant spatial-temporal relations. IEEE Access 2018, 6, 38186–38199. [Google Scholar] [CrossRef]
  18. Li, L.; Zhang, X.; Holt, J.B.; Tian, J.; Piltner, R. Spatiotemporal interpolation methods for air pollution exposure. In Proceedings of the Ninth Symposium of Abstraction, Reformulation, and Approximation, Catalonia, Spain, 17–18 July 2011. [Google Scholar]
  19. Wong, D.W.; Yuan, L.; Perlin, S.A. Comparison of spatial interpolation methods for the estimation of air quality data. J. Exp. Sci. Environ. Epidemiol. 2004, 14, 404–415. [Google Scholar] [CrossRef] [Green Version]
  20. Lin, Y.; Chiang, Y.Y.; Pan, F.; Stripelis, D.; Ambite, J.L.; Eckel, S.P.; Habre, R. Mining public datasets for modeling intra-city PM2.5 concentrations at a fine spatial resolution. In Proceedings of the 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Redondo Beach, CA, USA, 7–10 November 2017; pp. 1–10. [Google Scholar]
  21. Zhao, X.; Xu, T.; Fu, Y.; Chen, E.; Guo, H. Incorporating spatio-temporal smoothness for air quality inference. In Proceedings of the 2017 IEEE International Conference on Data Mining (ICDM), New Orleans, LA, USA, 18–21 November 2017; IEEE: Piscataway, NJ, USA; pp. 1177–1182. [Google Scholar]
  22. Wang, H. Air Pollution and Meteorological Data in Beijing 2016–2017. 2019. Available online: https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/RGWV8X (accessed on 10 November 2021).
  23. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–7 December 2017; pp. 5998–6008. [Google Scholar]
  25. Srivastava, N.; Mansimov, E.; Salakhudinov, R. Unsupervised learning of video representations using lstms. In Proceedings of the International Conference on Machine Learning. PMLR, Lille, France, 6–11 July 2015; pp. 843–852. [Google Scholar]
  26. Appice, A.; Ciampi, A.; Fumarola, F.; Malerba, D. Missing sensor data interpolation. In Data Mining Techniques in Sensor Networks; Springer: Berlin/Heidelberg, Germany, 2014; pp. 49–71. [Google Scholar]
  27. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of our proposed model. (a) Our feature extractors for processing grid-based features. The color green represents dynamic features. The color blue represents static features. The color cream represents grid extractors. (b) The inversed distance attention (color red) calculates the influences of labeled grids on unlabeled target grids. (c) Our proposed hybrid predictor(color grey) makes the final prediction.
Figure 1. The overall architecture of our proposed model. (a) Our feature extractors for processing grid-based features. The color green represents dynamic features. The color blue represents static features. The color cream represents grid extractors. (b) The inversed distance attention (color red) calculates the influences of labeled grids on unlabeled target grids. (c) Our proposed hybrid predictor(color grey) makes the final prediction.
Applsci 12 04268 g001
Figure 2. Technical data flows and components of the proposed framework. The color blue represents the data flows of unlabeled grid features. The color green represents the flows of labeled grid features. The color red represents the operations of inversed distance attention. The color grey represents the operations of the hybrid predictor.
Figure 2. Technical data flows and components of the proposed framework. The color blue represents the data flows of unlabeled grid features. The color green represents the flows of labeled grid features. The color red represents the operations of inversed distance attention. The color grey represents the operations of the hybrid predictor.
Applsci 12 04268 g002
Figure 3. The Beijing map. Each green mark indicates the location of a monitoring stations in Beijing.
Figure 3. The Beijing map. Each green mark indicates the location of a monitoring stations in Beijing.
Applsci 12 04268 g003
Figure 4. The method to collect available data in time Tc.
Figure 4. The method to collect available data in time Tc.
Applsci 12 04268 g004
Table 1. Comparison between the proposed and the baseline models. The values in the table are the MAE and RMSE of AQI values.
Table 1. Comparison between the proposed and the baseline models. The values in the table are the MAE and RMSE of AQI values.
MethodsMetrics T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
k-NN Nearest AverageMAE
RMSE
23.19
40.87
32.01
52.82
45.21
71.41
53.25
81.46
58.05
85.92
IDW InterpolationMAE
RMSE
22.80
40.24
31.77
52.44
45.08
71.25
53.43
81.33
57.94
85.78
GBDT RegressionMAE
RMSE
21.41
37.33
25.70
41.32
29.90
45.90
32.54
49.83
34.04
51.10
Linear RegressionMAE
RMSE
24.93
39.15
29.00
43.68
35.14
51.42
39.09
57.14
40.30
58.38
MLPMAE
RMSE
24.95
43.28
26.64
45.22
31.54
52.23
33.74
56.89
36.02
59.91
LSTMMAE
RMSE
52.66
88.60
53.73
91.33
57.30
97.18
60.90
102.53
62.53
103.55
LSTM-IDWMAE
RMSE
22.80
40.24
31.87
52.61
42.43
68.72
51.29
77.48
57.85
81.22
Our ModelMAE
RMSE
19.35
33.78
19.41
32.08
19.72
33.14
20.41
35.88
22.25
39.95
Table 2. Ablation analysis on Inversed Distance Attention (IDA) networks. The values in the table are the MAE and RMSE of AQI values.
Table 2. Ablation analysis on Inversed Distance Attention (IDA) networks. The values in the table are the MAE and RMSE of AQI values.
MethodsMetrics T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
Without AttentionMAE
RMSE
21.53
36.74
20.88
35.40
22.00
37.35
22.12
39.94
24.91
45.75
Without IDAMAE
RMSE
20.42
33.28
19.67
31.84
20.68
33.26
21.21
36.12
23.14
40.65
With IDAMAE
RMSE
19.35
33.78
19.41
32.08
19.72
33.14
20.41
35.88
22.25
39.95
Table 3. The MAE with only MLP as the predictor for all 10 times. The values in the table are the MAE of AQI values.
Table 3. The MAE with only MLP as the predictor for all 10 times. The values in the table are the MAE of AQI values.
No. of Data T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
Data 118.5318.5020.7919.8221.85
Data 222.1120.1419.2319.0920.30
Data 324.4122.5423.6323.4029.16
Data 417.8818.3118.9020.4823.59
Data 518.6821.7726.8522.7624.47
Data 623.2819.8322.0823.1725.53
Data 723.8023.1925.0723.8024.80
Data 822.1319.7619.9723.9723.45
Data 919.8117.7318.7820.0521.54
Data 1018.0217.6118.7919.3922.16
AVERAGE20.8619.9421.4121.5923.68
Table 4. The MAE with only LSTM AutoEncoder as the predictor for all 10 times. The values in the table are the MAE of AQI values.
Table 4. The MAE with only LSTM AutoEncoder as the predictor for all 10 times. The values in the table are the MAE of AQI values.
No. of Data T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
Data 125.4125.5928.4531.0033.14
Data 231.7027.8625.0126.7732.33
Data 328.7329.0131.1534.4543.32
Data 427.0028.7630.6534.6837.84
Data 525.5227.9133.8733.6734.52
Data 628.2827.6431.2033.9537.78
Data 730.7530.8830.5429.2829.05
Data 833.2531.2334.6939.4842.04
Data 928.0026.7727.8128.8032.11
Data 1022.7722.9725.3225.7729.80
AVERAGE28.1427.8629.8731.7835.19
Table 5. The results of aggregating two data flow with simple summation in all 10 times. The values in the table are the MAE of AQI values.
Table 5. The results of aggregating two data flow with simple summation in all 10 times. The values in the table are the MAE of AQI values.
No. of Data T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
Data 1 16.8817.3118.3919.6620.61
Data 2 20.4617.9017.4618.8019.83
Data 3 23.1220.8622.3923.7428.83
Data 4 17.5019.1619.3721.2923.29
Data 5 18.6021.9026.1522.0025.08
Data 6 20.2818.3320.4121.3123.84
Data 7 21.4022.6724.7923.4624.08
Data 8 23.4120.6119.2422.2723.29
Data 9 20.1717.7618.7219.7120.22
Data 10 15.4215.9317.1617.4118.78
AVERAGEMAE
RMSE
19.72
33.90
19.24
32.76
20.41
34.76
20.97
37.46
22.78
41.71
Table 6. The detailed results of our proposed model with weighted summation for all 10 times. The values in the table are the MAE of AQI values.
Table 6. The detailed results of our proposed model with weighted summation for all 10 times. The values in the table are the MAE of AQI values.
No. of Data T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
Data 1 16.9217.4918.7819.3620.18
Data 2 19.9818.2418.3818.8319.93
Data 3 21.0220.1720.3121.1122.26
Data 4 19.8622.6323.7724.8930.67
Data 5 16.1818.1518.3317.8319.18
Data 6 20.9220.8319.1620.9323.08
Data 7 21.9622.3623.8122.9725.89
Data 8 22.3420.7820.3423.5223.80
Data 9 19.1617.4917.5417.9219.34
Data 10 15.2115.9816.7616.7518.17
AVERAGEMAE
RMSE
19.35
33.78
19.41
32.08
19.72
33.14
20.41
35.88
22.25
39.95
Table 7. Comparison results for our proposed model with the different number k of labeled grids. The values in the table are the MAE and RMSE of AQI values.
Table 7. Comparison results for our proposed model with the different number k of labeled grids. The values in the table are the MAE and RMSE of AQI values.
Number of Labeled Grids (k)Metrics T 0 T 1 T 6 T 7 T 12 T 13 T 18 T 19 T 24
k = 10MAE
RMSE
20.89
36.71
23.22
40.73
27.94
47.36
27.76
47.34
32.22
56.61
k = 15MAE
RMSE
19.98
35.86
22.62
39.74
27.46
45.67
26.63
44.94
30.43
52.54
k = 20MAE
RMSE
19.43
35.01
21.83
39.05
25.72
43.63
25.33
42.59
28.8
50.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hsieh, H.-P.; Wu, S.; Ko, C.-C.; Shei, C.; Yao, Z.-T.; Chen, Y.-W. Forecasting Fine-Grained Air Quality for Locations without Monitoring Stations Based on a Hybrid Predictor with Spatial-Temporal Attention Based Network. Appl. Sci. 2022, 12, 4268. https://doi.org/10.3390/app12094268

AMA Style

Hsieh H-P, Wu S, Ko C-C, Shei C, Yao Z-T, Chen Y-W. Forecasting Fine-Grained Air Quality for Locations without Monitoring Stations Based on a Hybrid Predictor with Spatial-Temporal Attention Based Network. Applied Sciences. 2022; 12(9):4268. https://doi.org/10.3390/app12094268

Chicago/Turabian Style

Hsieh, Hsun-Ping, Su Wu, Ching-Chung Ko, Chris Shei, Zheng-Ting Yao, and Yu-Wen Chen. 2022. "Forecasting Fine-Grained Air Quality for Locations without Monitoring Stations Based on a Hybrid Predictor with Spatial-Temporal Attention Based Network" Applied Sciences 12, no. 9: 4268. https://doi.org/10.3390/app12094268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop