Next Article in Journal
Sensing of Soil Organic Matter Using Laser-Induced Breakdown Spectroscopy Coupled with Optimized Self-Adaptive Calibration Strategy
Next Article in Special Issue
Graph-Powered Interpretable Machine Learning Models for Abnormality Detection in Ego-Things Network
Previous Article in Journal
A Software Defined Radio Based Anti-UAV Mobile System with Jamming and Spoofing Capabilities
Previous Article in Special Issue
Infrequent Pattern Detection for Reliable Network Traffic Analysis Using Robust Evolutionary Computation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Graph Neural Network with Spatio-Temporal Attention for Multi-Sources Time Series Data: An Application to Frost Forecast †

Inria Chile Research Center, Las Condes 7550268, Chile
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in Lira, H.; Martí, L.; Sanchez-Pi, N. In Proceedings of the AI: Modeling Oceans and Climate Change Workshop at ICLR 2021, May 2021, Santiago, Chile.
Sensors 2022, 22(4), 1486; https://doi.org/10.3390/s22041486
Submission received: 29 November 2021 / Revised: 21 January 2022 / Accepted: 3 February 2022 / Published: 15 February 2022
(This article belongs to the Special Issue Data Fusion and Machine Learning in Sensor Networks)

Abstract

:
Frost forecast is an important issue in climate research because of its economic impact on several industries. In this study, we propose GRAST-Frost, a graph neural network (GNN) with spatio-temporal architecture, which is used to predict minimum temperatures and the incidence of frost. We developed an IoT platform capable of acquiring weather data from an experimental site, and in addition, data were collected from 10 weather stations in close proximity to the aforementioned site. The model considers spatial and temporal relations while processing multiple time series simultaneously. Performing predictions of 6, 12, 24, and 48 h in advance, this model outperforms classical time series forecasting methods, including linear and nonlinear machine learning methods, simple deep learning architectures, and nongraph deep learning models. In addition, we show that our model significantly improves on the current state of the art of frost forecasting methods.

1. Introduction

Generating accurate weather forecasts from reliable localized data is a key feature of precision agriculture that enables farmers to improve their resources in terms of efficiency, productivity, sustainability, etc. Moreover, it is a tool for countering weather uncertainty by reducing the risks posed by extreme weather that can impact the overall quality of the production. Frost is one such threat that kills plant tissue, causing low production and economic losses since it prevents the normal development of crops. During some periods of the year, temperatures can drop considerably between day and night, with temperatures reaching below freezing. A low temperature can cause the crop to flower early or if there is frost it can cause a considerable reduction in production [1]. These negative consequences could be prevented or mitigated with a frost forecast model that provides information to the farmer regarding the probability of a frost event several hours in advance, so the farmer can take action to protect the crops.
From a forecasting perspective, the prediction of frost events presents some challenges. For instance, they are a complex meteorological phenomenon influenced by a combination of environmental factors, including air temperature, humidity, radiation, and wind, and local factors, including topography and field orientation. Since there are local factors involved, frost events can occur in small areas even within the same crop field [2]. Therefore, a method to collect high-resolution weather data is required.
Typical weather data sources such as open weather, used for global weather forecasts, or national weather stations are useful to understand the weather dynamics across big areas. Previous research used frost forecast models from these sources, but they are limited to forecasting in the vicinity of the weather station [3]. Therefore, these methods do not provide the resolution required when seeking to forecast the weather in a specific location/field. As a consequence, farmers use these forecasts as a reference, but most of their decisions to prevent frost events are based on experience and intuition.
To collect data specifically from a field, we developed a low-cost and easy-to-install IoT platform, with environmental sensors and cloud technologies for collecting and storing data. Specifically, we measure air temperature, air humidity, and relevant metadata. In addition, we use meteorological data collected from weather stations located near the field [4]. The purpose is to provide the farmer with relevant information so that they can make informed decisions. The frost forecasting model requires an intelligent component that uses field data to capture specific conditions about the field combined with weather stations data to capture weather dynamics and possible future scenarios.
The majority of research relating to frost forecasting is based on simulating partial differential equations or traditional statistical models to predict future weather conditions [5]. This approach is computationally expensive as it requires a recurrent theoretical upgrade to incorporate weather and atmospheric assumptions. On the contrary, machine learning algorithms do not make any assumptions about weather behavior. Instead, they use historical weather data as an input and train a model to predict future weather values [3].
There are some challenges related to machine learning models. Since frost can be highly variable across a small area, the collection of temperature data usually from weather stations is not available with sufficient frequency. In addition, the number of frost events during the year is relatively small, making it difficult to build an accurate prediction model due to limited available data [2]. Finally, by viewing the model as a binary classification problem, i.e., Frost/No Frost, we need to consider that both errors are unwanted. If no frost is predicted and the frost occurs, it may impact on the partial or total loss of production. On the contrary, if frost is predicted and the frost does not occur, unnecessary resources such as fuel and electricity used to mitigate the frost will be wasted.
In light of the aforementioned constraints, especially the scarcity of data and small datasets, a range of machine learning models were evaluated, including models to learn time series data as well as advanced deep learning architectures. In particular, graph neural networks (GNN) and attention mechanisms were considered suitable for this problem, since they incorporate spatial knowledge that can model field and environmental interactions [6,7]. In addition, since the occurrence of frost is caused by a prior movement of environmental factors, GNNs can be naturally extended to model this type of temporal interaction. Therefore, in this paper we discuss a time series forecasting problem using GNNs and attention.
In this study, we collected air temperature and humidity data from an experimental site and from 10 weather stations. In particular, we propose GRAST-Frost, a GNN with spatio-temporal attention architecture for frost forecast. We map weather stations’ locations to nodes on a graph and construct the edges based on geographical proximity. Furthermore, the adjacency matrix is optimized during the training phase, therefore other interactions can be learned. We utilize spatio-temporal attention to incorporate similar locations and time.
To the best of our knowledge, deep learning and graph neural networks were so far not applied to the frost forecasting problem. Although considerable research was devoted to this area, most of the research focused on developing an IoT platform for collecting in-place weather data. Consequently, little attention was paid to developing a model that can take advantage of multiple data sources and spatio-temporal dynamics. However, GNN models were recently adopted for traffic forecasting, epidemiology, and fraud detection. The implementation of GNN models in these scenarios showed the potential for predicting multivariate time series [8,9] and graph attentional networks for capturing spatio-temporal dynamics [7,10,11].
Our contributions to this field are two-fold. First, we propose a full pipeline for the frost problem, including the development of an IoT platform, data collection, and forecasting model using two data sources. Second, we approach the frost problem by proposing a multivariate time series forecasting method, GRAST-Frost, that computes each time series at the same time. This method is based on a graph data structure, and it uses a spatio-temporal attention mechanism for weighted relevance according to time and/or space.
Moreover, this paper seeks to answer the following research questions:
  • Is a GNN capable of improving the time series forecasting of data sources from different locations in comparison to state-of-the-art frost forecasting methods?
  • Does the spatio-temporal attention mechanism improve forecasts?
  • Does the combined use of different data sources improve forecasts?
To answer the first question, we compare our approach with classical time series, machine learning, and deep learning methods by classification and regression metrics, which are currently the state of the art for frost forecasting. With regard to the second question, we compare our approach with GNN models proposed in previous studies that do not use spatio-temporal attention mechanisms. For the final question, we compare the forecasts using separate data sources.
The proposed model outperforms current state-of-the-art methods for predicting temperature and classifying frost from 6 h, 12 h, 24 h and 48 h in advance. Furthermore, graph-based modeling and spatio-temporal attention mechanisms are key factors to perform a more accurate prediction and minimize classification errors.
The rest of the paper is organized into the following sections: Section 2 examines previous studies regarding machine learning for time series, weather, and frost forecasting. Section 3 describes the platform and our proposed method. After that, Section 4 presents the experiments and results, and then we offer a discussion of the results in Section 5. Finally, Section 6 concludes the paper and presents possible directions for future work.

2. Related Work

Mort et al. [12] and Verdes et al. [13] are two research studies that address the frost phenomenon using machine learning. In these studies, the authors use artificial neural networks to create temperature prediction models based on weather time series data and apply them to agricultural applications. Their principal objectives are to predict the next day’s minimum temperature using historical data. In recent years, there were great advances in the field of internet of things (IoT) systems and deep learning. Thanks to these advancements, it was possible to install a wide variety of sensors and collect data from almost any place of interest with the purpose of building accurate prediction models. Regarding IoT systems designed for weather forecasting, Muck et al. [14] designed an IoT based weather station using a Raspberry Pi, which provides short-term weather forecasting. Similarly, Levin et al. [15] presents a weather forecasting system based on a Raspberry Pi 3 Model B+ with environmental sensors and a weather forecasting algorithm. Their systems monitor air temperature, humidity, pressure, and altitude at experimental locations. Their weather forecast algorithms are based on a linear regression model. Other studies such as Diedrichs et al. [3] and Castaneda–Miranda et al. [16] used IoT devices to extract weather data from selected locations as well, but instead of focusing on weather, they used classic applied machine learning techniques to predict frost events. Likewise, the research group of Guillén–Navarro et al. [1,17] developed over the years an IoT platform to predict frost events. This platform appears to be more robust than the previous studies in terms of engineering and technological components. Although these studies made interesting progress in terms of IoT and sensor data collection systems, the development of their machine learning models was limited. Therefore, the resulting prediction results are unsatisfactory in terms of error rates and/or classification metrics. In addition, the data sources were constrained to the experimental field where the system is located.
Few studies attempted to focus on the development of the frost forecast model itself. For example, Ding et al. [2] concentrated their efforts on the development of a causal-effect machine learning model that uses locally collected temperature, humidity, and radiation data to create frost prediction. They were able to describe causal relationships between variables and outputs; however, their model requires improvements to minimize the false-positive predictions. Another example is the study by Cadenas et al. [18], which was based on a soft computing framework that collects and stores weather data. They propose a data preprocessing technique to build fuzzy time series from raw data and serve it as an input to classification and regression problems. In addition, Guillén–Navarro et al. [19] used a simple long short-term memory (LSTM) architecture to produce frost forecasts from data collected using their IoT system. Although these studies provide interesting methods to address the frost forecasting problem, there is still a wide range of solutions to investigate. For example, exploring current developments within deep learning models that could improve forecasts and include different data sources.
In contrast to frost forecasting, weather-related forecasting has plenty of studies that use advanced deep learning techniques. For instance Shi et al. [20] proposed a fully connected convolutional LSTM network to predict short-term future rainfall intensity in a local area and extract spatio-temporal dynamics of the data. Likewise, Mehrkanoon et al. [21] proposed a model for predicting temperature and wind speed 1 to 10 days in advance using a convolutional neural network. They introduced an architecture based on 1D-CNNs to process tensor 3D data and to extract spatio-temporal relations. In addition, Hewage et al. [5] presented a weather forecasting model that uses an LSTM and a temporal convolution network. The results obtained are better than classical time series forecasting and classical machine learning. However, these models do not capture the complexities of our specific problem. For instance, we need to deal with multivariate time series forecasting of several time series at different locations. In the aforementioned approaches, there are no spatial relations between the entities (e.g., different cities), the interaction is determined by the entity order in the tensor. To capture the spatio-temporal dynamics of different entities, a promising approach is to use GNNs, which have the capacity to model temporal dynamics of nodes and spatial dynamics between them at the same time.
GNNs showed recent progress in the area of time series forecasting and spatio-temporal relations. Moreover, a GNN can extract greater insights compared to that of networks that can only analyze data in isolation. This is achieved by obtaining structural relationships between the data [22]. There are a number of domains in which GNNs were successfully applied in recent years, such as traffic flow forecasting, fraud detection, epidemiology, and forecasting weather-related events. In regards to the latter, Wilson et al. [23] addressed the spatio-temporal correlation in the data by proposing a deep learning model based on a weighted graph convolutional LSTM. The general goal was to capture temporal autocorrelation with the LSTM and the spatial relationships with the graph convolution. Similarly, Khodayar et al. [24] presented a spatio-temporal Graph convolutional network (GCN) for short-term wind speed forecasting. Another example is the study by Wang et al. [25], which proposed a graph-based model to predict PM2.5 particle concentration and capture the spatio-temporal dependencies.
The most recent advances in GNN were applied to other domains. For instance, a study by Cheng et al. [7] presented a model using GNN for fraud detection in credit card transactions. They implemented a spatio-temporal attention mechanism which produces the input for a 3D convolution network. As per the previous study, Gao et al. [26] proposed a GNN with spatio-temporal attention mechanism and a GRU architecture to forecast the number of infected cases in a pandemic by considering local disease status and demographic and transmission dynamics. There were several GNN models applied to traffic flow forecasting and urban planning. In particular, Song et al. [27] developed a spatio-temporal GCN with a synchronous temporal mechanism to predict the flow of a network. Likewise, the studies by Lu et al. [28], Kong et al. [10] and Li et al. [11] proposed different model versions of a spatio-temporal GNN with attention mechanisms for urban sensor value forecasting, traffic flow forecasting, and segment-level traffic prediction, respectively.
The goal of this study is to predict frost events by using air temperature and humidity data obtained from an IoT system installed on an experimental field and weather stations located around the field. We have two main sources of inspiration. First, we are inspired by recent advancements in GNN models as detailed previously, especially the spatio-temporal attention mechanism. Second, we are inspired by the study of Wu et al. [9], which is a model for multivariate time series forecasting using a GNN, and by the recent study of Shang et al. [8], who proposed a model for multivariate time series forecasting using a GNN in which they consider pairwise interactions between features in a node representation. Therefore, the contribution of this paper is a multivariate frost forecasting model based on a GNN with a spatio-temporal attention mechanism.

3. Proposal

In this paper, we propose a bivariate time series forecasting model to predict the occurrence of frost. The benefit of such a model is that it can be trained with historical time series data X = { x t 1 , x t 2 , , x t m } with x t R 2 the value of the bivariate variable at time t is used to forecast future values of the variable in a certain time-window r, Y = { x t m + 1 , x t m + 2 , , x t m + r } . Then, the goal is to create a mapping function from X to Y and minimize the loss, typically using a l 2 regularization [9].
In particular, given a set of input data P = { n , t e m p , h u m } and a derived label F = { 1 , 0 } which indicates the presence (1) or absence (0) of frost for each one of the records. We aim to forecast the minimum temperature and the frost class for future time windows { t m , , t m + r } based on the historical time window { t 1 , , t m 1 } .
In addition to bivariate time series forecasting, we model the spatial and temporal relationships of weather data from multiple locations. For that purpose we utilize GNNs to describe and formalize those relationships. The following are important definitions for graph modeling [9].
  • Graph: a graph is represented as G = ( V , E ) where V represents the set of nodes and E represents the set of edges. There are n number of nodes in a graph.
  • Node neighborhood: describes a set of nodes connected by an edge. A singular node v V and an edge e = ( v , u ) E maps from v to u describes the connection between nodes. The neighborhood of v is defined as N ( v ) = u V ( v , u ) E .
  • Adjacency matrix: states the connections between nodes in a graph. It is denoted by a matrix A R n × n with A i j = q > 0 if ( v i , v j ) E and A i , j = 0 if ( v i , v j ) E .
Then, the graph network is formally defined as G = ( V , E , A ) , which represents the relationships between nodes in the spatial dimension.

3.1. Data Sources

The model was trained and tested using data collected from our IoT platform and 10 meteorological stations, which are all located in the central region of Chile. Appendix A lists the stations and their geographical location.
The 10 meteorological stations, which are within close proximity of the orchard, provide structured temperature and humidity data in the form of a graph. Here, the data are collected every hour. The 10 meteorological stations considered for this study and their representation as a graph are shown in Figure 1.
The IoT platform was developed with both air temperature and air humidity sensors and consists of 12 low-power wireless sensor nodes (motes). The latter are divided into eight sensor data nodes and four repeaters, which are connected to a gateway through a SmartMesh IP manager. The wireless sensor network is exposed directly to environmental conditions (sun, dust, rain, and snow), and therefore all sensors are protected by an Internal Protection 65 (IP65) rating enclosure. The IoT platform was installed in a local orchard where data were collected every 10 s at four different heights above ground (one, two, three, and four meters). An image of the IoT platform in situ in the orchard is shown in Figure 2.
Data were collected between 4 September 2020 until 5 April 2021. Training data were taken from the months of September to February and the last two months of data were used for training. It is assumed that environmental factors can have an accumulated impact on frost; therefore, the prediction was performed using time series data. For this reason, we need the model to learn time-related patterns, which is why we do not randomly split training and testing data.
All the data are preprocessed for missing values and outliers. Data from the IoT platform is downsampled using 1-min time-windows. In addition, to map the frequency of both data sources, data from weather stations are linearly interpolated. Finally, in case of the classification of frost, a label is created in the training data with two classes (Frost/No Frost) and given the imbalance between them, the synthetic minority oversampling technique (SMOTE) method was applied.

3.2. Forecasting Model

In this study, we develop a GRAST-Frost model, the architecture of which is shown in Figure 3. First, we convert our input P into a 3D (spatial, temporal and features), high-order tensor representation X . The tensor is then fed into the graph neural network, which aims to correlate the IoT platform data with that of the meteorological stations. For each t, we create a graph of nodes that relate to each of the stations (meteorological and the IoT platform) and their corresponding edges. Then, we apply a spatio-temporal attention mechanism and a 3D convolution layer to obtain a feature-learned tensor X C . With spatio-temporal attention and 3D convolution, the idea is to weigh the importance of different dimensions and find hidden patterns from the input data. Finally, X C is fed through a recurrent neural network to produce a forecast with two different loss functions depending on the type of task. We present details of each part of this architecture in the following subsections.
To forecast local weather conditions at the orchard using the aforementioned sensory data, we predict at differing time intervals into the future (6, 12, 24, and 48 h). At a single time step, the geographical locations of the nodes are graphed as can be seen in Figure 1, where the blue dots represent the meteorological sites and the green dot depicts the experimental site. For multiple time steps, the graph is expanded into a spatial-temporal graph where feature values for a given node are related to its previous and future values and its spatial neighbors. A schematic view of this idea is shown in Figure 4.

3.2.1. Feature Engineering

To represent the input data as a tensor X R T × S × F where T, S, and F denote the temporal, spatial, and feature dimensions. In particular, for each spatio-temporal pair composed of a certain time horizon and location, a feature vector is built ( f R F ) based on the measurements collected in that pair. There are a total of T × S spatio-temporal pairs, each one of them are features of F dimensions. Based on [7], the feature vector is composed of two parts: feature measurements and graph features. For measurements, we include raw temperature, humidity data, mean, standard deviation, median, maximum, and minimum values for a specific time horizon. For the graph related features, we include several metrics obtained from the graph neural network processing described in the following section.

3.2.2. Graph Construction at a Single Time Step

Given the tensor X, the geographical location (latitude, longitude) of the weather stations and experimental field is obtained. This data are used to calculate the geographical proximity, which in turn contributes to the development of the adjacency matrix. Nodes are simply a subset of the tensor in time t. We utilize an aggregation function to reduce the edge updates to a single element. Therefore, for a single node, we summarize the interactions with other nodes.
The nodes in the graph G ( V , E ) are constructed to represent input data, by associating features such as ( t e m p , h u m ) and location ( l a t , l o n ) . In total we have 11 nodes. Generally, we have two types of nodes: v is the node corresponding to the experimental field, and u is the node corresponding to the weather stations. Thus, f v and f u are the corresponding feature vectors for each one of them.
The edges are constructed based on geographical distance between nodes. We construct a weighted graph where the weight is inversely proportional to distance. Formally, in a weighted graph G = ( V , E , w ) [6], that do not contain any cycle of negative weight, the distance between node u and v is defined as d : V × V R as
d ( u , v ) = 0 if   u = v ,   and d ( u , v ) = ( l o c a t i o n u , l o c a t i o n v ) otherwise ,
that corresponds to the Euclidean distance between the location of nodes u and v.
Therefore, the weight is,
w u v = 1 d ( u , v )
For this study, all nodes are integrated, thus, there are edges between all the nodes. Therefore, the feature vector of the edge is f e = w u v f g n n , which is the weighted measurement of f g n n , features to be obtained by the graph neural network (GNN).

3.2.3. Spatial-Temporal Graph Construction

To represent the relationship between each node and its neighbors across time, inspired by [27], we connect all nodes with themselves sequentially for each time step. This allows us to create a spatial-temporal graph by sequentially connecting nodes from previous to current and future time steps as shown in Figure 5. Therefore, it is possible to understand the relationships between nodes through time. In practice, we create a new adjacency matrix A T R n × T for the spatial-temporal graph. The new adjacency matrix can be formulated simply as
A i , j T = 1 if   v i   is   related   with   v j , 0 otherwise .
As illustrated in Figure 5 the new adjacency matrix has dimension N × T , where its diagonal represents the adjacency matrix for a single time step and a relation of the nodes through time.
For each time step a GNN consumes the graph described on the previous section. The information of interactions between nodes is described in terms of nodes involved u , v , time t, location l and measurement m. Based on it, the edges are constructed. Features for nodes f v , f u and edges f e are initialized accordingly. Then, we iteratively include graph constructed in subsequent time steps by updating the representation of nodes and edges based on the previous time step and weighted by the adjacency matrix A T . Edge features are updated following
f e u v t = F C e ( f e u v t 1 , i d u , i d v , A t , t 1 T ) ,
where f e u v t is the edge feature between nodes u and v in time t, F C e is a fully connected neural network with a R e L U activation function which computes all the edges updates. In addition, i d u , i d v are the location on graph of node u and v. Then, for each time step we proceed to update the node features f u and f v . For each node, we sum all the edges that are connected to that node. Regarding f v the equation is
f v t = F C v ( i N f e i t , f v t 1 , A t , t 1 T ) .
Here f v t is the node feature in time t, N is the total number of nodes and F C v is a fully connected neural network.
Finally, we calculate the feature vector of the graph f g which is the average of all the updates of f e and f v for every time step.
The output of the GNN are the updated values of f v , f u , f e and f g . We use these values to complete the feature vector F on the tensor X regarding the graph part, as described in the previous section.

3.2.4. Spatio-Temporal Attention Mechanism

The idea of the attention network is to weigh the importance of spatial and temporal values from the current measurement for a specific node and time. We use the GAT model [6] to extract Spatio-temporal similarity features. The idea is to update the embedding information of each node using the aggregate data from its neighbors. Therefore, weather stations and the IoT platform receive prior temperature and humidity data from nearby areas. This allows us to hypothesize that a specific prediction is more likely due to a higher importance being given to meteorological sites in close proximity to the orchard, rather than those located further away, and recent measurements being given a higher priority compared to that of older measurements. Given the feature tensor X , we can query the temporal values t X t s f ( t { 1 , 2 , . . . , T } ) to extract the time horizon, and the spatial values s X t s f ( s { 1 , 2 , . . . , S } ) to extract the location coordinates.

Temporal Attention

In the temporal dimension, there exist correlations between temperature and humidity values in different time steps. Since weather is highly dynamic, temporal correlations are variable under different conditions. Therefore, to have a mechanism for capturing those correlations, we use attention to adaptively obtain the importance of previous data in relation to current data.
Based on [7], considering a tensor X , the temporal attention layer is the temperature and humidity values for a specific time multiplied by a weighted sum of the matrix representation of all temporal values. Formally, it is described as
X t = t = 1 T a t X t s f   with
a t = exp ( ( 1 λ t ) · F C t ( W t , X t s f ) ) t = 1 T exp ( ( 1 λ t ) · F C t ( W t , X t s f ) ) .
Here a t is the weight for each time step, F C t is a fully connected layer with R e L U activation and the weight vector W t ; λ t [ 0 , 1 ] is the temporal penalty factor to control the importance of temporal attention; X t , the output of the temporal attention layer is a tensor with X R T × S × F .

Spatial Attention

The spatial dimension, temperature, and humidity values from different nodes have varying influences, and due to the weather, the behavior of these influences are highly dynamic. In particular, we are interested in capturing correlations between the nodes on the weather values in the spatial dimension. Therefore, we want to explicitly capture the relationships between close and distant nodes.
Given the output from the temporal attention layer X t , the spatial attention mechanism is applied. Formally, it is described as
X s t = s = 1 S a s X t s f t   with
a s = exp ( ( 1 λ s ) · F C s ( W s , X t s f t ) ) s = 1 S exp ( ( 1 λ s ) · F C s ( W s , X t s f t ) ) ,
where W s is the weight of the fully connected spatial network F C s , and X s t is the output of both self-attention layers. The output is reshaped into a tensor format with the same order as X ; a s is the weight for each spatial step; and λ s [ 0 , 1 ] is the spatial penalty factor to control the importance of spatial attention.

3.2.5. Convolution Process

Several studies stated the benefits of applying convolutional neural networks to feature tensors based on GNN modeling [29,30,31]. In our case, we used a 3D convolutional network with the X s t tensor as an input. The idea is to extract hidden patterns from spatio-temporal features by stacking multiple layers in the architecture.
The 3D convolution is represented as
X k c = k X c 1 ( d t c a , d s c b , d f c q ) K i l ( c a , c b , c q ) .
Here, K i l is a 3D kernel in the l th layer and i th kernel in a convolution with feature X c 1 . In particular, the 3D convolution kernel is ( c a , c b , c q ). The first layer of X c 1 is the output of our attention mechanism X s t . d t , d s , and d f are the dimensions of X s t considering temporal, spatial, and feature components, which equals to T , S , F of the first convolution layer.
Finally, the output feature X c is
X c = σ ( k X k c + b c ) ,
where b c is the bias parameter, and σ is the sigmoid function.

3.2.6. GNN Forecasting

In the last part of the forecasting model, we apply a recurrent neural network to capture the sequential aspect of the problem and produce forecasts based on historical data. Given the tensor X c we use a sequence-to-sequence model (seq2seq) [32] over each node, i.e., s X t s f c . Thus, we extract the transformed series from the experimental field and weather stations. The reason for using seq2seq is that in a graph structure, we can perform recurrent graph convolution to handle all series simultaneously [8]. In practice, for each series we used { t 1 , . . . , t m 1 } time values to train the model and { t m , . . . , t r } to forecast the weather using r { 6 , 12 , 24 , 48 } h in the future. Specifically, for each time t, the seq2seq model takes X t s f c s for all series and updates the internal hidden state from H t 1 to H t . The encoder recurrently updates the training data to be included, producing H t + r as a summary. The decoder takes that input and continues the recurrence to include all the testing data for the forecasting phase.
Finally, we use two loss functions, one for classification and one for regression. The former loss function is defined as
L = 1 N i = 1 N [ y i l o g ( H i ) + β ( 1 y i ) l o g ( 1 H i ) ] ,
where N is the total number of samples in the series, β is a sample weight regarding the distribution of Frost/No Frost; y i { 0 , 1 } is the real label and H i is the value score produce by the forecasting.
The regression loss function uses mean squared error
L = 1 N i = 1 N y i H i 2 .

4. Results

4.1. Baselines and Evaluation Metrics

We use our model to solve a regression problem, to predict the minimum temperature in all the nodes, and a classification problem, to predict two classes Frost and No Frost. In case of regression, we compare with the following forecasting methods:
  • Non-deep learning methods: historical average (HA), ARIMA with Kalman filter (ARIMA), vector auto-regression (VAR), and support vector regression (SVR). The historical average accounts for weekly seasonality and predicts for a day by using the weighted average of the same day in the past few weeks;
  • Deep learning methods that produce forecasting for each series separately (not graph-based) such as feed-forward neural network (FNN) and LSTM;
  • Autoencoder forecasting method with attention mechanism (AC-att);
  • Graph convolutional network applied to the given graph without spatio-temporal attention mechanism (GCN);
  • Variants of this architecture using convolutions [7] and GRU [26].
In case of classification, we do not use the non-deep learning methods described above, but we use support vector machines (SVM) and tree-based classification algorithms such as naïve Bayes and XGBoost.
For regression, all methods are evaluated with three metrics: mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE). For classification, all methods are evaluated with precision, recall, and F1 metrics.

4.2. Hyperparameters

Several hyperparameters are tuned through grid search: initial learning rate in 0.1 , 0.01 , 0.001 , dropout rate in 0.1 , 0.2 , 0.3 , embedding size of the LSTM layer was set in 32 , 64 , 128 , 256 , the k value in kNN in 5 , 10 , 20 , 30 , and the weight of regularization in 0 , 1 , 2 , 5 , 10 , 20 . For other hyperparameters, the convolution kernel size in the feature extractor is 10 and the decay ratio of learning rate is 0.1 . After tuning, the best initial learning rate for our dataset is 0.001 . The optimizer is Adam.
All models are implemented in PyTorch and ran in the Google Colaboratory platform.

4.3. Results for Regression Problem

Table 1 and Figure 6 show the evaluation of the proposed GSTA-RCN model compared with that of the baselines. The tasks are to forecast the minimum temperature of the experimental field with 6 h, 12 h, 24 h and 48 h in advance.
The proposed model outperforms all the compared baselines for frost forecast in 6 h, 12 h, 24 h and 48 h tasks. A nongraph model such as an autoencoder with an attention mechanism outperforms GNN with spatio-temporal attention using convolution and GRU. To improve these results, it is necessary to collect more weather data and weather variables and to use more weather stations for modeling geographical and temporal interactions.
In general terms, we can separate the results from non-deep learning models and FNN from LSTM, autoencoder and graph-based neural network architectures. Non-deep learning models (HA, ARIMA) and the simple deep learning architecture FNN only have similar error scores with the other architectures on the 6 h task. For a greater time-window, their performance drastically decreases. Then, we can compare results from LSTM with the results from STA-C, the variation of the spatio-temporal GNN with attention mechanism, and a convolution-based forecasting method. For this dataset, the complexity added by STA-C does not have an impact on the performance of the model and a simple LSTM is preferable, especially for 6 h, 12 h, and 24 h predictions. As mentioned previously, the autoencoder architecture with attention mechanism performs better than STA-C and similar to STA-GRU for 6 h, 12 h, and 24 h tasks. Finally, our model outperforms all the previous baselines. In our case, the complexity of the modeling successfully increases the performance of the prediction for each time-window task.
Figure 7 shows model’s results in terms of RMSE for the experimental field node and the three geographically closest node neighbors. By focusing on the prediction for each node, instead of the average of all nodes, the results description remains the same. In addition, the Pearson correlation statistic is calculated to compare model prediction with time windows. The performance of all models decreases when the time-window increases, and its results are variable in the different nodes. Therefore, the behavior of models regarding time-window is different in each node, which is a result that could be worth to continue studying.
In addition, Figure 8 and Figure 9 show real and predicted data for a specific day. In Figure 9 we present the variation of our model prediction regarding 6 h, 12 h, 24 h, and 48 h tasks. Our model’s best performance is in the 6 h window task, and then the performance decreases gradually. Although the performance decreases for larger time windows, our model is capable of detecting the temperature trend. The range of the error interval increases if we want to make the prediction point by point, but if that range is detected, we can be sure that the prediction is reliable. Figure 8 shows our model prediction and the second-best prediction (STA-GRU) in the 6 h task. Both models detect the temperature trend, but in our model the error range is smaller.

4.4. Results for Classification Problem

Table 2 and Figure 10 show the evaluation of the proposed GSTA-RCN model compared with the baselines as a classification problem. In this problem, two classes are predicted: Frost (temperature below 0 °C) and No Frost (temperature above 0 °C).
In this case our model also outperforms all the baselines for the 6 h, 12 h, 24 h, and 48 h tasks. In addition, similar to regression, models performance decreases when time-window increases.
For this dataset, Naive Bayes and SVM models provide the worst predictions, especially for recall score, which implies a high value of false-negative predictions. Compared with regression, in classification, FNN and LSTM perform similar. The autoencoder with attention mechanism is slightly better that the previous ones but performs worse than all the graph-based models and even the XGBoost model. Figure 10 shows that for classification in this dataset the graph-based models are preferable. In this case, the spatio-temporal attention mechanism in the GNN, which captures the dynamics between the nodes, is a key factor to improve the predictions. In particular, in our model the strategy to use the spatio-temporal vectors to perform the prediction at the same time obtains the best results by minimizing classification errors, which could be valuable for real-world applications.

5. Discussion

5.1. GNN and Spatio-Temporal Attention Applied to Frost Forecast

The first research question pretends to evaluate the performance of the GNN against current state-of-the-art models for frost forecasting. In other words, does the use of a deep learning model with graph-based data improve the performance of a predictor?
The overall results indicate that our model outperforms previous frost forecasting models and most of the variations presented for graph neural networks. To statistically check whether the models’ performance difference is significant, we conducted a t-test between our model and each previously proposed frost forecasting model and other baselines presented. Table 3 and Table 4 shows the statistics results for regression and classification, respectively.
As can be seen in Table 3 and Table 4, there is statistical significance for the results of our model against all previously proposed frost forecasting models for 6 h, 12 h, 24 h, and 48 h in advance predictions. More generally, for all the baselines presented, there are some exceptions in the 48 h time-window prediction since there is no evidence of a difference between our model and STA-GRU. This result can be used for future research with the purpose of improving the performance in larger time-windows.
The second research question “Does the spatio-temporal attention mechanism improves the forecast?” is approached similarly to the previous one. In this case, we are interested to statistically check whether there is a difference between the performance of our model with GNN models that do not use spatio-temporal attention mechanisms. In particular, Table 5 shows the results of a t-test applied between our model, GCNreg and GCNclas separately for regression and classification problems. As a result, we can demonstrate that for our dataset, results are statistically different between our model and non-spatio-temporal attention graph-based models.
The last research question, regarding whether the combined use of different data sources improve the forecast in comparison with using a single data source, can be answered based on the first two questions. By answering the two first research questions, we are implicitly answering the last one since the use of different data sources imply on the use of weather data collected from different locations. Therefore, it is possible to construct the graph and collect the spatial relationships between the nodes, which is crucial for the model’s performance.
Finally, in terms of an evaluation for the IoT-based system in comparison with the weather stations used, as shown in Figure 7 the overall results indicate that there is no difference between sites. Concretely, Table 6 presents statistics by comparing the prediction results of using our model in the experimental field against each weather station. There is no statistical significance that indicates a performance difference. Therefore, data collected from our system are reliable.

5.2. Limitations

The main limitation of this model is concerned with the decrease in performance with larger time-windows. Despite that it is a problem for all the models studied in this paper, the performance difference of our model in short-term windows with all models cannot be established in 48 h time-windows. We suspect that this behavior is similar for even larger time-windows. The lack of statistical significance, in those specific cases, is probably a consequence of a constrained period of time in which we collected the data from the experimental site and weather stations. Consequently, model performance should be revisited when the amount of data collected is larger, one year and more. However, deeply studying the deep learning architecture used and investigating whether further improvements can be made to obtain better results for long-term predictions are of interest.

6. Conclusions

We presented our frost forecasting model that uses and optimizes a graph structure between multiple time series using a graph neural network (GNN) architecture with a recurrent graph convolution mechanism to process each series simultaneously. The model concludes with spatio-temporal attention to consider spatial relations and extract temporal dynamics.
Frost forecast is an important area of climate research because of its economic impact on several industries. In this study, a GNN with spatio-temporal architecture was proposed to predict minimum temperatures at an experimental site. The model considers spatial and temporal relations and processes multiple time series simultaneously. Performing predictions of 6, 12, 24, and 48 h this model outperforms statistical and nongraph deep learning models.
To further improve this model, we will continue our research by studying deep learning architectures to specifically adapt to different time-window forecasts. In addition, we aim to include domain knowledge from climate sciences that could help in the construction of the graph to transit from a statically defined graph to a dynamically defined one. Finally, by including domain knowledge or by applying new methods, we want to extract the influences of the nodes with each other for the purposes of explaining the dynamics of the graph and, as a consequence, to provide better practical insights to users of the system.

Author Contributions

Conceptualization, H.L., L.M. and N.S.-P.; Data curation, H.L.; Formal analysis, H.L. and L.M.; Funding acquisition, L.M. and N.S.-P.; Investigation, H.L., L.M. and N.S.-P.; Methodology, H.L. and L.M.; Project administration, N.S.-P.; Supervision, L.M. and N.S.-P.; Validation, H.L.; Writing—original draft, H.L.; Writing—review and editing, L.M. and N.S.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by CORFO/ANID International Centers of Excellence Program 10CEII-9157 Inria Chile and CORFO “Crea y Valida” Project 19CV-107497.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Weather Stations Considered in the Study

The following public weather stations were used in the paper:
  • Campanacura (36°12′48.0″ S, 71°44′15.0″ W)
  • Copihue (36°04′32.0″ S, 71°45′29.0″ W)
  • Longavi Norte (36°02′13.0″ S, 71°40′60.0″ W)
  • Los Despachos (36°03′44.0″ S, 72°22′17.0″ W)
  • Monte-flor Tucapel (36°14′41.0″ S, 71°56′03.0″ W)
  • Parral (36°12′48.0″ S, 71°44′15.0″ W)
  • Odjfell (36°08′24.0″ S, 72°16′12.0″ W)
  • Parral Norte (36°13′49.0″ S, 71°43′56.0″ W)
  • Ñiquen (36°17′37.0″ S, 71°53′17.0″ W)
  • CE Arroz (36°24′33.0″ S, 72°00′24.0″ W)

References

  1. Guillén-Navarro, M.Á.; Pereñíguez-García, F.; Martínez-España, R. IoT-based System to Forecast Crop Frost. In Proceedings of the 2017 International Conference on Intelligent Environments (IE), Seoul, Korea, 21–25 August 2017. [Google Scholar]
  2. Ding, L.; Noborio, K.; Shibuya, K. Modelling and learning cause-effect—Application in frost forecast. Procedia Comput. Sci. 2020, 176, 2264–2273. [Google Scholar] [CrossRef]
  3. Diedrichs, A.L.; Bromberg, F.; Dujovne, D.; Brun-Laguna, K.; Watteyne, T. Prediction of frost events using machine learning and IoT sensing devices. IEEE Internet Things J. 2018, 5, 4589–4597. [Google Scholar] [CrossRef] [Green Version]
  4. Lira, H.; Martí, L.; Sanchez-Pi, N. Frost forecasting model using graph neural networks with spatio-temporal attention. In Proceedings of the AI: Modeling Oceans and Climate Change Workshop at ICLR 2021, Virtual Event, Austria, 3–7 May 2021. [Google Scholar]
  5. Hewage, P.; Trovati, M.; Pereira, E.; Behera, A. Deep learning-based effective fine-grained weather forecasting model. Pattern Anal. Appl. 2021, 24, 343–366. [Google Scholar] [CrossRef]
  6. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  7. Cheng, D.; Wang, X.; Zhang, Y.; Zhang, L. Graph Neural Network for Fraud Detection via Spatial-temporal Attention. IEEE Trans. Knowl. Data Eng. 2020. [Google Scholar] [CrossRef]
  8. Shang, C.; Chen, J.; Bi, J. Discrete Graph Structure Learning for Forecasting Multiple Time Series. arXiv 2021, arXiv:2101.06861. [Google Scholar]
  9. Wu, Z.; Pan, S.; Long, G.; Jiang, J.; Chang, X.; Zhang, C. Connecting the dots: Multivariate time series forecasting with graph neural networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, CA, USA, 23–27 August 2020; pp. 753–763. [Google Scholar]
  10. Kong, X.; Xing, W.; Wei, X.; Bao, P.; Zhang, J.; Lu, W. STGAT: Spatial-temporal graph attention networks for traffic flow forecasting. IEEE Access 2020, 8, 134363–134372. [Google Scholar] [CrossRef]
  11. Li, D.; Lasenby, J. Spatiotemporal Attention-Based Graph Convolution Network for Segment-Level Traffic Prediction. IEEE Trans. Intell. Transp. Syst. 2021. [Google Scholar] [CrossRef]
  12. Mort, N.; Chia, C.L. Minimum temperature prediction in agricultural area using artificial neural networks. In Proceedings of the IEE Colloquium on Neural Networks for Systems: Principles and Applications, London, UK, 25 January 1991. [Google Scholar]
  13. Verdes, P.F.; Granitto, P.M.; Navone, H.D.; Ceccatto, H.A. Frost prediction with machine learning techniques. In VI Congreso Argentino de Ciencias de la Computación; UNLP: Usuhuaia, Argentina, 2000. [Google Scholar]
  14. Muck, P.Y.; Homam, M.J. Iot Based Weather Station Using Raspberry Pi 3. Int. J. Eng. Technol. 2018, 7, 145. [Google Scholar] [CrossRef] [Green Version]
  15. Varghese, L.; Deepak, G.; Santhanavijayan, A. An IoT Analytics Approach for Weather Forecasting using Raspberry Pi 3 Model B+. In Proceedings of the 2019 Fifteenth International Conference on Information Processing (ICINPRO), Bengaluru, India, 20–22 December 2019. [Google Scholar]
  16. Castañeda-Miranda, A.; Castaño-Meneses, V.M. Internet of things for smart farming and frost intelligent control in greenhouses. Comput. Electron. Agric. 2020, 176, 105614. [Google Scholar] [CrossRef]
  17. Guillén-Navarro, M.A.; Martínez-España, R.; López, B.; Cecilia, J.M. A high-performance IoT solution to reduce frost damages in stone fruits. Concurr. Comput. Pract. Exp. 2021, 33, e5299. [Google Scholar] [CrossRef]
  18. Cadenas, J.M.; Garrido, M.; Martínez-España, R.; Guillén-Navarro, M.A. Making decisions for frost prediction in agricultural crops in a soft computing framework. Comput. Electron. Agric. 2020, 175, 105587. [Google Scholar] [CrossRef]
  19. Guillén-Navarro, M.A.; Martínez-España, R.; Llanes, A.; Bueno-Crespo, A.; Cecilia, J.M. A deep learning model to predict lower temperatures in agriculture. J. Ambient. Intell. Smart Environ. 2020, 12, 21–34. [Google Scholar] [CrossRef]
  20. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. Adv. Neural Inf. Process. Syst. 2015, 2015, 802–810. [Google Scholar]
  21. Mehrkanoon, S. Deep shared representation learning for weather elements forecasting. Knowl.-Based Syst. 2019, 179, 120–128. [Google Scholar] [CrossRef]
  22. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph convolutional networks: A comprehensive review. Comput. Soc. Netw. 2019, 6, 1–23. [Google Scholar] [CrossRef] [Green Version]
  23. Wilson, T.; Tan, P.N.; Luo, L. A Low Rank Weighted Graph Convolutional Approach to Weather Prediction. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; pp. 627–636. [Google Scholar]
  24. Khodayar, M.; Wang, J. Spatio-temporal graph deep neural network for short-term wind speed forecasting. IEEE Trans. Sustain. Energy 2018, 10, 670–681. [Google Scholar] [CrossRef]
  25. Wang, S.; Li, Y.; Zhang, J.; Meng, Q.; Meng, L.; Gao, F. PM2.5-GNN: A Domain Knowledge Enhanced Graph Neural Network For PM2.5 Forecasting. In Proceedings of the 28th International Conference on Advances in Geographic Information Systems, Seattle, WA, USA, 3–6 November 2020; pp. 163–166. [Google Scholar]
  26. Gao, J.; Sharma, R.; Qian, C.; Glass, L.M.; Spaeder, J.; Romberg, J.; Sun, J.; Xiao, C. STAN: Spatio-temporal attention network for pandemic prediction using real-world evidence. J. Am. Med Inform. Assoc. 2021, 28, 733–743. [Google Scholar] [CrossRef]
  27. Song, C.; Lin, Y.; Guo, S.; Wan, H. Spatial-temporal synchronous graph convolutional networks: A new framework for spatial-temporal network data forecasting. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 914–921. [Google Scholar]
  28. Lu, Y.J.; Li, C.T. AGSTN: Learning Attention-adjusted Graph Spatio-Temporal Networks for Short-term Urban Sensor Value Forecasting. In Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy, 17–20 November 2020; pp. 1148–1153. [Google Scholar]
  29. Seo, Y.; Defferrard, M.; Vandergheynst, P.; Bresson, X. Structured sequence modeling with graph convolutional recurrent networks. In Proceedings of the International Conference on Neural Information Processing, Siem Reap, Cambodia, 13–16 December 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 362–373. [Google Scholar]
  30. Yu, B.; Yin, H.; Zhu, Z. Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting. arXiv 2017, arXiv:1709.04875. [Google Scholar]
  31. Zhao, L.; Song, Y.; Zhang, C.; Liu, Y.; Wang, P.; Lin, T.; Deng, M.; Li, H. T-GCN: A temporal graph convolutional network for traffic prediction. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3848–3858. [Google Scholar] [CrossRef] [Green Version]
  32. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. arXiv 2014, arXiv:1409.3215. [Google Scholar]
Figure 1. Graph modeling based on location of experimental site and weather stations. (a) Map of weather stations and experimental site; (b) Schematic version of a graph constructed at a single time step.
Figure 1. Graph modeling based on location of experimental site and weather stations. (a) Map of weather stations and experimental site; (b) Schematic version of a graph constructed at a single time step.
Sensors 22 01486 g001
Figure 2. Internet of things (IoT) platform installed on experimental field composed by temperature and humidity sensors.
Figure 2. Internet of things (IoT) platform installed on experimental field composed by temperature and humidity sensors.
Sensors 22 01486 g002
Figure 3. Overview of GRAST-Frost model.
Figure 3. Overview of GRAST-Frost model.
Sensors 22 01486 g003
Figure 4. Schematic view of spatio-temporal graph network for modeling relationships between experimental field and weather stations through time.
Figure 4. Schematic view of spatio-temporal graph network for modeling relationships between experimental field and weather stations through time.
Sensors 22 01486 g004
Figure 5. Spatio-temporal graph construction and adjacency matrix for including multiple time steps.
Figure 5. Spatio-temporal graph construction and adjacency matrix for including multiple time steps.
Sensors 22 01486 g005
Figure 6. MAE, MAPE, and RSME metric results for temperature forecasting on 6, 12, and 24 h in advance.
Figure 6. MAE, MAPE, and RSME metric results for temperature forecasting on 6, 12, and 24 h in advance.
Sensors 22 01486 g006
Figure 7. RMSE and Pearson statistic results regarding forecasting on the experimental field and three nearest weather stations.
Figure 7. RMSE and Pearson statistic results regarding forecasting on the experimental field and three nearest weather stations.
Sensors 22 01486 g007
Figure 8. On left, there is a comparison of GRAST-Frost model with real data; on right, a comparison of STA-GRU model with real data.
Figure 8. On left, there is a comparison of GRAST-Frost model with real data; on right, a comparison of STA-GRU model with real data.
Sensors 22 01486 g008
Figure 9. GRAST-Frost model results compared with that of real data for 6, 12, 24, and 48 h forecasts.
Figure 9. GRAST-Frost model results compared with that of real data for 6, 12, 24, and 48 h forecasts.
Sensors 22 01486 g009
Figure 10. Receiver operating characteristic (ROC) curve comparing GRAST-Frost with other models implementations.
Figure 10. Receiver operating characteristic (ROC) curve comparing GRAST-Frost with other models implementations.
Sensors 22 01486 g010
Table 1. Average mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) metrics of time series forecasting models applied to frost forecast. Best values are highlighted in bold.
Table 1. Average mean absolute error (MAE), root mean square error (RMSE), and mean absolute percentage error (MAPE) metrics of time series forecasting models applied to frost forecast. Best values are highlighted in bold.
MAERMSEMAPE
Model6 h12 h24 h6 h12 h24 h6 h12 h24 h
VAR FNN4.284.765.747.888.6010.757.218.779.98
LSTM4.034.765.466.937.359.116.956.898.61
AC-att3.413.794.265.896.647.445.745.847.05
GCN3.613.894.286.017.097.465.555.897.03
STA-C3.774.024.836.558.028.716.117.878.30
STA-GRU3.523.734.176.147.207.395.825.997.05
GRAST-Frost3.083.253.865.425.796.195.395.446.04
Table 2. Average precision, recall and F1 metrics for classification of frost events according different models implementations. Best values are highlighted in bold.
Table 2. Average precision, recall and F1 metrics for classification of frost events according different models implementations. Best values are highlighted in bold.
PrecisionRecallF1
Model6 h12 h24 h6 h12 h24 h6 h12 h24 h
SVM0.5690.5350.5130.5650.5290.5070.5550.5310.502
XGBoost0.7540.7400.7130.7490.7300.7010.7480.7160.711
FNN0.6440.6430.6380.6630.6580.6390.6900.6690.658
LSTM0.6750.6660.6520.6950.6500.6320.6820.6470.641
AC-att0.7130.7090.6900.7050.7000.6960.6990.6910.671
GCN0.7920.7870.7800.7840.7780.7690.8060.8020.787
STA-C0.8140.7970.7940.8370.8100.8090.8300.8260.784
STA-GRU0.8450.8440.8240.8500.8210.8050.8630.8450.842
GRAST-Frost0.8910.8760.8480.8980.8820.8450.9080.8690.863
Table 3. t-test for comparing our regression model with baselines for 6 h, 12 h, 24 h and 48 h forecasts.
Table 3. t-test for comparing our regression model with baselines for 6 h, 12 h, 24 h and 48 h forecasts.
6 h12 h24 h48 h
Modeltp-Valuetp-Valuetp-Valuetp-Value
HA9.6620.0117.7160.0227.3980.0287.2760.031
ARIMA9.6730.0129.4430.0268.5340.00368.1390.033
VAR10.3300.0178.2470.0417.8310.0457.1720.044
FNN5.9460.0205.7220.0255.5950.0304.2390.041
LSTM6.5400.0296.2600.0245.1440.0394.1360.045
AC-att4.2930.0173.6740.0203.5620.0253.1790.035
STA-C3.3140.0112.9800.0142.5930.0192.2840.041
STA-GRU1.3870.0301.2710.0381.0090.0410.7660.081
Table 4. t-test for comparing our classification model with baselines for 6 h, 12 h, 24 h, and 48 h forecasts.
Table 4. t-test for comparing our classification model with baselines for 6 h, 12 h, 24 h, and 48 h forecasts.
6 h12 h24 h48 h
Modeltp-Valuetp-Valuetp-Valuetp-Value
SVM9.6100.0159.0420.0367.9460.0387.9390.049
NB9.5270.0238.1290.0327.1390.0457.0510.046
XGBoost3.7860.0133.7120.0223.4670.0413.3780.043
FNN6.7650.0166.0020.0264.7580.0304.2060.036
LSTM6.0950.0185.2450.0184.5860.0264.4880.046
AC-att4.8860.0254.3980.0293.6650.0343.4790.040
STA-C3.0180.0232.5790.0292.1520.0472.0840.048
STA-GRU1.4760.0391.0940.0360.9860.0470.7130.071
Table 5. t-test for comparing our model with graph-based models without spatio-temporal attention for 6 h, 12 h, 24 h and 48 h forecasts.
Table 5. t-test for comparing our model with graph-based models without spatio-temporal attention for 6 h, 12 h, 24 h and 48 h forecasts.
6 h12 h24 h48 h
Modeltp-Valuetp-Valuetp-Valuetp-Value
GCNreg3.6310.0133.4120.0182.4560.0332.4160.035
GCNclas4.0090.0123.8610.0312.8590.0322.7870.041
Table 6. t-test for comparing our model forecast results from experimental site with weather stations.
Table 6. t-test for comparing our model forecast results from experimental site with weather stations.
6 h12 h24 h48 h
Modeltp-Valuetp-Valuetp-Valuetp-Value
WS10.9630.0750.8100.0700.7640.0980.6650.093
WS20.8430.0710.6830.0610.6290.0690.5400.084
WS30.8020.0750.6760.0750.5680.0810.5010.083
WS40.8210.0640.7310.0640.6910.0780.5930.088
WS50.9150.090.0250.1023.9210.1364.3260.119
WS60.8170.1110.6600.0810.6070.1050.5350.116
WS70.9300.0870.7030.0990.6690.0920.6530.099
WS80.9150.0780.0250.0833.9210.1184.3260.139
WS90.9240.0600.9180.0860.7400.1110.7230.118
WS100.9030.0940.7940.0970.7410.1260.6020.137
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lira, H.; Martí, L.; Sanchez-Pi, N. A Graph Neural Network with Spatio-Temporal Attention for Multi-Sources Time Series Data: An Application to Frost Forecast. Sensors 2022, 22, 1486. https://doi.org/10.3390/s22041486

AMA Style

Lira H, Martí L, Sanchez-Pi N. A Graph Neural Network with Spatio-Temporal Attention for Multi-Sources Time Series Data: An Application to Frost Forecast. Sensors. 2022; 22(4):1486. https://doi.org/10.3390/s22041486

Chicago/Turabian Style

Lira, Hernan, Luis Martí, and Nayat Sanchez-Pi. 2022. "A Graph Neural Network with Spatio-Temporal Attention for Multi-Sources Time Series Data: An Application to Frost Forecast" Sensors 22, no. 4: 1486. https://doi.org/10.3390/s22041486

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop