Next Article in Journal
Evaluation and Optimization of the Oil Jet Lubrication Performance for Orthogonal Face Gear Drive: Modelling, Simulation and Experimental Validation
Previous Article in Journal
Optimization of an Organic Rankine Cycle System for an LNG-Powered Ship
Previous Article in Special Issue
Prediction of the Optimal Vortex in Synthetic Jets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Study of Time Series Forecasting Methods for Short Term Electric Energy Consumption Prediction in Smart Buildings

by
Federico Divina
1,*,
Miguel García Torres
1,
Francisco A. Goméz Vela
1 and
José Luis Vázquez Noguera
2
1
Division of Computer Science, Universidad Pablo de Olavide, ES-41013 Seville, Spain
2
Ingeniería Informática, Universidad Americana, Asunción PY-1429, Paraguay
*
Author to whom correspondence should be addressed.
Energies 2019, 12(10), 1934; https://doi.org/10.3390/en12101934
Submission received: 9 April 2019 / Revised: 14 May 2019 / Accepted: 15 May 2019 / Published: 20 May 2019
(This article belongs to the Special Issue Data Science and Big Data in Energy Forecasting with Applications)

Abstract

:
Smart buildings are equipped with sensors that allow monitoring a range of building systems including heating and air conditioning, lighting and the general electric energy consumption. Thees data can then be stored and analyzed. The ability to use historical data regarding electric energy consumption could allow improving the energy efficiency of such buildings, as well as help to spot problems related to wasting of energy. This problem is even more important when considering that buildings are some of the largest consumers of energy. In this paper, we are interested in forecasting the energy consumption of smart buildings, and, to this aim, we propose a comparative study of different forecasting strategies that can be used to this aim. To do this, we used the data regarding the electric consumption registered by thirteen buildings located in a university campus in the south of Spain. The empirical comparison of the selected methods on the different data showed that some methods are more suitable than others for this kind of problem. In particular, we show that strategies based on Machine Learning approaches seem to be more suitable for this task.

1. Introduction

The rapid economic growth and the increasing population in developing countries are causing the global energy consumption to increase. Despite efficiency gains over the last decades, it is estimated that by 2040 the global energy demand will likely increase nearly 25 percent [1]. According to the International Energy Agency (IEA) [2], developing regions are expected to experience large demand growth between 2017 and 2040, especially countries from Asia and Africa. Under the studied scenario, CO 2 emissions from energy will rise around 10 % (from 32.6 to 35.9 gigatons). While in developed countries CO 2 emissions are expected to drop by 23 % , in developing regions they are supposed to rise by 27 % .
Energy consumption has a significant impact on the environment. CO 2 emissions have been identified as responsible of most of the progressive warming of the Earth [3]. Furthermore, current predictions show that this growing trend will continue. Therefore, the reduction of CO 2 emissions has become the focus of international political, economic and environmental research. To lower the energy consumption and help protect the environment, new energy efficiency policies are being implemented. For example, the current energy plan developed by the European Commission requires European Union’s countries adopt a set of measures to reach an energy efficiency of at least 20 % [4]. Despite the fact that many industries contribute to such emissions, the building and building construction sectors are responsible for 36 % of total energy consumption and nearly 40 % of global CO 2 emissions [5].
In an efficient scenario, several building sub-sectors, such as space heating and cooling, water heating, lighting, etc., have been identified for potential energy savings [5]. In this context, the development of energy consumption prediction models are growing in importance for decision-making to implement energy policies. The forecasting problem is generally divided into three categories, based on the prediction horizon: short-term, medium-term and long-term load forecasting. Short-term forecasting involves prediction horizons going from one hour up to a week while medium-term refers to predictions from one moth up to a year. Finally, long-term predictions are characterized by a prediction horizon of more than a year [6]. Most works address short-term prediction horizon since it achieves higher accuracy than the other horizons.
Nowadays, more and more buildings are equipped with sensors that can measure various aspects of the functioning of a building, including the electric energy consumption registered. Such buildings are called smart buildings. Such measurement can be stored, and thus can provide valuable historical data, which can be organized as time series, and thus could be used to predict future energy demands of the buildings.
Traditionally, time series have been tackled using conventional methods, such as statistical analysis, smoothing techniques and exponential smoothing and regression-based approaches. For example, Sen et al. [7] applied the Auto Regressive Integrated Moving Average (ARIMA) approach for forecasting energy consumption and greenhouse gas emission in Indian pig iron manufacturing organizations. In [8], Chujai et al. analyzed the household electric consumption using the Autoregressive Moving Average (ARMA) and ARIMA models. However, despite their popularity, these models are not able to capture complex interactions from non-linear data. In this case, Machine Learning (ML) approaches have arisen as a suitable approach to handle such complexity. For instance, Bonetto and Rossi [9] applied Support Vector Machines and Neural Networks to energy consumption in households.However, despite the good performance achieved with Machine Learning techniques, they may get stuck in a local optimum, a phenomenon that negatively affects their performances. To overcome this problem, in recent works, many authors are exploring new approaches. In Divina et al. [10], Divina et al. proposed to apply an ensemble approach, while, in [11], Meira and Oliveira introduced a bagging approach of conventional methods (ARIMA and exponential smoothing methods).
In this work, we propose a comparative empirical evaluation of different time series forecasting strategies. In particular, we consider both statistical and ML based approaches to this problem. To validate the techniques, we applied them on a dataset regarding the electricity consumption registered by thirteen buildings located at the Pablo de Olavide (UPO) University campus in Seville, Spain, collected over five and a half years. In the experiments conducted, we aimed at predicting the electric energy consumption with a one day horizon. Predictions were based on historical data, and another objective of the experimentation proposed was to assess an optimal value of the amount of historical data that should be used for this kind of problem. Therefore, we can summarize the contributions of this work as follows:
  • Analysis and comparison of the performance of statistical and ML based strategies;
  • Establishing the size of the historical window to be used to optimize the predictions; and
  • Analysis of the electricity consumption data collected from the smart buildings considered.
Results obtained show that Machine Learning approaches could achieve better results, in particular methods based on bagging and boosting ensemble schemes obtained the best results. Moreover, we found that using historical data relating to more than seven days yielded better predictions.
The rest of the paper is organized as follows. In Section 2, some related works are introduced. Then, the data and the techniques used are described in Section 3. In Section 4, the analysis and discussion of the experiments conducted are presented. Finally, the conclusions and future works are presented in Section 5.

2. Related Works

In the scientific community, there is a growing interest in addressing the forecasting problem in energy-related problems such as electricity consumption, loading and demand for various reasons. For example, as pointed out in [12], to create a healthy and sustainable economy, it is necessary to measure the socio-economic and environmental impact of energy production. Another important aspect is the identification of the different electricity consumption behaviors, which may allow adopting policies according to demand response scenarios [13]. Being able to predict future energy demands would provide several concrete benefits, at both economic and social levels. For example, the ability to balance fluctuations in renewable energy generation would facilitate a greater penetration of renewable resources into the electricity system [14]. On the other hand, we may also improve the economic efficiency by applying real-time prices and reducing costs connected to the generation capacity requirements, reducing at the same time the related CO 2 emissions [15,16].
In the literature, time series analysis is the most popular approach for forecasting demands [17]. Among the most widely used time series forecasting methods, we can find the autoregressive moving average (ARMA) and the autoregressive integrated moving average (ARIMA). ARMA and ARIMA are statistical approaches that have been widely applied to short-term forecasting problem. For example, Abdel-Aal and Al-Garni [18] used, in an early work, ARIMA to forecast monthly electric consumption in the Eastern Province of Saudi Arabia. Chujai et al. [8] compared ARIMA with ARMA on household electric power consumption. ARIMA model was also applied by Shirpa and Shashadri [19] to develop a short-term electric load forecasting model of Karnataka State, India.
Various extensions of ARMA and ARIMA have been proposed to include different aspects affecting a time series: the Auto Regressive Integrated Moving Average with eXternal (or eXogenous) input (ARIMAX), which add explanatory variables; the Seasonal Autoregressive Integrated Moving Average (SARIMA), which supports the direct modelling of the seasonal component of the series; or the Multiplicative Seasonal Autoregressive Integrated Moving Average (MSARIMA), which incorporates independent explanatory variables to SARIMA. The ARIMAX was used by Newsham and Birt in [20] to forecast several hours ahead the power demand for an office building with the objective of obtaining a better response to utility signals and reducing the overall energy use. MSARIMA was applied in [21] by Rallapalli and Ghosh to forecast monthly peak electricity demand from five different regions of India with the aim of achieving a better load management.
Regression based techniques represent another branch of classical methods used in time series forecasting. In particular, linear regression has been extensively applied. For instance, Schrock and Claridge [22] used a linear regression approach to investigate the hourly and daily electrical consumption pattern of a supermarket located 100 miles north-northwest of Houston, Texas. Nowotarski et al. [23] applied a family of regression models trained on different subsets of variables. More specifically, data generated from the Global Energy Forecasting Competition 2014 and ISO New England were used. In [24], Samarasinghe and Al-Hawani compared multiple linear regression with Gaussian Processes on power consumption data from 2008 to 2010 to forecast the values in the next 24 h in Norway. In a more recent work, Rahman et al. [25] proposed a high-precision methodology that included multiple linear regression and simple regression model along with other techniques to forecast the total energy consumption in India.
Other forecasting approaches are based on Machine Learning techniques. Within these approaches, the use of Artificial Neural Networks (ANN) have been extensively and successfully applied. In an early work, Park et al [26] used an ANN to induce the relationship among past, current and future temperatures and loads in the electric load forecasting problem. The model was tested on hourly temperature and load data from the Seattle and Tacoma areas in the period from 1 November 1988 to 30 January 1989, with the objective of predicting 1-h and 24-h ahead values. In another early work, Nizami and Ai-Garni [27] proposed a two-layered fed-forward ANN to study how weather-related features may affect the prediction of monthly electric energy consumption. This approach was tested on data collected from the Eastern Province of the Saudi Arabia between August 1987 and July 1993. In more recent works, ANNs are used to tackle the problem of short-term electrical load forecasting, e.g., [28,29,30,31]).
In addition to ANNs, other Machine Learning techniques have been successfully applied to the problem of energy consumption forecasting. For example, Support Vector Regression (SVR) was applied by Jain et al. [32] to develop an energy forecasting model for a multi-family residential building in New York City. The data used in this work span from 27 August 2012 to 19 December 2012. The model was tested using three different temporal granularities: daily, hourly and 10 min. SVR was also applied by Liu et al. [33] to show its effectiveness in predicting hourly energy consumption of buildings. The popular K-Nearest Neighbour (k-NN) technique was used in [34] to explore the performance of the model for short-term electric energy demand. Other approaches were based on Deep Learning [35,36], Random Forest [37] and Evolutionary Algorithms [38]. Finally, it is worth noting that many of these techniques have been adapted within the big data paradigm. For example, Talavera-Llames et al. [39] adapted the strategy proposed in [34] by Lora et al.
Another strategy that has been used is to combine different strategies to get advantage from the different approaches. For instance, Geng et al. [40] proposed a hybridization of SVR with Chaotic Sequence and Simulated Annealing to forecasting monthly electric load data from northeast China and New York City. Both datasets ranges from January 2004 to October 2005. Fan et al. [41] proposed a novel hybrid method that combines SVR, differential empirical mode decomposition (DEMD) strategy and auto regression (AR) for electric load forecasting. The resulting strategy was tested on two datasets. The first consisting of half-hourly electric load data from the New South Wales while the second one of hourly electric load data from the New York Independent System Operator.
Lately, ensembles approaches have also received considerable attention. Jetcheva et al. [42] proposed an ensemble model based on ANNs for day-ahead electricity load and generation forecasting. The model was tested on data from commercial and industrial sites collected over a period of 10 months ranging from July 2011 to April 2012. In another work, Khairalla et al. [43] introduced a stacking multi-learning ensemble scheme to address the problem of short-term energy consumption forecasting. The proposal was tested on the Global Oil Consumption dataset collected over 52 years (1965–2016). Another example of a stacking ensemble approach can be found in [10]. In this case, Divina et al. applied such an approach to the global short-term energy consumption forecasting problem in Spain.
To get a deeper insight on this topic, we refer the reader to [44], where an exhaustive review of Machine Learning techniques for time series forecasting is presented. Reviews of conventional and Artificial Intelligence methods are presented in [15,45].

3. Materials and Methods

3.1. Data

The time series used in this paper refer to data collected by different sensors installed in 13 buildings located on a university campus in the south of Spain. The selected buildings are equipped with sensors that are able to collect the energy consumption every 15 min. However, for the study presented in this paper, a daily consumption resolution was used. In particular, the data used in this paper cover the daily electric energy consumption registered for the buildings over a period of five and a half years, namely from 1 March 2012 to 31 October 2017.
The campus where the buildings are located consists of 33 buildings. We used only 13 buildings since not all buildings are equipped with sensors for registering the electricity consumption.
Table 1 shows a general description of the selected buildings, including the year of construction (and year of refurbishment where applicable), the size and a description of the main purpose of the building.
The geographical location of the campus is characterized by extremely high temperatures, with maximum temperatures exceeding 40 ° C during the summer months, and by relatively mild winters, with minimum temperatures of approximately 0–2 ° C. This climate is termed Mediterranean hot summer climate (Csa) according to the Köppen–Geiger classification. Table 2 shows average temperatures, average maximum and minimum temperatures, the amount of rainfall and average annual wind speed recorded for the years considered in this paper.
The raw data collected presented missing values. To overcome this issue, the raw data were preprocessed. For each building, we computed the daily average electric energy consumption in Kwh. However, raw data presented missing values on some days and we proceeded to apply a preprocessing step to handle this issue. The steps are summarized as follows: (a) days with missing values corresponding to periods of more than 8 h (either continuous or discontinuous) were discarded; (b) days with missing values corresponding to a continuous period of time corresponding to more than 4 h were discarded; and (c) in the rest of the cases, the missing values were approximated by means of a linear regression method. The same approach was used in the case of missing values corresponding to a period of time of less than 8 h for the period from 00:00 to 08:00. In total, 260 days were discarded, thus the final dataset consists of 1810 days. The final dataset is available upon request.
Figure 1 shows the electric consumption and the general trend of each building over the period of time analyzed, scaled using a logarithmic scale in base 10. We used a logarithmic transformation since some significant outliers were present and would have rendered the graphs relative to the original data difficult to read. More specifically, significant outliers are present in buildings 3, 24 and 25. To further analyze such anomalous cases, we checked the temperature registered by Spanish National Agency of Meteorology [46] on such dates. We verified that no sudden changes in the temperatures can justify these anomalous values. Moreover, we verified that there are no records of anomalous operation in the buildings. We decided to maintain such anomalous values to simulate the real functioning of the sensors, and to check in this way the robustness of the techniques to the presence of outliers.
In this way, we could assess the robustness of the various strategies applied to these time series against the presence of outliers, since they are likely to being produce in the normal functioning of sensors.
Figure 2 shows the distribution of the values in the final dataset for the buildings considered in this paper. From the results this figure we could confirm the presence of the outliers in the buildings highlighted above. However, we can also notice that there are anomalous values in almost all the buildings, even if their values do not exceed by too much the values considered to be normal. These values may be due to a malfunctioning of the sensors as well as extreme peaks of temperatures. We can also noticed that Building 25 presents a wider distribution of values with a higher average. These aspects could be partially explained by the fact that Building 25 hosts the university’s library, which consists of a huge open space. This, and the particular climatic conditions of south of Spain, implies that this building uses more energy for the heating, ventilation, and air conditioning (HVAC) than other buildings that consist of smaller spaces.
To complete our study of the time series, we performed an analysis aimed at establishing whether the time series used presented stationarity behavior or not. In particular, we performed the Ljung–Box (LB) statistical tests for stationarity [47]. The p-value returned by this test is, in all cases, lower than 0.01 except for Building 3. To further investigate this case, we present in Figure 3 the the Partial AutoCorrelation Function (PACF) and the AutoCorrelation Function (ACF) pertaining to the data of Building 3. We can notice that in both plots there are no lags that exceed the confidence interval of the ACF and PACF plots, respectively. We can then conclude that Building 3 shows a stationary behavior.

3.2. Methods

In this section, we provide a basic description of the methods used in the experimentation, starting from the classical statistical approaches, namely linear regression and ARIMA. After that, we describe the Machine Learning techniques used in our empirical study.
Linear Regression (LM). Ref. [48] is statistical technique commonly used for time series forecasting. The basic idea behind linear regression is to try to find the relationship between two variables. In its simplest formulation, a linear equation, Y = a + b X , is used to represent the association between the independent variable (X) and the dependent one (Y), i.e., the variables to be predicted. In the case where multiple independent variables are used to determine the value of a dependent variable, as is the case of this work, the process is called multiple linear regression. In this case, the goal is to model the linear relation between multiple independent, or explanatory, variables and a dependent variable using an equation such as Y = a + b 1 X 1 + + b n X n + ϵ , where ϵ is the residual (difference between the predicted and the observed value) and X i , 1 i n are the n explanatory variables. For this work, we used the implementation provided by the caret [49] R package.
Auto-Regressive Integrated Moving Average (ARIMA). is a classical method based on Box and Jenkins [50] widely used for time series forecasting, as shown in Section 2. ARIMA is an extension of the simpler AutoRegressive Moving Average (ARMA) that includes the concept of integration. We used this method since it can be used to forecast non-stationary time series. In [51], ARIMA is described as a stochastic model building. In particular, this method can be viewed as an iterative strategy that consists of three steps:
  • Identification. In this step, the data and all related information are used for selecting a sub-class of model that might best represent the data.
  • Estimation. In a second step, the parameters of the model are trained using the data.
  • Diagnostic Checking. Finally, the so-obtained model is validated using the data at disposal and areas where the model can be improved are identified.
ARIMA requires setting three parameters, namely the order of auto-regressive (AR) terms (p), the degree of differencing (d) and the order of moving-average (MA) terms (q). To estimate such parameters for each dataset and each historical window used, we used the arimapar function, part of the TSPred R package [52]. This function returns the parameters of an automatically fitted ARIMA model. The implementation provided by the caret.ts [53] package (an extension for time series models for caret package) was used.
Evolutionary Algorithms (EAs) for Regression Trees (EVTree). EAs [54] are optimization techniques based on the basic concepts of Darwinian evolution. EAs evolve a population of candidate solutions over a number of generations by means of the application of genetic operators, such as selection, crossover and mutation. Each candidate solution, i.e., an individual of the population, is assigned a quality score, or fitness. Usually, EAs start from a randomly initialized population, which is then evaluated to assign a fitness to each individual. A number of individuals are then selected, usually based on the fitness. Crossover and mutation operators are used to generate new individuals starting from the selected ones. The so-generated offspring are then inserted into the population. The whole process is repeated until a stopping criterion is met, e.g., when a maximum number of generations have been performed. The particular EA provided by the EVTree [55] R package was used. This algorithm evolves regression trees [56], i.e., trees used to predict a continuous value based on a set of predictor variables.
Generalized Boosted Regression Models (GBM). [57,58]. This is an ensemble algorithm, where a set of regression trees is trained following a sequential procedure. During this sequential procedure, GBM applies a gradient descend procedure, which is repeated until a given number of trees has been built or when no improvement is detected. The last condition is checked by using the set of current trees to produce predictions. Such predictions are also used to correct the model, so that mistakes obtained in previous iterations can be corrected. A known issue in gradient boosting is overfitting. GBM tackles this problem by using regularization methods, which basically penalize various parts of the algorithm. To produce the final prediction a voting mechanism, among all the tree obtained, is used. We used the GBM implementation provided by the caret R package.
Artificial Neural Networks (ANNs). In a rough analogy of biological learning system, ANNs consist of densely interconnected units, called neurons. Each neuron receives a number of real valued input, e.g., from other neurons of the network, and produces a single real valued output. The output depends on an activation function used in each unit, which introduce non-linearity to the output. However, the activation function is used only if the input received by a unit is higher than a given activation threshold. If this is not the case, then no output is produced. Normally, an ANN consists of different layers of neurons. Among such layers we can distinguish the input and output layers, and in between there may be one or more hidden layers. As said before, there is an activation threshold used in the network, which is determined by weights that are adjusted during a training phase. Several types of ANNs have been proposed. Among these, the simplest, and one of the most widely used one, is the so-called feedforward ANN. In these type of networks, neurons of adjacent layers are connected, and each of these connections is assigned a weight. The information advances from the input layer toward the output layer, which consists of only one unit. The output unit produces the final prediction of the network. The caret package implementation was used.
Random Forests (RF). were first proposed by Breinman and Cutle in [59]. Similar to GBM, RF is also an ensemble approach, where a set of trees are used to produce the final output, using a voting scheme. Each tree is induced from a randomly selected training subset and using also a randomly selected subset of features. This implies that the trees depend on the values of an independently sampled input dataset, using the same distribution for all trees. In the case of regression, the final prediction is represented by the average of the predictions of each induced tree. In addition, for this method, the implementation provided by caret was used.
Ensemble. This method was proposed by Divina et al. [10]. It is a stacking ensemble scheme. In particular, two layers are used to make a final prediction. In the first layer, EVTree, RF and NN are used, and the predictions made by these three methods are then combined on the top layer by GBM to produce the final output. This method shows excellent performances on a problem regarding the prediction of the global energy demands, this it is interesting to check its validity also in the case of a local energy demand setting, such as the one used in this paper.
Recursive Partitioning and Regression Trees (RPart). [60]. This method is based on the CART algorithm, and builds regression trees, using a two states strategy. The resulting model can be represented as a binary tree, which can be easily interpreted. In the first step, the algorithm determines the variable which best splits the data into two groups. For regression, as in this case, the attribute with the largest standard deviation reduction is considered the best and is used for building the decision node. The data are then separated according the the attribute selected, and the process is applied separately to each sub-group determined in the first phase, until the subgroups either get to a minimum size or no improvement can be achieved. In a second step, the obtained full tree is pruned using a cross-validation strategy. Again, the caret implementation was used.
Extreme Gradient Boosting (XGBoost). [61]. This method is similar to GBM, as it shares with it the principle of gradient boosting for building an ensemble of trees. There are, however, important differences. For instance, XGBoost controls over-fitting (a known issue in GBM) by adopting a more regularized model formalization. This allows XGBoost to generally obtain better performance than GBM. XGBoost has recently received much attention in the data science community, as it has been successfully applied in different domains. This popularity is mostly due to the scalability of the method. In fact, XGBoost can run up to ten times faster than other popular approaches on a single machine. Moreover, it is also capable of scaling to billions of instances in distributed or memory-limited settings. The latter feature is achieved via different systems and algorithmic optimizations. Some of this improvements include a novel tree induction algorithm for managing sparse data and a strategy that makes it possible to handle the weights of the instances in approximate tree learning. Parallel and distributed computing makes learning faster, allowing a faster model exploration. Additionally, XGBoost employs out-of-core computation, which enables to process massive date even of a simple desktop. We used the implementation provided by the caret R package.
All the methods used in this study have various configuration parameters. To set the different parameters of the methods, we ran a grid search, exploring the standard parameter values that are used in similar literature.

4. Results

In this section, we present the results obtained by the different techniques described in Section 3.2 on the datasets presented in Section 3.1 and draw the main conclusions. The aim of the experimentation performed was two folds. Firstly, we wanted to assess which technique is more suitable for predicting short term electric energy consumption in a local setting, such in smart buildings. Secondly, we wanted to find an optimal value for the amount of historical data that should be used to obtain good forecasts.
To measure the performances of the different strategies used, we used the Mean Absolute Error (MAE) and the Root Mean Squared Error (RMSE). The measures are defined as follows:
MAE = 1 n i = 1 n | Y i Y ^ i |
RMSE = 1 n i = 1 n ( Y i Y ^ i ) 2
where Y ^ i and Y i are the predicted and the real values, respectively. These two measures are commonly used when assessing the performances of a technique on time series forecasting (see, for instance [62,63]). The advantage of these two measures is that the average forecast error of a model is expressed in the same units of the variable to be predicted. The two measures can assume values greater than or equal to 0, and lower values are considered better. MAE expresses the absolute error, thus it is easy to understand. RMSE assigns high penalties to large errors, since the prediction errors are squared. It follows that the RMSE can be useful when we want abstain from large forecasting errors.
To be used in the proposed experimentation, the original datasets were transformed. The particular preprocessing steps are those used by Torres et al. [36], and are graphically shown in Figure 4. The preprocessing phase can be summarized in two steps. Firstly, we extracted the attribute representing the energy consumption, and as a result a vector of consumption data was obtained. Secondly, a matrix was built using the vector obtained in the first step and according to a historical window, w, and a prediction horizon, h. Notice that w represents the number of historical values that are taken into account so as to induce a predictive model. The obtained model was used to forecast subsequent values (h). In our study, we were interested in predicting the consumption of one day, thus h was set to 1.
Moreover, since one of the aims of the experimentation was to find an optimal value of w, different values of historical window were tested. Specifically, we used historical windows of 7 , 10 , 15 and 20 days. We also tested historical window sizes of multiples of 7, which are common in these kinds of problems, since they usually allow capturing seasonal patterns. However, in this case, the results obtained with such values are not better than those obtained with the above mentioned values, and thus are not included in this paper. This was probably due to the high number of days discarded in the preprocessing of raw data (described in Section 3.1).
The resulting datasets were split in two groups: training set and test set. In particular, we used 70 % of the data as training set, and the remaining 30 % as test set. Thus, the methods were used to induce a prediction model using the training set, and the forecasting capabilities of the models were assessed on the test set.
Table 3 presents the MAE and RMSE obtained on the different buildings for the different historical windows used. In particular, in the last row, the averages results obtained on all the buildings by the different methods are shown. The same results are presented graphically in Figure A1 and Figure A2.
One thing that we can notice is that, generally, ARIMA did not perform well on the datasets considered. In fact, on average, it obtained the worst results, in terms of both MAE and RMSE, on all buildings for all historical window values considered. In particular, in some cases, results achieved by this methods are significantly worse, as is the case, for example, for the results obtained on Buildings 2, 6, 7, 10 and 12.
We can also notice that, in general, better results were achieved when the historical window size used was greater than 7. Higher values of the historical window did not seem to yield better results, thus we can affirm that considering a historical window of more than 20 days would not lead to better predictions.
Looking more closely to each building, we can notice that the worst predictions were those related to Building 25, on which, for example, the mean absolute error achieved by the different strategies was always higher than 3. This could be due to the different features that this building presents, as can be seen in Table 1 and Figure 2. In the same figure, we can also notice that Buildings 3 and 24 presented noticeable outliers values, but, in this case, the forecasting methods did not seem to be affected by their presence. Thus, it appeared that the techniques used in this study are quite robust to the presence of outliers, and that the worst results obtained on Building 25 might be due more to the nature of the building.
Figure 5 and Figure 6 show the results, in terms of MAE and RMSE, obtained by the different methods on all the buildings when different sizes of the historical window were used. By analyzing these results, we can conclude that, in general, the best size for the historical window to use is 10 days. In fact, with this value, the results obtained are generally better, in terms of both MAE and of RMSE. This is an important result, since it may help researchers to set the historical window for other time series forecasting regarding data generated by smart buildings, especially in the case where various days present missing values and cannot be considered for inducing a forecasting model. We can also see that ARIMA failed to produce good forecasts for the datasets used in this study. Another difference between ARIMA and the other methods is that, in general, the historical window of 10 days did not yield better results when used with this method. From the graphs, we can also notice the presence of outliers for both MAE and RMSE. Such values are relative to the results obtained on Building 25, where all methods used in this study achieved the worst prediction results. This was probably due to the characteristics of this building, which is different from the others, as it contains the university’s library. It is a newer building, whose surface is also much bigger than that of the other buildings. Moreover, the average energy consumption is significantly higher than the other buildings, as shown in Figure 2. Outliers relative to the RMSE were also due to the results obtained on Building 3, meaning that the predictions errors were larger on this building. In this case, the building is similar to the other ones considered, in terms of both features and for its usage and average energy consumption. This suggest that a further analysis conducted with the mangers of the buildings is needed for this case to acquire more insights.
In Table 4 and Table 5, we reported the average performances of the methods for all historical windows values and for all the buildings, and the average results obtained when a historical window of size 10 was used, since this value was considered to be the best one in the experimentation presented. Moreover, in the tables, we also report results obtained with two simple baseline benchmark forecasting models. In particular, we considered the energy consumption to be the same as the one of the previous day and of the previous week, i.e., of seven days earlier. Results obtained by these two methods are, as expected, not competitive if compared with the other methods. Results are ordered according to the averages results. In this way, we can obtain a sort of ranking of the methods, which is useful to establish the overall performances of the methods.
To assess the significance of the differences of the results obtained by the various techniques, we applied a two-tailed t-test with confidence level of 1%. The result of such test are summarized in Table 6 and Table 7, for the MAE and the RMSE, respectively. In these tables, a “s” means that the difference of results between two methods was considered to be significant. For example, in Table 6, we can notice that the average MAE obtained by RF was significantly better than the average MAE achieved by EVTree.
In the tables, we can see that all methods, but ARIMA, obtained better results when w was set to 10, which also confirmed the result regarding the optimal historical window size reported above. We can see that the first three strategies, i.e., RF, GBM and XGBoost, basically achieved the same results, even if, when the MAE was considered, GMB and XGBoost achieved slightly better results when w was set to 10.
In general, we can notice that, according to both MAE and RMSE, the best performing forecasting methods on the datasets used in this study were RF, GMB and XGBoost. Neural Networks and linear regression also obtained competitive results. It is interesting also to notice that the ensemble scheme proposed in [10] did not perform well on the datasets considered in this study, while the performances of the same methods were very competitive in a global short-term electric energy forecasting scenario. This could mean that a stacking ensemble approach, such as the one used in [10], is suitable for global energy consumption predictions, but that it is not suitable for a localized problem such as the one tackled in this study. ARIMA was confirmed to be the worst performing methods, suggesting that perhaps this method is not adequate for these kinds of problems.
Finally, Figure 7 shows four examples of the observed and forecasted values from ARIMA and RF methods obtained on different buildings over a period of 10 days. Recall that RF is the method that achieved better results, while ARIMA is the worst performing technique. We can notice how RF better approximated the real values, while the errors made by ARIMA were in some cases significant. This fact caused ARIMA to perform poorly on the datasets used in this paper. From these graphs, we can conclude that the main difficulty of the datasets presented is to predict the energy consumption corresponding to the peaks represented in the graphs. However, methods such as RF can generally obtain predictions that are close to the observed values.

5. Conclusions and Future Work

In this paper, we have presented a comparative study of different time series forecasting techniques for predicting one-day electric energy consumption in non-residential buildings. In particular we were interested in both determining which strategy is more suitable for this kind of problem, and also to establish the optimal value for the historical window to be considered so as to produce the most accurate predictions. For such purpose, we compared nine forecasting methods on time series regarding data collected from thirteen smart buildings located in a university campus in Spain. The datasets used in this paper have never been used before, and are available on request. This represents an added value to the paper, since the datasets could represent a relevant resource for researches in this area.
From the experiments conducted, we can conclude that the best performing methods were Machine Learning based approaches. In particular, the Random Forests, Generalized Boosted Regression Mode and Extreme Gradient Boosting achieved, in general, the best performances on the data used in this work, since they achieved the lowest prediction errors. As such, they could represent the first choice for any further study on similar problems.
As far as the optimal historical window value is concerned, we found that, when the size of the window used was higher than seven, the results obtained improved. Moreover, the best results were achieved when the window size was set to 10, i.e., when ten days were considered to predict the electric energy consumption of the next day. This could help researchers facing similar forecasting scenarios.
As for future work, we are planning to apply a Deep Learning approach to the time series presented in this paper, since this strategy has obtained good results in other time series forecasting problems. A problem with this approach is represented by the setting of the parameters used to build the Deep Neural Network. In fact, a wrong setting can prevent such approach from obtaining good performances. To overcome this problem, we are also planning to use a Neurovolution approach [64,65,66]. In this approach, an Evolutionary Algorithm is used to establish a sub-optimal set of hyper parameters for the Deep Neural Network.
We are also planning to extend this study to other kinds of time series, for instance time series regarding the water and/or gas usage in smart buildings, and to include in the experimentation weather-related data. We also plan to to extend the experimentation to different buildings located in other regions to compare results obtained in different environmental conditions.

Author Contributions

F.D. proposed the concept of this research and design and was involved in the whole experimentation phase. F.A.G.V. was involved in the visualization and interpretation of results and in the data preparation, while M.G.T. provided overall guidance and was involved in the experimentation. All authors were involved in preparing the manuscript.

Funding

This research was funded by Spanish Ministry of Economic and Competitiveness and the European Regional Development Fund, grant number TIN2015-64776-C3-2-R (MINECO/FEDER).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Graphical Representation of Results

A graphical representation of the results obtained on the datasets used in this paper out is depicted in Figure A1 and Figure A2. In these figures, all results from all methods and all buildings are presented. Additionally, the the average results obtained on all the buildings are also reported in Figure A1n and Figure A2n.
Figure A1. Comparison of the MAE obtained on the various buildings.
Figure A1. Comparison of the MAE obtained on the various buildings.
Energies 12 01934 g0a1
Figure A2. Comparison of the RMSE obtained on the different buildings.
Figure A2. Comparison of the RMSE obtained on the different buildings.
Energies 12 01934 g0a2

References

  1. ExxonMobil. 2018 Outlook for Energy: A View to 2040. Available online: https://corporate.exxonmobil.com/en/~/media/Global/Files/outlook-for-energy/2018-Outlook-for-Energy.pdf. (accessed on 25 March 2019).
  2. International Energy Agency. World Energy Outlook 2018; International Energy Agency Publications: Paris, France, 2018. [Google Scholar]
  3. Ledu, M.; Matthews, H.D.; de Elía, R. Regional estimates of the transient climate response to cumulative CO2 emissions. Nat. Clim. Chang. 2016, 6, 474–478. [Google Scholar] [CrossRef]
  4. Energy 2020, A Strategy for Competitive, Sustainable and Secure Energy. Available online: http://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:52010DC0639&from=EN (accessed on 12 March 2019).
  5. Agency, T.I.E. Energy Efficiency: Buildings. Available online: https://www.iea.org/topics/energyefficiency/buildings (accessed on 20 March 2019).
  6. Raza, M.Q.; Khosravi, A. A review on artificial intelligence based load demand forecasting techniques for smart grid and buildings. Renew. Sustain. Energy Rev. 2015, 50, 1352–1372. [Google Scholar] [CrossRef]
  7. Sen, P.; Roy, M.; Pala, P. Application of ARIMA for forecasting energy consumption and GHG emission: A case study of an Indian pig iron manufacturing organization. Energy 2016, 116, 1031–1038. [Google Scholar] [CrossRef]
  8. Chujai, P.; Kerdprasop, N.; Kerdprasop, K. Time Series Analysis of Household Electric Consumption with ARIMA and ARMA Models. In Proceedings of the International MultiConference of Engineers and Computer Scientists, Hong Kong, China, 13–15 March 2013; Volume 1. [Google Scholar]
  9. Bonetto, R.; Rossi, M. Machine Learning Approaches to Energy Consumption Forecasting in Households. arXiv 2017, arXiv:1706.09648. [Google Scholar]
  10. Divina, F.; Gilson, A.; Goméz-Vela, F.; García Torres, M.; Torres, J.F. Stacking Ensemble Learning for Short-Term Electricity Consumption Forecasting. Energies 2018, 11, 949. [Google Scholar] [CrossRef]
  11. Oliveira, E.M.; Oliveira, F.L.C. Forecasting mid-long term electric energy consumption through bagging ARIMA and exponential smoothing methods. Energy 2018, 144, 776–788. [Google Scholar] [CrossRef]
  12. Medina, A.; Cámara, A.; Monrobel, J.R. Measuring the Socioeconomic and Environmental Effects of Energy Efficiency Investments for a More Sustainable Spanish Economy. Sustainability 2016, 8, 1039. [Google Scholar] [CrossRef]
  13. Campillo, J.; Wallin, F.; Torstensson, D.; Vassileva, I. Energy Demand Model Design For Forecasting Electricity Consumption And Simulating Demand Response Scenarios In Sweden. In Proceedings of the 4th International Conference in Applied Energy 2012, Suzhou, China, 5–8 July 2012; pp. 1–7. [Google Scholar]
  14. Pinson, P.; Madsen, H.; O’Malley, M.; O’Connell, N. Benefits and challenges of electrical demand response: A critical review. Renew. Sustain. Energy Rev. 2014, 39, 686–699. [Google Scholar]
  15. Deb, C.; Zhang, F.; Yang, J.; Lee, S.E.; Shah, K.W. A review on time series forecasting techniques for building energy consumption. Renew. Sustain. Energy Rev. 2017, 74, 902–924. [Google Scholar] [CrossRef]
  16. Changhong, C.; Bingyan, W.; Qingyan, F.; Green, C.; Streets, D.G. Reductions in emissions of local air pollutants and co-benefits of Chinese energy policy: A Shanghai case study. Energy Policy 2006, 34, 754–762. [Google Scholar] [CrossRef]
  17. Gooijer, J.G.D.; Hyndman, R.J. 25 years of time series forecasting. Int. J. Forecast. 2006, 22, 443–473. [Google Scholar] [CrossRef]
  18. Abdel-Aal, R.E.; Al-Garni, A.Z. Forecasting monthly electric energy consumption in eastern Saudi Arabia using univariate time-series analysis. Energy 1997, 22, 1059–1069. [Google Scholar] [CrossRef]
  19. Shilpa, G.N.; Sheshadri, G.S. Short-Term Load Forecasting Using ARIMA Model For Karnataka State Electrical Load. Int. J. Eng. Res. Dev. 2017, 13, 75–79. [Google Scholar]
  20. Newsham, G.R.; Birt, B.J. Building-level Occupancy Data to Improve ARIMA-based Electricity Use Forecasts. In Proceedings of the 2nd ACM Workshop on Embedded Sensing Systems for Energy-Efficiency in Building (BuildSys ’10), Zurich, Switzerland, 2 November 2010; ACM: New York, NY, USA, 2010; pp. 13–18. [Google Scholar]
  21. Rallapallia, S.R.; Ghosh, S. Forecasting monthly peak demand of electricity in India—A critique. Energy Policy 2012, 45, 516–520. [Google Scholar] [CrossRef]
  22. Schrock, D.W.; Claridge, D.E. Predicting Energy Usage in a Supermarket. In Proceedings of the 6th Symposium on Improving Building Systems in Hot and Humid Climates, Dallas, TX, USA, 3–4 October 1989; Available online: https://core.ac.uk/download/pdf/79624985.pdf (accessed on 20 May 2019).
  23. Nowotarski, J.; Liu, B.; Weron, R.; Hong, T. Improving short term load forecast accuracy via combining sister forecasts. Energies 2016, 98, 40–49. [Google Scholar] [CrossRef]
  24. Samarasinghe, M.; Al-Hawani, W. Short-Term Forecasting of Electricity Consumption using Gaussian Processes. Master’s Thesis, University of Agder, Kristiansand, Norway, 2012. [Google Scholar]
  25. Rahman, H.; Selvarasan, I.; Begum, J. Short-Term Forecasting of Total Energy Consumption for India-A Black Box based Approach. Energies 2018, 11, 3442. [Google Scholar] [CrossRef]
  26. Park, D.; El-Sharkawi, M.; Marks, R.J.; Atlas, L.; Damborg, M. Electric Load Forecasting Using An Artificial Neural Network. IEEE Transadions Power Syst. 1991, 6, 442–449. [Google Scholar] [CrossRef]
  27. Nizami, J.; Ai-Garni, A.Z. Forecasting electric energy consumption using neural networks. Energy Policy 1995, 23, 1097–1104. [Google Scholar] [CrossRef]
  28. Hong, T. Short Term Electric Load Forecasting. Ph.D. Thesis, Graduate Faculty of North Carolina State University, Raleigh, NC, USA, 2010. [Google Scholar]
  29. Kelo, S.; Dudul, S. A wavelet Elman neural network for short-term electrical load prediction under the influence of temperature. Int. J. Electr. Power Energy Syst. 2012, 43, 1063–1071. [Google Scholar] [CrossRef]
  30. Chitsaz, H.; Shaker, H.; Zareipour, H.; Wood, D.; Amjady, N. Short-term electricity load forecasting of buildings in microgrids. Energy Build. 2015, 99, 50–60. [Google Scholar] [CrossRef]
  31. Zheng, H.; Yuan, J.; Chen, L. Short-Term Load Forecasting Using EMD-LSTM Neural Networks with a Xgboost Algorithm for Feature Importance Evaluation. Energies 2017, 10, 1168. [Google Scholar] [CrossRef]
  32. Jaina, R.K.; Smith, K.M.; Culligan, P.J.; Taylor, J.E. Forecasting energy consumption of multi-family residential buildings using support vector regression: Investigating the impact of temporal and spatial monitoring granularity on performance accuracy. Appl. Energy 2014, 123, 168–178. [Google Scholar] [CrossRef]
  33. Liu, D.; Chen, Q.; Mori, K. Time series forecasting method of building energy consumption using support vector regression. In Proceedings of the 2015 IEEE International Conference on Information and Automation, Lijiang, China, 8–10 August 2015; pp. 1628–1632. [Google Scholar]
  34. Lora, A.T.; Santos, J.M.R.; Riquelme, J.C.; Expósito, A.G.; Ramos, J.L.M. Time-Series Prediction: Application to the Short-Term Electric Energy Demand. In Lecture Notes in Computer Science, Proceedings of the Current Topics in Artificial Intelligence TTIA 2003, San Sebastian, Spain, 12–14 November 2004; Springer: Berlin, Germany, 2004; Volume 3040, pp. 577–586. [Google Scholar]
  35. Zheng, J.; Xu, C.; Zhang, Z.; Li, X. Electric load forecasting in smart grids using Long-Short-Term-Memory based Recurrent Neural Network. In Proceedings of the 51st Annual Conference on Information Sciences and Systems (CISS), Baltimore, MD, USA, 22–24 March 2017; pp. 1–6. [Google Scholar]
  36. Torres, J.F.; Fernández, A.M.; Troncoso, A.; Martínez-Álvarez, F. Deep Learning-Based Approach for Time Series Forecasting with Application to Electricity Load. In Biomedical Applications Based on Natural and Artificial Computing; Springer International Publishing: Cham, Switzerland, 2017; pp. 203–212. [Google Scholar]
  37. Galicia, A.; Torres, J.F.; Martínez-Álvarez, F.; Troncoso, A. Scalable Forecasting Techniques Applied to Big Electricity Time Series. In Advances in Computational Intelligence, Proceedings of the 14th International Work-Conference on Artificial Neural Networks, IWANN 2017, Cadiz, Spain, 14–16 June 2017; Springer International Publishing: Cham, Switzerland, 2017; Part II; pp. 165–175. [Google Scholar]
  38. Castelli, M.; Vanneschi, L.; De Felice, M. Forecasting short-term electricity consumption using a semantics-based genetic programming framework: The South Italy case. Energy Econ. 2015, 47, 37–41. [Google Scholar] [CrossRef]
  39. Talavera-Llames, R.L.; Pérez-Chacón, R.; Martínez-Ballesteros, M.; Troncoso, A.; Martínez-Álvarez, F. A Nearest Neighbours-Based Algorithm for Big Time Series Data Forecasting. In Hybrid Artificial Intelligent Systems, Proceedings of the 11th International Conference, HAIS 2016, Seville, Spain, 18–20 April 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 174–185. [Google Scholar]
  40. Geng, J.; Huang, M.L.; Li, M.W.; Hong, W.C. Hybridization of seasonal chaotic cloud simulated annealing algorithm in a SVR-based load forecasting model. Neurocomputing 2015, 151, 1362–1373. [Google Scholar] [CrossRef]
  41. Fan, G.F.; Peng, L.L.; Hong, W.C.; Sun, F. Electric load forecasting by the SVR model with differential empirical mode decomposition and auto regression. Neurocomputing 2016, 173, 958–970. [Google Scholar] [CrossRef]
  42. Jetcheva, J.G.; Majidpour, M.; Chen, W.P. Neural network model ensembles for building-level electricity load forecasts. Energy Build. 2014, 84, 214–223. [Google Scholar] [CrossRef]
  43. Khairalla, M.A.; Ning, X.; AL-Jallad, N.T.; El-Faroug, M.O. Short-Term Forecasting for Energy Consumption through Stacking Heterogeneous Ensemble Learning Model. Energies 2018, 11, 1605. [Google Scholar] [CrossRef]
  44. Martínez-Álvarez, F.; Troncoso, A.; Asencio-Cortés, G.; Riquelme, J.C. A Survey on Data Mining Techniques Applied to Electricity-Related Time Series Forecasting. Energies 2015, 8, 13162–13193. [Google Scholar] [CrossRef]
  45. Daut, M.A.M.; Hassan, M.Y.; Abdullah, H.; Rahman, H.A.; Abdullah, M.P.; Hussin, F. Building electrical energy consumption forecasting analysis using conventional and artificial intelligence methods: A review. Renew. Sustain. Energy Rev. 2017, 70, 1108–1118. [Google Scholar] [CrossRef]
  46. Spanish-Government. Agencia Estatal de Meteorologia—AEMET. Gobierno de España. 2019. Available online: https://opendata.aemet.es/ (accessed on 25 April 2019).
  47. Kleiber, C.; Zeileis, A. Applied Econometrics with R; Springer: New York, NY, USA, 2008; ISBN 978-0-387-77316-2. [Google Scholar]
  48. Neter, J.; Kutner, M.H.; Nachtsheim, C.J.; Wasserman, W. Applied Linear Statistical Models; Irwin: Chicago, IL, USA, 1996. [Google Scholar]
  49. Kuhn, M. Building Predictive Models in R Using the caret Package. J. Stat. Softw. Artic. 2008, 28, 1–26. [Google Scholar]
  50. Box, G.; Jenkins, G. Time Series Analysis: Forecasting and Control; John Wiley and Sons: New York, NY, USA, 2008. [Google Scholar]
  51. Box, G.E.P.; Jenkins, G. Time Series Analysis, Forecasting and Control; Holden-Day, Inc.: San Francisco, CA, USA, 1990. [Google Scholar]
  52. Salles, R.; Assis, L.; Guedes, G.; Bezerra, E.; Porto, F.; Ogasawara, E. A Framework for Benchmarking Machine Learning Methods Using Linear Models for Univariate Time Series Prediction. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017. [Google Scholar]
  53. Time Series Models for “Caret” Package. Available online: https://github.com/sfeuerriegel/caret.ts (accessed on 3 April 2019).
  54. Eiben, A.E.; Smith, J.E. Introduction to Evolutionary Computing; Springer: New York, NY, USA, 2003. [Google Scholar]
  55. Grubinger, T.; Zeileis, A.; Pfeiffer, K.P. evtree: Evolutionary Learning of Globally Optimal Classification and Regression Trees in R. J. Stat. Softw. 2014, 61, 1–29. [Google Scholar] [CrossRef]
  56. Loh, W.Y. Classification and regression trees. In Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery; Wiley: New York, NY, USA, 2011; Volume 1, pp. 14–23. [Google Scholar]
  57. Friedman, J.H. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2000, 29, 1189–1232. [Google Scholar] [CrossRef]
  58. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  59. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  60. Breiman, L.; Friedman, J.; Olshen, R.; Stone, C. Classification and Regression Trees; Wadsworth and Brooks: Monterey, CA, USA, 1984. [Google Scholar]
  61. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’16), San Francisco, CA, USA, 13–17August 2016; ACM: New York, NY, USA, 2016; pp. 785–794. [Google Scholar]
  62. Talavera-Llames, R.; Pérez-Chacón, R.; Troncoso, A.; Martínez-Álvarez, F. MV-kWNN: A novel multivariate and multi-output weighted nearest neighbours algorithm for big data time series forecasting. Neurocomputing 2019, 353, 56–73. [Google Scholar] [CrossRef]
  63. Galicia, A.; Talavera-Llames, R.; Troncoso, A.; Koprinska, I.; Martínez-Álvarez, F. Multi-step forecasting for big data time series based on ensemble learning. Knowl.-Based Syst. 2019, 163, 830–841. [Google Scholar] [CrossRef]
  64. Floreano, D.; Dürr, P.; Mattiussi, C. Neuroevolution: From architectures to learning. Evol. Intell. 2008, 1, 47–62. [Google Scholar] [CrossRef]
  65. Stanley, K.O.; Clune, J.; Lehman, J.; Miikkulainen, R. Designing neural networks through neuroevolution. Nat. Mach. Intell. 2019, 1, 24–35. [Google Scholar] [CrossRef]
  66. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized Evolution for Image Classifier Architecture Search. arXiv 2018, arXiv:1802.01548. [Google Scholar]
Figure 1. Graphical representation of the time series of all the buildings. The values are shown in logarithmic scale.
Figure 1. Graphical representation of the time series of all the buildings. The values are shown in logarithmic scale.
Energies 12 01934 g001
Figure 2. Distribution of values for the different buildings.
Figure 2. Distribution of values for the different buildings.
Energies 12 01934 g002
Figure 3. ACF and PACF graphs for the time series of Building 3.
Figure 3. ACF and PACF graphs for the time series of Building 3.
Energies 12 01934 g003
Figure 4. Dataset pre-processing. w determines the amount of the historical data used, while h determines the prediction horizon.
Figure 4. Dataset pre-processing. w determines the amount of the historical data used, while h determines the prediction horizon.
Energies 12 01934 g004
Figure 5. MAE obtained on all the buildings for different values of the historical window w.
Figure 5. MAE obtained on all the buildings for different values of the historical window w.
Energies 12 01934 g005
Figure 6. RMSE obtained on all the buildings for different values of the historical window w.
Figure 6. RMSE obtained on all the buildings for different values of the historical window w.
Energies 12 01934 g006
Figure 7. Four examples of real data against the predictions from ARIMA and Random Forest for w = 10.
Figure 7. Four examples of real data against the predictions from ARIMA and Random Forest for w = 10.
Energies 12 01934 g007
Table 1. Description of the selected buildings. The first column presents the year of construction and the year of the refurbishment if was made. The second column present the size in m 2 , and finally a short description.
Table 1. Description of the selected buildings. The first column presents the year of construction and the year of the refurbishment if was made. The second column present the size in m 2 , and finally a short description.
BuildingYear (Refurbishment)Size (m 2 )Description
11956 (2005)75,323university canteen
21956 (2001)527,277a four story building with several classrooms and offices
31956 (1999)5231.01a four story building with several classrooms and offices
61956 (1999)5231.01a four story building with four large classrooms
71958 (2003)5231.01a four story building with several classrooms and offices
81956 (1999)1948.16a two story building with four large classrooms
101956 (1999)5231.01a four story building with several classrooms and offices
121956 (1999)5231.01a four story building with several classrooms and offices
141956 (1990)5231.01a four story building with several classrooms and offices
241956 (2007)9134.93a three story building with several classrooms and offices
252012 (-)16,281.71a newly two story building hosting the university library and some classrooms
321956 (2005)2530.78Administration offices
442008 (-)1752.83a newly three story built building with research labs, classrooms and offices.
Table 2. Average temperatures, rainfall and average wind speed in Seville.
Table 2. Average temperatures, rainfall and average wind speed in Seville.
YearAvg. Temp.Avg. Max. Temp.Avg. Min. Temp.Rainfall (mm)Avg. Wind Sp. (Km/h)
201218.725.812.4324.139.0
201318.525.312.6425.679.9
201418.925.613.3636.509.4
201519.126.813.1318.468.6
201619.126.013.5626.319.4
201719.627.013.2321.08.8
Table 3. Results obtained by all the methods on all the buildings for each value of the historical window considered.
Table 3. Results obtained by all the methods on all the buildings for each value of the historical window considered.
ARIMALMEvTreeRFNNGMBrpartENSEMBLEXGBoost
wMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSEMAERMSE
Edificio 170.570.710.500.580.480.580.480.560.490.570.490.570.510.600.550.670.490.57
100.520.630.460.540.500.610.470.550.460.540.460.550.500.580.510.640.460.55
150.520.630.460.540.480.580.480.550.460.540.470.550.520.590.540.660.470.55
200.550.680.470.550.490.590.480.560.470.540.470.550.520.600.530.650.480.56
Edificio 272.102.571.371.551.411.611.351.531.401.611.351.531.391.571.411.691.351.54
101.732.161.311.501.341.571.321.501.321.501.301.501.381.571.411.701.311.51
151.521.891.321.531.381.631.331.511.331.531.301.531.391.571.391.721.331.54
201.662.051.321.531.361.611.331.511.331.531.321.531.391.571.351.641.321.54
Edificio 371.074.721.074.721.064.671.014.641.094.671.034.651.054.661.044.561.024.65
101.064.921.064.920.974.630.934.640.914.650.914.641.004.661.024.660.924.64
151.165.141.165.140.964.680.944.660.934.680.914.661.004.670.984.670.934.66
201.165.141.165.140.994.680.944.670.934.690.934.670.994.681.024.680.944.67
Edificio 671.832.340.961.250.981.290.941.240.971.260.971.280.951.231.001.350.981.28
101.321.740.881.150.881.200.871.150.871.140.871.160.891.170.911.260.861.16
151.131.480.881.160.911.250.871.170.891.170.861.180.911.200.931.280.891.19
201.171.510.881.170.951.260.871.170.891.170.881.190.911.200.931.260.871.18
Edificio 772.342.891.131.321.161.411.091.281.121.311.091.301.121.311.131.361.091.30
102.252.761.061.251.081.321.041.221.061.271.031.211.081.291.131.391.041.24
151.061.261.061.261.081.281.051.221.081.261.031.211.091.261.111.341.041.22
201.061.301.061.251.091.321.061.231.071.241.061.241.101.271.121.371.051.22
Edificio 870.390.490.360.440.370.460.350.440.360.440.360.450.360.440.390.520.350.45
100.470.600.340.430.360.460.340.430.340.430.330.430.360.440.380.530.340.43
150.360.450.340.430.360.470.340.430.340.430.340.430.360.440.370.500.340.43
200.340.430.340.430.350.460.340.430.340.430.340.430.360.440.380.510.340.43
Edificio 1071.381.711.021.251.021.281.021.261.021.261.011.241.071.371.071.321.021.26
101.231.520.951.160.961.210.961.190.941.160.941.171.011.291.021.270.941.18
151.011.270.971.170.971.210.971.190.971.180.941.151.011.271.031.300.951.17
201.181.480.981.180.991.230.971.190.971.180.951.161.011.271.041.270.961.18
Edificio 1270.520.650.390.450.390.460.370.430.380.450.370.430.390.450.400.500.360.43
100.420.520.370.430.390.470.360.430.370.430.350.430.390.460.390.490.350.43
150.430.530.370.430.380.460.360.430.370.430.360.430.390.460.380.480.360.44
200.510.640.370.430.390.470.360.430.370.430.360.420.390.460.390.490.360.44
Edificio 1471.872.211.191.451.231.511.091.331.171.431.121.361.271.521.241.521.151.39
101.111.370.981.191.101.390.971.181.011.230.951.141.211.431.031.290.951.15
150.901.090.891.071.001.230.941.140.901.080.891.091.041.260.941.200.891.09
201.061.310.891.071.001.230.941.140.901.080.901.101.041.261.031.310.911.10
Edificio 2471.201.560.941.220.981.270.931.200.951.230.921.200.941.221.021.320.941.21
101.121.470.881.140.981.300.881.150.851.110.881.140.921.201.001.310.881.15
151.251.630.871.140.951.250.881.140.881.160.901.170.921.180.941.220.881.16
200.991.280.881.141.001.320.891.150.881.160.921.190.901.170.961.270.911.18
Edificio 2574.345.543.484.293.864.943.434.243.734.563.444.283.514.294.015.073.414.26
106.417.733.294.063.624.563.244.053.474.233.264.093.384.183.824.783.244.07
153.834.843.334.123.844.833.344.163.494.323.404.293.764.733.704.793.414.29
204.275.233.334.123.754.733.334.143.474.323.384.213.764.743.724.803.404.25
Edificio 3270.660.800.520.630.560.690.520.630.530.640.520.640.530.630.590.750.520.63
100.861.060.510.620.560.690.510.620.510.620.510.630.520.630.580.750.510.62
150.630.780.520.620.560.700.510.620.520.620.510.630.530.640.550.700.520.65
200.590.720.520.630.540.660.520.630.520.630.510.630.530.640.570.720.530.64
Edificio 4470.550.680.460.560.470.570.450.540.470.570.450.550.460.550.490.610.460.55
100.600.740.450.540.470.580.450.530.450.540.440.540.460.550.490.610.450.54
150.440.530.450.530.480.590.450.530.450.530.450.530.470.550.500.630.450.54
200.670.820.450.530.490.590.450.530.450.530.450.530.460.550.480.610.450.54
Average71.452.071.031.521.081.601.001.491.051.541.011.501.041.531.101.631.011.50
101.472.090.961.461.021.540.951.430.971.450.941.431.011.501.051.590.941.44
151.101.660.971.471.031.550.961.440.971.460.951.451.031.531.031.580.961.46
201.171.740.971.471.031.550.961.440.971.460.961.451.031.531.041.580.961.46
Table 4. Average MAE obtained by the methods for all the values of historical window and on all the buildings. The standard deviation is shown in brackets.
Table 4. Average MAE obtained by the methods for all the values of historical window and on all the buildings. The standard deviation is shown in brackets.
MethodAverageAverage for w = 10
RF0.97 (0.75)0.95 (0.75)
GBM0.97 (0.76)0.94 (0.75)
XGBoost0.97 (0.76)0.94 (0.75)
LM0.98 (0.76)0.96 (0.75)
NN0.99 (0.81)0.97 (0.81)
rpart1.03 (0.75)1.01 (0.79)
EVTree1.04 (0.86)1.02 (0.85)
ENSEMBLE1.06 (0.87)1.05 (0.90)
ARIMA1.30 (1.15)1.29 (1.14)
previous day1.44 (1.18)1.43 (1.22)
previous week1.33 (1.12)1.32 (1.14)
Table 5. Average RMSE obtained by the methods for all the values of historical window and on all the buildings.The standard deviation is shown in brackets.
Table 5. Average RMSE obtained by the methods for all the values of historical window and on all the buildings.The standard deviation is shown in brackets.
MethodAverageAverage for w = 10
RF1.45 (1.32)1.43 (1.35)
GBM1.46 (1.33)1.43 (1.35)
XGBoost1.46 (1.33)1.44 (1.35)
NN1.47 (1.36)1.45 (1.38)
LM1.48 (1.39)1.46 (1.40)
rpart1.52 (1.37)1.50 (1.36)
EVTree1.56 (1.41)1.54 (1.41)
ENSEMBLE1.59 (1.42)1.58 (1.45)
ARIMA1.89 (1.67)1.92 (1.85)
previous day2.07 (1.90)2.14 (1.99)
previous week2.05 (1.91)2.04 (1.96)
Table 6. Significance of the results regarding MAE according to two-tailed t-test with confidence level of 1%.
Table 6. Significance of the results regarding MAE according to two-tailed t-test with confidence level of 1%.
MethodEVTreeRFNNGBMARIMALMrpartENSEMBLEXGBoost
EVTree sssss s
RF s ss
NN ss ss
GBM s ss
ARIMA ssss
LM ss
rpart s
ENSEMBLE s
Table 7. Significance of the results regarding RMSE according to two-tailed t-test with confidence level of 1%.
Table 7. Significance of the results regarding RMSE according to two-tailed t-test with confidence level of 1%.
MethodEVTreeRFNNGBMARIMALMrpartENSEMBLEXGBoost
EVTree sssss ss
RF s s sss
NN s ss
GBM s sss
ARIMA ssss
LM s
rpart ss
ENSEMBLE s

Share and Cite

MDPI and ACS Style

Divina, F.; García Torres, M.; Goméz Vela, F.A.; Vázquez Noguera, J.L. A Comparative Study of Time Series Forecasting Methods for Short Term Electric Energy Consumption Prediction in Smart Buildings. Energies 2019, 12, 1934. https://doi.org/10.3390/en12101934

AMA Style

Divina F, García Torres M, Goméz Vela FA, Vázquez Noguera JL. A Comparative Study of Time Series Forecasting Methods for Short Term Electric Energy Consumption Prediction in Smart Buildings. Energies. 2019; 12(10):1934. https://doi.org/10.3390/en12101934

Chicago/Turabian Style

Divina, Federico, Miguel García Torres, Francisco A. Goméz Vela, and José Luis Vázquez Noguera. 2019. "A Comparative Study of Time Series Forecasting Methods for Short Term Electric Energy Consumption Prediction in Smart Buildings" Energies 12, no. 10: 1934. https://doi.org/10.3390/en12101934

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop