Next Article in Journal
Financial Support for Neighborhood Regeneration: A Case Study of Korea
Previous Article in Journal
Factors That Affect Digital Innovation Sustainability among SMEs in the Middle East Region
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Interval Energy-Forecasting Method for Sustainable Building Management Based on Deep Learning

School of Law, Huazhong University of Science and Technology, Wuhan 430074, China
Sustainability 2022, 14(14), 8584; https://doi.org/10.3390/su14148584
Submission received: 27 June 2022 / Revised: 11 July 2022 / Accepted: 12 July 2022 / Published: 13 July 2022
(This article belongs to the Topic Short-Term Load Forecasting)

Abstract

:
Energy conservation in buildings has increasingly become a hot issue for the Chinese government. Compared to deterministic load prediction, probabilistic load forecasting is more suitable for long-term planning and management of building energy consumption. In this study, we propose a probabilistic load-forecasting method for daily and weekly indoor load. The methodology is based on the long short-term memory (LSTM) model and penalized quantile regression (PQR). A comprehensive analysis for a time period of a year is conducted using the proposed method, and back propagation neural networks (BPNN), support vector machine (SVM), and random forest are applied as reference models. Point prediction as well as interval prediction are adopted to roundly test the prediction performance of the proposed model. Results show that LSTM-PQR has superior performance over the other three models and has improvements ranging from 6.4% to 20.9% for PICP compared with other models. This work indicates that the proposed method fits well with probabilistic load forecasting, which could promise to guide the management of building sustainability in a future carbon neutral scenario.

1. Introduction

Energy consumption in the Chinese office building sector has reached nearly one-fifth of China’s total energy consumption in 2015 [1]. This makes the study of energy use of buildings an essential part of the implementation of energy conservation policies in China. Moreover, with the development of information technology, extensive building operation data are available to be recorded and collected. It, therefore, allows the data-driven methods to be widely applied in the field of building energy efficiency enhancement. Among the different data driven applications, building energy consumption prediction is the most widespread one [2].
In general, the time horizons of office building energy prediction can be roughly classified into three categories, that is, short term for one day or less, medium term for one week to one month and long term for one year. Due to the applications in real-time control and anomaly detection, short-term forecasting has aroused great interest among researchers, while the study of medium- and long-term forecasting has received far less attention. Since the aim of the medium- and long-term forecasting study is to assess building energy consumption and formulate energy plans, ensuring the accuracy of the forecasting results is a priority at the current stage, which is also a fairly difficult objective [3].
Most interval forecasting is presented in the form of probabilistic density function, and one of the most commonly used methods in this field is quantile regression (QR) [4]. Quantile regression has been mainly used in the field of risk assessment and electricity price forecasting, and has shown good results in literature [5,6,7]. However, to the best of the author’s knowledge, few papers are available in the technical literature considering the possibility of application in the field of building energy consumption prediction. In Ref. [8], D-vine copula quantile regression is applied to predict conditional quantile of heating energy consumption after retrofitting, and it is the first to use quantile regression to analyze residential building energy consumption.
The main contributions of this paper can be summarized as the following: (1) analyze the prediction performance for daily and weekly indoor load in an office building by combining the deep-learning method with penalized quantile regression, as quantile regression has so far not been adequately studied in the field of building energy consumption prediction; (2) verify the advantages of deep learning in medium- and long-term forecasting by choosing BPNN, SVM, and random forest as the basic point estimation models.
The remainder of this paper is organized as follows. Section 2 reviews the building energy prediction methods. Section 3 introduces the data set and the research synopsis of this article and the LSTM model as well as the penalized quantile regression method used in the study. The results and discussion are detailed in Section 4. Concluding remarks are presented at the end of the paper.

2. Literature Review

2.1. Related Work for Point Forecasting

The development of short-term forecasting based on the data-driven method evolved from simplicity to complexity. In the past decades, the commonly used machine-learning method, i.e., artificial neural network (ANN) [9], SVM [10], decision tree (DT) [11] have been extensively studied in the literature. For example, Gonzalez and Zamarreno [12] predicted hourly energy consumption with a feedback neural network, and it improved the prediction performance of the traditional ANN. Chen et.al [13] proposed a short-term load prediction method based on the support vector regression (SVR) model and compared it with the other seven traditional prediction models; results showed that the prediction accuracy of the SVR model is much better than the other models. With the deepening of theoretical research, optimization algorithms have been applied to optimize the parameter selection of the data-driven methods. Common algorithms include particle swarm optimization (PSO) [14], principal component analysis (PCA) [15], differential evolution (DE) [16], and fuzzy methods [17], and most research so far has corroborated the superiority of optimization algorithms in improving predictive performance.
More recently, deep-learning methods are also introduced in the field of short-term forecasting [18]. By their ascendancy in two aspects, predictive practicability and performance and deep learning promptly became research hotspots. Fu [19] used a deep-belief-network-based method to predict cooling load in real buildings hours ahead, and the results indicate that the deep-belief network performed the best when compared to the prediction algorithms, such as persistence, BPNN, and SVM. Li and Liu [20] adopted deep recurrent neural networks to make the real-time state estimation of room-cooling load. In [21], the deep-learning method is compared with eight commonly used machine-learning methods, authors make a comprehensive analysis of the day-ahead energy demand prediction, and infer that the deep-learning method outperforms other popular traditional machine-learning methods with higher accuracy.
On the whole, the study of short-term forecasting is relatively mature, and the existing research findings can reach the level of engineering application. Despite novel algorithms popping up all the time, the core issue lies in improving accuracy and essentially reducing computational costs. Table 1 exhibits the time horizon and methods used in short-term forecasting articles in recent years, and it can be concluded that most of the short-term forecasting focuses on the day-ahead prediction problem at different time intervals. As pointed out above, short-term forecasting can be regarded as the key link of building energy savings, and with the rapid development of artificial intelligence, it will receive more and more attention.
Unlike short-term forecasting, of which the prediction accuracy can reach more than 90%, previous literature shows that the relative errors corresponding to medium- and long-term forecasting at hourly resolutions often are in excess of 40% [29,30,31,32]. Several studies have focused on forecasting the aggregate building energy usage to improve prediction performance. Williams and Gomez [33] found that monthly energy consumption at the household level can be predicted with 74% accuracy, while for aggregate monthly energy consumption, the prediction accuracy can reach nearly 94%. Cai et al. [34] divided the monthly electricity energy consumption of residential buildings into different categories, and the prediction accuracy of the proposed model vastly exceeds those of conventional methods.

2.2. Related Work for Interval Forecasting

The above literature belongs to the application scope of the point estimation methods, which are suitable for the application of real-time control based on short-term forecasting. However, it is impossible to find a high-accuracy and reliability point estimation method in medium- and long-term forecasting, as there exists too many uncertain factors in the medium- and long-term operations of the systems to affect their energy consumption. Therefore, interval forecasting has become increasingly important in recent years, as interval forecasting provides prediction range and confidence rather than a point estimation, which would be more meaningful to the building managers compared with the point estimation. Walter et al. [35] carried out the interval prediction of energy consumption for 17 different commercial buildings and transformed the determined prediction information into the uncertain interval estimation, the authors believe that this interval estimation can provide reliable decision information for the investment of energy-saving measures. Ahmad and Chen [36] studied the problem of the smart grid energy demand prediction, and they predicted the energy demand one month, one quarter, and one year ahead, respectively. They gave a high-precision prediction range and provided a reference for managers to optimize resource planning. Martinez, Soto, and Jentsch [37] proposed a residential energy model that can be applied to different countries. The model can predict the monthly and annual residential energy demand of a given country or region and give the approximate energy consumption in different confidence regions.

3. Methodology

3.1. Data Description

The data used in this study comprises the temperature, humidity, and flow rate of air inlet and outlet of all air-conditioning units in an office building from Shanghai, China, and the indoor load can be calculated by these data. In addition, we used the worldmet package in the open-source software R to obtain the meteorological data from the airport which is near the target building, and the distance between the target building and the corresponding airport is no more than 7 km. Therefore, the meteorological data of the airport can be used as the actual meteorological data of the target building. All the data set were recorded with 15 min resolution from 17 April 2017 to 3 May 2018. In this paper, we considered that LSTM can learn the order dependence between items in a sequence, and it has the promise of being able to learn the context required to make predictions in time-series forecasting problems. Therefore, we ignore the date information of the input variables in the expectation of further demonstrating the predictive advantage of the LSTM model.
In total, the dataset contains four different variables, each with 36,672 observations. The raw data set is divided into two parts, training set (70%) and testing set (30%), the former consists of the indoor load and weather data from April 2017 to January 2018, and the latter contains the rest of the data set. Meanwhile, missing values are processed by interpolation, and outliers do not need to be removed or replaced, as there is no explicit outlier label in the data set, and some outlier detection methods based on statistical rules may destroy the authenticity of the data set.
Figure 1 presents the probability distribution of the indoor load and weather data. The probability distribution of the indoor load exhibited in Figure 1A reveals that it is close to skewed normally distributed. In addition, it can be concluded from other subgraphs that the weather in Shanghai may be humid and hot in summer, which is in line with its subtropical monsoon climate characteristics. The occasional extreme weather in Figure 1B could cause a sharp increase in indoor load, and this situation is also reflected in Figure 1A. In addition, the most likely distributions of each dataset in Figure 1 are evaluated by the commonly used Kolmogorov–Smirnov test, and the results are shown in Table 2:

3.2. Research Outline

Figure 2 illustrates the outline of the research in this paper. It is divided into three main steps. The first step is data preparation, including energy consumption data and weather data, where the energy consumption data is calculated based on the temperature, humidity, and flow rate of the indoor air-conditioning inlet and outlet, while the weather data comes from airport data as described in Section 2.1.
The second step is model training. In this paper, the deep-learning method, LSTM, is chosen as the basic point estimation algorithm, and the three commonly used machine-learning methods, BPNN, SVM, and random forest (RF), are used as reference models to facilitate comparisons with other models. Based on the prediction results of the above point estimation algorithms, penalized quantile regression (PQR) was chosen to construct the prediction intervals. Specifically speaking, as energy consumption data and weather data are input variables, we defined them as x t ( t = 1 , 2 , , n ) . Thus, the outputs of the aforementioned machine-learning methods can be expressed as:
(1)
BPNN
We can describe the training process of BPNN by the following equations:
n o d e j = t = 1 n ω t j x t , j = 1 , 2 , , m
z j = f H ( n o d e j ) , j = 1 , 2 , , m
Here, n o d e j is the activation value of the jth node, z j is the output of the hidden layer, and f H is the activation function of a node, which is usually a sigmoid function:
f H ( x ) = 1 1 + exp ( x )
The outputs of all neurons in the output layer are given as follows:
y t = f t ( j = 1 m ω t j z j )
Here, f t is the activation function, usually a line function.
In this paper, the single hidden layer BPNN model is chosen for predictive modelling, considering its generalizability and relatively low computational cost.
(2)
SVM
SVM uses the following form to approximate the relationship between input and output variables:
f ( x i ) = ω · φ ( x i ) + b
where φ ( x i ) represents a high-dimensional feature space mapped from the low-dimensional space. In terms of the ω and b, the following regularized risk function is used to find them.
1 2 ω 2 + C 1 n i = 1 n L ε ( y i , f ( x i ) )
where the ω 2 is called the regularized term. The L ε ( y i , f ( x i ) ) is the empirical error measured by the ε -insensitive loss function, which is defined as follows:
L ε ( y i , f ( x i ) ) = { 0 ,     y i f ( x i ) ε | y i f ( x i ) | ε ,       o t h e r w i s e
This defines the range of ε values such that if the predicted value is within the range, the loss is zero, and if the predicted point is outside the range, then the loss is the magnitude of the difference between the predicted value and the distance ε of the range. C is a penalty factor.
To get the estimation of ω and b, Equation (5) is transformed into objective function (8) by introducing the positive slack variables ,   ξ , ξ * .
M i n i m i z e 1 2 ω 2 + C 1 n i = 1 n L ε ( y i , f ( x i ) ) Subject   to   y i ω · φ ( x i ) b ε + ξ , ω · φ ( x i ) + b y i ε + ξ * , ξ ,   ξ * 0
By introducing four Lagrange multipliers,   α i , α i * ,   ξ i , ξ i * , we can turn Equation (8) to the Lagrangian function. The Lagrange function becomes:
L = 1 2 ω 2 + C i = 1 n (   ξ i + ξ i * ) i = 1 n ( η i   ξ i + η i * ξ i * ) i = 1 n α i ( ε + ξ i y i + ω · φ ( x i ) + b ) i = 1 n α i * ( ε + ξ i * + y i ω · φ ( x i ) b )
By introducing kernel function,   K ( x , x i ) , the SVM forecasting model can be obtained by quadratic programming:
f ( x ) = i = 1 n ( α i α i * ) K ( x , x i ) + b
where K ( x , x i ) is the kernel function of the SVR model. The commonly used kernel functions of the SVR models contain Gaussian, polynomial, and sigmoid.
(3)
RF
RF constructs a large number of decision trees during training and averages the values of output from each tree as the final output. Firstly, the training data are randomly selected and used to build the   B decision tree. Then, RF is generated by adding in parallel these B trees { T 1 ( x ) , T 2 ( x ) , T B ( x ) } , where x = { x 1 , x 2 , x n , } is an n-dimensional feature vector. The ensemble produces B outputs { Y 1 = T 1 ( x ) , Y 2 = T 2 ( x ) , Y B = T B ( x ) } , where Y is the value predicted by the decision tree; here, the estimation process of each tree is totally independent. The final prediction, y t , is made by averaging the predicted values of each tree.
y t = Y 1 + Y 2 + + Y B B
(4)
LSTM
The long short-term memory network is derived from recurrent neural networks (RNN), and it can handle the long-term dependency problems of RNN. Hochreiter and Schmidhuber [38] mathematically elaborated the architecture of LSTM, and the article showed that LSTM can keep or delete information by three controlling gates: input gate, output gate, and forget gate. Figure 2 shows a schematic diagram of a LSTM network; y t 1 represents the output of the previous step, and x t is the input of the current step.
When an input, x t , enters the LSTM, the forget gate,   f , decides what information is deleted. This decision can be computed as:
f t = σ ( W f · [ y t 1 , x t ] + b f )
Here, σ is the sigmoid activation function applied to elements inside the parentheses, W f is the weight matrix of the forget gate, and b f is the bias of the forget gate.
While for the input gate, i , it is used to decide what information to keep and update in the LSTM, and i can be expressed as:
i t = σ ( W i · [ y t 1 , x t ] + b i )
Here, W i is the weight matrix of the input gate, and b i is the bias of the input gate.
Then, activation function, C ˜ t , is used to describe the cell state under the current input:
C ˜ t = tan h ( W C · [ y t 1 , x t ] + b C )
After information selecting, the previous LSTM cell state, C t 1 , can be updated to the current LSTM cell state C t as shown in Figure 2, and C t can be computed as:
C t = f t * C t 1 + i t * C ˜ t
Here, * is an element-wise multiplier.
The output gate can be used to scale the output of the LSTM activation function, and it is expressed as:
o t = σ ( W o · [ y t 1 , x t ] + b o )
Finally, the output, y t , is calculated based on the LSTM cell state and output gate, o , and can be expressed as:
y t = o t * tanh ( C t )
(5)
Penalized quantile regression
Quantile regression is a modelling approach that models the quantile of the distribution of the variables. While traditional regression models focus on calculating the mean value of the target variable, quantile regression calculates the median value of it. Quantile regression utilizes a quantile loss function to provide information about future uncertainty:
L q , t ( y t , p q t ) = i : y t p q t N ( 1 q ) ( p q t y t ) + i : y t > p q t N q ( y t p q t )
where q denotes the quantile, y t is the output at time   t obtained from the machine learning-methods, p q t denotes the estimated q t h quantile at time t , and L q , t denotes the quantile loss for the q t h quantile at time t .
For the smoothness of the fitting result of quantile regression, we adopt an appropriate penalty function which is designed to penalize the difference of consecutive regression coefficients [39]. Suppose that the estimated q t h quantile function is given as a linear function, j = 1 M β j x i j , and the penalty function can be expressed as j = 1 M ( β j + 1 β j ) 2 . This penalty function gives a smoothing effect of the regression coefficients by penalizing the high quantity of the difference of two adjacent coefficients. Then, the objective loss function is represented by:
L q , t ( y t , β x ) = i : y t β x N ( 1 q ) ( β x y t ) + i : y t > β x N q ( y t β x ) + j = 1 M ( β j + 1 β j ) 2
The third step is model performance evaluation, which is divided into two main aspects: firstly, the traditional point forecast evaluation criteria are used to assess the daily and weekly indoor-load-forecasting performance of each model, i.e., MAE, RMSE, and CV (RMSE). Secondly, to assess the interval forecasting results of the models, PICP and PINAW were used for quantitative comparison, and all interval forecasts were within the 95% prediction interval. The following subsection provides a brief description of the evaluation metrics used in this study.

3.3. Optimizing Hyperparameters for Machine-Learning Methods

In this section, a widely used optimisation algorithm, grid search, is introduced to find the optimal hyperparameters for machine-learning methods. In the hyperparametric optimisation process, the performance of the model needs to be tested by comparing the predicted energy consumption with actual energy data. However, a fixed partition of the training and testing sets may lead to severe overfitting problems. To eliminate possible overfitting, we use the k-fold cross-validation method (k = 5) for optimization. This technique obtains a comprehensive evaluation metric of the prediction model by transforming the training and testing sets multiple times, reflecting the predictive performance and generalization ability of the model.
In this study, the parameters in the BPNN, SVM, RF, and LSTM are selected to be optimized to improve the accuracy of the energy prediction models. Table 3 lists the range and optimal combination of hyperparameters for using the grid search with the cross validation.

3.4. Evaluation Metrics

To evaluate the model accuracy of point estimation, we use three metrics: the root mean square error (RMSE), mean absolute error (MAE), and coefficient of variation of RMSE (CV(RMSE)). All these error evaluation indexes have been extensively applied in the forecasting model estimation. RMSE and MAE can be expressed as below:
R M S E = 1 n i = 1 n ( p i o i ) 2
M A E = 1 n i = 1 n | p i o i |
where, p i is the predicted value, o i is the observed value, and n is the number of measured data.
CV(RMSE) is the measure of the accumulated magnitude of error and is calculated as follows:
C V ( R M S E ) = 1 n i = 1 n ( p i o i ) 2 1 n i = 1 n o i
For assessing the interval prediction results, two evaluation metrics obtained from [40] can be utilized in this work. Firstly, the prediction interval coverage probability (PICP) evaluates whether the actual value of energy consumption is within the predicted interval. In addition, the prediction interval normalized average width (PINAW) is used to measure the width of the predicted interval. The PINAW is related to the informativeness of PICP or equivalently to the sharpness of the predictions. If PINAW is large, the PICP will have little value as it is meaningless to say that future energy consumption will lie within its possible extreme range. Ideally, prediction intervals should have PICPs close to the expected coverage rate and low PINAWs. These two metrics are defined as below:
P I C P = 1 n i = 1 n α i
P I N A W = 1 n E i = 1 n ( U i L i ) ,
where α i = 1 if the actual energy consumption lies within the prediction interval, and α i = 0 otherwise. L i and U i represent the minimum and the maximum values of the predicted interval, respectively. E is the difference between them.
To comprehensively compare the different evaluation metrics, the entropy weights method (EWM), which is a commonly used information-weighting method in decision making [41], is introduced in this paper. The EWM is mainly divided into the following steps:
Step 1: Construction of the initial matrix
The initial matrix is constructed as follows:
X = [ x 11 x 12 x 21 x 22 x 1 m x 2 m x n 1 x n 2 x n m ]
where X is the initial matrix, m is the number of factors, n is the number of data sample, x i j is the analyzed values of each sample parameter, i = 0, 1, 2, …, m and j = 0, 1, 2, …, n .
Step 2: Normalization of the matrix
The normalized matrix can be expressed as following:
X = [ x 11 x 12 x 21 x 22 x 1 m x 2 m x n 1 x n 2 x n m ] x i j = ( x i j ) m a x j x i j ( x i j ) m a x j ( x i j ) m i n j
where X is the normalized decision matrix.
Step 3: Calculation of the entropy
The entropy of each factor is calculated as the following:
e j = 1 ln n i = 1 n p i j ln p i j p i j = x i j / i = 1 n x i j
where e j is the entropy of each factor.
Step 4: Calculation of the weight
The weight of each factor is calculated as the following:
w j = ( 1 e j ) / j = 1 m ( 1 e j )
where w j is the weight of each factor.
Step 5: Calculation of the composite score
The composite score of factors can be calculated as the following:
s i = j = 1 m w j x i j
where s i is the composite score of the factors.

4. Results and Discussion

4.1. Model Training

The first step to setup a prediction model is to determine the number of lag periods, which can also be understood as input to the time-series model. AIC and BIC have been widely used in literature [42,43]; these two criteria can find the best balance between the complexity of the model and the accuracy of the model. Therefore, we select AIC and BIC to be the selection criteria for the number of lag periods, and they are implemented using MATLAB. Figure 3 shows the variations of AIC and BIC in the different number of lag periods.
As seen in Figure 3, both criteria present a gradual decline in the trend, while the value of AIC rises slowly after each sharp drop when the number of lag periods is greater than 100. The software EViews is also applied to ensure the reliability of the results as it is very convenient to calculate AIC and BIC. After comprehensive consideration of two results, 292 is chose to be the optimal number of lag periods, which is indicated by a dotted purple line in Figure 3.
After obtaining the optimal lag periods, we adopt an iterative approach for predicting the daily and weekly ahead indoor loads. The general idea is illustrated in Figure 4. Firstly, the prediction models are performed based on the historical values, and the predicted value at the next time step is generated by the prediction models. Then, this predicted value is combined with part of the historical values to form the new inputs. Finally, the process stops till all predictions of the daily or weekly indoor loads are generated. Therefore, the same lagged values are used for one week and one day ahead forecasting.

4.2. Point Prediction Results

In this section, the results of the point prediction performances of the models are evaluated on the test data. It is worth noting that the test data contains continuous and complete data for 112 days; therefore, we obtain 112 days of results for one day ahead forecasting, and each day contains 96 predicted values and 16 weeks of results for one week ahead forecasting, and each week contains 672 predicted values, respectively. It can be noticed that there are measurable differences between the results of the MAE of the test models for daily and weekly load forecasting in Figure 5. In general, RF performs the worst for one day ahead forecasting and SVM performs the worst for one week ahead forecasting in terms of MAE, and the performance difference between LSTM and BPNN is not obvious. Nevertheless, in Figure 5A,B, the median MAE of LSTM is still lower than that of BPNN. It can be indicated that the LSTM model is more advantageous and reliable than the other models.
Figure 5 and Figure 6 depict the RMSE results of the proposed models for daily and weekly load forecasting. Obviously, the performances of the other three models are still not as good as the LSTM model. It is interesting to note that there are some extremely high values in the RF and LSTM models, especially in Figure 6A. A possible reason for this might be due to the failure of the model in some situation as the predicted error increases with the number of iterations. We also observe that the difference between the RMSE of LSTM and other models are amplified in Figure 6, which further indicate that the LSTM model outperforms other models due to the smaller dispersion degree of RMSE distribution.
Regarding the CV(RMSE), as illustrated in Figure 7, the differences between the proposed models are more evident. SVM performs badly in one day ahead load forecasting, as the CV(RMSE) of this model is much higher than that of other models. In addition, an interesting finding can further be noticed. In general, the overall distribution of CV(RMSE) in one day ahead load forecasting is higher than that in one week ahead load forecasting, which indicates that the prediction of the energy consumption trend in the short-term time is relatively accurate.
Based on the point prediction results above, it can be concluded that the LSTM models perform the best among these commonly used prediction models. The reason for this may be that LSTM learned the dependencies between input variables in sequence by its chain structure, i.e., the periodic characteristics of the energy consumption data in the time domain. Therefore, in terms of periodic prediction time, LSTM will have an advantage over other commonly used models. It is worth noting that although the LSTM model obtains the best prediction performance among the three proposed models, it still has a large gap in its performance compared with short-time load forecasting [44]; it is inevitable because the prediction errors add up as the number of prediction steps increases. Table 4 indicates that the LSTM models perform the best among these commonly used prediction models as well.

4.3. Interval Prediction Results

In this section the results of the interval prediction are obtained by the aforementioned four models integrated with PQR. The forecast results of 16 weeks of the test data are presented in Figure 8 and Figure 9B. By comparing the four proposed methods, it can be seen that the width of the prediction interval of RF is the largest, whereas the probability that the actual value falling into the calculated prediction interval limits of SVM is the lowest. On the contrary, LSTM obtains the highest PICP and lowest PINAW among the proposed models, which means this model is suitable for weekly interval load forecasting. Furthermore, the numerical results for the weekly interval load forecasting are described in Table 5; it is found that the prediction interval of LSTM and RF can cover nearly all the actual load in some weeks’ predictions, as the maximum PICP of these models is very close to 1. In terms of PICP in Table 5, after comprehensive comparison, the LSTM model still has the best performance, although the standard deviation of the PICP results is not the smallest. Moreover, in the aspect of the width of the prediction interval, there is a slight advantage in viewing the PINAW results in the LSTM model compared with the BPNN model. However, our main concern is still the accuracy of the prediction, and thus, PICP accounts for a larger share of the forecast results. Therefore, LSTM-PQR is proven to be a more convincing model for weekly interval load forecasting.
Similar to the weekly interval load forecasting, the results of the daily interval load forecasting are shown in Figure 8 and Figure 9A. In the mass, the prediction performance of the proposed models for daily load is improved in terms of PICP; it is reasonable due to the relative simplicity of the short-term model. Likewise, LSTM-PQR maintains the best predictive performance, while SVM has the lowest value of PICP, and RF gets the highest value of PINAW.
As exhibited in Table 6, LSTM performs better than the other three proposed models, and the mean PICP of LSTM is 6.4% and 20.9% higher than that of RF and SVM, respectively. Despite that, BPNN has the highest value of maximum PICP among these models; it is still not an optimal model as it has the highest value of standard deviation PICP. In addition, the standard deviation PICP of the LSTM model is similar to the RF model but inferior to RF model. This is due to the stable point prediction performance of the RF model, which can be seen in Figure 7A. It can be concluded that the stability of the point prediction affects the stability of the interval prediction to some extent.
The prediction performance advantage also further illustrates that the LSTM-PQR model can accurately predict the indoor load of the building range within a period of time. Similar to the weekly interval load forecasting, LSTM-PQR get the highest value of PICP and the lowest value of PINAW simultaneously; this indicates that LSTM-PQR can lead to accurate and reliable results in the daily interval load forecasting. The results in Table 7 also confirm this conclusion.

5. Conclusions

In this paper, we applied the deep-learning method, LSTM, with penalized quantile regression for interval indoor load forecasting in an office building. The commonly used machine-learning methods, BPNN, SVM, and random forest, are adopted as reference models. The point prediction performance and interval prediction performance of the proposed models have been comprehensively studied in the paper. We can draw the following conclusions:
  • The proposed LSTM-PQR model has better performance than BPNN, SVM, and RF for interval indoor load forecasting in an office building.
  • For point prediction performance, the distribution of MAE and RMSE in different models are quite different. Generally, the LSTM model maintains optimal performance, while the SVM and RF models have relatively poor prediction results.
  • For interval prediction performance, the LSTM-PQR model is still the most suitable choice, especially in daily interval load forecasting, and it has improvements ranging from 6.4% to 20.9% for PICP comparing with the other three models.
Future works should focus more on the ability to capture sudden sharp increments of the load, as it is quite challenging work and there is still no literature on this area, and the prediction of negative energy consumption, such as in eco-parks where buildings receive and deliver energy to factories [45].

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to commercial privacy.

Conflicts of Interest

The author declares no conflict of interest.

Nomenclature

ANNArtificial neural network
SVMSupport vector machine
DTDecision tree
PSOParticle swarm optimization
PCAPrincipal component analysis
MAEmean absolute error
RMSEroot mean square error
AICAkaike Information Criterion
BICBayesian Information Criterion
CV(RMSE)Coefficient of variation of the RMSE
PQRPenalized quantile regression
RFRandom forest
PICPPrediction interval coverage probability
PINAWPrediction interval normalized average width
DEDifferential evolution

References

  1. Ma, M.; Cai, W.; Wu, Y. China Act on the Energy Efficiency of Civil Buildings (2008): A decade review. Sci. Total Environ. 2019, 651, 42–60. [Google Scholar] [CrossRef]
  2. Fan, C.; Wang, J.; Gang, W.; Li, S. Assessment of deep recurrent neural network-based strategies for short-term building energy predictions. Appl. Energy 2019, 236, 700–710. [Google Scholar] [CrossRef]
  3. Somu, N.; Raman, M.R.G.; Ramamritham, K. A hybrid model for building energy consumption forecasting using long short term memory networks. Appl. Energy 2020, 261, 114131. [Google Scholar] [CrossRef]
  4. van der Meer, D.W.; Widén, J.; Munkhammar, J. Review on probabilistic forecasting of photovoltaic power production and electricity consumption. Renew. Sustain. Energy Rev. 2018, 81, 1484–1512. [Google Scholar] [CrossRef]
  5. Chen, C.W.S.; Gerlach, R.; Hwang, B.B.K.; McAleer, M. Forecasting Value-at-Risk using nonlinear regression quantiles and the intra-day range. Int. J. Forecast. 2012, 28, 557–574. [Google Scholar] [CrossRef] [Green Version]
  6. Maciejowska, K.; Nowotarski, J.; Weron, R. Probabilistic forecasting of electricity spot prices using Factor Quantile Regression Averaging. Int. J. Forecast. 2016, 32, 957–965. [Google Scholar] [CrossRef]
  7. Bayer, S. Combining Value-at-Risk forecasts using penalized quantile regressions. Economet. Stat. 2018, 8, 56–77. [Google Scholar] [CrossRef]
  8. Niemierko, R.; Toppel, J.; Trankler, T. A D-vine copula quantile regression approach for the prediction of residential heating energy consumption based on historical data. Appl. Energy 2019, 233, 691–708. [Google Scholar] [CrossRef]
  9. Kalogirou, S.A. Applications of artificial neural-networks for energy systems. Appl. Energy 2000, 67, 17–35. [Google Scholar] [CrossRef]
  10. Dong, B.; Cao, C.; Lee, S.E. Applying support vector machines to predict building energy consumption in tropical region. Energy Build. 2005, 37, 545–553. [Google Scholar] [CrossRef]
  11. Chou, J.S.; Bui, D.K. Modeling heating and cooling loads by artificial intelligence for energy-efficient building design. Energy Build. 2014, 82, 437–446. [Google Scholar] [CrossRef]
  12. Gonzalez, P.A.; Zamarreno, J.A. Prediction of hourly energy consumption in buildings based on a feedback artificial neural network. Energy Build. 2005, 37, 595–601. [Google Scholar] [CrossRef]
  13. Chen, Y.B.; Xu, P.; Chu, Y.Y.; Li, W.L.; Wu, Y.T.; Ni, L.Z.; Bao, Y.; Wang, K. Short-term electrical load forecasting using the Support Vector Regression (SVR) model to calculate the demand response baseline for office buildings. Appl. Energy 2017, 195, 659–670. [Google Scholar] [CrossRef]
  14. Chen, J.L.; Yang, J.X.; Zhao, J.W.; Xu, F.; Shen, Z.; Zhang, L.B. Energy demand forecasting of the greenhouses using nonlinear models based on model optimized prediction method. Neurocomputing 2016, 174, 1087–1100. [Google Scholar] [CrossRef]
  15. Azar, E.; Al Ansari, H. Framework to investigate energy conservation motivation and actions of building occupants: The case of a green campus in Abu Dhabi, UAE. Appl. Energy 2017, 190, 563–573. [Google Scholar] [CrossRef]
  16. Awan, S.M.; Aslam, M.; Khan, Z.A.; Saeed, H. An efficient model based on artificial bee colony optimization algorithm with Neural Networks for electric load forecasting. Neural Comput. Appl. 2014, 25, 1967–1978. [Google Scholar] [CrossRef]
  17. Lee, H.S.; Chou, M.T. Fuzzy forecasting based on fuzzy time series. Int. J. Comput. Math. 2004, 81, 781–789. [Google Scholar] [CrossRef]
  18. Fan, C.; Xiao, F.; Zhao, Y. A short-term building cooling load prediction method using deep learning algorithms. Appl. Energy 2017, 195, 222–233. [Google Scholar] [CrossRef]
  19. Fu, G.Y. Deep belief network based ensemble approach for cooling load forecasting of air-conditioning system. Energy 2018, 148, 269–282. [Google Scholar] [CrossRef]
  20. Li, Z.P.; Liu, T.Z. Improved particle filter based soft sensing of room cooling load. Energy Build. 2017, 142, 56–61. [Google Scholar]
  21. Paterakis, N.G.; Mocanu, E.; Gibescu, M.; Stappers, B.; van Alst, W. Deep Learning Versus Traditional Machine Learning Methods for Aggregated Energy Demand Prediction. In Proceedings of the 2017 IEEE PES Innovative Smart Grid Technologies Conference Europe (Isgt-Europe), Torino, Italy, 26–29 September 2017. [Google Scholar]
  22. Dahl, M.; Brun, A.; Andresen, G.B. Using ensemble weather predictions in district heating operation and load forecasting. Appl. Energy 2017, 193, 455–465. [Google Scholar] [CrossRef]
  23. Chitsaz, H.; Shaker, H.; Zareipour, H.; Wood, D.; Amjady, N. Short-term electricity load forecasting of buildings in microgrids. Energy Build. 2015, 99, 50–60. [Google Scholar] [CrossRef]
  24. Bedi, J.; Toshniwal, D. Deep learning framework to forecast electricity demand. Appl. Energy 2019, 238, 1312–1326. [Google Scholar] [CrossRef]
  25. Chae, Y.T.; Horesh, R.; Hwang, Y.; Lee, Y.M. Artificial neural network model for forecasting sub-hourly electricity usage in commercial buildings. Energy Build. 2016, 111, 184–194. [Google Scholar] [CrossRef]
  26. Behl, M.; Smarra, F.; Mangharam, R. DR-Advisor: A data-driven demand, response recommender system. Appl. Energy 2016, 170, 30–46. [Google Scholar] [CrossRef] [Green Version]
  27. Li, Z.X.; Dong, B. Short term predictions of occupancy in commercial buildings-Performance analysis for stochastic models and machine learning approaches. Energy Build. 2018, 158, 268–281. [Google Scholar] [CrossRef]
  28. Tascikaraoglu, A.; Sanandaji, B.M. Short-term residential electric load forecasting: A compressive spatio-temporal approach. Energy Build. 2016, 111, 380–392. [Google Scholar] [CrossRef]
  29. Mocanu, E.; Nguyen, P.H.; Gibescu, M.; Kling, W.L. Deep learning for estimating building energy consumption. Sustain. Energy Grids 2016, 6, 91–99. [Google Scholar] [CrossRef]
  30. Catalina, T.; Iordache, V.; Caracaleanu, B. Multiple regression model for fast prediction of the heating energy demand. Energy Build. 2013, 57, 302–312. [Google Scholar] [CrossRef]
  31. Ji, Y.; Xu, P.; Duan, P.F.; Lu, X. Estimating hourly cooling load in commercial buildings using a thermal network model and electricity submetering data. Appl. Energy 2016, 169, 309–323. [Google Scholar] [CrossRef]
  32. Yun, K.; Luck, R.; Mago, P.J.; Cho, H. Building hourly thermal load prediction using an indexed ARX model. Energy Build. 2012, 54, 225–233. [Google Scholar] [CrossRef]
  33. Williams, K.T.; Gomez, J.D. Predicting future monthly residential energy consumption using building characteristics and climate data: A statistical learning approach. Energy Build. 2016, 128, 1–11. [Google Scholar] [CrossRef]
  34. Cai, H.L.; Shen, S.P.; Lin, Q.C.; Li, X.F.; Xiao, H. Predicting the Energy Consumption of Residential Buildings for Regional Electricity Supply-Side and Demand-Side Management. IEEE Access 2019, 7, 30386–30397. [Google Scholar] [CrossRef]
  35. Walter, T.; Price, P.N.; Sohn, M.D. Uncertainty estimation improves energy measurement and verification procedures. Appl. Energy 2014, 130, 230–236. [Google Scholar] [CrossRef] [Green Version]
  36. Ahmad, T.; Chen, H.X. Potential of three variant machine-learning models for forecasting district level medium-term and long-term energy demand in smart grid environment. Energy 2018, 160, 1008–1020. [Google Scholar] [CrossRef]
  37. Martinez-Soto, A.; Jentsch, M.F. A transferable energy model for determining the future energy demand and its uncertainty in a country’s residential sector. Build. Res. Inf. 2019. [Google Scholar] [CrossRef]
  38. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  39. Yeon, K. A study on repeat sales house price index based on penalized quantile regression. Procedia Comput. Sci. 2016, 91, 260–267. [Google Scholar] [CrossRef] [Green Version]
  40. Shepero, M.; van der Meer, D.; Munkhammar, J.; Widén, J. Residential probabilistic load forecasting: A method using Gaussian process designed for electric load data. Appl. Energy 2018, 218, 159–172. [Google Scholar] [CrossRef]
  41. Qu, W.; Li, J.; Song, W.; Li, X.; Zhao, Y.; Dong, H.; Qi, Y. Entropy-Weight-Method-Based Integrated Models for Short-Term Intersection Traffic Flow Prediction. Entropy 2022, 24, 849. [Google Scholar] [CrossRef]
  42. Tosunoglu, F. Accurate estimation of T year extreme wind speeds by considering different model selection criterions and different parameter estimation methods. Energy 2018, 162, 813–824. [Google Scholar] [CrossRef]
  43. Ulmeanu, A.P.; Barbu, V.S.; Tanasiev, V.; Badea, A. Hidden Markov Models revealing the household thermal profiling from smart meter data. Energy Build. 2017, 154, 127–140. [Google Scholar] [CrossRef]
  44. Xu, C.L.; Chen, H.X.; Xun, W.D.; Zhou, Z.X.; Liu, T.; Zeng, Y.K.; Ahmad, T. Modal decomposition based ensemble learning for ground source heat pump systems load forecasting. Energy Build. 2019, 194, 62–74. [Google Scholar] [CrossRef]
  45. Butturi, M.A.; Sellitto, M.A.; Lolli, F.; Balugani, E.; Neri, A. A model for renewable energy symbiosis networks in eco-industrial parks. IFAC-PapersOnLine 2020, 53, 13137–13142. [Google Scholar] [CrossRef]
Figure 1. Probability distribution of the indoor load and weather data.
Figure 1. Probability distribution of the indoor load and weather data.
Sustainability 14 08584 g001
Figure 2. Research outline.
Figure 2. Research outline.
Sustainability 14 08584 g002
Figure 3. Variations of AIC and BIC.
Figure 3. Variations of AIC and BIC.
Sustainability 14 08584 g003
Figure 4. Iterative process for daily and weekly ahead indoor load prediction.
Figure 4. Iterative process for daily and weekly ahead indoor load prediction.
Sustainability 14 08584 g004
Figure 5. MAE results of proposed models for daily and weekly load forecasting.
Figure 5. MAE results of proposed models for daily and weekly load forecasting.
Sustainability 14 08584 g005
Figure 6. RMSE results of proposed models for daily and weekly load forecasting.
Figure 6. RMSE results of proposed models for daily and weekly load forecasting.
Sustainability 14 08584 g006
Figure 7. CV(RMSE) results of proposed models for daily and weekly load forecasting.
Figure 7. CV(RMSE) results of proposed models for daily and weekly load forecasting.
Sustainability 14 08584 g007
Figure 8. PICP results of proposed models for daily and weekly interval load forecasting with 95% prediction interval.
Figure 8. PICP results of proposed models for daily and weekly interval load forecasting with 95% prediction interval.
Sustainability 14 08584 g008
Figure 9. PINAW results of proposed models for daily and weekly interval load forecasting with 95% prediction interval.
Figure 9. PINAW results of proposed models for daily and weekly interval load forecasting with 95% prediction interval.
Sustainability 14 08584 g009
Table 1. The statistics of the short-term forecasting articles.
Table 1. The statistics of the short-term forecasting articles.
ReferencesTemporal GranularityTime HorizonMethod
[22,23,24]Hourly24 hARX, SRWNN, LSTM
[25]15 min24 hANN
[26]5 min24 hmbCRT
[27]15 min15–60 minMarkov model, ANN, SVR
[28]Hourly6 hAR, WT-AR, NM-AR
Table 2. Distribution analysis of the indoor load and weather data.
Table 2. Distribution analysis of the indoor load and weather data.
DataDistributionParameters
Energy consumptionLognormalMean of logarithmic values: 4.284
Standard deviation of logarithmic values: 0.706
Ambient temperatureNormalMean:18.481
Standard deviation: 9.451
Relative humidityExtreme valueLocation parameter: 77.860
Scale parameter:14.660
Wind speedWeibullScale parameter: 4.564
Shape parameter: 2.473
Table 3. Optimal combination of hyperparameters using grid search with cross validation.
Table 3. Optimal combination of hyperparameters using grid search with cross validation.
ModelHyperparametersRangeStep SizeOptimal Value
BPNNNumber of neurons[10, 100]540
Activation function‘sigmoid’, ‘tanh’, ‘relu’None‘sigmoid’
Initial learning rate[0.001, 0.01]0.0010.002
SVMKernel function‘Gaussian’, ‘linear’, ‘poly’None‘Gaussian’
C[1, 5]0.53
Epsilon[0.1, 1]0.10.2
RFNumber of the trees[5, 100]550
Depth of the trees[10, 120]1050
Minimum sample number in leaf node[1, 5]11
LSTMNumber of neurons[10, 100]540
Activation function‘softsign’, ‘sgdm’, ‘tanh’None‘sgdm’
Initial learning rate[0.001, 0.01]0.0010.002
Table 4. Composite score of proposed models for daily and weekly load forecasting.
Table 4. Composite score of proposed models for daily and weekly load forecasting.
ModelComposite Score for Daily Load ForecastingComposite Score for Daily Load Forecasting
LSTM99.6099.60
RF16.5260.59
BPNN57.3281.14
SVM8.940.20
Table 5. One week ahead prediction accuracy.
Table 5. One week ahead prediction accuracy.
ModelPICP (mean)PICP (max)PICP (min)PICP (SD)PINAW (mean)PINAW (max)PINAW (min)PINAW (SD)
LSTM0.7390.9880.4500.1710.5650.7570.3860.123
RF0.6460.9820.4000.1790.6371.0040.4110.194
BPNN0.6870.9640.4700.1580.6180.9630.3990.192
SVM0.6690.9490.5230.1270.6591.0630.4130.186
Table 6. One day ahead prediction accuracy.
Table 6. One day ahead prediction accuracy.
ModelPICP (mean)PICP (max)PICP (min)PICP (SD)PINAW (mean)PINAW (max)PINAW (min)PICP (SD)
LSTM0.7810.9350.5220.1130.3560.5190.1840.121
RF0.7310.9130.5040.1110.4140.6080.2070.142
BPNN0.6940.9780.4020.1970.4030.5920.2180.126
SVM0.6180.9110.4130.1510.4100.7660.2090.158
Table 7. Composite score of proposed models for daily and weekly interval load forecasting.
Table 7. Composite score of proposed models for daily and weekly interval load forecasting.
ModelComposite Score for Weekly Interval Load ForecastingComposite Score for Daily Interval Load Forecasting
LSTM99.6099.60
RF11.9521.27
BPNN43.7827.45
SVM12.374.96
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Duan, Y. A Novel Interval Energy-Forecasting Method for Sustainable Building Management Based on Deep Learning. Sustainability 2022, 14, 8584. https://doi.org/10.3390/su14148584

AMA Style

Duan Y. A Novel Interval Energy-Forecasting Method for Sustainable Building Management Based on Deep Learning. Sustainability. 2022; 14(14):8584. https://doi.org/10.3390/su14148584

Chicago/Turabian Style

Duan, Yun. 2022. "A Novel Interval Energy-Forecasting Method for Sustainable Building Management Based on Deep Learning" Sustainability 14, no. 14: 8584. https://doi.org/10.3390/su14148584

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop