Next Article in Journal
Diffusion Length Mapping for Dye-Sensitized Solar Cells
Next Article in Special Issue
Year Ahead Demand Forecast of City Natural Gas Using Seasonal Time Series Methods
Previous Article in Journal
Flashover Characteristics of Silicone Rubber Sheets under Various Environmental Conditions
Previous Article in Special Issue
Comparative Study of Hybrid Models Based on a Series of Optimization Algorithms and Their Application in Energy System Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Application of Non-Linear Autoregressive Neural Networks to Predict Energy Consumption in Public Buildings

by
Luis Gonzaga Baca Ruiz
,
Manuel Pegalajar Cuéllar
,
Miguel Delgado Calvo-Flores
and
María Del Carmen Pegalajar Jiménez
*
Department of Computer Science and Artificial Intelligence, University of Granada, Granada 18071, Spain
*
Author to whom correspondence should be addressed.
Energies 2016, 9(9), 684; https://doi.org/10.3390/en9090684
Submission received: 14 July 2016 / Revised: 18 August 2016 / Accepted: 23 August 2016 / Published: 26 August 2016
(This article belongs to the Special Issue Energy Time Series Forecasting)

Abstract

:
This paper addresses the problem of energy consumption prediction using neural networks over a set of public buildings. Since energy consumption in the public sector comprises a substantial share of overall consumption, the prediction of such consumption represents a decisive issue in the achievement of energy savings. In our experiments, we use the data provided by an energy consumption monitoring system in a compound of faculties and research centers at the University of Granada, and provide a methodology to predict future energy consumption using nonlinear autoregressive (NAR) and the nonlinear autoregressive neural network with exogenous inputs (NARX), respectively. Results reveal that NAR and NARX neural networks are both suitable for performing energy consumption prediction, but also that exogenous data may help to improve the accuracy of predictions.

Graphical Abstract

1. Introduction

According to the International Energy Agency, the building sector is the largest energy-consuming sector, accounting for one-third of total global energy consumption, as well as an equally important source of CO2 emissions [1]. Around 90% of building emissions are produced to operate the HVAC—heat, ventilation, and air-conditioning—and lighting systems [2]. Thus, energy efficiency currently represents a highly relevant and important issue.
During the last few years, the development of new sensing technologies has led to the successful installation of monitoring devices in different environments and, consequently, the wide availability of real-time data being acquired from rather varied elements. However, due to the heterogeneity and diversity of these information flows, coming from very different sensors and sources, appropriate management is difficult. Buildings with energy management systems have much lower operating costs than those without. Therefore, forecasting models for energy consumption represent an essential factor for controlling energy costs and reducing environmental impact. The first target of the models is knowledge extraction, which means the extraction of behavioural patterns and detection of anomalies. Accordingly, specific applications for building efficient energy management systems have been recently explored, such as prediction of the energy demand and detection of consumption profiles [3].
There is plenty of research literature related to time series prediction. Therefore, there are a multitude of models and methodologies related to this field, applied to energy efficiency. Nagy et al. [4] suggest a generalized additive tree ensemble approach to predict solar and wind power generation. Yuan et al. [5] use autoregressive moving average model (ARIMA), grey model GM(1,1), and a hybrid model to forecast China’s primary energy consumption. Fumo and RafeBiswas [6] apply linear regression analysis to predict energy consumption in single-family homes. Mordjaoui et al. present a short-term electric load forecasting model using an adaptive neuro-fuzzy inference system [7]. Guo et al. [8] write about energy consumption prediction of heat pump water heaters based on grey system. Zhang et al. [9] apply support vector regression (SVR) to building energy consumption.
Due to the excellent results and noteworthy success achieved in real applications [10,11,12,13,14,15], artificial neural networks (ANNs) are held up as one of the most popular models. Several studies have illustrated that ANNs have produced better results when compared to other techniques [16,17,18,19]. One of the great advantages of ANNs is their potential to model non-linear data relationships. Furthermore, once trained, ANNs provide swift performance of the network. Karatasou et al. [20] employed neural networks to estimate energy consumption. Deb et al. [21] present a methodology to forecast diurnal cooling load energy consumption for institutional buildings. Gonzalez and Zamorreño [22] predict short-term electricity by means of feedback ANN. Likewise, Camara et al. [23] apply feedforward networks to forecast energy consumption in the United States.
In the real world, there are a huge amount of non-linear systems, or those whose behaviour is dynamic and depends on their current state. The dynamic recurrent neural network (RNN), the nonlinear autoregressive (NAR), and the nonlinear autoregressive neural network with exogenous inputs (NARX) are neural network structures that can be useful in these cases [24,25]. The first advantage of these networks is that they can accept dynamic inputs represented by time series sets. Time series forecasting using neural network (NN) is a non-parametric method, which means that knowledge of the process that generates the time series is not indispensable. On one hand, the NAR model uses the past values of the time series to predict future values. On the other hand, the RNN model does not need past time series values as inputs nor delays, since it has recurrent connections within its own structure.
In this study, we focus on an experiment using NAR and NARX neural networks to predict energy consumption in public buildings. Our goal is to provide a methodology framework to analyze the time series historical data of energy consumption, and to know if the mid-term prediction of such energy consumption can be achieved with these models. Moreover, we test both NAR and NARX models to study if the energy consumption depends solely on historical data, or if further external variables can be used to increase the accuracy of predictions. This study in public buildings (faculties and research centers) of the University of Granada (UGR) has been carried out as a test case. At this University, innovative management systems have been installed in the last few years; as a consequence we can obtain sufficient energy consumption data to carry out the experimentation. A minimum of the last two years of data will be used for each building analyzed.
The main goal of this study is to analyze different forecasting methods for energy consumption. To that end, we are also interested in the study of the impact of exogenous data on the forecasting. The results could be used to avoid unnecessary consumption: reducing costs in the long term, saving energy and decreasing the emission of harmful by products (the most dangerous of which is carbon monoxide) and providing an initial framework on which to build more complex decision-support systems.
The manuscript is organized as follows: Section 2 introduces the proposed methodology and the description of the artificial neural network methods. After that, Section 3 describes the phases of the methodology used in the study, comprising the processing of data, noise data management, transformations, and the process of training, test, and validation. Section 4 provides a description of the real data used, obtained from buildings at the University of Granada. Section 5 and Section 6 shows the experimentation with the proposed models and their results. Finally, Section 7 gathers the main conclusions obtained and makes proposals for future work.

2. Methodology

The current research was developed in three stages: firstly, the data collection and pre-processing; secondly, ANN modelling and, finally, the analysis of performance and comparison between two different network prediction models: NAR and NARX networks.
Previous work has shown that classic feedforward ANN may provide outstanding results in time series prediction tasks [7,10,11,12,13,14,24,25]. In this study, we have selected two neural network architectures aimed specifically at this problem to be used as prediction models for energy consumption. The first model is non-linear autoregressive neural networks (NAR), which are to forecast samples framed in a one-dimensional time series. We have also selected non-linear autoregressive networks with exogenous inputs (NARX), which expand multidimensional time series using external information to enhance time series prediction performance.
Each network model has its benefits and costs: NAR methods are simpler than NARX. Nevertheless, the latter model allows the use of additional information that may improve prediction accuracy. In a real situation, not every building has the same management system nor the same number of features: some only register consumption whilst others handle more information, such as external and internal temperature. The selected ANN models allow us to work with both approaches. The following subsections describe these models to solve the problem of energy consumption time series prediction.

2.1. NAR Model

In the majority of cases, time series applications are characterized by high variations and fleeting transient periods. This fact makes it difficult to model time series using a liner model, therefore a nonlinear approach should be suggested. A nonlinear autoregressive neural network [26,27], applied to time series forecasting, describes a discrete, non-linear, autoregressive model that can be written as follows [28]:
y ( t ) = h ( y ( t 1 ) ,   y ( t 2 ) ,   ,   y ( t p ) ) + ϵ ( t )
This formula describes how a NAR network is used to predict the value of a data series y at time t, y(t), using the p past values of the series. The function h(·) is unknown in advance, and the training of the neural network aims to approximate the function by means of the optimization of the network weights and neuron bias. Finally, the term є(t) stands for the error of the approximation of the series y at time t.
The topology of a NAR network is shown in Figure 1. The p features y(t−1), y(t−2), …, y(t−p), are called feedback delays. The number of hidden layers and neurons per layer are completely flexible, and are optimized through a trial-and-error procedure to obtain the network topology that can provide the best performance. Nevertheless, it is important to bear in mind that increasing the number of neurons makes the system more complex, while a low number of neurons may restrict the generalization capabilities and computing power of the network.
The most common learning rule for the NAR network is the Levenberg-Marquardt backpropagation procedure (LMBP) [29,30,31,32]. This training function is often the fastest backpropagation-type algorithm. The LMBP algorithm was designed to approximate the second-order derivative with no need to compute the Hessian matrix, therefore increasing the training speed. When the performance function has the form of a sum of squares (frequently in feedforward network training), then the Hessian matrix can be approximated as shown in Equation (2) and the gradient can be computed as described in Equation (3).
H = J T J
g = J T e
In Equations (2) and (3), J is the Jacobian matrix which contains the first derivatives of the network errors with respect to the weights and biases, and e is a vector of network errors in all training samples. To estimate the Jacobian matrix, the study in [32] uses a standard backpropagation algorithm to approximate the Hessian matrix. This approach is simpler than computing the Hessian matrix. The LMBP algorithm uses this approach in the Newton-like update described in Equation (4).
x k + 1 = x k [ J T J + μ I ] 1 J T e
It should be noted that this method uses the Jacobian matrix for calculations, assuming that the performance function is the mean of the sum of the squared errors. Hence, networks must use either the mean square error (MSE) or error sum of squares (SSE), stated in Equations (5) and (6), where yi stands for the i-th data sample, y i ^ is the approximated data obtained by the network for the value yi, and n the number of data samples for the network training.
S S E = i = 1 n ( y i ^ y i ) 2
M S E = S S E n
In this study, we use NAR neural networks to model energy consumption time series as follows: the network structure contains one input (corresponding to the energy consumption at time t−1, y(t−1)), and one output (the next value of the time series, y(t), to be predicted). The number of delays is set experimentally after a data pre-processing and analysis stage.

2.2. NARX Model

In many real applications, there is an important correlation between the modelled time series and additional external data. In [4] it is indicated that a usual feature of renewable energy is that the output of power plants largely depends on weather conditions. Thus, the integration of knowledge or data about weather could benefit the time series modelling process to provide an accurate forecast instead of a unique approach with one single value corresponding to the consumption data [17,33].
The model nonlinear autoregressive with exogenous (External) inputs (NARX) is proposed in [34]. NARX predict series y ( t ) given p past values of series y and another external series x ( t ) , which can be single or multidimensional. The equation that models the NARX network behaviour for time series prediction is shown in Equation (7).
y ( t ) = h ( x ( t 1 ) ,   x ( t 2 ) ,   ,   x ( t k ) ,   y ( t 1 ) ,   y ( t 2 ) ,   ,   y ( t p ) ) + ϵ ( t )
The NARX is a nonlinear model which estimates the future values of the time series based on its last outputs and external data. In this study, we use NARX with one input for the energy consumption time series at time t-1, y(t-1), and another input with exogenous data at time t−1, x(t−1), to provide a single output data y(t), corresponding to the value of the energy consumption one step forward. Figure 2 shows this architecture, similar to the NAR network. The only difference is in input; henceforth, output y ( t ) takes account of the external data as it appears in the Equation (7). The learning rule LMBP, previously explained, is also used to train this model. Despite the flexibility of NARX to model exogenous input to help improve results by modelling external dependencies, NAR models are a good alternative because of their simplicity, as discussed in [35]. In the experimental section, we make a comparison of both approaches.

3. Energy Time Series Modelling and Forecasting

This section describes the process used to obtain an energy time series model, along with the forecasting methodology to predict the consumption in a building. The process overview is shown in Figure 3. There is no learning model indicated, so that it can support any of them without distinction. In other words, the methodology is general and can be utilized for both NAR and NARX neural networks. Each stage is detailed as follows:
(a)
Building automation system: software designed to support building energy management, which is made up of several components that provide coordinated control. It collects, summarizes and presents building data in usable way.
(b)
Historical energy consumption database: stores the building’s energy use (consumption, energy and power recorded and monitored by diverse sensors), as well as some external information such as temperature. In particular, we study the relationship between temperature and energy consumption using NARX models. The experimental section shows the results and conclusions obtained.
(c)
Data transformation: the building automation system is responsible for the storage of raw data. This information must be processed in order to change the consumption time scale and to extract the consumption data from the raw sensor storage. Regarding the time scale, some manuscripts propose a model for forecasting hourly [15,22,36]. However, most short-term forecasting problems involve predicting events for only a few time periods in a scale of days, weeks, or months [36]. Hence, our experimentation is focused in this context.
(d)
Noise treatment: the data captured from the physical world by sensor devices are sometimes incomplete, noisy and unreliable. This noise is mainly due to fault detection, a broken device or a connection failure. Two main strategies are pursued: (1) moving-average filter [37]—see Equation (8)—to remove outliers; and (2) linear interpolation, based on the values at neighbouring grid points to fill missing values.
y ( n ) = 1 w i n d o w s S i z e · ( x ( n ) + x ( n 1 ) + + x ( n ( w i n d o s S i z e 1 ) ) )
(e)
Train and test: the database is divided into 70% training data and 30% test data. The training and test data are selected randomly from the initial consumption dataset.
(f)
Training time series forecasting model: its aim is to fit the model using the training data. NAR and NARX models have been trained using LMBP as the training function and hyperbolic tangent sigmoid transfer function [38] for the neurons of intermediate layers. The parameters will be detailed in the experimental section. The models are trained with 70% of the days collected, and are tested with the remaining 30% of the data.
(g)
Learned time series forecasting model: once the model is trained, we test the prediction models to know their performance in training data sets and we store these results in a database.
(h)
Validating time series forecasting model: the final step is to verify that accurate results have been obtained, comparing the test data prediction with test data isolated in step 5. This dataset has not previously been used to train the network and can be used to know if the networks have correctly learned the time dependencies of the consumption data.

4. Data Collection Process

Smart management systems collect a multitude of information, monitored in real-time. In this paper, data are obtained from an actual building automation system (BAS). BAS is an automatic centralized control of the building’s heating, ventilation, air conditioning and lighting systems. This BAS is used by the University of Granada (UGR) (Granada, Spain) to monitor its buildings’ consumption, in order to analyze the information coming from several sources with the ultimate objective of better understanding of how and when energy is consumed in the distributed facilities. Data are generated and stored in a fine-grained fashion, which hinders the processes of data analysis and decision-making. The information should, therefore, be summarized and simplified in order to make useful forecasting.
The UGR has access to energy consumption and climatic data gathered in the buildings. In each campus, different kinds of buildings and sensor setups can be found. Teaching—units’ faculties, schools, and departments—is divided throughout five campuses—Centro, Cartuja, Fuentenueva, Aynadamar, and Ciencias de la Salud—which are distributed in different areas of the city of Granada. In addition, the UGR has centres in the autonomous cities of Ceuta and Melilla which are located in separate campus. Overall, the UGR is associated with 22 faculties, 5 schools, 8 training centres and five culture, sport, and service centres.
Each building consumption dataset is grouped by area. Therefore, it is assumed that there is a relationship between those buildings. Eight buildings belonging to the same compound have been selected for this energy consumption experimentation. Due to anonymity, the building consumptions are labelled as Building 1, Building 2, Building 3, and Building 8. These consumptions have been selected in the following way: the first and second buildings are chosen from the same compound, the third and fourth in the same way, etc. We have selected two representative buildings from four campuses to study the general energy consumption at the UGR.

4.1. Data Selection

Data selection was carried out by a preliminary pre-processing of all data coming from several faculties and determined buildings. After obtaining the data to be used to model the ANN, the subsequent step was to separate the data into two parts: training dataset and test dataset. All of the data were normalized to have the same range of values for each of the inputs to the ANN model. This guarantees that there are no attributes which are more important than others because of data ranges, and also eases a stable convergence of network weights and biases.
All data associated with energy consumption are stored hourly in the Historical Energy Consumption Database, and these data are the raw values obtained from energy counters. Thus, one sample of the raw consumption data describes the aggregate sum of all previous consumption until that moment. In this piece of research, data have been transformed in consumption registered per hour, so that one sample of the transformed data shows the consumption that took place precisely in that hour.

4.2. Sample Building

This subsection aims to summarize one sample building structure and to describe in detail the main characteristics and parameters monitored in the building compound. The selected building is comprised of four floors: basement, ground floor, first floor, and second floor (see Figure 4).
Each floor monitors three different types of information: lighting, electricity, heating and air conditioning. This building has a network automation engine—NAE—which delivers comprehensive equipment monitoring and control through features such as scheduling, alarm and event management, energy management, data exchange, data trending, and data storage. This tool helps lower efficiency costs for buildings. The MS-NAE5520-2 is the specific model used in this building (see Figure 5).
The MS-NAE5522-2 device supports a LonWorks trunk, two N2 Bus trunks or two BACnet MS/TP (RS-485), or a mixture of one N2 Bus trunk and one BACnet MS/TP trunk. It supports a maximum number of 255 devices on the LonWorks trunk and a maximum of 100 devices on each N2 Bus or BACnet MS/TP trunk. Its technical characteristics are summarized as follows:
  • Power requirement: Dedicated nominal 24 VAC, Class 2 power supply (North America), SELV power supply (Europe), at 50/60 Hz (20 VAC minimum to 30 VAC maximum).
  • Power consumption: 50 VA maximum.
  • Ambient operating conditions: 0 to 50 °C (32 to 122 °F); 10% to 90% RH, 30 °C (86 °F) maximum dew point.
  • Ambient storage conditions: −40 to 70 °C (−40 to 158 °F); 5% to 95% RH, 30 °C (86 °F) maximum dew point.
  • Data protection battery: Supports data protection on power failure. Rechargeable gel cell battery: 12 V, 1.2 Ah, with a typical life of 3 to 5 years at 21 °C (70 °F); Product Code Number: MS-BAT1010-0.
  • Clock battery: Maintains real-time clock through a power failure. Onboard cell; typical life 10 years at 21 °C (70 °F).
  • Processor: 1.6 GHz Intel® Atom™ processor.
  • Memory: 4 GB (2 GB partitioned) flash nonvolatile memory for operating system, configuration data, and operations data storage and backup, 1 GB SDRAM for operations data dynamic memory.
  • Operating system: Johnson Controls OEM Version of Microsoft Windows Standard 2009.
  • Network and serial interfaces: One Ethernet port; connects at 10 Mbps, 100 Mbps, or 1 Gbps; eight-pin RJ-45 connector. Two optically isolated RS-485 ports; 9.6 k, 19.2 k, 38.4 k, or 76.8 k baud; pluggable and keyed four-position terminal blocks (RS-485 ports available on NAE55 models only). Two RS-232-C serial ports, with standard nine-pin sub-D connectors, that support all standard baud rates. Two USB serial ports; standard USB connectors support an optional, user-supplied external modem. Options: One telephone port for internal modem; up to 56 kbps; six-pin modular connector. One LONWORKS port; FTT10 78 kbps; pluggable, keyed three-position terminal block (LONWORKS port available on NAE552x-x models only).
  • Housing: Plastic housing with internal metal shield. Plastic material: ABS + polycarbonate UL94-5VB Protection: IP20 (IEC 60529).
  • Mounting: On flat surface with screws on four mounting feet or on dual 35 mm DIN rail.
  • Dimensions (Height × Width × Depth): 226 × 332 × 96.5 mm including mounting feet. Minimum space for mounting: 303 × 408 × 148 mm.
The energy management system in this building monitors different features, such as voltage, intensity, and power, which are collected in their different phases:
  • Voltage phase R, S and T-neutral (V).
  • Intensity phase R, S and T (A).
  • Active power phase R, S and T (kW).
  • Reactive power phase R, S and T (kVAR).
  • Power factor phase R, S and T.
  • Three-phase active power (kW).
  • Frequency (Hz).
  • Voltage phases: R-S, S-T and T-R (V).
  • Energy consumed (kWh).
  • Neutral intensity (A).
  • Maximum demand (kW).
  • Total harmonic distortion (%).
Furthermore, there are other devices used to monitor the lighting state. Figure 6 shows the state of a lamp in a room. A green state indicates that a light is on, and a blue state indicates that a light is off. Moreover, there is a pair of indicators for each room, the left one specifies the order—for example, if a light is going to power off in 30 seconds it appears in blue—and the right box shows the real state.
This figure shows how all the lamps are on in all the rooms, except in the corridor and the stairs.
All of these counters are gathering information 24 h a day, every day of the year. However, not all data are stored permanently; some features, such as lamp states or the state of a valve, are not saved in the long term. This research is concentrated mainly on the energy consumption registered and the temperature measured.
Regarding the temperature monitoring system, Figure 7 shows an example of a complete control ventilation system. This system senses external temperature and internal building temperature. It has several valves to control optimal air conditions, such as the temperature in degrees Celsius in the environment, which leads the system to close or open valves and, thus, introduce appropriate air into the system. One sensor indicates if there is any problem in the system, and another sensor is used to distinguish between summer and winter. Furthermore, there is a de-icing device, which is turned on when the system detects that the temperature is too low to prevent freezing during the winter.
In this research paper, temperature will be a useful and distinctive characteristic because of its relation with energy consumption. These results will be discussed in more detail in the experimental section. It is reasonable to believe that if the temperature becomes too high or too low air conditioners will have more use. Figure 8 shows the consumption and the temperatures registered over time. Both time series have been normalized because they have different data measurement units. Temperature is in degrees Celsius and consumption is in kW. This figure illustrates interesting trends which have been described previously. For instance, low temperatures, such as those between December and February, have a high consumption and the air conditioners were used more during those dates.

5. Results

This section describes the entire series of tests performed and their experimental settings. Figure 9 shows the consumption of a sample building during a period of two years, including all week days. Figure 9a illustrates the building’s consumption during working days; and Figure 9b shows the consumption on Saturdays and Sundays. It is important to remark that the consumption is similar for days that are consecutive, both on the working and weekend days. Thus, the weekdays in the graph are almost indistinguishable from each other. Nevertheless, this behavior is very interesting and is useful to adapt the parameters of the forecasting models. In other words, one example of relevant information which can be mined of this data is if a week starts with an increase in consumption, the previous and next days will have the same behavior.
Figure 9a,b show similar behavior between the same types of day. Likewise, the behavior presents similarity between different types of days. On one hand, the increases and decreases in consumption have the same shape in the Figure 9c, although, as noted in the figure, consumption is almost always lower on non-working days.
Unfortunately, this behavior is not the same for all buildings. Figure 10 shows consumption registered in another building on working and weekend days. This time, these two time-series follow two different progresses: non-working days and working days seem to follow a completely different trend depending on the building type. We are not able to establish a relation between these two behaviors according to the building use (offices, classrooms, data centers, research centers, etc.); the mining of the underlying data relations that produce these types of consumption will be a subject of further research. As a consequence, in the current study, we make no initial hypotheses about the data relationship between buildings.
According to this information and analysis, it is important to adjust the delay parameters of the ANN for each building separately, in order to obtain an accurate prediction model in each case. Delay parameters concern the number of days the model is going to use to perform the consumption prediction. In other words, the model is trained with the last p days as delays. We have experimented and run 10 tests with ten different delays to find the most appropriate value for each building. Among these are: the day before ( p = 1 ); the complete last week ( p = 7 ); the last two weeks ( p = 14 ); and the last three weeks ( p = 21 ). To find the best delay, we set all parameters to a fixed value and modified the delay in a trial-and-error procedure to find the value that could provide the network with the best performance.
Table 1 illustrates the average mean square error of 10 executions for each delay value and building, using the NAR model. From this table, we can acquire the best delay obtained for each building (marked as bold values). In all cases, the worst error is acquired with only 1 delay (marked as crossed-out values). The minimum delay required to get the most accurate prediction performance is nine days; although two buildings extend to two weeks to get the best result. From these results, conclusions can then be made relative to the delay information: a delay of between nine to 14 previous days is needed in order to obtain the best NAR forecasting model.
Once the delays have been set, the next step is to adjust the number of neurons needed to train the NAR network. In this case, the delay parameters have been set to a fixed value, obtained from the best MSE attained in Table 1. With these delays, we have run 10 different executions of each network configuration to find the average MSE of each setting, in order to find the best network structure for each energy consumption time series. Table 2 collects experiments with a different number of hidden neurons, ranging from 2 to 20. This experimentation is aimed at knowing the network structure that can provide the best performance. It emerges that more neurons result in a worse prediction, due to local optimum optimization results using the training algorithm, and also to overtraining. From Table 2, we conclude that the best average errors are obtained using between six and 12 neurons, depending on the building in question.
After this experimentation, we acquired the best NAR networks trained to predict the energy consumption of the eight buildings. Figure 11 shows an example of the prediction of energy consumption for Building 5, for a period of 100 days. This figure contains both the best and the worst trained networks in the 10 runs. Figure 11a presents the best forecasting curves captured, with a MSE of 0.68, and the second shows the worst forecasting, with a MSE of 2.23.
Since these experiments have been run using NAR models, the results have been obtained using consumption information only, without external data. NARX models allow the use of extra information that could be useful for the prediction. In this study, we use temperature data as an external variable.
For the experimentation with NARX, we followed the same methodology as with NAR models. Table 3 shows the results regarding the delays required in order to obtain the best model. In this case, the best delay horizons are under one week, whereas the worst delay horizons are always over two weeks.
In the same way as was done for the NAR model, Table 4 shows the results of the test with the number of hidden neurons. The best results achieved with the NARX model have been found with a fewer number of neurons since the best average MSE involve a minimum of two neurons in most of cases.
According to this information, we can conclude that using exogenous data may help to reduce the time horizon of past values to acquire an accurate prediction, but also that the network topology requires less computing units to perform the time series modelling. While the average number of neurons required by NAR to get the best forecasting is six, NARX uses an average of only two neurons.
Figure 12 shows the forecasting curves using NARX models for the consumption data of Building 5. The best result obtained is plotted in Figure 12a, with a MSE of 0.66. Likewise, the worst prediction curve (Figure 12b) is better than NAR, with an MSE of 1.67. NARX models fit the consumption data better, due to the effect of exogenous data.
Finally, Figure 13 compares the means obtained with NAR and NARX networks in respect to the models’ complexity (number of hidden neurons). Both graphs demonstrate that NARX’s outcomes are better than NAR’s. The curve of MSE in the case of NARX network is below the NAR curve in the majority of cases, especially when the network complexity is simpler. The next section discusses all of the results and outcomes obtained from this experimentation.

6. Discussion

Our experimentation has focused on the comparison of neural network models for energy consumption forecasting, considering those cases where there are exogenous data available and those where there are not. Thus, the first issue to be taken into consideration is the availability of data, so that we can perform the model selection. The relative novelty of sensing technologies means that there are still buildings without automated management systems and, consequently, little information can be used. In these cases, the only information stored is the energy consumption. However, there is an increase in the installation and use of complex energy monitoring systems, which enables the possibility of using additional data to model the energy consumption. In the former case, only the consumption data can be used to make the prediction, whilst in the latter case we are able to include external information, such as weather conditions, to design a more accurate prediction model.
According to the analysis of results carried out in the previous section, it is important to draw attention to the significance of the delay parameter. The results collected in Table 1 and Table 3 show that each building’s consumption data requires its own delay parameter, and its value cannot be generalized to the whole building compound. This delay parameter is important because each set of buildings has different behavior, as is observed in Figure 9 and Figure 10. However, we have found that the use of exogenous data may help to reduce the time horizon of past values to be used for the prediction in all cases, since NARX requires fewer delays than NAR, therefore obtaining simpler prediction models.
The experimentation using NAR models demonstrates that this network architecture is useful when only energy consumption information is available, and can provide accurate mid-term predictions. This alternative can be used with a simple dataset containing consumption information only. Table 1 shows that a NAR model with little knowledge, such as one day of delay, is not enough to model the data and provides the worst results. In our dataset, a NAR model requires at least nine days of delay horizon to fit a good predictor. Including a larger delay horizon is not necessary to improve results, since the performance stabilizes after nine days in the delay parameter, according to the results of Figure 13a.
Regarding the network complexity, Figure 13b shows that a simple model with fewer neurons is not capable of modeling the time series complexity, but it also demonstrates that a large number of neurons provides unsuitable results due to local optima in the network parameters’ optimization process during the training stage. Indeed, the worst outcomes result from a large number of neurons.
Finally, the accuracy of these models provides a suitable prediction, as is shown in the sample of Figure 11a.
Regarding NARX neural networks experimentation, results show that the inclusion of simple external data, such as temperature, can help explain sudden changes in the trend of the consumption data and, therefore, obtain a more accurate prediction. For instance, during a hot summer week, if the weather remains constant, it is normal that the consumption is considered to be similar on all days of the week. In the NAR model these changes are not considered, but NARX can model them and use this information to improve the prediction accuracy.
As a consequence, the NARX model requires less past information than the NAR model to achieve an accurate prediction. According to Table 1 and Table 3, the number of delays required in the same dataset decreases abruptly from one model to another. The study of average MSE in respect to the delay horizon in Figure 13a also supports this analysis. The curve of NAR’s MSE regarding the delay parameter quickly decreases as the delays increase, yet these changes are not as radical in the NARX model. In this case, the exogenous temperature data help to model the energy consumption, so that fewer past consumption data are required for the prediction. Half of the outcomes provide the best results using the past five days as delay, and the rest using seven days.
With regard to the NARX model complexity, this approach only needs two neurons to get the best results in five buildings (Table 4). Although the consumption of Building 4 obtains the best MSE with 14 neurons, the second best results are obtained with two neurons. Similar results are drawn for the consumption data of Building 3, where the best results are obtained with six neurons and the second best results with two neurons. However, the consumption of Building 1 needs four neurons and no simpler model provides better results.
According to Figure 13b, the network complexity is always lower when using exogenous data. Since fewer delays are required to obtain the best prediction, fewer computing units (neurons) are needed in the network. When compared to the NAR model, NARX is able to obtain a better prediction accuracy using less historical data, and also with less model complexity. This analysis is supported by Figure 12, where it is shown that the prediction obtained with the NARX model adjusts considerably well to the curve of the real data, with respect to the NAR model.
In summary, Figure 13 provides clear data about the benefits and costs of the proposed network models. The NAR model is a good option when no exogenous information is available. It works with a simpler dataset, but requires a more complex model to get a good predictor. On the other hand, the NARX model can work with external information, and this alternative allows us to feed the model with additional data to obtain simpler predictors. The NAR neural network needs a higher delay horizon and more neurons to adjust a good model; NARX takes advantage of exogenous information to achieve better outcomes with less historical data. Figure 13 shows that the NARX model improves the results of the NAR network; in fact, the worst outcomes achieved by NARX are close to the best results of the NAR in the majority of cases.
As a final remark, Table 5 summarizes the results obtained for the prediction of all buildings, using NAR (column 2) and NARX networks (column 3). According to these results, the use of exogenous data may provide better results in most cases, in terms of prediction accuracy. However, if no external data is available, only the energy consumption might be used to obtain accurate predictions.

7. Conclusions

In this study we have provided a methodology to predict energy consumption in a set of public buildings using neural networks. In our approach, we assume that each building is equipped with an automation system containing energy consumption sensors that store consumption data in a shared database, but we also perform experimentation considering that other external data, such as temperature, are available. Using the sensor data in this database, we applied data pre-processing techniques to transform and normalize the data and remove the noise. After that, we tested different prediction models coming from the neural network area: NAR and NARX. Although both approaches provide suitable prediction accuracy, NARX networks furnish a better performance thanks to the exogenous data. Other advantages of using these exogenous data concern the decrease in the network complexity and the amount of historical data or delays to be used in the forecasting.
According to the dataset, we have discovered that building consumptions at the UGR can be aggregated into two main categories whose main feature is the similarity/dissimilarity of the energy consumption during the working and weekend days. Having knowledge of relations between the buildings’ consumption in advance, and the features that explain these relations, could enable us to use the consumption data of one or more buildings on a campus as exogenous data to predict the energy consumption of the remaining buildings—using the hypothesis that there is a correlation between those data, as the analysis of Figure 9 suggests. In a future piece of research, we will test these assumptions to study the increase in the prediction accuracy.
On the other hand, future research will be also targeted at developing an automatic procedure to find time relations in the energy consumption of the buildings, using data mining techniques to cluster data. We believe that the study of the effect of intra-cluster and inter-cluster relations, and information fusion techniques, will be key aspects to building decision-support systems for energy efficiency.

Acknowledgments

This work has been developed with the support of the Department of Computer Science and Artificial Intelligence of the University of Granada, TIC111, and the project TIN201564776-C3-1-R.

Author Contributions

All authors contributed equally to this work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial neural network
ARIMAAutoregressive integrated moving average
BASBuilding automation system
HVACHeat, ventilation and air-conditioning
LMBPLevenberg-Marquardt backpropagation
MLMachine learning
MSEMean square error
NAENetwork Automation Engine
NARNonlinear autoregressive
NARXNonlinear autoregressive with external input
NNNeural network
RNNRecurrent neural network
SVRSupport vector regression
UGRUniversity of Granada

References

  1. International Energy Agency. Transition to Sustainable Buildings—Strategies and Opportunities to 2050. 2013. Available online: http://www.iea.org/publications/freepublications/publication/transition-to-sustainable-buildings.html (accessed on 21 July 2016).
  2. United Nations Environment Programme. Buildings and Climate Change—Summary for Decision-Makers. 2009. Available online: http://www.unep.org/sbci/resources/Publications.asp (accessed on 21 July 2016).
  3. Molina-Solana, M. Data mining for building energy management. In Proceedings of Sustainable Places 2014, Nice, France, 1–3 October 2014.
  4. Nagy, G.I. GEFCom2014: Probabilistic solar and wind power forecasting using a generalized additive tree ensemble approach. Int. J. Forecast. 2016, 32, 1087–1093. [Google Scholar] [CrossRef]
  5. Yuan, C.; Liu, S.; Fang, Z. Comparison of China’s primary energy consumption forecasting by using ARIMA (the autoregressive integrated moving average) model and GM(1,1) model. Energy 2016, 100, 384–390. [Google Scholar] [CrossRef]
  6. Fumo, N.; Biswas, M.R. Regression analysis for prediction of residential energy consumption. Renew. Sustain. Energy Rev. 2015, 47, 332–343. [Google Scholar] [CrossRef]
  7. Haydari, Z.; Kavehnia, F.; Askari, M.; Ganbariyan, M. Time-series load modelling and load forecasting using neuro-fuzzy techniques. In Proceedings of the 9th International Conference on Electrical Power Quality and Utilisation, Bushehr, Iran, 9–11 October 2007.
  8. Guo, J.J.; Wu, J.Y.; Wang, R.Z. A new approach to energy consumption prediction of domestic heat pump water heater based on grey system theory. Energy Build. 2011, 43, 1273–1279. [Google Scholar] [CrossRef]
  9. Zhang, F.; Lee, S.E.; Yang, J.J.; Shah, K.W. Time series forecasting for building energy consumption using weighted Support Vector Regression with differential evolution optimization technique. Energy Build. 2016, 126, 94–103. [Google Scholar] [CrossRef]
  10. Doucoure, B.; Agbossou, K.; Cardenas, A. Time series prediction using artificial wavelet neural network and multi-resolution analysis: Application to wind speed data. Renew. Energy 2016, 92, 202–211. [Google Scholar] [CrossRef]
  11. Macas, M.; Moretti, F.; Fonti, A.; Giantomassi, A.; Comodi, G.; Annunziato, M.; Pizzuti, S.; Capra, C.A. The role of data sample size and dimensionality in neural network based forecasting of building heating related variables. Energy Build. 2016, 111, 299–310. [Google Scholar] [CrossRef]
  12. Ding, N.; Benoit, C.; Foggia, G. Neural network-based model design for short-term load forecast in distribution systems. IEEE Trans. Power Syst. 2016, 31, 72–81. [Google Scholar] [CrossRef]
  13. Kiartzis, S.J.; Bakirtzis, A.G.; Petridis, V. Short-term forecasting using NNs. Electric Power Syst. Res. 1995, 33, 1–6. [Google Scholar] [CrossRef]
  14. Panapakidis, I.P.; Dagoumas, A.S. Day-ahead electricity price forecasting via the application of artificial neural network based models. Appl. Energy 2016, 172, 132–151. [Google Scholar] [CrossRef]
  15. Chae, Y.T. Artificial neural network model for forecasting sub-hourly electricity usage in commercial buildings. Energy Build. 2016, 111, 184–194. [Google Scholar] [CrossRef]
  16. Neto, A.H.; Fiorelli, F.A.S. Comparison between detailed model simulation and artificial neural network for forecasting building energy consumption. Energy build. 2008, 40, 2169–2176. [Google Scholar] [CrossRef]
  17. Aydinalp-Koksal, M.; Ugursal, V.I. Comparison of neural network, conditional demand analysis, and engineering approaches for modeling end-use energy consumption in the residential sector. Appl. Energy 2008, 85, 271–296. [Google Scholar] [CrossRef]
  18. Belshaw, M.; Taati, B.; Snoek, J.; Mihailidis, A. Towards a single sensor passive solution for automated fall detection. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Toronto, ON, Canada, 30 August–3 September 2011.
  19. Mehr, H.D.; Polat, H.; Cetin, A. Resident activity recognition in smart homes by using artificial neural networks. In Proceedings of 2016 4th International Istanbul Smart Grid Congress and Fair (ICSG), Ankara, Turkey, 20–21 April 2016.
  20. Karatasou, S.; Santamouris, M.; Geros, V. Modeling and predicting building's energy use with artificial neural networks: Methods and results. Energy Build. 2006, 38, 949–958. [Google Scholar] [CrossRef]
  21. Deb, C. Forecasting diurnal cooling energy load for institutional buildings using Artificial Neural Networks. Energy Build. 2016, 121, 284–297. [Google Scholar] [CrossRef]
  22. González, P.A.; Zamarreno, J.M. Prediction of hourly energy consumption in buildings based on a feedback artificial neural network. Energy Build. 2005, 37, 595–601. [Google Scholar] [CrossRef]
  23. Camara, A.; Feixing, W.; Xiuqin, L. Energy Consumption Forecasting Using Seasonal ARIMA with Artificial Neural Networks Models. Int. J. Business Manag. 2016, 11. [Google Scholar] [CrossRef]
  24. Cao, Q.; Ewing, B.T.; Thompson, M.A. Forecasting wind speed with recurrent neural networks. Eur. J. Oper. Res. 2012, 221, 148–154. [Google Scholar] [CrossRef]
  25. Mohanty, S.; Patra, P.K.; Sahoo, S.S. Prediction of global solar radiation using nonlinear auto regressive network with exogenous inputs (narx). In Proceedings of 2015 39th National Systems Conference (NSC), Odisha, India, 14–16 December 2015.
  26. Nyanteh, Y.D.; Srivastava, S.K.; Edrington, C.S.; Cartes, D.A. Application of artificial intelligence to stator winding fault diagnosis in Permanent Magnet Synchronous Machines. Electric Power Syst. Res. 2013, 103, 201–213. [Google Scholar] [CrossRef]
  27. López, M. Application of SOM neural networks to short-term load forecasting: The Spanish electricity market case study. Electric Power Syst. Res. 2012, 91, 18–27. [Google Scholar] [CrossRef]
  28. Ibrahim, M.; Jemei, S.; Wimmer, G.; Hissel, D. Nonlinear autoregressive neural network in an energy management strategy for battery/ultra-capacitor hybrid electrical vehicles. Electric Power Syst. Res. 2016, 136, 262–269. [Google Scholar] [CrossRef]
  29. Alwakeel, M.; Shaaban, Z. Face Recognition based on haar wavelet transform and principal component analysis via Levenberg-Marquardt backpropagation neural network. Eur. J. Sci. Res. 2010, 42, 25–31. [Google Scholar]
  30. Marquardt, D.W. An algorithm for least-squares estimation of nonlinear parameters. J. Society Ind. Appl. Math. 1963, 11, 431–441. [Google Scholar] [CrossRef]
  31. Hagan, M.T.; Demuth, H.B.; Beale, M.H. Neural Network Design; PWS Pub. Co.: Boston, MA, USA, 1996; p. 3632. [Google Scholar]
  32. Hagan, M.T.; Menhaj, M.B. Training feed forward networks with the Marquardt algorithm. IEEE Trans. Neural Netw. 1994, 5, 989–993. [Google Scholar] [CrossRef] [PubMed]
  33. Hong, T.; Pinson, P.; Fan, S.; Zareipour, H.; Troccoli, A.; Hyndman, R.J. Probabilistic energy forecasting: Global energy forecasting competition 2014 and beyond. Int. J. Forecast. 2016, 32, 896–913. [Google Scholar] [CrossRef] [Green Version]
  34. Lin, T.; Horne, B.; Tino, P.; Giles, C.L. Learning long-term dependencies in NARX recurrent neural networks. IEEE Trans. Neural Netw. 1996, 7, 1329–1338. [Google Scholar] [PubMed]
  35. Safavieh, E.; Andalib, S.; Andalib, A. Forecasting the unknown dynamics in NN3 database using a nonlinear autoregressive recurrent neural network. In Proceedings of the IEEE International Joint Conference on Neural Networks, Tehran, Iran, 12–17 August 2007.
  36. Alvarez, F.M. Energy time series forecasting based on pattern sequence similarity. IEEE Trans. Knowl. Data Eng. 2011, 23, 1230–1243. [Google Scholar] [CrossRef]
  37. Smith, W.S. Moving average filters. In The Scientist and Engineer’s Guide to Digital Signal Processing; California Technical Publishing: San Diego, CA, USA, 1999; pp. 277–284. [Google Scholar]
  38. Vogl, T.P.; Mangis, J.K.; Rigler, A.K.; Zink, W.T.; Alkon, D.L. Accelerating the convergence of the back-propagation method. Biol. Cybern. 1988, 59, 257–263. [Google Scholar] [CrossRef]
Figure 1. Non-linear autoregressive (NAR) network.
Figure 1. Non-linear autoregressive (NAR) network.
Energies 09 00684 g001
Figure 2. Non-linear autoregressive with exogenous (External) inputs (NARX).
Figure 2. Non-linear autoregressive with exogenous (External) inputs (NARX).
Energies 09 00684 g002
Figure 3. Flowchart of the energy time-series model.
Figure 3. Flowchart of the energy time-series model.
Energies 09 00684 g003
Figure 4. Sample building architecture.
Figure 4. Sample building architecture.
Energies 09 00684 g004
Figure 5. NAE55 Network Automation Engine.
Figure 5. NAE55 Network Automation Engine.
Energies 09 00684 g005
Figure 6. Lighting state of a sample floor.
Figure 6. Lighting state of a sample floor.
Energies 09 00684 g006
Figure 7. Sample of control ventilation system.
Figure 7. Sample of control ventilation system.
Energies 09 00684 g007
Figure 8. Temperature and consumption time series over time (normalized).
Figure 8. Temperature and consumption time series over time (normalized).
Energies 09 00684 g008
Figure 9. Example of a building’s consumption for every day of the week. The first graph (a) shows the registered consumption (kW) of each working day for a two-year period. The second graph (b) shows the consumption (kW) during the weekend days. The third graph (c) shows a comparison between the average consumption (kW) on working and weekend days, over the same two-year period.
Figure 9. Example of a building’s consumption for every day of the week. The first graph (a) shows the registered consumption (kW) of each working day for a two-year period. The second graph (b) shows the consumption (kW) during the weekend days. The third graph (c) shows a comparison between the average consumption (kW) on working and weekend days, over the same two-year period.
Energies 09 00684 g009
Figure 10. Example of a building average consumption with different behavior between working days and non-working days.
Figure 10. Example of a building average consumption with different behavior between working days and non-working days.
Energies 09 00684 g010
Figure 11. Best prediction (a) and worst prediction (b) of the next 100 days for the normalized data of Building 5 using the NAR model.
Figure 11. Best prediction (a) and worst prediction (b) of the next 100 days for the normalized data of Building 5 using the NAR model.
Energies 09 00684 g011
Figure 12. Best prediction (a) and worst prediction (b) of the next 100 days for the normalized data of energy consumption of Building 5 using NARX model.
Figure 12. Best prediction (a) and worst prediction (b) of the next 100 days for the normalized data of energy consumption of Building 5 using NARX model.
Energies 09 00684 g012aEnergies 09 00684 g012b
Figure 13. Comparison of average MSE in respect to the delay parameter (a) and the network complexity—number of neurons (b).
Figure 13. Comparison of average MSE in respect to the delay parameter (a) and the network complexity—number of neurons (b).
Energies 09 00684 g013
Table 1. MSE of the delay parameter obtained with the NAR model.
Table 1. MSE of the delay parameter obtained with the NAR model.
ConsumptionDelays
135791214161821
Building 15.802.502.282.132.062.022.012.032.222.05Worst
Building 23.822.281.801.531.531.421.521.611.491.57Best
Building 34.802.242.091.831.531.301.661.511.691.76
Building 44.872.682.161.781.951.851.721.751.821.83
Building 51.701.211.170.850.710.780.740.790.750.73
Building 62.951.751.080.800.620.810.800.780.850.88
Building 75.413.121.871.671.581.351.511.661.621.67
Building 83.291.971.591.080.941.061.081.091.031.02
Mean4.082.221.761.461.371.321.381.401.431.44
Bold values: the best delay; crossed-out values: the worst error.
Table 2. MSE of the neurons parameter obtained with the NAR model.
Table 2. MSE of the neurons parameter obtained with the NAR model.
ConsumptionNeurons
2468101214161820
Building 12.271.831.821.872.102.202.132.082.142.15Worst
Building 21.631.781.481.541.641.441.621.611.941.67Best
Building 31.811.691.541.671.631.681.591.581.911.64
Building 41.781.901.751.731.701.901.851.771.931.87
Building 50.720.700.700.670.740.690.750.720.760.79
Building 60.750.780.850.770.740.640.760.800.900.87
Building 71.821.491.701.891.621.361.731.581.711.52
Building 81.201.100.960.960.931.000.951.020.941.03
Mean1.491.411.351.391.391.361.421.401.531.44
Bold values: the best delay; crossed-out values: the worst error.
Table 3. MSE of the delay parameter obtained with the NARX model.
Table 3. MSE of the delay parameter obtained with the NARX model.
ConsumptionDelay MSE
135791214161821
Building 11.832.121.631.501.521.822.142.061.842.07Worst
Building 21.771.691.451.391.531.481.441.401.581.59Best
Building 31.591.471.511.351.581.461.551.901.661.76
Building 41.691.581.581.711.711.831.801.991.942.16
Building 50.910.790.760.790.830.870.900.930.951.01
Building 61.080.830.771.031.121.161.131.241.441.31
Building 71.331.441.171.271.361.351.711.541.571.65
Building 81.031.091.050.990.991.061.301.141.341.34
Mean1.401.381.241.261.331.381.501.521.541.61
Bold values: the best delay; crossed-out values: the worst error.
Table 4. MSE of the neurons parameter obtained with the NARX model.
Table 4. MSE of the neurons parameter obtained with the NARX model.
ConsumptionNeurons MSE
2468101214161820
Building 11.701.331.661.881.701.731.701.591.691.86Worst
Building 21.311.471.501.441.651.741.771.541.661.78Best
Building 31.241.451.211.271.281.521.341.371.491.64
Building 41.451.741.631.701.851.771.401.621.711.80
Building 50.710.820.740.860.850.820.780.820.820.90
Building 60.790.870.840.851.080.941.011.031.001.10
Building 71.171.391.271.351.311.311.241.671.591.30
Building 80.881.051.031.011.011.061.081.111.121.01
Mean1.161.271.241.301.341.361.291.341.391.42
Bold values: the best delay; crossed-out values: the worst error.
Table 5. Summary of NAR and NARX network performance (MSE) in all buildings.
Table 5. Summary of NAR and NARX network performance (MSE) in all buildings.
ConsumptionNARNARX
Building 11.821.33
Building 21.441.31
Building 31.541.21
Building 41.701.40
Building 50.670.71
Building 60.640.79
Building 71.361.17
Building 80.930.88
Mean1,261,10

Share and Cite

MDPI and ACS Style

Ruiz, L.G.B.; Cuéllar, M.P.; Calvo-Flores, M.D.; Jiménez, M.D.C.P. An Application of Non-Linear Autoregressive Neural Networks to Predict Energy Consumption in Public Buildings. Energies 2016, 9, 684. https://doi.org/10.3390/en9090684

AMA Style

Ruiz LGB, Cuéllar MP, Calvo-Flores MD, Jiménez MDCP. An Application of Non-Linear Autoregressive Neural Networks to Predict Energy Consumption in Public Buildings. Energies. 2016; 9(9):684. https://doi.org/10.3390/en9090684

Chicago/Turabian Style

Ruiz, Luis Gonzaga Baca, Manuel Pegalajar Cuéllar, Miguel Delgado Calvo-Flores, and María Del Carmen Pegalajar Jiménez. 2016. "An Application of Non-Linear Autoregressive Neural Networks to Predict Energy Consumption in Public Buildings" Energies 9, no. 9: 684. https://doi.org/10.3390/en9090684

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop