Next Article in Journal
An Effective Ensemble Convolutional Learning Model with Fine-Tuning for Medicinal Plant Leaf Identification
Next Article in Special Issue
Reducing the Power Consumption of Edge Devices Supporting Ambient Intelligence Applications
Previous Article in Journal
Hierarchical Keyword Generation Method for Low-Resource Social Media Text
Previous Article in Special Issue
Advancements in On-Device Deep Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Network-Based Solar Irradiance Forecast for Edge Computing Devices

by
Georgios Venitourakis
1,†,
Christoforos Vasilakis
1,†,
Alexandros Tsagkaropoulos
1,†,
Tzouma Amrou
2,†,
Georgios Konstantoulakis
2,†,
Panagiotis Golemis
2,† and
Dionysios Reisis
1,*,†
1
Electronics Lab, Physics Department, National & Kapodistrian University of Athens, 15772 Athens, Greece
2
Inaccess Networks, 12 Sorou Str., 15125 Maroussi, Greece
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Information 2023, 14(11), 617; https://doi.org/10.3390/info14110617
Submission received: 18 October 2023 / Revised: 9 November 2023 / Accepted: 15 November 2023 / Published: 18 November 2023
(This article belongs to the Special Issue Artificial Intelligence on the Edge)

Abstract

:
Aiming at effectively improving photovoltaic (PV) park operation and the stability of the electricity grid, the current paper addresses the design and development of a novel system achieving the short-term irradiance forecasting for the PV park area, which is the key factor for controlling the variations in the PV power production. First, it introduces the Xception long short-term memory (XceptionLSTM) cell tailored for recurrent neural networks (RNN). Second, it presents the novel irradiance forecasting model that consists of a sequence-to-sequence image regression NNs in the form of a spatio-temporal encoder–decoder including Xception layers in the spatial encoder, the novel XceptionLSTM in the temporal encoder and decoder and a multilayer perceptron in the spatial decoder. The proposed model achieves a forecast skill of 16.57% for a horizon of 5 min when compared to the persistence model. Moreover, the proposed model is designed for execution on edge computing devices and the real-time application of the inference on the Raspberry Pi 4 Model B 8 GB and the Raspberry Pi Zero 2W validates the results.

1. Introduction

Artificial intelligence (AI) is an ever-expanding technology that has spread in unconventional scientific and industrial fields and has been integrated in smart systems [1,2,3] in order to execute notorious tasks such as predicting the future state of a system and decision making. These tasks often appear in the Smart Grid (SG) concept [3,4], where power grids are supported by the information provided by Internet of Things (IoT) devices, which are constantly monitoring the environment and the interaction between the energy provider and the client. SGs are essential to power production involving renewable energy sources (RES) because the RES and especially the photovoltaic (PV) parks have the disadvantage of not producing energy at a constant rate.
In PV parks, the energy production depends heavily on the global horizontal irradiance (GHI), the diffuse horizontal irradiance (DHI), the direct normal irradiance (DNI), the cloud cover (CC) and other meteorological parameters. AI-enabled smart PV parks utilize machine learning (ML) tools to forecast the future values of these parameters [5]. These forecasting results lead the PV park controller to improve the power production capabilities and henceforth the grid balancing [6]. The problem in many PV parks is the lack of historical data that prevents a neural network (NN) from making reliable predictions. It is also the case that numerical values are often not sufficient to depict the current state of a system such as the weather conditions in the atmosphere. For this reason, researchers and engineers considered sky images taken from the ground as an attractive solution. This is because sky images carry significantly more information compared to numerical data, and moreover, they are able to provide even further detailed information by the use of advanced ML techniques and convolutional NN (CNN)-based models. In the last decade, the approaches with image-regression-based techniques have shown improved results and these advances have made edge computing applications possible by the use of lightweight models such as ShuffleNet [7] and the MobileNet [8].
The motivation of this research is the Archon project [9] that designs and develops a controller that is efficient with respect to the implementation cost, the computational load and the deployment, which will manage the PV park power production [9]. The controller refers to all the aspects of the controlling mechanism, ranging from the underlying infrastructure to the controller’s software. The infrastructure includes the sensors that generate data and the hardware solutions for deploying the irradiance forecasting system (IFS) on the edge. The cost-efficient controller design excludes sensors that generate numeric data in order to reduce the cost of the equipment and the complexity of the interconnection of all the contributing devices in the system. For the same purpose, the design excluded the pyranometer, a device that provides essential information regarding current weather conditions, because it is in high demand and expensive. Therefore, the entire design of the proposed IFS depends solely on image sequences captured by low-cost cameras. Moreover, the implementation targets edge computing devices that require less energy and are of lesser cost compared to a work station.
Aiming at an improved solution to the GHI forecasting problem executable on edge devices, the current article exploits image sequence regression techniques to introduce the following two novel entities: (a) the Xception long short-term memory (XceptionLSTM), a recurrent NN (RNN) for image sequence parsing and generation, and (b) a model for a complete solution to the GHI forecasting that uses the proposed XceptionLSTM to improve the forecasting results. The design of the proposed XceptionLSTM cell is based on convolutional LSTM (ConvLSTM) cells. The proposed GHI forecasting model is a sequence-to-sequence (Seq2Seq) image regression NN in the form of a spatio-temporal encoder–decoder [10] that consists of Xception layers (XL) [11] in the spatial encoder, the novel XceptionLSTM cells proposed by this work in the temporal encoder and decoder and a multilayer perceptron (MLP) in the spatial decoder. The novel XceptionLSTM has the following advantages compared to the ConvLSTMs: (a) its design allows the parallelized execution of ConvLSTM cells with different kernel sizes, (b) it is significantly more lightweight and (c) it showcases significantly improved usage of the data and kernel tensors. The novel GHI forecasting model has the following improvements with respect to the reported Seq2Seq architectures based on ConvLSTMs: (a) the proposed model converges faster and (b) it requires less memory in order to infer data, making it ideal for executing inference on-the edge devices.
The development and evaluation of the proposed model employs a custom dataset of red green blue (RGB) 180 field-of-view sky images and GHI measurements collected over a full callendar year period during the development of the the Archon project [9]. The proposed model is trained and evaluated on an NVIDIA GeForce RTX 3080. The design of the entire model has as target the execution on edge computing devices. For the time performance evaluation we opted the Raspberry Pi 4 Model B 8 GB and the Raspberry Pi Zero 2W as low-power, edge computing devices. The development of the model was based on Pytorch [12], which is a Python package and a framework for NNs with a relatively high Graphics Processing Unit (GPU) acceleration. Pvlib [13] is a Python package for PV park performance simulation; in this work, it was the basis for estimating the position of the sun in an image and for the generation of sun masks [14].
The paper is organised as follows. First, Section 2 introduces the XceptionLSTM cell and the model for the short-term irradiance forecasting. Section 3 reports the evaluation results of CNN models for irradiance forecasting. Section 4 follows with a discussion regarding the results presented in this and other related work reported in the literature. Finally, Section 5 concludes the article.

2. Materials and Methods

Making decisions proactively for a system that we monitor relies on the following assumption: the system’s next state depends on the sequence of states in the recent history of that system. Accordingly, we need the accurate forecasts for the forthcoming GHI state in order to improve the PV park’s operation. In order to accomplish the latter task, a sky camera captures a sequence of images consecutively with a constant time interval (referred to as horizon from now on). Then, the PV park’s controller forwards the images to an NN. The prediction model calculates a new sequence of values that corresponds to consecutive GHI values with the same horizon as the input sequence. The input and output sequence length and the horizon used for the prediction are the major model’s hyperparameters that can be tuned to achieve the most accurate possible outcome. Other hyperparameters are the model’s structure, the training schemes and any data preprocessing.

2.1. Model Structure

Most often, NNs need to be quite complex because of the large systems they are tasked to simulate. Therefore, in highly complex systems such as the Earth’s atmosphere, the traditional models that can respond to such fast-changing parameters may consist of dozens of sequentially connected layers. Such models are most probably time consuming in the tasks of training and inference and they are considered non-optimal for time-critical applications and on-the-edge inference. These facts show that such systems require efficient state-of-the-art ML algorithms and novel techniques with improved complexity and lesser requirements for computational resources.

2.1.1. Xception Layer

The proposed prediction model utilizes XLs [11], which are a type of CNN that combines the characteristics of Inception Modules [15] and depthwise separable convolutions (DWSC) [11,16]. A DWSC extracts the parallelism of a traditional convolutional layer (CL) by partitioning the operation in two simpler operations: a depthwise convolution and a pointwise convolution. The former is a convolution in each frame of the channels of the input tensor, while the latter is a convolution in each pixel of the input tensor. Combining the depthwise and pointwise convolution sequentially results in a CL with the same result-producing capabilities but with a significantly reduced number of parameters and computational complexity. The computational graph of a DWSC is depicted in Figure 1. An Inception Module consists of a nested CL, where all nested layers process the same input in parallel, and all the results are concatenated, added or in general reduced to a new output tensor. The layers’ parallelization leads to a greater degree of data usage, making it an ideal layer for inference on low-power edge devices. The benefit of employing Inception Modules is also the attenuation of the vanishing gradient problem [17]. This problem refers to those models that have a great number of sequentially connected layers, and during the training scheme in these models, the gradient often becomes insignificant during backpropagation.
The combination of DWSC and Inception Modules results in the XLs, which execute in parallel depthwise operations such as the depthwise convolutions or the pooling operations. Then, the model concatenates these results and forwards them to a pointwise convolution. The latter scheme exploits two forms of parallelism: the inter-task parallelism (parallel execution of nested layers in the Inception Module) and intra-task parallelism (parallel execution of convolutions in all channels of the input tensor in a depthwise convolution). Although the CL allows for an intra-task parallelism because of the different kernels that can be applied independently in parallel (called inter-kernel parallelism), the proposed model’s design can achieve a greater degree of intra-task parallelism with an XL and the depthwise convolutions; the latter allow the partition of the input tensor into the channels of the tensor. Hence, we can operate on each of these channels independently. The advantage of the XL compared to the traditional CL is less overhead in terms of memory access and smaller number of parameters and operations. We can add to the above a further improvement. This is the partitioning of the depthwise convolution in two asymmetrical depthwise convolutions, one for each dimension of the frame, meaning convolutions with kernel size 1 × N and N × 1 [18], where N × N is the kernel size of the traditional CL. This fact allows even further parallelism of the computations.
This work considers the structure of the XL shown in Figure 2. It consists of four nested layers, two of which are depthwise convolutions with kernel sizes of 3 and 5, a max pooling layer and the identity function. The identity function is ultimately used as a pointwise convolution of the XL’s input tensor. Also, including the identity function improves the gradient descent; the gradient is broadcasted to and propagates through all four nested layers, but is unaffected by the identity function. The gradients of all the nested layers are then added and (back)propagated to the previous layer. As a result, the gradient of an XL is mostly affected by the pointwise convolution and less affected by the nested depthwise convolutions (they can be viewed as small adjustments in the output gradient). Thus, the layers in the later stages of the backpropagation are less likely to experience the vanishing gradient descent problem [15]. We can optionally add an activation function between the depthwise and pointwise operations. However, early results have shown that an activation before the pointwise convolution diminishes the accuracy of the results.

2.1.2. XceptionLSTM

ConvLSTM cells [19] are RNNs that utilize convolutions and operate on tensors, in contrast to fully-connected LSTM (FC-LSTM) cells [20] that operate on vectors. The use of LSTM cells is either to process or generate a sequence of data; their most usual applications are the deep learning (DL) based natural language processing (NLP) applications [21] and time-series predictions [22,23]. ConvLSTM cells are able to process more complex tasks, such as next frame prediction [24] and other time-series predictions with feature extraction [25]. The proposed model utilises XLs [11] as a substitute for convolutions in ConvLSTM. XLs used in the proposed model’s LSTM cells are structured as shown in Figure 3. Equations (1)–(10) and Figure 3 describe XceptionLSTM cells:
x = Concat ( X t , H t 1 ) ,
i c = XL i x ,
f c = XL f x ,
c c = XL c x ,
o c = XL o x ,
i g = σ i c + C t 1 W h i ,
f g = σ f c + C t 1 W h f ,
o g = σ o c + C t 1 W h o ,
C t = f g C t 1 + i g act c c ,
H t = o g act C t ,
where the new input and the hidden state of the previous iteration of the cell are concatenated (Equation (1)) and forwarded to four XLs (input, forget, cell and output convolutions, Equations (2)–(5)). The input, forget and output gates (Equations (6)–(8)) are calculated by first summing the results of the corresponding convolutions with the Hadamard products of the cell state of the previous iteration ( C t 1 ) with their corresponding parameters and then applying the sigmoid function. The new cell state ( C t ) is a combination of the previous cell state, the input and forget gate and the cell convolution (Equation (9)). The new hidden state ( H t ) is the Hadamard product of the output state with the activated new cell state (Equation (10)). We note here that the XL operations in the XceptionLSTM cell do not have a nested max pooling operation.
The proposed RNN has significant advantages compared to ConvLSTMs:
  • Parallelized execution of multiple ConvLSTM cells with different kernel sizes in a single XceptionLSTM cell.
  • Significantly more lightweight when compared to the ConvLSTMs that have similar structural elements.
  • Improved utilization of the data and kernel tensors: k times less input data calls in the depthwise convolution and w 2 times less kernel calls in the pointwise convolution when compared to traditional convolutions, where k is the number of kernels and w is the window size of the tradithonal CL.

2.1.3. Proposed Model

The proposed model is a spatio-temporal encoder/decoder (Figure 4) with an Xception-based spatial encoder, an XceptionLSTM temporal encoder and decoder and an MLP as the spatial decoder. The structures of the spatial encoder and the decoder are depicted in Figure 5 and Figure 6, respectively.
The spatial encoder contains 6 layers. The first two layers are the two input XLs that extract data from the input image. Then, the two middle residual layers refine the data. Each of these two residual layers consists of a nested sequential module of two XLs. Each residual layer’s output is the sum of the nested module’s output and its input. Note here that the number of output tensor channels of the second residual layer is double compared to its input tensor channels. To match the above in the second residual layer, the sum is calculated with the result of a 1 × 1 convolution of the input in order to match the number of channels in all nested layers. The nested module can be interpreted as an input corrector that refines the input tensor’s data. The last two XLs compress the data to an encoded state. The spatial encoder finally normalizes the encoded state. The spatial decoder involves three linear layers, which are represented by the three last layers in Figure 6. The first three layers in Figure 6 reduce the encoded state produced by the temporal decoder to a fixed sized tensor with adaptive average pooling and flatten it to a vector. The third stage normalizes the vector and forwards it to the 3-layer MLP.
All layers are followed by leaky rectified linear unit (LeakyReLU) with a negative slope of 0.1125. ReLU is the output activation function in the inference, but the training scheme uses LeakyReLU with a near-zero negative slope ( 10 3 ) to allow backpropagation when a negative value is produced in the early stages of training. The proposed model can accept any image of frame size of at least 64 × 64 pixels. This work uses frames of size 128 × 128 . We note here that during the development stages of the proposed model, we studied a variety of models; all these models yielded the best metrics and results with the 128 × 128 size frames.

2.2. Dataset

The current study has conducted research for datasets appropriate for the model. During the early stages, we considered the Folsom, CA, dataset [26], which was based on a camera that did not provide constant orientation over the time that the dataset was produced; therefore, there was no method to locate the sun systematically. Other datasets include the WSISEG database [27], WILLIAM Meteo Database [28], SKIPP’D [29] and SRRL BMS [30], which are limited with respect to the number of images compared to our needs for training the proposed model and for the target forecasting horizon. Hence, we proceeded by organizing and developing the Archon–Athens, Greece Dataset, a custom dataset for the purposes of the project Archon [9]. The dataset consists of approximately 250 thousand sky images from the All Sky Imager (ASI-16), an automatic full-sky camera system with fisheye lens for a 180 field of view, and GHI measurements of 1 min intervals. The instruments are placed in the rooftop of the Inaccess office (38.04 N, 23.81 E, Sorou Str, Athens, GR) and have gathered data since 25 October 2022. The captured images depict the various weather conditions that occur in the greek capital. Most of the images show partly cloudy or clear sky conditions. There are also plenty of days with precipitation and thin cloud ceilings. Also, the halo phenomenon appears often in images from morning and evening hours. Samples of the dataset are provided in Figure 7.
We keep all images from January, April and July 2023 as evaluation data and use the rest of the dataset for training and validation with a 90% split. January represents a set of days with bad power yields and frequent weather changes, while April and July represent days with good yields with frequent and infrequent weather changes, respectively. To achieve a more objective validation loss, the training and validation sets are created by splitting the dataset based on the days in a month and then extracting the valid input image and target irradiance sequences from each subset for all months. This guarantees that the model has not processed an image during both training and validation and that the two subsets cover the weather conditions from all the available months. Overall, we used 67% of the dataset for training, 8% for validation and fine-tuning and 25% for the evaluation presented in Section 3.

2.2.1. Input Images

The proposed model accepts as input a sequence of consecutive images captured one frame per minute. All images are first resized from 1536 × 1536 RGB images of 8-bit depth to 128 × 128 × 3 tensors of single floats. An extra channel called sun mask [14] highlights the sun disk in an image at clear sky conditions. The introduction of this binary mask is one way to provide the models with useful data related to the image. Such data include the solar azimuth and the solar elevation, and by using this binary mask, they can be easily correlated with the image. It also hints that the highlighted areas are the region in the image expected to correspond to the sky fragments providing the larger fraction of GHI.

2.2.2. Output Irradiance

The proposed model outputs irradiance (GHI) values as single floats. These GHI target values are integers and range from one (1) to around 1460 W m 2 . We have discarded all possible sequences that have sky images captured before sunrise or after sunset.

2.3. Metrics

The literature provides a variety of metrics for evaluating solar forecasts, which are envisaged from different perspectives [31]. In this article, we evaluate the results of the tested models with mean bias error (MBE), mean absolute error (MAE), mean absolute percentage error (MAPE), root mean square error (RMSE), and forecast skill (FS) shown in Equations (11)–(15), respectively.
Mean bias error : MBE = 1 N i = 1 N y i ^ y i ,
Mean absolute error : MAE = 1 N i = 1 N y i ^ y i ,
Mean absolute percentage error : MAPE = 1 N i = 1 N y i ^ y i y i ,
Root mean square error : RMSE = 1 N i = 1 N y i ^ y i 2 ,
Forecast skill : FS = 1 RMSE RMSE pers ,
where N is the size of the test dataset, y i ^ is the forecast and y i is the target value for a horizon H. The MBE highlights whether a model shows bias when forecasting and hence, whether the results tend to consistently under- or overestimate the target value. The MAE and RMSE show the measured deviation of the results in respect to the target values. We can interpret the former as the expected deviation in the lower range of the GHI values, whereas the latter refers to the expected deviation in the upper range of the GHI values. The FS provides a more dataset-independent way to evaluate models [32]. This is accomplished by comparing the models to the Persistence Model, which forecasts that no change will occur to the target value after a horizon. The Persistence Model is a baseline model that often appears in short- and ultra short-term irradiance forecasting solutions, where the forecast horizon ranges from 15 s to 2 min. The MAPE metric indicates the normalized deviation of the forecasts from the target values, which also helps in assessing the models’ performance more comprehensively.

3. Results

This section presents the evaluation results of the proposed model for the task of short-term irradiance forecasting. Moreover, it presents the results of the study on the performance of models with various temporal encoders/decoders and it compares their results to that of the proposed model. The comparison models include ConvLSTMs, stacked ConvLSTMs, bidirectional ConvLSTMs and their respective depthwise separable (DWSConvLSTM) and Xception versions. The presented benchmark also compares the temporal encoders and decoders that are based on convolutional gated recurrent Units (ConvGRU) [33], an RNN initially intended for spatio-temporal feature learning from videos. We note here that bidirectionality only applies to the temporal encoder, as it is the only module that accepts a sequence as an input, and the results of the forward and backward pass of the input sequence are summed and forwarded to the temporal decoder. All layers of the tested temporal models accept and generate tensors of size 8 × 8 × 128 . Table 1 is an overview of the models evaluated in this work.
All the models are trained and evaluated in a Linux workstation with an Intel(R) Core(TM) i7-9700K CPU @ 3.60   G Hz and a NVIDIA GeForce RTX 3080 GPU. We deploy a Raspberry Pi 4 Model B 8 GB and a Raspberry Pi Zero 2W for time performance tests as devices on the edge, configured as a Linux workstation with a quad core Cortex-A72 (ARM v8) 64-bit SoC @ 1.5   G Hz for the former and as a Linux workstation with a quad-core Arm Cortex-A53 64-bit SoC @ 1 G Hz for the latter device. We use Python 3.9.13 and Pytorch 2.0.0+cuda11.7 for the development of the evaluated models.

3.1. Training Scheme

In order to reduce the total training time of all models that are evaluated in this article, we used transfer learning and partitioned the models’ training in two stages, as shown in Figure 8. In the first stage, we trained the spatial encoder and decoder in the training dataset for the problem of irradiance estimation. Specifically, the spatial model accepts an image and estimates the GHI value for this particular image. This stage is common to all the models we tested; therefore, the spatial encoder and decoder were trained only once. We train the second stage’s spatio-temporal model using as initial weights: (a) for the spatial encoder and decoder those resulting of the first stage and (b) for the temporal encoder and decoder arbitrary weights. This scheme allows us to test whether the spatial encoder and decoder can effectively forecast GHI values.
Table 2 lists all the hyperparameters that we examined and chose for the presented benchmark. For the first stage, we used RMSProp with decay of 0.9 and ϵ = 1.0 . We used a learning rate of 0.001 decaying every epoch using an exponential rate of 0.94 and the MSELoss as the criterion for calculating the loss. The spatial model was trained for 29 epochs and achieved an RMSE of 35.3 W m 2 for the solar irradiance estimation. For the second stage, we used the same scheduler, optimizer and loss function with an initial learning rate of 5 × 10 5 . With this training scheme we were able to reduce the total epochs from 30 to just 6 epochs for the spatio-temporal models. Moreover, all the models achieved better metrics when they were trained with this scheme compared to training the whole spatio-temporal model without any transfer learning.

3.2. Model Evaluation

Table 1 presents the implementation details of the models evaluated in this work: the ConvLSTM, the DWSConvLSTM and the XceptionLSTM-based models in their single-cell, bidirectional and double-stacked versions and the ConvGRU-based models. Table 3 presents the evaluation results of the metrics for the tested models. All the models considered for this comparison infer a sequence of five images with a horizon of 1 min and output 15 GHI values that correspond from 1 to 15-min forecasts. Note here that the XceptionLSTM cell in the three versions we examined prevails with respect to the RMSE and FS scores. Specifically, the single XceptionLSTM cell scores 68.8 W m 2 , 94.6 W m 2 and 116 W m 2 RMSE for the horizons of 1, 5, 15 min with a mean score of 99.8 W m 2 and the double-stacked XceptionLSTM scores 69.2 W m 2 , 94.4 W m 2 and 115.8 W m 2 RMSE with a mean score of 99.8 W m 2 . Moreover, the proposed cell in its bidirectional form scores 68.7 W m 2 , 94.5 W m 2 and 117.3 W m 2 in the RMSE metric for the same horizons and reports a mean score of 100.3 W m 2 . The bidirectional XceptionLSTM is the best-performing model based on the MAPE metric, achieving a mean score of 24.5% across the examined horizons. Furthermore, the XceptionLSTM cell achieves low MAE scores for the 1 min horizon, but the stacked DWSConvLSTM cells report lower MAE metrics in all the other horizons.
The MBE metric reports no noticable bias for ConvLSTM and XceptionLSTM-based spatio-temporal models, which means that these two kinds of models have no tendency to over- or underestimate forecasts. On the other hand, most DWSConvLSTM-based models tend to systematically underestimate the forecasts of the first few horizons. The same tendency appears in the bidirectional versions of all models. In contrast to the above, the ConvGRU based models tend to overestimate forecasts. Moreover, the models with bidirectional temporal encoders appear to forecast more accurately for greater horizons when compared to their unidirectional counterparts, but less accurately when one more layer is added to the unidirectional models. As we can conclude by the metrics of Table 3, the increased accuracy of the models with stacked temporal encoders and decoders seems to be more intense in the cells that use depthwise operations. In addition, given that all the benchmarked models score RMSE and MAE values within 4.6 W m 2 from one another, we conclude that the structure of the temporal encoder and decoder does not significantly change the reported metric scores.

3.3. Timing Reports

Figure 9 is a diagram of the mean execution time of the evaluated models on the edge devices. The proposed model, in its three benchmarked versions, executes as a single cell in 2.69 s and 7.54 s, as a bidirectional cell in 2.91 s and 7.99 s and as a double-stacked cell 3.31 s and 8.78 s for the Raspberry Pi 4 Model B and the Raspberry Pi Zero 2W. We notice that the inference time of all models are within 10% of the slowest recorded time on both edge devices, which corresponds to the model that includes stacked XceptionLSTMs in its temporal encoder and decoder. This is because a major fraction of the complexity derives from the repeated execution of the spatial encoder and decoder, that is, once per element of the input and output sequences. The graph also shows the expected behavior for the models that are based on the same LSTM cells, a fact deducing that the model with less parameters executes faster. That behaviour is not true for cells of different structure; despite the great difference in the amount of parameters and total operations, XceptionLSTM cells appear to have comparable execution times with ConvLSTM cells due to the optimizations that the Pytorch library performs in convolutions. We believe that the reason behind this is that the operation of concatenating the channels of the results of all the nested depthwise operations in an XL to form a single tensor for the pointwise convolution to process causes the reported execution time overhead. One way to cope with this is to avoid concatenation by executing the pointwise operation first, then splitting the intermediate tensor and finally, forwarding the chunks to the nested depthwise operations. These modifications improve the reported execution times but they result in degraded metrics. During the evaluation of the inference times, both devices reported a constant power consumption measured at 5.1   W for the Raspberry Pi 4 Model B and 0.7   W for the Raspberry Pi Zero 2W.

4. Discussion

Researchers and engineers in the renewable energy field are keen for solutions to the short-term irradiance forecasting problem [31]. Especially in the last two decades, they focus on computer vision and ML-based systems, which often include image processing for satellite imagery [34,35,36,37]. Given that the satellite images cover a vast area of the Earth’s land, measuring the GHI in different areas of an image may provide significantly different values. The alternative is the ground-based imagery [38,39,40], which clearly depicts the current weather conditions in the area of interest with a notable application example the case of large PV parks. The PV park controllers use multiple ground-based sensors and they can yield more accurate results [41].
As Ziyabari et al. [10] suggest, researchers often consider spatio-temporal architectures as a solid base for their models because the dimensionality of the input data does not constrain considerably the final structure of their models. This holds whether the input data is multiple time-series of environmental measurements from multiple sensors that are spread in a wide area or, as in this article, a time-series of images from a single sky camera. ConvLSTM-based solutions are reported in image regression related techniques that target irradiance forecasting [42,43,44]. This is because CLs are effective in modelling the complex dynamics of the environmental variables, such as the cloud and the wind movement. Moreover, as the name suggests, ConvLSTMs can capture the long-term evolution of the irradiance values. More accurately, ConvLSTMs excel in modeling the long-term dependences of the target data and extract the correlation among the input data [40]. It is quite common for researchers to utilize image segmentation for cloud cover estimations as a means to enhance the results of ML-based forecasting models [45,46].
Zhang et al. [47] compare the results of MLP, CNN and LSTM models that are trained to predict PV power differences by using PV power data and sky images. They conclude that a hybrid model using both PV power data and images has a better-balanced performance across different types of weather conditions. Sun et al. [48] present the SUNSET, a deep CNN architecture that accepts an image sequence and other data produced by the PV park and it outputs PV power and clear sky index (CSI) predictions. The input image sequence is in the form of a single hyperspectral image. Ajith et al. [49] developed a multi-modal fusion network for ultra-short irradiance forecasting using infrared images and past irradiance data. They explain that infrared images of the sky can better capture the cloud dynamics in the small horizon of 15 s. Kumari et al. [40] discuss the advantages and drawbacks of using LSTMs, gated reccurrent units (GRU), CNNs, deep belief networks (DBN), RNNs and hybrid artificial neural networks (ANN) for solar irradiance forecasting. Basmile et al. [50] review and compare eight different AI models for horizons of a minute, an hour and for daily average forecasts of GHI, DHI and DNI values. Nie et al. [44] explore training tactics for heterogeneous datasets and how transfer learning contributes to reducing the training effort and improving the results of a model. Lyu et al. [51] use deep reinforcement learning (DRL) in order to dynamically change between optimal features of a model by recognising weather patterns. We note here that the FS reported in the bibliography ranges from 2.4 33.2 % for the lower range to 14.4 25.2 % for the upper range of the examined horizons [14,43,52,53].
The proposed model targets implementations for short-term irradiance forecasting on low-power devices. In this article, we presented a benchmarking of known ConvLSTM based spatio-temporal models [10,19,40] for the latter task with the evaluation of the models in terms of metric scores and execution times. The proposed XceptionLSTM cell and spatio-temporal model show notable performance for horizons over 10 min and improved forecasting skills for smaller horizons. We noticed that, when taking into account the evaluation results of other related works [43,54], the proposed model exhibit a lower drop of forecasting skill as the horizon increases. This means that our model constitutes a very attractive solution for short-term irradiance forecasting; moreover, it can be integrated in SG systems that are based on ultra short-term forecasts. Furthermore, the proposed RNN is significantly lightweight when compared to traditional RNNs and models from the bibliography [43,44], as it requires half of the memory that the weights of the ConvLSTM-based spatio-temporal models need.
Focusing on the edge devices, the proposed model is optimized for inference on low-power devices and can process up to 22.27 sequences per minute in the low power device Raspberry Pi 4 Model B and up to 7.81 sequences per minute in the ultra low-power device Raspberry Pi Zero 2W. Hence, less powerful IoT devices suffice for the execution of forecasting tasks that were once considered power intensive and computationally demanding. Considering that the number of PV parks added to the power grids constantly increases and that low-end devices are easy to maintain and cost-effective, the proposed model offers a very tempting alternative to the high performance and high cost controllers.
We note here that, on one hand, the real-time execution of the inference computations on the edge will impose limitations on the NN design and consequently on the number of respective NN layers. On the other hand, the models in the bibliography that consider mainly the improvement of feature extraction and not the execution efficiency consist of dozens of sequentially connected layers [44,52,53]. Compared to the later works, the proposed model has limited NN layers but it achieves competitive results and conforms to the specified execution times. An addition to the above cost and execution related considerations in the proposed NN is the exclusion of the injection of numeric data in the final NN layers. This injection [40,44,47,48,49,51,52] gives more accurate forecasts but increases the overall complexity. Finally, one of the key factors of this research is the response time of the PV park controller. Although the presented IFS is statically configured, the controller must be able to dynamically configure the IFS in order to adjust to the environment and satisfy the time constraints. The controller has to choose the optimal input and output sequence lengths and the horizon for improved PV park management. Therefore, the proposed solution has to be configurable and also maintain a certain level of accuracy across the allowed configurations.

5. Conclusions

The current paper introduced a novel irradiance forecasting model that is efficient with respect to the computational complexity and specifically designed for edge computing devices. It also introduced the basis for the design of the forecast model, which is an innovative RNN called XceptionLSTM. The advantages of the proposed model lie in its reduced computational complexity, the achievement of Forecast Skill of 16.57% for a horizon of 5 min when compared to the Persistence Model and finally, the execution time results on known edge computing devices. The results, which are accomplished on a Raspberry Pi 4 Model B 8 GB and on a Raspberry Pi Zero 2W, validate the real-time performance of the model. The future work will first include further optimizations of the XL that will allow for real-time deployment for under a minute horizons. The second very interesting target includes the combination of the XLs with other RNNs, as well as experiments on NN quantization for efficient mapping on very large-scale integration (VLSI) and edge-oriented field programmable gate arrays (FPGA).

Author Contributions

Conceptualization, G.V., C.V. and D.R.; methodology, G.V., C.V., A.T. and D.R.; software, G.V., C.V. and A.T.; validation, T.A. and P.G.; formal analysis, G.V. and D.R.; investigation, C.V. and G.K.; resources, D.R. and G.K.; writing—original draft preparation, G.V. and C.V.; writing—review and editing, G.V., C.V., A.T., T.A. and D.R.; supervision, G.K. and D.R.; project administration, G.K.; funding acquisition, G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been co-financed by the European Regional Development Fund of the European Union and Greek national funds through the Operational Program Competitiveness, Entrepreneurship and Innovation, under the call RESEARCH—CREATE—INNOVATE (project name “ARCHON” and project code: T2EDK-00864).

Data Availability Statement

The Archon–Athens, Greece Dataset targeting short-term irradiance forcasting is first introduced in the current work and will soon be available at the project’s site http://archonproject.eu/english.html (accessed on 29 August 2023).

Acknowledgments

The authors would like to thank Vasilios Kalekis for his work in the early models that led to this study.

Conflicts of Interest

Authors Tzouma Amrou, Georgios Konstantoulakis and Panagiotis Golemis are employed by the company Inaccess Networks. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial intelligence
ANNArtificial neural network
CCCloud cover
CLConvolutional layer
CNNConvolutional neural network
ConvLSTMConvolutional long short-term memory
ConvGRUConvolutional gated recurrent unit
CPUCentral processing unit
CSIClear sky index
DBNDeep belief network
DLDeep learning
DHIDiffuse horizontal irradiance
DNIDirect normal irradiance
DRLDeep reinforcement learning
DWSCDepthwise separable convolution
DWSConvLSTMDepthwise separable convolutional long short-term memory
FC-LSTMFully connected long short-term memory
FPGAField programmable gate array
FSForecast skill
GHIGlobal horizontal irradiance
GPUGraphics processing unit
GRUGated recurrent unit
IFSIrradiance forecasting system
IoTInternet of Things
LeakyReLULeaky rectified linear unit
LSTMLong short-term memory
MAEMean absolute error
MBEMean bias error
MLMachine learning
MLPMultilayer perceptron
NLPNatural language processing
NNNeural network
PVPhotovoltaic
ReLURectified linear unit
RESRenewable energy source
RGBRed green blue
RMSERoot mean square error
RNNRecurrent neural network
Seq2SeqSequence-to-sequence
SGSmart grid
SoCSystem-on-chip
VLSIVery large-scale integration
XceptionLSTM  Xception long short-term memory
XLXception layer

References

  1. Molinara, M.; Bria, A.; De Vito, S.; Marrocco, C. Artificial intelligence for distributed smart systems. Pattern Recognit. Lett. 2021, 142, 48–50. [Google Scholar] [CrossRef]
  2. Aguilar, J.; Garces-Jimenez, A.; R-Moreno, M.; García, R. A systematic literature review on the use of artificial intelligence in energy self-management in smart buildings. Renew. Sustain. Energy Rev. 2021, 151, 111530. [Google Scholar] [CrossRef]
  3. Alahi, M.E.E.; Sukkuea, A.; Tina, F.W.; Nag, A.; Kurdthongmee, W.; Suwannarat, K.; Mukhopadhyay, S.C. Integration of IoT-Enabled Technologies and Artificial Intelligence (AI) for Smart City Scenario: Recent Advancements and Future Trends. Sensors 2023, 23, 5206. [Google Scholar] [CrossRef]
  4. Jiao, J. Application and prospect of artificial intelligence in smart grid. IOP Conf. Ser. Earth Environ. Sci. 2020, 510, 022012. [Google Scholar] [CrossRef]
  5. Khandakar, A.E.H.; Chowdhury, M.; Khoda Kazi, M.; Benhmed, K.; Touati, F.; Al-Hitmi, M.; Gonzales, A.S.P., Jr. Machine Learning Based Photovoltaics (PV) Power Prediction Using Different Environmental Parameters of Qatar. Energies 2019, 12, 2782. [Google Scholar] [CrossRef]
  6. Diagne, M.; David, M.; Lauret, P.; Boland, J.; Schmutz, N. Review of solar irradiance forecasting methods and a proposition for small-scale insular grids. Renew. Sustain. Energy Rev. 2013, 27, 65–76. [Google Scholar] [CrossRef]
  7. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Computer Vision—ECCV 2018, Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Cham, Switzerland, 2018; pp. 122–138. [Google Scholar] [CrossRef]
  8. Sandler, M.; Howard, A.G.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef]
  9. Archon Project. Available online: http://archonproject.eu/english.html (accessed on 29 August 2023).
  10. Ziyabari, S.; Du, L.; Biswas, S.K. Multibranch Attentive Gated ResNet for Short-Term Spatio-Temporal Solar Irradiance Forecasting. IEEE Trans. Ind. Appl. 2022, 58, 28–38. [Google Scholar] [CrossRef]
  11. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar] [CrossRef]
  12. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Curran Associates, Inc.: Red Hook, NY, USA, 2019; pp. 8024–8035. [Google Scholar] [CrossRef]
  13. Holmgren, W.F.; Hansen, C.W.; Mikofski, M.A. pvlib python: A python package for modeling solar energy systems. J. Open Source Softw. 2018, 3, 884. [Google Scholar] [CrossRef]
  14. Papatheofanous, E.A.; Kalekis, V.; Venitourakis, G.; Tziolos, F.; Reisis, D. Deep Learning-Based Image Regression for Short-Term Solar Irradiance Forecasting on the Edge. Electronics 2022, 11, 3794. [Google Scholar] [CrossRef]
  15. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar] [CrossRef]
  16. Kaiser, L.; Gomez, A.N.; Chollet, F. Depthwise Separable Convolutions for Neural Machine Translation. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar] [CrossRef]
  17. Basodi, S.; Ji, C.; Zhang, H.; Pan, Y. Gradient amplification: An efficient way to train deep neural networks. Big Data Min. Anal. 2020, 3, 196–207. [Google Scholar] [CrossRef]
  18. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Singh, S., Markovitch, S., Eds.; AAAI Press: NW Washington, DC, USA, 2017; pp. 4278–4284. [Google Scholar] [CrossRef]
  19. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.; Wong, W.; Woo, W. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In Advances in Neural Information Processing Systems 28, Proceedings of the Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; Cortes, C., Lawrence, N.D., Lee, D.D., Sugiyama, M., Garnett, R., Eds.; MIT Press: Cambridge, MA, USA, 2015; pp. 802–810. [Google Scholar] [CrossRef]
  20. Staudemeyer, R.C.; Morris, E.R. Understanding LSTM—A tutorial into Long Short-Term Memory Recurrent Neural Networks. arXiv 2019, arXiv:1909.09586. [Google Scholar]
  21. Young, T.; Hazarika, D.; Poria, S.; Cambria, E. Recent Trends in Deep Learning Based Natural Language Processing. IEEE Comput. Intell. Mag. 2018, 13, 55–75. [Google Scholar] [CrossRef]
  22. Zhang, X.; Liang, X.; Zhiyuli, A.; Zhang, S.; Xu, R.; Wu, B. AT-LSTM: An Attention-based LSTM Model for Financial Time Series Prediction. IOP Conf. Ser. Mater. Sci. Eng. 2019, 569, 052037. [Google Scholar] [CrossRef]
  23. Ghany, K.K.A.; Zawbaa, H.M.; Sabri, H.M. COVID-19 prediction using LSTM algorithm: GCC case study. Inform. Med. Unlocked 2021, 23, 100566. [Google Scholar] [CrossRef] [PubMed]
  24. Zhou, Y.; Dong, H.; El-Saddik, A. Deep Learning in Next-Frame Prediction: A Benchmark Review. IEEE Access 2020, 8, 69273–69283. [Google Scholar] [CrossRef]
  25. Lin, C.W.; Lin, M.; Yang, S. SOPNet Method for the Fine-Grained Measurement and Prediction of Precipitation Intensity Using Outdoor Surveillance Cameras. IEEE Access 2020, 8, 188813–188824. [Google Scholar] [CrossRef]
  26. Pedro, H.T.C.; Larson, D.P.; Coimbra, C.F.M. A comprehensive dataset for the accelerated development and benchmarking of solar forecasting methods. J. Renew. Sustain. Energy 2019, 11, 036102. [Google Scholar] [CrossRef]
  27. Xie, W.; Liu, D.; Yang, M.; Chen, S.; Wang, B.; Wang, Z.; Xia, Y.; Liu, Y.; Wang, Y.; Zhang, C. SegCloud: A novel cloud image segmentation model using a deep convolutional neural network for ground-based all-sky-view camera observation. Atmos. Meas. Tech. 2020, 13, 1953–1961. [Google Scholar] [CrossRef]
  28. Krauz, L.; Janout, P.; Blažek, M.; Páta, P. Assessing Cloud Segmentation in the Chromacity Diagram of All-Sky Images. Remote Sens. 2020, 12, 1902. [Google Scholar] [CrossRef]
  29. Nie, Y.; Li, X.; Scott, A.; Sun, Y.; Venugopal, V.; Brandt, A. 2017–2019 Sky Images and Photovoltaic Power Generation Dataset for Short-Term Solar Forecasting (Stanford Benchmark); Stanford Digital Repository: Stanford, CA, USA, 2022. [Google Scholar] [CrossRef]
  30. Stoffel, T.; Andreas, A. NREL Solar Radiation Research Laboratory (SRRL): Baseline Measurement System (BMS); NREL: Golden, CO, USA, 1981. [CrossRef]
  31. Chu, Y.; Li, M.; Coimbra, C.F.; Feng, D.; Wang, H. Intra-hour irradiance forecasting techniques for solar power integration: A review. iScience 2021, 24, 103136. [Google Scholar] [CrossRef]
  32. Marquez, R.; Coimbra, C.F.M. Proposed Metric for Evaluation of Solar Forecasting Models. J. Sol. Energy Eng. 2012, 135, 011016. [Google Scholar] [CrossRef]
  33. Ballas, N.; Yao, L.; Pal, C.; Courville, A.C. Delving Deeper into Convolutional Networks for Learning Video Representations. In Proceedings of the 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, 2–4 May 2016; Conference Track Proceedings. Bengio, Y., LeCun, Y., Eds.; arXiv: Ithaca, NY, USA, 2016. [Google Scholar] [CrossRef]
  34. Miller, S.D.; Rogers, M.A.; Haynes, J.M.; Sengupta, M.; Heidinger, A.K. Short-term solar irradiance forecasting via satellite/model coupling. Solar Energy 2018, 168, 102–117. [Google Scholar] [CrossRef]
  35. Ayet, A.; Tandeo, P. Nowcasting solar irradiance using an analog method and geostationary satellite images. Solar Energy 2018, 164, 301–315. [Google Scholar] [CrossRef]
  36. Ordoñez Palacios, L.E.; Bucheli Guerrero, V.; Ordoñez, H. Machine Learning for Solar Resource Assessment Using Satellite Images. Energies 2022, 15, 3985. [Google Scholar] [CrossRef]
  37. Boussif, O.; Boukachab, G.; Assouline, D.; Massaroli, S.; Yuan, T.; Benabbou, L.; Bengio, Y. What if We Enrich day-ahead Solar Irradiance Time Series Forecasting with Spatio-Temporal Context? arXiv 2023, arXiv:2306.01112. [Google Scholar]
  38. Nie, Y.; Li, X.; Paletta, Q.; Aragon, M.; Scott, A.; Brandt, A. Open-Source Ground-based Sky Image Datasets for Very Short-term Solar Forecasting, Cloud Analysis and Modeling: A Comprehensive Survey. arXiv 2022, arXiv:2211.14709. [Google Scholar]
  39. Juncklaus Martins, B.; Cerentini, A.; Mantelli, S.L.; Loureiro Chaves, T.Z.; Moreira Branco, N.; von Wangenheim, A.; Rüther, R.; Marian Arrais, J. Systematic review of nowcasting approaches for solar energy production based upon ground-based cloud imaging. Sol. Energy Adv. 2022, 2, 100019. [Google Scholar] [CrossRef]
  40. Kumari, P.; Toshniwal, D. Deep learning models for solar irradiance forecasting: A comprehensive review. J. Clean. Prod. 2021, 318, 128566. [Google Scholar] [CrossRef]
  41. Kumar, D.S.; Yagli, G.M.; Kashyap, M.; Srinivasan, D. Solar irradiance resource and forecasting: A comprehensive review. IET Renew. Power Gener. 2020, 14, 1641–1656. [Google Scholar] [CrossRef]
  42. Rajagukguk, R.A.; Kamil, R.; Lee, H.J. A Deep Learning Model to Forecast Solar Irradiance Using a Sky Camera. Appl. Sci. 2021, 11, 5049. [Google Scholar] [CrossRef]
  43. Paletta, Q.; Hu, A.; Arbod, G.; Lasenby, J. ECLIPSE: Envisioning Cloud Induced Perturbations in Solar Energy. arXiv 2021, arXiv:2104.12419. [Google Scholar] [CrossRef]
  44. Nie, Y.; Paletta, Q.; Scott, A.; Pomares, L.M.; Arbod, G.; Sgouridis, S.; Lasenby, J.; Brandt, A. Sky-image-based solar forecasting using deep learning with multi-location data: Training models locally, globally or via transfer learning? arXiv 2022, arXiv:2211.02108. [Google Scholar]
  45. Park, S.; Kim, Y.; Ferrier, N.J.; Collis, S.M.; Sankaran, R.; Beckman, P.H. Prediction of Solar Irradiance and Photovoltaic Solar Energy Product Based on Cloud Coverage Estimation Using Machine Learning Methods. Atmosphere 2021, 12, 395. [Google Scholar] [CrossRef]
  46. Manandhar, P.; Temimi, M.; Aung, Z. Short-term solar radiation forecast using total sky imager via transfer learning. Energy Rep. 2023, 9, 819–828. [Google Scholar] [CrossRef]
  47. Zhang, J.; Verschae, R.; Nobuhara, S.; Lalonde, J.F. Deep photovoltaic nowcasting. Solar Energy 2018, 176, 267–276. [Google Scholar] [CrossRef]
  48. Sun, Y.; Venugopal, V.; Brandt, A.R. Short-term solar power forecast with deep learning: Exploring optimal input and output configuration. Solar Energy 2019, 188, 730–741. [Google Scholar] [CrossRef]
  49. Ajith, M.; Martínez-Ramón, M. Deep learning based solar radiation micro forecast by fusion of infrared cloud images and radiation data. Appl. Energy 2021, 294, 117014. [Google Scholar] [CrossRef]
  50. Bamisile, O.; Cai, D.; Oluwasanmi, A.; Ejiyi, C.; Ukwuoma, C.C.; Ojo, O.; Mukhtar, M.; Huang, Q. Comprehensive assessment, review, and comparison of AI models for solar irradiance prediction based on different time/estimation intervals. Sci. Rep. 2022, 12, 9644. [Google Scholar] [CrossRef] [PubMed]
  51. Lyu, C.; Eftekharnejad, S.; Basumallik, S.; Xu, C. Dynamic Feature Selection for Solar Irradiance Forecasting Based on Deep Reinforcement Learning. IEEE Trans. Ind. Appl. 2023, 59, 533–543. [Google Scholar] [CrossRef]
  52. Paletta, Q.; Lasenby, J. Convolutional Neural Networks applied to sky images for short-term solar irradiance forecasting. arXiv 2020, arXiv:2005.11246. [Google Scholar]
  53. Mercier, T.M.; Rahman, T.; Sabet, A. Solar Irradiance Anticipative Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2023—Workshops, Vancouver, BC, Canada, 17–24 June 2023; pp. 2065–2074. [Google Scholar] [CrossRef]
  54. Choi, M.; Rachunok, B.; Nateghi, R. Short-term solar irradiance forecasting using convolutional neural networks and cloud imagery. Environ. Res. Lett. 2021, 16, 044045. [Google Scholar] [CrossRef]
Figure 1. Depthwise separable convolution breakdown.
Figure 1. Depthwise separable convolution breakdown.
Information 14 00617 g001
Figure 2. Structure of the implemented Xception Layer (XL).
Figure 2. Structure of the implemented Xception Layer (XL).
Information 14 00617 g002
Figure 3. The structure of an Xception long short-term memory (XceptionLSTM) cell.
Figure 3. The structure of an Xception long short-term memory (XceptionLSTM) cell.
Information 14 00617 g003
Figure 4. Spatio-temporal encoder/decoder breakdown.
Figure 4. Spatio-temporal encoder/decoder breakdown.
Information 14 00617 g004
Figure 5. Spatial encoder.
Figure 5. Spatial encoder.
Information 14 00617 g005
Figure 6. Spatial decoder.
Figure 6. Spatial decoder.
Information 14 00617 g006
Figure 7. The Archon–Athens, Greece Dataset.
Figure 7. The Archon–Athens, Greece Dataset.
Information 14 00617 g007
Figure 8. Training scheme.
Figure 8. Training scheme.
Information 14 00617 g008
Figure 9. Timing reports for inference on low-cost, edge computing devices.
Figure 9. Timing reports for inference on low-cost, edge computing devices.
Information 14 00617 g009
Table 1. Overview of spatio-temporal models’ number of parameters and operations and the training time per epoch for an input sequence of five 128 × 128 × 4 images and an output sequence of fifteen irradiance values.
Table 1. Overview of spatio-temporal models’ number of parameters and operations and the training time per epoch for an input sequence of five 128 × 128 × 4 images and an output sequence of fifteen irradiance values.
Temporal ModelKernel SizeTemporal Encoder/DecoderSpatio-Temporal Encoder/DecoderTraining
Time per
Epoch (min)
Param.OPs (MAC)Param.OPs (MAC)
Spatial Encoder---833 K1.03 G3.26
Spatial Decoder---658 K0.66 M
ConvGRU31.79 M1.13 G3.90 M6.27 G19.16
54.93 M3.15 G7.05 M8.28 G20.54
ConvLSTM32.43 M1.51 G4.94 M6.65 G19.33
56.62 M4.19 G9.13 M9.33 G20.79
bi-ConvLSTM34.85 M1.89 G6.80 M7.02 G20.89
513.2 M5.24 G15.8 M10.4 G21.68
Stacked
ConvLSTM
3, 36.06 M3.02 G8.57 M8.16 G21.67
3 , 5 12.3 M5.71 G14.9 M10.8 G23.54
5 , 5 16.5 M8.39 G19.1 M13.5 G25.45
DWSConvLSTM3334 K172 M2.85 M5.31 G19.66
5342 K177 M2.85 M5.31 G19.73
bi-
DWSConvLSTM
3668 K215 M3.18 M5.35 G20.13
5684 K221 M3.20 M5.36 G20.23
Stacked
DWSConvLSTM
3 , 3 826 K343 M3.34 M5.48 G20.64
3 , 5 839 K349 M3.35 M5.49 G20.75
5 , 5 847 K354 M3.36 M5.49 G20.87
XceptionLSTMXL871 K516 M3.38 M5.65 G19.63
bi-XceptionLSTMXL1.74 M645 M4.25 M5.78 G19.75
Stacked XceptionLSTM 2   ×   XL2.17 M1.03 G4.68 M6.17 G20.79
Table 2. The hyperparameters that were examined and chosen for the presented benchmark.
Table 2. The hyperparameters that were examined and chosen for the presented benchmark.
HyperparameterTested OptionsFinal
Input Sequence Length{5, 10, 15}5
Output Sequence Length{5, 10, 15}15
Image Frame Size{64, 128, 256}128
Concatenate Sunmask{True, False}True
Removed Foreign Objects{True, False}True
Encoded State’s Channels{16, 32, 64, 128, 256}128
Optimizer{Adam, RMSProp}RMSProp
Scheduler ReduceOnPlateau , Exponential , Step Exponential
Learning Rate 5 × 10 2 , 10 3 , 5 × 10 4 , 10 4 , 5 × 10 5 , 10 5 1st Stage: 10 3
2nd Stage: 5 × 10 5
Loss Function L 1 , SmoothL 1 Huber , MSE MSE Loss
Batch Size{8, 12, 16, 20, 24}16
Table 3. Evaluation results of the models in Table 1 for the horizons of 1, 5, 15 min and the average for the first 15 min. The best models for each metric and horizon are bolded.
Table 3. Evaluation results of the models in Table 1 for the horizons of 1, 5, 15 min and the average for the first 15 min. The best models for each metric and horizon are bolded.
ModelKernel
Size
MBE (W m 2 )MAE (W m 2 )MAPE (%)
1 min5 min15 minMean1 min5 min15 minMean1 min5 min
Persistence<0.025−0.08410.1795<0.00120.6144.174.852.96.2115.89
ConvGRU3−3.199.639.098.8334.647.259.650.424.6632.5
5−1.5087.868.177.5233.046.259.049.722.7229.34
ConvLSTM3−0.6865.683.823.6033.147.059.649.921.4127.73
52.0801.898−1.3210.146533.047.160.950.821.7327.87
bi-ConvLSTM33.022.520<0.0250.174134.149.263.353.519.8926.69
5−4.24−5.07−7.30−6.1034.347.460.050.819.9124.87
Stacked
ConvLSTM
3 , 3 3.103.261.3041.74133.547.359.550.322.0829.31
3 , 5 0.24650.313−4.12−1.50033.047.060.050.421.4927.68
5 , 5 −2.170−1.474−4.08−2.82435.047.759.951.019.7626.35
DWSConvLSTM3−14.13−5.37−2.234−4.4935.546.659.850.322.1127.69
5−11.49−5.97−3.01−4.7534.846.359.049.821.5626.20
bi-
DWSConvLSTM
3−15.90−5.630.988−3.9437.046.058.549.522.7126.88
53.543.81−4.481.33534.546.660.150.123.0728.81
Stacked
DWSConvLSTM
3, 3−6.380.6782.4811.02235.046.458.549.623.1429.92
3 , 5 −8.51−3.711.160−1.83134.745.458.349.020.6026.18
5 , 5 1.9940.452−6.07−1.51834.546.759.650.123.0128.48
XceptionLSTMXL−3.011.139−4.87−0.92432.546.159.949.720.4624.43
bi-XceptionLSTMXL−7.03−2.992−8.16−4.5232.745.759.049.219.5722.98
Stacked XceptionLSTM 2 × XL−2.650−0.963−3.35−1.65733.045.860.449.919.3823.27
RMSE (W m−2)FS (%)MAPE (%)
1 min5 min15 minMean1 min5 min15 minMean15 minMean
Persistence75.2113.3146.6122.438.530.7
ConvGRU369.494.8113.199.97.7016.3722.8817.7738.530.7
568.395.4114.6100.09.1215.7721.8317.7436.831.4
ConvLSTM369.595.5114.5100.07.5015.6921.9217.6738.530.7
570.098.2116.2102.16.9413.3420.7515.9337.230.5
bi-ConvLSTM370.296.5115.5101.36.6114.8521.2216.5833.228.38
570.195.9115.6100.96.7915.3621.1316.9431.126.53
Stacked
ConvLSTM
3 , 3 69.796.3115.1100.77.2715.0621.4717.0936.731.0
3 , 5 70.297.8116.5102.26.6713.6620.5215.8734.329.43
5 , 5 70.796.5115.0101.15.9314.8821.5616.7231.927.64
DWSConvLSTM372.296.7117.0101.73.9614.6520.1716.1537.630.5
571.495.2115.2100.35.0215.9721.3917.3134.928.99
bi-
DWSConvLSTM
372.395.0115.1100.03.7916.1421.5017.5135.729.55
571.095.7114.8100.25.5715.5721.7117.3738.831.6
Stacked
DWSConvLSTM
3 , 3 73.395.9113.7100.32.5115.3822.4317.1636.731.8
3 , 5 72.195.2114.3100.04.0916.0122.0117.5535.228.72
5 , 5 71.195.5114.4100.25.4115.7121.9917.4239.431.4
XceptionLSTMXL68.894.6116.099.88.5316.5220.8817.8532.327.09
bi-XceptionLSTMXL68.794.5117.3100.38.5216.5720.0117.5228.9124.50
Stacked XceptionLSTM 2 × XL69.294.4115.899.87.9416.6921.0417.8531.325.66
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Venitourakis, G.; Vasilakis, C.; Tsagkaropoulos, A.; Amrou, T.; Konstantoulakis, G.; Golemis, P.; Reisis, D. Neural Network-Based Solar Irradiance Forecast for Edge Computing Devices. Information 2023, 14, 617. https://doi.org/10.3390/info14110617

AMA Style

Venitourakis G, Vasilakis C, Tsagkaropoulos A, Amrou T, Konstantoulakis G, Golemis P, Reisis D. Neural Network-Based Solar Irradiance Forecast for Edge Computing Devices. Information. 2023; 14(11):617. https://doi.org/10.3390/info14110617

Chicago/Turabian Style

Venitourakis, Georgios, Christoforos Vasilakis, Alexandros Tsagkaropoulos, Tzouma Amrou, Georgios Konstantoulakis, Panagiotis Golemis, and Dionysios Reisis. 2023. "Neural Network-Based Solar Irradiance Forecast for Edge Computing Devices" Information 14, no. 11: 617. https://doi.org/10.3390/info14110617

APA Style

Venitourakis, G., Vasilakis, C., Tsagkaropoulos, A., Amrou, T., Konstantoulakis, G., Golemis, P., & Reisis, D. (2023). Neural Network-Based Solar Irradiance Forecast for Edge Computing Devices. Information, 14(11), 617. https://doi.org/10.3390/info14110617

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop