Previous Article in Journal
A Multimodal Learning Approach for Protecting the Metro System of Medellin Colombia Against Corrupted User Traffic Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cloud-Enabled Hybrid, Accurate and Robust Short-Term Electric Load Forecasting Framework for Smart Residential Buildings: Evaluation of Aggregate vs. Appliance-Level Forecasting

by
Kamran Hassanpouri Baesmat
*,
Emma E. Regentova
and
Yahia Baghzouz
Department of Electrical and Computer Engineering, University of Nevada, Las Vegas, NV 89154, USA
*
Author to whom correspondence should be addressed.
Smart Cities 2025, 8(6), 199; https://doi.org/10.3390/smartcities8060199
Submission received: 1 October 2025 / Revised: 23 November 2025 / Accepted: 25 November 2025 / Published: 27 November 2025

Highlights

What are the main findings?
  • A cloud-enabled hybrid framework combining Seasonal ARIMAX, Random For-est, and LSTM achieves high accuracy for a short-term residential load forecasting.
  • Appliance-level (AL) and aggregate forecasts (AF) are compared, showing that AF outperforms AL in reliability and accuracy.
What are the implications of the main findings?
  • Multi-cloud parallel calculations with a two-out-of-three voting scheme enhances forecasting robustness by mitigating single-cloud failures.
  • The proposed scalable solution supports smart-city energy management, demand response, and grid reliability.

Abstract

Accurate short-term load forecasting is vital for smart-city energy management, enabling real-time grid stability and sustainable demand response. This study introduces a cloud-enabled hybrid forecasting framework that integrates Seasonal Autoregressive Integrated Moving Average with Exogenous variables (SARIMAX), Random Forest (RF), and Long Short-Term Memory (LSTM) models, unified through a residual-correction mechanism to capture both linear seasonal and nonlinear temporal dynamics. The framework performs fine-grained 5 min forecasting at both appliance and aggregate levels, revealing that the aggregate forecast achieves higher stability and accuracy than the sum of appliance-level predictions. To ensure operational resilience, three independent hybrid models are deployed across distinct cloud platforms with a two-out-of-three voting scheme, that guarantees continuity if a single-cloud interruption occurs. Using a real residential dataset from a house in Summerlin, Las Vegas (2022), the proposed system achieved a Root Mean Squared Logarithmic Error (RMSLE) of 0.0431 for aggregated load prediction representing a 35% improvement over the next-best model (Random Forest) and maintained consistent prediction accuracy during simulated cloud outages. These results demonstrate that the proposed framework provides a scalable, fault-tolerant, and accurate energy forecasting.

1. Introduction

In the modern era, energy management is shaping the world: nearly every sector, whether directly or indirectly, depends on efficient energy utilization, from AI-driven data centers and the protection of the ozone layer to the preservation of ecosystems, advancements in construction, industry, medicine, and tourism.
Energy management in smart cities ensures a balance between production and demand, prevents resource waste, improves air quality, and safeguards the planet.
“Smart” turns appliances into measuring, labeled, and controllable assets, and that radically improves AL data collection and forecasting accuracy/efficiency.
“Smart” appliances offer multiple advantages for accuracy of data, and hence for forecasting. Smart appliances can natively stream instantaneous power, voltage, current, and cumulative energy, not only periodic kWh. Also, they secure machine-readable semantics, not raw points. Devices can expose states (e.g., “wash cycle”, “defrost”, “setpoint”) with consistent naming/units. That makes feature engineering automatic and portable across vendors/homes. Forecasting upgrades to “state-aware”, that is one can move from generic time-series to state-space/HSMM or sequence models (RNN/Transformer) that understand cycles and dwell times (e.g., washer, fridge, heat pump). Papers on bottom-up/AL STLF show markedly better fidelity when appliance characteristics and states are included [1].
If devices are truly smart, one can read the appliance directly. This means less dependencies on NILM: Non-Intrusive Load Monitoring (a.k.a. energy disaggregation): Authors of [2] show sampling, transient features, and deep learning, including federated learning push accuracy, however under limits of overlapping loads.
Accurate load forecasting, enables companies to optimally scheduling their activities and determining the appropriate amount of energy to generate in order to maintain stability.

1.1. Load Forecasting Categories and Challenges

Electrical load consumption (ELC) forecasting is typically categorized into: short-term load forecasting (STLF), mid-term load forecasting (MTLF), and long-term load forecasting (LTLF). This study focuses on developing an efficient framework for STLF in residential buildings and addresses several critical research questions (this STLF framework directly supports smart-city energy operations by using smart-meter and weather streams to forecast demand at fine granularity, enabling demand response, EV-charging coordination, and optimal dispatch of solar-plus-storage across neighborhoods):
  • Are statistical methods (SM) more effective for load prediction in residential buildings compared to machine learning (ML) techniques?
  • Does a hybrid approach that combines SM and ML provide superior results?
  • Is it more accurate to predict the overall consumption of a residential building (RB), or to forecast the consumption of individual appliances and then aggregate the results?
Furthermore, this paper examines the impact of feature-selection techniques, methods for combining statistical and machine learning (ML) models, residual (remainder) prediction (RP), and cloud-based computing.
Table 1 and Table 2 summarizes challenges for statistical modeling and ML modeling.

1.2. Feature Selection

Feature selection techniques have long been employed in scientific research, and their effectiveness has been widely studied [3,4,5,6]. Each method has its distinct advantages. For instance, the ANOVA F-test assesses whether the mean of a continuous (numeric) feature differs significantly across the categories of the target, by comparing between-class to within-class variance [7], mutual information measures nonlinear dependencies between features and the target variable [8], information gain and other entropy-based scores are commonly applied in classification tasks [9], LASSO suppresses irrelevant features and thus performing implicit feature selection [10], and genetic algorithms offer evolutionary optimization for selecting feature subsets [11]. Research in [12] investigates the role of correlation matrices in prediction and showed that features with high correlation can enhance dataset performance. In [13], the authors evaluated the Chi-Square statistical test as a feature selection method, demonstrating its ability to assess whether features have a meaningful relationship with the target. Furthermore, the authors of [14] highlighted the strength of the Maximum Relevance Minimum Redundancy (MRMR) method in identifying features that are both relevant and non-redundant, thereby mitigating overfitting caused by redundant information.
This research proposes a novel framework that integrates a voting-based system for feature selection. By combining statistical methods with a hybrid approach, the framework leverages correlation analysis, feature independence, and redundancy avoidance to maximize predictive performance in residential load forecasting.

1.3. Statistical Methods (SM)

Statistical methods [15,16], machine learning (ML) techniques [17,18], and hybrid models [19,20] have continually evolved over the past decades, leading to significant advances in prediction accuracy.
For example, research in [21] demonstrated the effectiveness of the autoregressive integrated moving average (ARIMA) model for time-series forecasting, particularly in scenarios driven by temporal dependencies. However, ARIMA struggles to capture seasonality, a limitation that has been addressed by seasonal ARIMA (SARIMA) [22] and SARIMAX (seasonal ARIMA with exogenous regressors) [23]. Despite these improvements, statistical methods generally lack the scalability and flexibility of ML techniques when dealing with large datasets and complex nonlinear relationships.

1.4. Machine Learning (ML)

Machine learning methods, on the other hand, have proven highly powerful and are now widely applied across nearly all domains of science. In [24], the authors highlighted the accuracy of support vector machines (SVMs), which are supervised learning models used in both classification and regression tasks. Tree-based and sequence-based models have also emerged as dominant approaches in forecasting. Decision trees (DTs) [25] produce hierarchical decision rules by recursively splitting on feature values, offering interpretability but often suffering from overfitting. Random Forests (RFs) [26] overcome this drawback by aggregating multiple randomized trees, thereby reducing variance and improving robustness. Extreme Gradient Boosting (XGBoost) [27] further enhances ensemble learning by constructing trees sequentially to correct residual errors, incorporating gradient optimization, regularization, and efficient computation for superior predictive performance. In parallel, Long Short-Term Memory networks (LSTMs) [28] are designed to capture long-range temporal dependencies using gated memory cells, making them especially effective for sequential and time-series forecasting.
Nevertheless, a limitation of many ML-based approaches is their tendency to focus on identifying patterns without explicitly incorporating seasonality or linear temporal dynamics, areas where statistical models such as ARIMA, SARIMA, and SARIMAX excel. This contrast motivates development of hybrid frameworks, which combine the nonlinear modeling power of ML with trend- and seasonality-capturing strengths of statistical approaches. By integrating these complementary capabilities, hybrid systems achieve higher accuracy, reliability, and generalizability.
Authors in [29] present a short-term residential load forecasting framework that integrates evolutionary algorithms for neural architecture search with transfer learning to adapt models across multiple households. The method demonstrates superior performance over benchmark techniques, with an average accuracy gain of 3.17 percentage points compared to the state-of-the-art LSTM one-shot approach. Researchers in [30] introduced a hybrid forecasting framework for urban power load that identifies and quantifies abnormal variation periods via residual decomposition, augments scarce abnormal samples using WGAN-GP, and employs TimesNet to capture complex nonlinear patterns. The method boosts forecasting accuracy for abnormal load periods by about 38.9%. Although it incurs only a minor drop (2.65%) across all periods, it outperforms baseline deep-learning models. The authors of [31] proposed a hybrid deep-learning framework (GNN-ViGNet) enhanced by the use of IoT sensor data and novel preprocessing, feature engineering, and imbalance correction methods, coupled with a Coati–Northern Goshawk Optimization algorithm for real-time route optimization in EV fleets. Their model achieves 98.9% accuracy in charging load forecasting and reduces route distance to 511 km, significantly outperforming baseline techniques such as PSO and Firefly in both convergence speed and prediction error metrics. The authors of [32] introduced radian scaling, a new normalization method that transforms differences of consecutive electricity load values into radians to maintain normalized distributions under seasonal concept drift without retraining machine learning models. Their experiments using lightweight ANN models show radian scaling reduces RMSE dramatically from 158.63 to 43.38 and cuts training epochs by about 67%, statistically outperforming traditional normalization methods. The work in [33] develops a learning framework for forecasting electrical load in renewable energy communities with dynamic portfolios, synthesizing non-stationary, discontinuous, and NON-IID time series, adapting preprocessing and model input (past, present, future features) to achieve domain-invariant feature learning.
Their approach achieves forecasting accuracy comparable to that of centrally trained models, while outperforming single-source models by more than 18% in terms of MAE, and demonstrating strong transferability to unseen data. This research proposes a hybrid forecasting framework that integrates RF, LSTM, and SARIMAX models to address these challenges and enhance short-term load forecasting performance in residential buildings. As hybrid models have demonstrated high performance, this paper also investigates this approach and extends it by incorporating additional techniques to further enhance forecasting accuracy and robustness. Specifically, the study integrates efficient feature-selection methods to improve predictive performance and employs concurrent cloud computing together with a voting-based strategy to increase the model’s resilience and scalability.

1.5. The Contributions of the Presented Work Can Be Summarized as Follows

  • Integration across paradigms: A unified forecasting architecture that synergistically combines Seasonal ARIMAX, Random Forest, and LSTM models through a residual-correction layer, effectively merging statistical trend modeling with nonlinear and sequential learning.
  • Feature–model customization: Each sub-model employs an optimized feature selection strategy (Correlation Matrix, Chi-Square, or MRMR), enabling the framework to capture diverse predictive relationships with minimal redundancy.
  • Multi-cloud redundancy and voting: A two-out-of-three cloud voting ensures continuous operation under individual cloud or model failures, an aspect rarely addressed in prior forecasting frameworks.
  • Fine-grained temporal resolution: The framework is validated on 5 min interval residential load data, demonstrating that aggregate forecasts outperform the summation of appliance-level predictions (sum of predictions per appliances) in both stability and accuracy.
  • Demonstrated accuracy and robustness: The system achieves a 35% reduction in RMSLE relative to the best single-model benchmark and maintains reliable performance under simulated cloud outages.
A detailed comparison between aggregate prediction and appliance-level forecasting in terms of accuracy and reliability. Altogether builds a scalable, fault-tolerant, and statistically grounded hybrid methodology for short-term load forecasting in smart residential environments.
The rest of the paper is organized as follows: Section 2: Data description and preparation, Section 3: Feature selection, Section 4: Methodology, Section 5: Discussion and comparison of the results, Section 6: Conclusions.

2. Methodology

This research proposes a novel method that integrates both statistical approaches and machine learning techniques, termed SSRLR (Sparse voting feature-selection technique, Seasonal Autoregressive Integrated Moving Average with Exogenous variables model, Random Forest, Long Short-Term Memory, and Remainder Prediction). This integration significantly enhances the overall predictive accuracy of the proposed framework. The machine learning component is responsible for capturing complex, non-linear patterns and intricate relationships within the data, while the statistical component simultaneously models seasonality effects and time-dependent correlations. By operating in parallel, these two components complement and reinforce each other: the statistical models address temporal regularities that machine learning alone might overlook, and the machine learning algorithms uncover high-dimensional interactions that are beyond the capacity of purely statistical methods. This mutual coverage enables the framework to extract the maximum amount of useful information from the data while mitigating the risk of overfitting. Furthermore, the synergy between the two paradigms ensures that weaknesses in one component are compensated for by the strengths of the other, thereby improving robustness and generalization performance across diverse scenarios.
Figure 1 illustrates the flow of the embedded procedures. After collecting data on electricity consumption and relevant weather variables, preprocessing is performed by applying the Holiday/Weekend factor and modeling the seasonal effects. Then, the process proceeds to the feature selection phase using three feature selection techniques: Chi-Square test, MRMR, and CM. These methods are applied to identify and retain the most informative features. By systematically selecting only the most relevant variables, this step extracts an optimal set of features, thereby improving generalization and reducing the risk of overfitting. Consequently, the feature selection phase plays a pivotal role in enhancing the robustness and accuracy of the forecasting framework.
Next, the process of primary prediction (PP) employs both machine learning and statistical methods, specifically, the Long Short-Term Memory (LSTM), the Seasonal Autoregressive Integrated Moving Average with Exogenous variables (SARIMAX) model, and the Random Forest (RF) algorithm. The SARIMAX model is utilized to capture and exploit the most informative seasonal effects and other time-dependent features, thereby maximizing the extraction of temporal patterns from the data. The RF method enables the exploration of diverse forecasting scenarios by leveraging its ensemble-based decision-making capability, which enhances robustness against overfitting. In parallel, the LSTM component contributes to the modeling of both short-term and long-term temporal dependencies, thus complementing the strengths of SARIMAX and RF. This hybrid integration substantially enhances the learning potential of the proposed method, enabling it to adapt effectively to complex and dynamic patterns present in the dataset.
A primary trend is first established to capture the general direction of the target variable. This trend enables the derivation of an additional and highly informative feature the Remainder Prediction (RP) in which deviations from the trend are referred to as remainders. Once the primary trend is obtained, the deviations of the actual values from the underlying forecasted trajectory are calculated. These deviations are then treated as a separate time series, which can be independently modeled to predict the future remainders. The final forecast, i.e., overall consumption, all individual appliance-level forecasting, and also their total consumption. that is subsequently obtained by summing the predicted primary consumption values and the predicted remainders. This decomposition-based strategy reduces forecasting error because it allows the model to focus on different components of data separately: the trend model specializes on capturing long-term patterns, while the remainder model focuses on short-term fluctuations and irregularities. By isolating these components, the approach mitigates the risk of error propagation that can occur when a single model attempts to fit both large-scale and fine-grained variations simultaneously.
This research further enhances the reliability, processing speed, and real-time computing capability of the proposed framework through the use of cloud computing. In this setup, three independent cloud instances operate in parallel. The final forecast is obtained through an averaging mechanism, while a majority-voting mechanism is employed to ensure that a correct prediction is selected. Specifically, if at least two of the three instances exhibit the same level of error (the difference is below a threshold), their output is accepted as correct. If the third instance shows a significantly larger error, it is excluded from the forecasting process and flagged for expert review and retuning. This redundancy significantly improves the accuracy and robustness, allowing the system to operate effectively even in the event of a single instance failure, and it minimizes the risk of erroneous outputs caused by transient faults or network disruptions. The adoption of this dual-stage forecasting methodology trend plus Remainder Prediction within a fault-tolerant cloud infrastructure offers several benefits. From a modeling perspective, separating long-term and short-term patterns allows for more specialized learning and reduces cumulative forecasting errors. From a system architecture perspective, the two-out-of-three cloud framework ensures high availability, operational continuity, and consistent performance under varying computational loads. Leveraging cloud computing brings additional advantages. The elastic scalability of cloud resources allows the system to handle sudden increases in data volume without sacrificing processing speed. Geographic distribution of cloud servers reduces latency, enabling near real-time decision-making. Furthermore, cloud platforms facilitate seamless integration of advanced analytics tools, automated backups, and security protocols, ensuring that forecasting models remain both robust and secure while being accessible from anywhere. This combination of advanced forecasting methodology and resilient cloud-based deployment provides a high-performance, dependable solution for real-world operational environments.

2.1. Long Short-Term Memory (LSTM) Networks

Long Short-Term Memory (LSTM) [34] networks, sketched in Figure 2, are a form of Recurrent Neural Network (RNN) specifically built to capture dependencies that span long intervals in sequential data. Unlike standard RNNs, which struggle to learn from distant inputs due to vanishing gradients during backpropagation, LSTMs introduce a dedicated memory cell that can preserve and carry information for many time steps. This design makes them especially powerful for applications like forecasting time series, understanding speech, and processing text.
At each time step, the memory cell’s content is carefully managed by three gates forget, input, and output. The forget gate decides which past information should be removed, the input gate controls what new information enters the cell, and the output gate determines what part of the memory contributes to the next hidden state. By adjusting these gates dynamically, an LSTM can flexibly balance learning short-term fluctuations and long-term trends within the data.
LSTMs have demonstrated strong performance in numerous areas. In natural language processing, they learn the flow of sentences to generate or translate text. In speech recognition, they interpret audio sequences to convert spoken words into text. They are also effective in spotting unusual patterns in time-series data for anomaly detection. Despite their benefits, LSTMs require substantial computational resources and their inherently sequential updates limit the ability to parallelize training compared to architectures like Convolutional Neural Networks (CNNs). Nevertheless, their capacity to handle both isolated observations and entire sequences makes them a cornerstone model in deep learning research involving temporal or ordered inputs.

2.2. Random Forest (RF)

The Random Forest [35] algorithm (see Figure 3) is an ensemble-based technique that leverages the simplicity of individual decision trees while dramatically improving their reliability. Instead of relying on a single tree, it grows dozens or even hundreds of them, each trained on a different “bootstrap” sample of the data (i.e., random draws with replacement). At each split in a tree, only a randomly chosen subset of the available features is considered, which encourages diversity among trees and helps prevent overfitting. When making a prediction, the forest either tallies a vote among all trees (for classification) or computes the mean of their outputs (for regression), yielding a consensus result that is typically more accurate and stable than any single tree.
One of Random Forest’s greatest assets is its ability to estimate the relative importance of each feature: by measuring how much the model’s error increases when a given variable’s values are permuted, insight is gained into which inputs truly drive predictions. This makes RF a valuable tool both for high-stakes applications such as detecting fraudulent transactions, forecasting electrical loads, or diagnosing diseases and for exploratory analysis, where understanding the data is as important as prediction itself. On the downside, training and storing a large collection of trees can demand considerable computation and memory, and making real-time predictions may be slower if the forest is very deep. Moreover, while each decision tree is easy to follow on its own, the combined forest becomes something of a “black box”, making it harder to explain individual decisions than with simpler, single-tree models. Despite these trade-offs, Random Forest remains a go to choose whenever robustness to noise, missing values, and outliers is required together with strong predictive power.

2.3. Seasonal Autoregressive Integrated Moving Average with Exogenous Variables (SARIMAX)

The SARIMAX [36] model extends the familiar ARIMA approach by explicitly handling seasonal fluctuations and allowing for outside influences. It is ideal when the data repeat patterns at regular intervals such as monthly sales cycles or annual weather variations while also being affected by additional variables like temperature or economic indicators. In essence, SARIMAX merges three pieces:
  • Non-seasonal ARIMA component:
Autoregressive (AR)p)):
This component models the current value of the series by referencing its own previous observations. The order parameter p designates how many past points are used to predict the present.
Integrated (I)d)):
To stabilize the series, the data are differenced d times. This process removes trends and helps ensure that statistical characteristics such as the mean and variance remain roughly constant throughout the sequence.
Moving Average (MA)q)):
Here, the forecast is formed by combining current and past forecast errors. The parameter q indicates how many of these lagged error terms are included in the smoothing equation.
Mathematically, this is written as:
ϕ ( B )   ( 1 B ) d y t = θ ( B )   ϵ t
where
  • ϕ(B): is the non-seasonal AR polynomial
  • θ(B): is the non-seasonal MA polynomial
  • B: is the backshift operator
  • yt: is the series at time t, and
  • ϵ t : is a white-noise error term
2.
Seasonal Component (s)
To capture repeating patterns every “s” period, SARIMAX uses seasonal AR and MA polynomials of orders P and Q, together with seasonal differencing D. This portion is given by:
Φ B s 1 B s D   y t = Θ B s   ϵ t
where
  • Φ( B s ): Seasonal autoregressive polynomial of order P.
  • Θ( B s ): Seasonal moving average polynomial of order Q.
  • S: Seasonality period
  • D: Number of seasonal differences.
3.
Exogenous regressors ( X )
y t = β X t + ϵ t
where
  • X t : External (exogenous) regressors at time t.
  • β : Coefficient vector for the exogenous variables.
Combining all components, the SARIMAX model can be represented as:
ϕ ( B )   ( 1 B )   d   Φ B s ( 1 B s ) D   y t = θ B   Θ B s ϵ t +   β X t
By combining these three elements non-seasonal dynamics, periodic behavior, and outside influences SARIMAX delivers a flexible framework for forecasting. It is widely used in examples like projecting retail sales with marketing spend as an exogenous input, estimating electricity demand driven by temperature and humidity, or any scenario where patterns repeat and external drivers matter.

2.4. SSRLR

The pseudocode of the proposed SSRLR is illustrated below (Algorithm 1):
SSRLR, a novel hybrid method that integrates statistical and machine learning techniques including a Sparse voting feature selection technique, Seasonal Autoregressive Integrated Moving Average Exogenous model, Random Forest, Long Short-Term Memory, and a Remainder Prediction step to leverage both linear trends and complex nonlinear relationships.
Algorithm 1. Pseudocode of proposed algorithm
Input: Historical load data, Weather features, Forecast horizon = 12
Step 1: Data Preprocessing
  Normalize data using MinMaxScaler to range [0,1]
Step 2: Feature Selection
  Apply CM
  Select relevant features → CM_Output
  Use this for LSTM
  Apply Chi-Square
  Select relevant features → Chi-Square_Output
  Use this for SARIMAX
  Apply MRMR
  Select relevant features → MRMR_Output
  Use this for RF
Step 3: Model Training and Prediction
  Train SARIMAX model with seasonal and non-seasonal components
  Predict next 12 points → SARIMAX_Output
  Train Random Forest model using weather and historical data
  Predict next 12 points → RF_Output
  Train LSTM model for time-series forecasting
  Predict next 12 points → LSTM_Output
Step 4: Remainder Prediction (RP)
  Perform additional adjustments if needed:
  Use SARIMAX for LSTM on cloud 1
  Use LSTM for RF on cloud 2
  Use RF for SARIMAX on cloud 3
Step 5: Ensemble Averaging
  SARIMAX Final Output (PP+RP) → cloud 1_Output
  RF Final Output (PP+RP) → cloud 2_Output
  LSTM Final Output (PP+RP) → cloud 3_Output
Step 6: Ensemble Averaging
  Ensemble_Output = (cloud 2_Output + cloud 2_Output + cloud 3_Output)/3
Step 7: Output Final Prediction
  Return Ensemble_Output
The dataset is divided into 80% for training and 20% for testing, after which the following procedures are applied:
  • The process begins with feature selection using the CM method, followed by primary forecasting with LSTM. A trend is then identified for the remainders, and SARIMAX is applied to forecast the subsequent remainders. The final output for this stage is obtained by summing the LSTM predictions with the forecasted remainders.
  • Feature selection is performed using the MRMR method, followed by primary forecasting with Random Forest (RF). A trend is then identified for the remainders, and LSTM is applied to forecast the subsequent remainders. The final output for this stage is obtained by summing the RF predictions with the LSTM-forecasted remainders.
  • Feature selection is performed using the Chi-Square method, followed by primary forecasting with SARIMAX. A trend is then identified for the remainders, and Random Forest (RF) is applied to forecast the subsequent remainders. The final output for this stage is obtained by summing the SARIMAX predictions with the RF-forecasted remainders.
Figure 4 presents the detailed schematic of the proposed method.

2.5. Two-Out-of-Three Multi-Cloud Voting Architecture

Single cloud has been used in load forecasting in previous research [37]. To enhance the robustness, accuracy, and real-time performance of the forecasting system, the three pipelines described in Section 2.4 are deployed across three independent cloud instances. As illustrated in Figure 5, each cloud instance executes one of the proposed hybrids forecasting strategies:
  • Cloud 1: Implements the CM → LSTM → SARIMAX pipeline.
  • Cloud 2: Implements the MRMR → RF → LSTM pipeline.
  • Cloud 3: Implements the Chi-Square → SARIMAX → RF pipeline.
Figure 5. Proposed two-out of-three cloud-based system.
Figure 5. Proposed two-out of-three cloud-based system.
Smartcities 08 00199 g005
Each cloud processes the same input dataset in parallel, but with different feature-selection methods, primary forecasting models, and remainder forecasting techniques. This methodological diversity significantly reduces the risk of correlated model errors and increases the resilience of the overall system.
Preliminary Averaging: The forecasts from all three clouds are averaged to produce an initial consensus prediction.
Error-Based Majority Voting: The errors from each cloud are evaluated against the actual values. If at least two clouds yield similar and low error magnitudes, their results are accepted as correct. The third cloud, if it shows significantly higher error, is excluded from the current forecasting decision and flagged for expert review and model retuning.
This redundancy ensures that the system maintains robustness remaining operational and accurate even in the presence of network disruptions, computational errors, or degraded model performance in one instance. By leveraging agreement between at least two independent pipelines, the architecture minimizes the likelihood of passing erroneous forecasts to the end user or decision-making process. Figure 5 shows this cloud-based architecture.
The deployment of the two-out of-three architecture in a cloud environment provides several operational and strategic benefits:
  • High Availability and Robustness: The redundancy of multiple clouds ensures continuous operation even if one instance experiences downtime or degraded performance.
  • Elastic Scalability: Cloud resources can be dynamically allocated to handle varying computational demands, enabling real-time or near real-time processing for large-scale datasets.
  • Parallel Computation: Hosting clouds in different regions improving forecast delivery speed and responsiveness.
  • Security and Data Integrity: Cloud platforms provide encryption, controlled access, and automated backup systems, ensuring the safety and reliability of both input data and forecasting models.
  • Cost Efficiency: Pay-as-you-go billing models allow for optimal use of resources without the overhead of maintaining extensive on-premises infrastructure.
  • Seamless Integration and Maintenance: Cloud ecosystems support easy deployment of updates, integration of analytics tools, and automated monitoring of model performance with cheaper price compare to on- premise systems.
  • Collaboration and Accessibility: Authorized users can securely access the forecasting system from anywhere, enabling collaborative work across geographically distributed teams.
In the suggested two-out-of-three cloud-based redundancy model, reliability is expressed in terms of R, the probability of each cloud (e.g., AWS) functioning correctly:
  • R3 denotes the case where all three systems perform without error.
  • 3R2 (1 − R) corresponds to the case where two systems deliver correct outputs and one produces an error, which is automatically overruled by the majority decision.
The system failure occurs only if two or more clouds malfunction at the same time, a situation not covered by this formula.
By combining diverse forecasting models with a fault-tolerant two-out-of-three multi-cloud deployment, the proposed system achieves a high degree of accuracy, resilience, and computational efficiency. This design ensures that forecasts remain consistent, secure, and responsive under real-world operational conditions, even in the presence of hardware failures, software errors, or sudden spikes in data volume.

3. Data Description and Preparation

The dataset consists of power usage readings for a household in Summerlin, Las Vegas, Nevada, USA, sampled at five-minute intervals for 365 days of 2022. This dataset was officially collected from a residential customer of a local electric utility as part of its demand response program. Each measurement is given in kilowatts (kW) and reflects the consumption of the following household systems and appliances:
  • Forced-air heating unit on the lower level
  • Air conditioner on the lower level
  • Refrigerator
  • Air conditioner on the upper level of the house
  • General (miscellaneous) household load
  • Laundry machines (washer and dryer)
  • Dishwasher and garbage disposal
  • Microwave oven
  • Forced-air heating unit on the upper level of the house
  • Kitchen circuit (GFI) including the toaster
The dataset offers a full coverage with no gaps or missing entries, which makes it a solid foundation for any analysis. Electricity usage was logged every five minutes throughout 2022, giving a detailed picture that picks up even the smallest variations in consumption. Before feeding the data into the forecasting models, preprocessing is performed and relevant features are carefully selected. This step both smooths out irregularities and reduces the risk of overfitting. Figure 6 illustrates the data profile.

Estimating the Holiday/Weekend Adjustment Factor and Modeling the Seasonal Effect

Before feeding the data into the forecasting model, the dataset is preprocessed to adjust for unusual consumption patterns that occur during holidays and weekends, when usage often departs from typical weekday behavior. Social traditions, leisure activities, and economic regulations all play a role: some people travel or gather with friends, while others stay home using appliances or hosting events. Many people in Las Vegas work during the weekends, holidays, or change shifts frequently. In addition, manufacturing schedules may shift factories in some areas shut down for holidays, whereas others continue operating further altering the load profile. To quantify this effect, a Holiday Adjustment Factor is computed by first tagging each timestamp as either a “holiday/weekend” or a “weekday.” The data are then divided into two groups one representing holidays and weekends and the other representing regular weekdays and the mean energy consumption is calculated for each group. The adjustment factor is obtained by dividing the holiday/weekend average (1.89 kW) by the weekday average (1.94 kW), yielding a value of approximately 0.98. This value shows that, on average, consumption during holidays and weekends is about 98% of typical weekday use, and incorporating this factor into our model helps dampen the extra variability introduced by these irregular days, resulting in more accurate forecasts. (Holiday Coefficient = (Average Consumption on Holidays-Weekends)/(Average Consumption on Regular Days)).
Holidays included in 2022: 1 January (New Year’s Day, Saturday), 17 January (Martin Luther King Jr. Day, Monday), 21 February (Presidents Day, Monday), 30 May (Memorial Day, Monday), 4 July (Independence Day, Monday), 5 September (Labor Day, Monday), 10 October (Columbus Day, Monday), 11 November (Veterans Day, Friday), 24 November (Thanksgiving Day, Thursday), 25 December (Christmas Day, Sunday).
Next, each calendar month is standardized to a fixed length of 31 days. When a month contains fewer than 31 days, the readings from its final day are carried forward to fill the gap. This step creates a uniform time grid, enabling the model to learn genuine seasonal patterns without distortion from varying month lengths.
Seasonal effects are regular, repeating variations in energy consumption that occur over time hourly, daily, weekly, or monthly. By assigning each month a uniform 31-day span, artificial discontinuities in the data are eliminated, allowing the algorithm to detect genuine periodic patterns.
  • Applying this procedure to our 2022 dataset produces:
  • A total of 372 “normalized” days (12 months × 31 days)
  • 372 days × 24 h = 8928 h
  • At 5 min resolution: 8928 h × 12 readings per hour = 107,136 data points
To ensure temporal alignment across months, we standardized all monthly sequences to 31 days. This choice is consistent with prior studies in residential and urban load forecasting where month-length normalization is required. Importantly, the 31-day format preserves the complete set of real observations: months with fewer than 31 days are expanded using a small number of interpolated entries, while months with 31 days remain unchanged. In contrast, a 30-day standardization would require deleting one full day of real measurements from every 31-day month, which removes genuine load patterns and weakens the statistical richness of the dataset. Additionally, the 31-day format requires fewer temporal adjustments (7 days per year) compared to the 30-day approach (9 days per year), reducing distortion and maintaining closer fidelity to the original temporal structure. For these reasons, the 31-day standardization provides a more reliable and less intrusive alignment strategy for short-term load forecasting.
This expanded and smoothened dataset preserves the natural ebb and flow of consumption across hours, days, and months, enabling the forecasting model to capture seasonal trends with finer precision.

4. Applying Feature Selection

Selecting appropriate predictors is essential for improving model accuracy and avoiding redundant information. Common approaches include constructing a Correlation Matrix, applying Minimum Redundancy Maximum Relevance (MRMR), performing Analysis of Variance (ANOVA), and employing the Chi-Square test. For this analysis, seventeen atmospheric indicators were obtained from the Visual Crossing service, as illustrated in Figure 7. Several reputable weather data providers including the National Weather Service, the Weather Channel, AccuWeather, Weather Underground, and Visual Crossing were examined and compared. Based on a comprehensive assessment of data completeness, spatial coverage, and reliability, Visual Crossing was selected as the primary source of meteorological information, offering the most consistent dataset with 5 min resolution for the case study location.
These variables are: Temperature, the ambient air temperature reported in °C or °F; Apparent Temperature (Feels-Like), a composite measure that accounts for temperature, humidity, and wind chill; Dew Point, the temperature at which air becomes saturated and which indicates moisture content; Relative Humidity, the percentage of water vapor present relative to the maximum possible at that temperature; Precipitation, any form of water liquid or frozen falling from the sky; Precipitation Probability, the likelihood (as a percentage) of precipitation occurring; Snowfall Amount, the depth of snow accumulation; Snow Depth, the total height of snow on the ground; Wind Gust, short bursts of high wind speed exceeding the average; Wind Speed, the sustained wind velocity typically measured in km/h or mph; Wind Direction, the compass bearing (in degrees from true north) from which the wind originates; Sea Level Pressure, atmospheric pressure corrected to sea level; Cloud Cover, the fraction of the sky obscured by clouds expressed as a percentage; Visibility, the maximum distance at which objects or lights can be clearly seen; Solar Radiation, the instantaneous power of sunlight reaching the surface in W/m2; Solar Energy, the cumulative solar irradiance over a specified period in kWh/m2; and UV Index, a standardized scale indicating the strength of ultraviolet radiation.
Because our dataset is both large and subject to irregular fluctuations, it is difficult to discern the interdependencies among variables by simple inspection. Traditional statistical analyses often fall short when faced with the non-linear and multi-factor relationships present in meteorological data. To address this, machine learning algorithms were employed to detect subtle patterns, uncover complex associations, and highlight trends that would otherwise remain hidden.

4.1. Chi-Square Test

The Chi-Square Test [38] is a nonparametric technique used to determine whether two categorical variables are related. In load-forecasting applications, continuous weather measurements are first converted into discrete bins (for example, temperature grouped into “low”, “medium”, and “high”), and energy usage is similarly categorized into consumption ranges. The test then computes a Chi-Square statistic by comparing the observed frequencies in each combination of weather-bin and load-bin with the expected frequencies under the assumption of independence. A larger Chi-Square value indicates a stronger association between a given weather feature and energy consumption.
Procedure:
  • Split weather variables (temperature, UV index, etc.) into a set of categories.
  • Bin the target: Organize historical load values into energy consumption intervals.
  • Compute frequencies: Tally how often each weather-category/load-category pair occurs.
  • Calculate statistic: For each feature, sum over all bins using the term:
O b s e r v e d E x p e c t e d E x p e c t e d
5.
Rank features: Order variables by their Chi-Square scores higher scores indicate greater predictive value.
When applied to our dataset, the highest scores belong to temperature, feels-like temperature, UV index, sea level pressure, solar energy, and solar radiation (see Table 3). These results imply that these six weather factors most strongly influence when and how much energy is consumed for heating, cooling, and solar-powered operations. Dew point and humidity produced moderate Chi-Square values, whereas visibility and cloud cover scored relatively low, showing only weak associations with load. All examined variables yielded p-values below the usual significance threshold, confirming that none of these relationships arose by chance.
By selecting top-scoring features such as temperature and solar related variables leaner and more accurate forecasting models can be constructed, while safely omitting the least influential factors.

4.2. Minimum Redundancy Maximum Relevance (MRMR)

The MRMR [39] approach seeks a set of inputs that are both highly informative about the target (energy consumption) and as mutually non-redundant as possible. It does so by ranking features based on mutual information scores, which quantify how much knowing one variable reduces uncertainty about another.
Implementation steps:
  • Standardize each numeric feature:
S t d V = V μ ( V ) σ ( V )
where V is the original value, μ(V) its mean, and σ(V) its standard deviation. This puts all features on the same scale, preventing any single unit-scale variable from dominating the MI calculation.
2.
Compute mutual information between each feature and the energy-consumption target.
3.
Rank features by high relevance (large MI) and low redundancy (little overlap with already-selected features).
Table 4 shows the ranked results. In summary:
  • Temperature and feels-like temperature emerge as the strongest predictors.
  • Dew point, sea level pressure, and humidity have moderate importance.
  • Solar energy and visibility register relatively low relevance scores.
Table 4. Results of MRMR feature selection.
Table 4. Results of MRMR feature selection.
FeatureTemperatureFeels-LikeUV IndexSea Level PressureSolar EnergySolar RadiationVisibility
Score0.3860.3810.0520.2030.0730.1530.040
FeatureTemperatureFeels-LikeUV IndexSea Level PressureSolar EnergySolar Radiation
Score0.2420.1590.1390.1060.1140.085
Because building heating and cooling demand is sensitive to temperature-related inputs, those variables drive most of the predictive power. Less-relevant features can be dropped to simplify the model without sacrificing accuracy.

4.3. Correlation Matrix

A correlation matrix (CM) [40] in Figure 8 shows pairwise linear relationships between numeric features and the energy-consumption. Each entry is the Pearson correlation coefficient, ranging from −1 (perfect inverse) through 0 (no linear relationship) to +1 (perfect direct).
The analysis yields the following results:
  • Sea level pressure has the strongest negative correlation (−0.34), implying that higher pressures generally coincide with lower energy use (likely due to milder weather).
  • Temperature, feels-like temperature, and dew point have strong positive correlations (approximately +0.32, +0.31, and +0.22), reflecting increased cooling demand under warmer or more humid conditions.
  • Solar variables (UV index, solar radiation, solar energy) exhibit modest positive correlations (around +0.17).
  • Cloud cover and visibility feature near-zero correlations, indicating negligible direct effects on consumption.
This correlation analysis reinforces that temperature and related humidity measures are dominant drivers of energy demand, while factors such as cloud cover and visibility play only minor roles.
The top-ranked features for each selection method are: CM: sea level pressure, temperature, feels-like temperature, and dew point. MRMR: temperature, feels-like temperature, dew point, and sea level pressure. Chi-Square: temperature, feels-like temperature, UV index, and sea level pressure.
Each method serves a different role in the proposed approach and is paired with a different machine learning technique or statistical method.
The proposed framework is not dependent on any specific set of weather variables. All meteorological features used in this study (temperature, relative humidity, sea-level pressure, dew point, UV index, etc.) are publicly available through standard weather services and do not require specialized sensors. The pipeline can operate with smaller or alternative feature sets, depending on what is available in a given region. The feature selection component is included to demonstrate how the model behaves when richer predictor sets exist, but the framework remains fully functional even when only the commonly available variables (e.g., temperature, pressure, humidity, wind, precipitation) are present.

5. Discussion and Comparison of the Results: Overall Consumption Prediction vs. Aggregated Appliance Forecasts

This paper introduced a new cloud-enabled hybrid machine learning–statistical framework for short-term load forecasting in smart buildings. The proposed approach integrates the strengths of both machine learning algorithms and statistical models to maximize forecasting accuracy while simultaneously leveraging the reliability, high processing speed, and cost efficiency of cloud computing. By employing a parallel multi-cloud architecture with a two-out-of-three voting mechanism, the framework achieves even greater speed and reliability compared to conventional single-cloud deployments. This architecture also provides a scalable and fault-tolerant foundation that enables researchers to extend the methodology beyond load forecasting into other critical areas of the power grid, such as fault detection, protection, and predictive maintenance monitoring.
In this study, the next hour of electrical consumption was predicted, corresponding to 12 time steps, with each step representing a 5 min interval. The performance of the proposed method is compared against several well-known forecasting techniques including: Linear Regression (LR), Random Forest (RF), Decision Tree (DT), Support Vector Machine (SVM), SARIMAX, Extreme Gradient Boosting (XGBoost), Gated Recurrent Unit (GRU), Long Short-Term Memory (LSTM), and Deep Belief Network (DBN). These comparative results are presented in Table A1, Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10 and Table A11 (in the Appendix A), highlighting the improvements achieved through the dual-stage trend–remainder modeling process in combination with the robust cloud-based deployment.
Table A1 and Figure 9: Overall consumption, Table A2 and Figure 10: FAU (Forced Air Unit) Downstairs consumption, Table A3 and Figure 11: AC (Air Conditioning) Downstairs consumption, Table A4 and Figure 12: Refrigerator consumption, Table A5 and Figure 13: AC Upstairs consumption, Table A6 and Figure 14: Laundry/Dryer consumption, Table A7 and Figure 15: Dishwasher/Garbage Disposal consumption, Table A8 and Figure 16: Microwave consumption, Table A9 and Figure 17: FAU Upstairs consumption, Table A10 and Figure 18: Kitchen GFI (Ground Fault Circuit Interrupter)/Toaster consumption, Table A11 and Figure 19: General (miscellaneous) load consumption and Table A12 and Figure 20: Sum of results of Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10 and Table A11.
It is important to note that occasional negative consumption values may appear in the dataset. These values do not indicate that appliances such as air conditioners are generating energy; rather, they are a result of the building’s integration with rooftop solar panels. During periods of high solar production particularly when household energy demand is low excess electricity is exported back to the grid.
Five well-established evaluation metrics (Table 5) were employed to compare the results. Since all models were evaluated on the same single-building dataset, normalization measures (e.g., cvRMSE, NMBE) were not applied, as they become relevant only in multi-building comparisons, in accordance with ASHRAE Guideline 14 and IPMVP.
Although the model is trained and evaluated at a 5 min resolution, the output for each horizon is a 12-point vector covering the next hour. For deployment, building EMS/DR (EMS = Energy Management System, DR = Demand Response) controllers can map these to their operating cadence by forming block averages (10/15/30/60 min) or energy-integrated setpoints. This hierarchical use of the same forecasts attenuates short-term noise and matches actuator latencies, while retaining sub-hour detail for rapid interventions (e.g., staggered EV charging). Integrating the proposed forecasting framework with demand response (DR) controllers or optimization models could therefore enhance load flexibility, grid stability, and energy-cost savings an aspect identified as a promising direction for future work.
Aggregate vs. Sum of Appliances:
On the aggregate meter, SSRLR is the best model on every metric (RMSLE = 0.04311, RMSE = 0.07807, MAE = 0.05850). The next best RMSLE is RF = 0.06610, so SSRLR improves AF, RMSLE by ~34.8%.
For the sum of appliance forecasts, the best model is RF with RMSLE = 0.07128 (MAE = 0.13657, RMSE = 0.14152) and second is the SSRLR with RMSLE = 0.08009. Comparing the two strategies best vs. best, AF: SSRLR (0.04311) beats sum of appliances RF (0.07128) by ~39.5% in RMSLE (the summer series is ~65% worse). This behavior is expected: when many appliances operate intermittently (e.g., microwave, dishwasher, small plugs), their individual errors can accumulate when the loads are summed, even if each model performs well.
In short, direct aggregate forecasting uses stable daily/weekly patterns in the total load, while the AL approach must first learn many on/off behaviors and then add them. Without hierarchical reconciliation (to force the sum to match the meter) or special handling of zeros/events, the AL sum can penalize small timing errors and stack relative errors.
Results of Per Appliance Forecasting:
Averaged across all appliances, SSRLR ranks first (avg RMSLE ≈ 0.042; RMSE/MAE show the same trend). RF and XGBoost are the strongest baselines. SSRLR is the best method for every single appliance; examples:
  • FAU Downstairs: RMSLE 0.01783, RMSE 0.02536
  • AC Downstairs: RMSLE 0.04838, RMSE 0.04962
  • Refrigerator: RMSLE 0.04195, RMSE 0.04103
  • AC Upstairs: RMSLE 0.04732, RMSE 0.04878
  • Laundry/Dryer: RMSLE 0.04137, RMSE 0.04238
  • Dishwasher/Garbage: RMSLE 0.04063, RMSE 0.04165
  • Microwave: RMSLE 0.04116, RMSE 0.04207
  • FAU Upstairs: RMSLE 0.04192, RMSE 0.07004
  • Kitchen GFI/Toaster: RMSLE 0.03949, RMSE 0.04051
  • General/Misc: SSRLR is best but errors are higher (RMSLE 0.06854, RMSE 0.10892) because this circuit mixes many small, irregular uses.
By the appliance type:
  • Spiky/episodic loads (Microwave, Dishwasher/Garbage, Laundry) are handled well by SSRLR, which efficiently manages short bursts.
  • HVAC loads (AC and FAU) also favor SSRLR, showing that it captures both seasonality and start/stop behavior.
  • General/Misc is the hardest because it blends many small and random loads: SSRLR shows the best accuracy.
Model Ranking and Error Profiles:
A consistent ranking appears across aggregate, appliances, and the sum:
  • SSRLR is best overall (aggregate and per appliance averages), strong on both relative (RMSLE) and absolute (RMSE/MAE) errors.
  • RF/XGBoost are reliable runner ups; they capture nonlinear effects and are stable across devices.
  • SARIMAX is a good seasonal baseline, but weaker on sharp spikes.
  • LSTM and GRU do not dominate here, likely due to intermittency, short forecast horizons, and limited effective data per appliance.
  • LR/DT/DBN/SVM perform poorly for short-term load forecasting in this setting.
This order holds for the aggregate and appliance averages. In the sum, RF wins, but it still trails the aggregate SSRLR by a wide margin.
Key Takeaways:
  • Direct aggregate forecasting (SSRLR) is much more accurate than the sum of appliance forecasts on this dataset (~39.5% lower RMSLE, best vs. best).
  • SSRLR is the most consistent model at both levels; RF/XGBoost are strong baselines.
It may look surprising that SSRLR wins for every appliance, yet RF wins after summing the results. The reason for that is the shift of the error pattern when summing consumption of many appliance forecasts: small positively correlated residuals from several SSRLR appliance models can add up for the same time points (for example, around evening peaks). RF tends to produce smoother appliance predictions with less shared timing error across devices, so when added some errors cancel each other. In addition, hierarchical reconciliation was not applied; therefore, perfect per-device calibration does not guarantee that the aggregated total matches the main meter. Because RMSLE on the summed series is sensitive to synchronized timing errors, RF ends up with a lower summed RMSLE (0.07128) than SSRLR (0.08009). Adding reconciliation and aggregate aware training would likely reduce and may remove this gap.

6. Conclusions

This paper introduced a cloud-enabled hybrid ML–statistical framework for 5 min residential load forecasting that blends Seasonal ARIMAX, Random Forest, and LSTM within a residual-correction architecture (SSRLR) and deploys redundant instances across multiple clouds with a two-out-of-three voting scheme. Using sub-metered data from a residential home in Summerlin, Las Vegas (2022), aggregate forecasting was evaluated against appliance-level forecasting. The evidence is clear: on the aggregate series, SSRLR is best across all metrics (RMSLE = 0.04311, RMSE = 0.07807, MAE = 0.05850), outperforming the next-best model (RF, RMSLE = 0.06610) by ~34.8%. For the appliance-level results, the RF model performs best (RMSLE = 0.07128), followed by the SSRLR model (RMSLE = 0.08009). However, the aggregate SSRLR model still outperforms the appliance-level results by up to approximately 39.5% in terms of RMSLE. This gap aligns with expectations for many intermittent end-uses: per appliance models can produce positively correlated residuals that accumulate when added, while the aggregate model exploits stable daily/weekly total load patterns. Importantly, RMSLE is the most informative headline metric in this setting because many appliances operate near zero and fairly scores small absolute errors; RMSE/MAE confirm the same ranking in absolute terms.
Practically, if the goal is maximum accuracy with minimal complexity, a single SSRLR model for the aggregate prediction is the preferred choice. The proposed approach supports real-time, fault-tolerant forecasting for smart-city energy management. Its cloud redundancy mechanism ensures operational continuity during single-cloud failures, making it suitable for deployment in distributed residential and micro-grid environments. The framework’s modular design allows utilities to balance accuracy, latency, and cost by tuning cloud resources and model complexity.
Future work will extend the framework to multi-household, high-load buildings and cross-climate datasets, explore transfer learning for unseen dwellings, and integrate edge-cloud coordination for improved energy-efficiency and privacy. Further research will also focus on adaptive weighting and strengthen interpretability for decision-support systems.
Overall, this research demonstrates that hybrid multi-cloud forecasting can provide fine-grained, resilient, and scalable solutions for next-generation smart-residential infrastructures.

Author Contributions

Conceptualization, K.H.B., E.E.R. and Y.B.; methodology, K.H.B.; software, K.H.B.; validation, K.H.B., E.E.R. and Y.B.; formal analysis, K.H.B. and E.E.R.; investigation, K.H.B., E.E.R. and Y.B.; resources, K.H.B., E.E.R. and Y.B.; data curation, K.H.B. and Y.B.; writing original draft preparation, K.H.B., E.E.R. and Y.B.; writing review and editing, K.H.B., E.E.R. and Y.B.; visualization, K.H.B.; supervision, K.H.B., E.E.R. and Y.B.; project administration, K.H.B., E.E.R. and Y.B.; funding acquisition, K.H.B., E.E.R. and Y.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data will be made available upon reasonable request.

Acknowledgments

The authors would like to express their sincere appreciation to Zeinab Farrokhi (Z.F) for her significant contributions during the revision process, including her substantial assistance in responding to the reviewers’ comments and editing the manuscript.

Conflicts of Interest

The authors declare that they have no conflicts of interest to this work.

Abbreviations

The following abbreviations are used in this manuscript:
MLMachine Learning
MSEMean Squared Error
LFLoad Forecasting
MAPEMean Absolute Percentage Error
MAEMean Absolute Error
STLFShort-Term Load Forecasting
RMSLERoot Mean Squared Logarithmic Error
MTLFMid-Term Load Forecasting
RMSERoot Mean Squared Error
LTLFLong-Term Load Forecasting
LRLinear Regression
RFRandom Forest
DTDecision Tree
GRUGated Recurrent Unit
SVMSupport Vector Machine
DBNDeep Belief Network
ARIMAAutoregressive Integrated Moving Average
SSRLRSparse SARIMAX, Random Forest, LSTM, and Remainder Prediction
SARIMAXSeasonal Autoregressive Integrated Moving Average Exogenous
XGBoostExtreme Gradient Boosting
LSTMLong Short-Term Memory
CMCorrelation Matrix
MRMRMinimum Redundancy Maximum Relevance
RPRemainder Prediction
PPPrimary Prediction
ALAppliance-level
AFAggregate forecasts

Appendix A

Table A1. Next 12 predicted (KW) points by different methods: overall consumption.
Table A1. Next 12 predicted (KW) points by different methods: overall consumption.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
11.5690.9110.5921.6340.8530.9510.6210.9380.8710.7820.650
22.1040.8870.5451.6100.7030.2790.6020.9230.6990.7750.646
32.1041.0400.5161.5760.6430.5130.5690.9010.6370.7680.752
41.5690.9411.2941.9411.9633.1800.6651.1121.9681.7301.840
52.1040.8840.8121.6580.8920.9460.6370.9530.890.8830.846
61.5690.8880.5801.5730.6430.5130.5650.8990.6430.6120.590
71.5690.9200.5611.5770.6310.5130.5700.9020.6320.6120.590
81.5690.9270.5351.6240.8540.7880.6130.9320.8330.7080.625
92.6391.0500.5351.6340.8530.9510.6210.9380.8650.7950.664
102.1040.9581.2472.0041.7081.0310.6421.1431.711.7141.732
112.6390.9020.7891.6890.9881.0190.6550.9720.9880.9890.989
121.5690.9110.5751.5960.7380.7940.5890.9140.7370.7010.699
Table A2. Next 12 predicted (KW) points by different methods: FAU (forced air unit) downstairs.
Table A2. Next 12 predicted (KW) points by different methods: FAU (forced air unit) downstairs.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0175130.0089020.0071340.0124250.0069820.0091120.0076980.0082350.0070910.0067850.00601
20.0184250.0095030.007210.0128330.0069910.0096210.0078950.0085010.0072540.0069450.006043
30.0179010.0076250.0076250.0123320.0069510.0094320.009210.0071930.0083950.0068510.005963
41.2023310.7312990.5571280.9835720.6452210.712940.589320.6385210.5521180.5012370.477047
50.3912540.2513320.1735550.2884130.2108920.2391440.1998130.2203350.1766240.1612250.145373
60.0183110.0091080.0073240.0126410.0069140.0094050.0078090.008290.0071140.0068310.006037
70.0188120.009610.0075010.0130010.0072250.0098320.0080140.0085020.0073340.0070090.006147
80.0185010.0092150.0072540.0127210.0070010.0096110.0079210.0084110.0072090.0069820.00611
90.01890.00970.0076050.013220.0073320.0099550.0082250.008720.0074450.007120.00617
101.2104220.7402910.563210.9905120.6532220.7210080.595120.6428920.5572310.5031880.477707
110.390120.2509920.1728010.2871340.2099210.2385510.1989230.2198320.1758320.1605340.145263
120.018250.009020.0071110.012520.0069520.0092550.0077150.0081750.0070030.0068220.006033
Table A3. Next 12 predicted (KW) points by different methods: air conditioning downstairs.
Table A3. Next 12 predicted (KW) points by different methods: air conditioning downstairs.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0023010.0012050.0009240.001610.0008510.0011320.0009670.0010310.0009120.0008430.00078
20.0022980.0012110.0009190.0015990.0008450.0011270.0009630.0010280.0009080.0008390.00078
30.0021150.0010990.0008420.0015030.0007990.0010460.0008920.000950.0008380.0007720.0007
40.0020690.0010670.0008260.0014690.000780.0010210.000870.0009280.000820.0007550.00069
50.0019860.0010250.0007930.0014090.0007480.0009790.0008340.000890.0007870.0007240.00066
60.0020830.0010740.000830.0014530.0007720.0010090.000860.0009190.0008150.0007470.000693
70.0023520.0012150.0009380.0016440.0008690.0011540.0009850.0010510.0009330.0008650.00081
80.0018580.0009610.0007420.0012980.0006880.0009060.0007740.0008270.0007370.0006760.000643
90.0020670.0010720.000830.001450.000770.0010060.0008580.0009170.0008130.0007460.00071
100.0021570.001120.0008650.0015150.0008050.0010510.0008970.0009580.0008510.0007810.00074
110.0022760.0011830.0009140.0016190.0008610.0011240.000960.0010280.0009130.0008450.000773
120.0022960.0011960.0009220.0016350.0008710.0011370.000970.0010390.0009210.0008520.000783
Table A4. Next 12 predicted (KW) points by different methods: AC upstairs.
Table A4. Next 12 predicted (KW) points by different methods: AC upstairs.
Method PointDBNLSTMSARIMAXLRRFDTSVM GRUXGBOOSTSSRLRActual Data
1−0.00108−0.00052−0.00041−0.0008−0.00033−0.00049−0.00045−0.00047−0.00034−0.0003−0.00036
2−0.00106−0.00051−0.0004−0.00078−0.00033−0.00049−0.00044−0.00047−0.00033−0.00029−0.00035
3−0.00113−0.00054−0.00043−0.00082−0.00034−0.00051−0.00046−0.00049−0.00035−0.00031−0.00038
4−0.00137−0.00066−0.00052−0.00099−0.00042−0.00062−0.00056−0.00059−0.00042−0.00037−0.00046
5−0.00125−0.0006−0.00047−0.0009−0.00038−0.00056−0.00051−0.00054−0.00038−0.00034−0.00042
6−0.00123−0.00059−0.00046−0.00089−0.00037−0.00055−0.0005−0.00053−0.00038−0.00033−0.00041
7−0.00097−0.00046−0.00036−0.0007−0.00029−0.00044−0.00039−0.00042−0.0003−0.00026−0.00032
8−0.00142−0.00068−0.00054−0.00104−0.00043−0.00064−0.00058−0.00062−0.00044−0.00038−0.00048
9−0.00161−0.00077−0.00061−0.00117−0.00049−0.00072−0.00065−0.00069−0.00049−0.00043−0.00054
10−0.00164−0.00079−0.00062−0.00119−0.0005−0.00074−0.00067−0.00071−0.0005−0.00044−0.00055
11−0.00134−0.00064−0.00051−0.00098−0.00041−0.00061−0.00055−0.00059−0.00042−0.00036−0.00045
12−0.00155−0.00074−0.00059−0.00113−0.00047−0.0007−0.00063−0.00067−0.00048−0.00042−0.00052
Table A5. Next 12 predicted (KW) points by different methods: refrigerator.
Table A5. Next 12 predicted (KW) points by different methods: refrigerator.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSXLRActual Data
10.0189450.0091180.0072040.0133750.0070010.0094020.0080150.0084650.0072930.0069810.006727
20.0188320.0090650.0071660.0132880.0069420.0093510.0079840.0084320.0072590.0069510.006703
30.018840.0090710.007170.0132960.0069460.0093560.0079880.0084370.0072630.0069550.006703
40.0178580.0086040.006810.012640.0066010.0088920.00760.0080420.0068970.0065810.006343
50.0182010.0087710.006940.0128940.006740.0090860.0077690.008230.007060.0067320.00648
60.018260.00880.0069620.0129330.0067610.0091150.0077940.0082550.0070810.0067510.0065
70.0182730.0088070.0069680.0129440.0067680.0091220.00780.0082610.0070870.0067560.0065
80.0182840.0088120.0069710.012950.0067710.0091260.0078040.0082650.007090.0067590.006503
90.0183230.0088320.0069860.0129780.0067880.0091450.007820.008280.0071050.0067720.006517
100.0176710.0085070.0067180.0125060.006550.0087940.0075270.0079740.0068330.0065280.006193
110.0187660.009020.0071390.0132330.0069150.0093120.0079550.0084070.0072380.0069310.006593
120.01890.0090910.0072010.013350.0069860.0093810.008010.008460.007290.0069810.00664
Table A6. Next 12 predicted (KW) points by different methods: laundry/dryer.
Table A6. Next 12 predicted (KW) points by different methods: laundry/dryer.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0108250.0052730.0042420.0079170.0040080.0055360.0048210.0051290.0041810.0039560.003603
20.010910.0053160.0042770.0079880.0040450.0055880.0048660.0051750.0042170.003990.003637
30.0107890.0052560.0042270.0078780.0039930.0055220.0048040.0051130.004160.0039350.00358
40.0108330.0052780.0042440.0079130.0040120.0055390.0048220.0051330.0041760.0039510.003607
50.0105370.0051310.0041270.007690.0038980.0053820.0046860.0049830.0040580.0038360.00351
60.0106780.0052020.0041820.0078240.0039640.0054640.0047620.0050630.0041180.0038930.003563
70.0105790.0051520.0041420.0077380.0039210.0054140.0047180.0050180.0040790.0038560.003527
80.0105840.0051540.0041450.0077440.0039240.0054180.0047210.0050210.0040820.0038580.003527
90.0106060.0051660.0041540.007760.0039340.005430.0047310.0050320.0040910.0038670.00352
100.0105820.0051540.0041440.0077420.0039230.0054170.004720.005020.0040810.0038570.003527
110.0109190.005320.0042810.0080030.004050.0055930.0048720.0051830.0042240.0039960.003637
120.0107320.0052270.0042070.0078530.0039720.005490.0047870.005090.0041420.0039190.003567
Table A7. Next 12 predicted (KW) points by different methods: dishwasher/garbage disposal.
Table A7. Next 12 predicted (KW) points by different methods: dishwasher/garbage disposal.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0022560.0011210.0008740.0016610.0007520.0010510.0009420.0010030.0007890.0007050.00075
20.0024520.0012220.0009530.0017990.0008210.0011520.001030.0010980.0008630.0007730.000817
30.0022240.0011030.0008590.0016360.000740.0010330.0009260.0009850.0007750.0006930.00074
40.0048310.0023740.0018570.0035550.0015980.0022430.0020260.0021550.0016920.0015150.001643
50.0030740.0015140.0011860.0022630.0010220.0014350.0012960.0013790.0010820.0009680.001047
60.0023650.0011740.0009130.0017190.0007770.0010930.000980.0010490.0008230.0007390.000787
70.0025990.0012830.0009970.0018920.0008570.00120.0010770.0011510.0009020.0008090.000863
80.0021290.001050.0008140.0015520.0007040.0009850.0008820.0009440.000740.0006640.000707
90.0020970.0010340.0008020.0015280.0006930.000970.0008690.000930.0007290.0006540.000697
100.0047110.0023210.0018150.0034660.0015560.0021840.0019740.00210.001650.0014770.00161
110.0028570.0014140.0011060.0021020.0009450.0013260.001190.0012720.0010.0008970.000977
120.0022640.001120.0008740.0016620.0007520.001050.0009420.0010040.000790.0007050.00075
Table A8. Next 12 predicted (KW) points by different methods: microwave.
Table A8. Next 12 predicted (KW) points by different methods: microwave.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0032180.0018580.0014580.0023220.0011880.0015660.0015010.0015980.0013820.0013180.00108
20.0033120.0019270.001510.0023870.0012390.0016450.0015550.0016780.0014320.0013620.0011266
30.0030990.0017740.0013870.0022240.0011380.0015120.0014290.0015350.0013160.0012520.0010433
40.0031780.0018470.0014390.0022990.0011910.0015680.0014810.0015890.0013640.001310.0010733
50.0032750.0018980.0014870.0023640.0012210.001610.0015320.0016540.001410.0013430.00111
60.0032140.0018640.0014590.0023270.0012060.001590.0015030.0016240.0013950.0013390.0010966
70.0033290.001920.0015070.0024020.0012280.0016310.0015520.0016650.0014280.0013630.0011166
80.0031120.0018140.0014170.0022650.0011690.0015460.001460.0015770.0013420.0012980.0010733
90.00320.0018430.0014320.0022950.0011840.0015610.0014740.0015940.0013690.0013130.0010766
100.0032520.0018920.0014810.0023620.0012250.0016160.0015360.0016370.0014030.0013470.0011133
110.0032840.001910.0014970.0023890.0012280.0016310.001540.0016530.001420.0013630.0011166
120.0032860.0019090.0014990.0023870.0012320.0016340.0015430.0016540.0014210.0013540.00111
Table A9. Next 12 predicted (KW) points by different methods: FAU upstairs.
Table A9. Next 12 predicted (KW) points by different methods: FAU upstairs.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0238950.011130.0091720.0178820.0086120.0125440.0112760.0118690.0094420.0087720.00794
20.023660.0110160.009080.0177370.0085320.0124520.0111910.011780.0093630.0086980.007887
30.023750.0110620.0091150.0178010.0085680.0124930.0112270.0118210.0093920.0087280.007917
41.843650.8692250.6941181.3854270.6573110.9518820.8641130.9102250.7024470.6358220.62757
50.3205350.1557730.1241930.2441530.1158670.1681520.1525590.1610370.124910.1129670.109373
60.024040.0111910.0092180.0179750.0086650.0126050.0113320.0119290.0094910.0088190.008013
70.023710.0110430.0091070.0177680.0085540.0124760.0112120.0118020.0093760.0087130.007903
80.023990.0111770.0092070.017950.0086550.0125950.0113230.011920.0094810.008810.008007
90.023870.0111180.009160.0178680.0086080.0125440.0112760.0118710.0094370.0087710.007967
101.847820.8710250.6954521.3883620.6590140.9541930.8661740.9123560.7041230.637310.629273
110.3178340.1544620.1231640.2422730.1149520.1668830.1514230.1598220.1239910.1121420.10793
120.0235650.0109810.009050.0176960.0085130.0124260.011170.0117580.0093380.0086720.00785
Table A10. Next 12 predicted (KW) points by different methods: kitchen GFI (ground fault circuit interrupter)/toaster.
Table A10. Next 12 predicted (KW) points by different methods: kitchen GFI (ground fault circuit interrupter)/toaster.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
10.0000330.0000180.0000140.0000270.0000120.0000190.0000170.0000180.0000140.0000130.0000133
20.000060.000030.0000240.0000480.0000210.0000330.0000290.0000310.0000240.0000220.00002
30.0001080.0000520.0000410.0000860.0000380.0000580.0000510.0000550.0000430.0000390.0000367
40.0001760.0000840.0000670.000140.0000610.0000950.0000840.000090.0000710.0000650.00006
50.0001680.000080.0000640.0001340.0000590.000090.0000810.0000870.0000690.0000630.0000567
60.000170.0000810.0000650.0001360.0000590.0000910.0000820.0000880.000070.0000640.0000567
70.000130.0000620.000050.0001040.0000450.0000690.0000630.0000670.0000530.0000480.0000433
80.000330.0001550.0001240.0002640.0001180.0001760.0001590.0001680.0001330.000120.00011
90.0002490.000120.0000950.00020.0000890.0001330.000120.0001270.0001010.0000920.0000833
100.0002890.0001390.0001110.0002320.0001030.0001540.0001390.0001470.0001160.0001050.0000967
110.0002190.0001060.0000840.0001760.0000790.0001180.0001060.0001120.0000890.0000810.0000733
120.000090.0000440.0000350.0000720.0000320.000050.0000450.0000470.0000370.0000340.00003
Table A11. Next 12 predicted (KW) points by different methods: general (miscellaneous) household load.
Table A11. Next 12 predicted (KW) points by different methods: general (miscellaneous) household load.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
11.8221040.919240.6761181.3927810.2101250.5349250.613020.572480.2251940.1968330.624233
21.8073260.9104810.6698391.380350.2080120.5296890.6076230.567470.2232170.1950840.619827
32.11541.061390.7806261.614830.2435840.6205520.712280.6659820.2618830.2290290.72587
42.109371.058240.7783011.6100660.2429050.6187360.7101840.6640310.2608150.2280990.723123
51.690620.8484260.6240251.290160.1947170.4955860.5684770.5305340.2083090.1820780.579567
61.645580.8258610.607281.2558440.1895490.4824860.5539160.516620.2030050.1776660.56406
71.643750.824940.6066061.2542940.1893130.4818860.5532260.5159670.2026850.1773830.563447
81.747680.8770210.6446771.333290.2011330.5118070.5874890.5477830.214950.188150.59914
91.861490.9339680.6859211.429440.2154550.5478720.6289090.5861440.2314090.2025210.63872
101.784590.8946060.6570661.36910.2063240.5246640.6026420.5615010.2207270.1931420.612323
112.114041.060720.7800991.613760.2434480.6202240.7119050.6656440.2617530.2289130.723473
121.968190.9888080.7266221.504080.2268120.5781060.6636480.6198330.2432940.2132530.673247
Table A12. Next 12 predicted (KW) points by different methods: overall consumption—summation of Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11 and Table A12.
Table A12. Next 12 predicted (KW) points by different methods: overall consumption—summation of Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11 and Table A12.
Method PointDBNLSTMSARIMAXLRRFDTSVMGRUXGBOOSTSSRLRActual Data
14.4028090.9962311.0871944.9575310.7532810.9248441.0015380.8299091.0336350.7409850.650
24.5273831.0222951.1013315.0902350.7606270.9409431.0153610.8441711.0406050.7470440.646
34.8381781.0085711.2253775.3060840.8441951.012741.0832040.9187641.1621960.8314970.752
49.847921.6455981.9929399.408092.0220732.1114692.0549011.6592532.0456872.1832671.840
55.5800261.1137221.25695.8377240.9827091.2143481.2367531.0033341.2091441.0241850.846
64.5467160.9768131.059355.123110.7272870.9047050.9681670.8131760.9867950.7096720.590
74.6117371.044931.0699445.270210.7400320.9264681.0051230.8295260.9974190.7194650.590
84.9752470.9913441.1513285.5119580.7781070.957121.0091190.8652461.0565230.7556950.625
94.8540311.0043691.1599565.3926240.7892480.9618161.0234750.8700841.0782460.7737570.664
109.8112751.6714291.9062559.4980421.9591332.0573941.9989761.5986881.9399432.1112711.732
116.2113541.2466431.4654676.4822641.1196081.352831.3888541.1365811.4039611.1594550.989
124.6063891.004311.1493195.1539720.7910930.9610081.0349050.8687961.0919160.7838850.699

References

  1. Wang, S.; Deng, X.; Chen, H.; Shi, Q.; Xu, D. A bottom-up short-term residential load forecasting approach based on appliance characteristic analysis and multi-task learning. Electr. Power Syst. Res. 2021, 196, 107233. [Google Scholar] [CrossRef]
  2. Shabbir, N.; Vassiljeva, K.; Nourollahi Hokmabad, H.; Husev, O.; Petlenkov, E.; Belikov, J. Comparative Analysis of Machine Learning Techniques for Non-Intrusive Load Monitoring. Electronics 2024, 13, 1420. [Google Scholar] [CrossRef]
  3. Noroozi, Z.; Orooji, A.; Erfannia, L. Analyzing the impact of feature selection methods on machine learning algorithms for heart disease prediction. Sci. Rep. 2023, 13, 22588. [Google Scholar] [CrossRef]
  4. Hassanpouri Baesmat, K.; Farrokhi, Z.; Chmaj, G.; Regentova, E.E. Parallel Multi-Model Energy Demand Forecasting with Cloud Redundancy: Leveraging Trend Correction, Feature Selection, and Machine Learning. Forecasting 2025, 7, 25. [Google Scholar] [CrossRef]
  5. Farrokhi, Z. AI-Powered Grid Analytics for Sustainable Energy Management in ESG Frameworks. In Data-Driven ESG Strategy Implementation Through Business Intelligence; IGI Global Scientific Publishing: Hershey, PA, USA, 2026; pp. 279–318. [Google Scholar] [CrossRef]
  6. Mohtasham, F.; Pourhoseingholi, M.; Hashemi Nazari, S.S.; Kavousi, K.; Zali, M.R. Comparative analysis of feature selection techniques for COVID-19 dataset. Sci. Rep. 2024, 14, 18627. [Google Scholar] [CrossRef] [PubMed]
  7. Ali, M.; Islam, S.; Uddin, M.N.; Uddin, A. A conceptual IoT framework based on Anova-F feature selection for chronic kidney disease detection using deep learning approach. Intell.-Based Med. 2024, 10, 100170. [Google Scholar] [CrossRef]
  8. Souza, F.; Premebida, C.; Araújo, R. High-order conditional mutual information maximization for dealing with high-order dependencies in feature selection. Pattern Recognit. 2022, 131, 108895. [Google Scholar] [CrossRef]
  9. Qu, K.; Xu, J.; Hou, Q.; Qu, K.; Sun, Y. Feature selection using Information Gain and decision information in neighborhood decision system. Appl. Soft Comput. 2023, 136, 110100. [Google Scholar] [CrossRef]
  10. Muthukrishnan, R.; Rohini, R. LASSO: A feature selection technique in predictive modeling for machine learning. In Proceedings of the 2016 IEEE International Conference on Advances in Computer Applications (ICACA), Coimbatore, India, 24 October 2016; pp. 18–20. [Google Scholar] [CrossRef]
  11. Altarabichi, M.G.; Nowaczyk, S.; Pashami, S.; Mashhadi, P.S. Fast Genetic Algorithm for feature selection A qualitative approximation approach. Expert Syst. Appl. 2023, 211, 118528. [Google Scholar] [CrossRef]
  12. Luong, H.H.; Tran, T.T.; Van Nguyen, N.; Le, A.D.; Nguyen, H.T.T.; Nguyen, K.D.; Tran, N.C.; Nguyen, H.T. Feature Selection Using Correlation Matrix on Metagenomic Data with Pearson Enhancing Inflammatory Bowel Disease Prediction. In International Conference on Artificial Intelligence for Smart Community; Ibrahim, R., Porkumaran, K., Kannan, R., Mohd Nor, N., Prabakar, S., Eds.; Lecture Notes in Electrical Engineering; Springer: Singapore, 2022; Volume 758. [Google Scholar] [CrossRef]
  13. Rosidin, S.; Muljono; Fajar Shidik, G.; Zainul Fanani, A.; Al Zami, F.; Purwanto. Improvement with Chi Square Selection Feature using Supervised Machine Learning Approach on COVID-19 Data. In Proceedings of the 2021 International Seminar on Application for Technology of Information and Communication (iSemantic), Semarangin, Indonesia, 18–19 September 2021; pp. 32–36. [Google Scholar] [CrossRef]
  14. Zhao, Z.; Anand, R.; Wang, M. Maximum Relevance and Minimum Redundancy Feature Selection Methods for a Marketing Machine Learning Platform. In Proceedings of the 2019 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Washington, DC, USA, 5–8 October 2019; pp. 442–452. [Google Scholar] [CrossRef]
  15. Jamali, H.; Dascalu, S.M.; Harris, F.C. AI-Driven Analysis and Prediction of Energy Consumption in NYC’s Municipal Buildings. In Proceedings of the 2024 IEEE/ACIS 22nd International Conference on Software Engineering Research, Management and Applications (SERA), Honolulu, HI, USA, 30 May–1 June 2024; pp. 277–283. [Google Scholar] [CrossRef]
  16. Baesmat, K.H.; Shokoohi, F.; Farrokhi, Z. SP-RF-ARIMA: A sparse random forest and ARIMA hybrid model for electric load forecasting. Glob. Energy Interconnect. 2025, 8, 486–496. [Google Scholar] [CrossRef]
  17. Baesmat, K.H.; Regentova, E.E.; Baghzouz, Y. A Hybrid Machine Learning–Statistical Based Method for Short-Term Energy Consumption Prediction in Residential Buildings. Energy AI 2025, 21, 100552. [Google Scholar] [CrossRef]
  18. Alam, M.M.; Latifi, S. A Systematic Review of Techniques for Early-Stage Alzheimer’s Disease Diagnosis Using Machine Learning and Deep Learning. J. Data Sci. Intell. Syst. 2025. Vol. 00(00) 1–13. Available online: https://ojs.bonviewpress.com/index.php/jdsis/article/view/5037/1573 (accessed on 24 November 2025). [CrossRef]
  19. Shaon, M.A.R.; Baghzouz, Y. Day-Ahead Residential Customer Load Forecasting Using Prophet. In Proceedings of the 2023 IEEE International Conference on Environment and Electrical Engineering and 2023 IEEE Industrial and Commercial Power Systems Europe (EEEIC/I&CPS Europe), Madrid, Spain, 6–9 June 2023; pp. 1–4. [Google Scholar] [CrossRef]
  20. Sina, L.B.; Secco, C.A.; Blazevic, M.; Nazemi, K. Hybrid Forecasting Methods A Systematic Review. Electronics 2023, 12, 2019. [Google Scholar] [CrossRef]
  21. Fattah, J.; Ezzine, L.; Aman, Z.; El Moussami, H.; Lachhab, A. Forecasting of demand using ARIMA model. Int. J. Eng. Bus. Manag. 2018, 10, 1847979018808673. [Google Scholar] [CrossRef]
  22. Kumari, S.; Muthulakshmi, P. SARIMA Model: An Efficient Machine Learning Technique for Weather Forecasting. Procedia Comput. Sci. 2024, 235, 656–670. [Google Scholar] [CrossRef]
  23. Ram, C.S.; Raj, M.; Chaturvedi, R.K. Boosting Time-Series Forecasting Accuracy with SARIMAX Seasonal Interval Automation. Procedia Comput. Sci. 2025, 260, 814–821. [Google Scholar] [CrossRef]
  24. Li, J.; Lei, Y.; Yang, S. Mid-long term load forecasting model based on support vector machine optimized by improved sparrow search algorithm. Energy Rep. 2022, 8 (Suppl. S5), 491–497. [Google Scholar] [CrossRef]
  25. Nooruldeen, O.; Baker, M.R.; Aleesa, A.M.; Ghareeb, A.; Shaker, E.H. Strategies for predictive power: Machine learning models in city-scale load forecasting, e-Prime—Advances in Electrical Engineering. Electron. Energy 2023, 6, 100392. [Google Scholar] [CrossRef]
  26. Wang, Y.-F.; Wang, M.Y.-F.; Tu, L.-Y. An Evaluation of Machine Learning Models for Forecasting Short-Term U.S. Treasury Yields. Appl. Sci. 2025, 15, 6903. [Google Scholar] [CrossRef]
  27. Kanagaraj, R.; Revanth, N.; Vishnu, K.; Saxena, R. Demand Forecasting with XGBoost and LSTM: An Exploratory Analysis. In Proceedings of the 2025 International Conference on Inventive Computation Technologies (ICICT), Kirtipur, Nepal, 23–25 April 2025; pp. 1216–1220. [Google Scholar] [CrossRef]
  28. Abbasimehr, H.; Shabani, M.; Yousefi, M. An optimized model using LSTM network for demand forecasting. Comput. Ind. Eng. 2020, 143, 106435. [Google Scholar] [CrossRef]
  29. Gomez-Rosero, S.; Capretz, M.A.M.; Mir, S. Transfer Learning by Similarity Centred Architecture Evolution for Multiple Residential Load Forecasting. Smart Cities 2021, 4, 217–240. [Google Scholar] [CrossRef]
  30. Li, Y.; Gao, Z.; Zhou, Z.; Zhang, Y.; Guo, Z.; Yan, Z. Abnormal Load Variation Forecasting in Urban Cities Based on Sample Augmentation and TimesNet. Smart Cities 2025, 8, 43. [Google Scholar] [CrossRef]
  31. Aldossary, M. Enhancing Urban Electric Vehicle (EV) Fleet Management Efficiency in Smart Cities: A Predictive Hybrid Deep Learning Framework. Smart Cities 2024, 7, 3678–3704. [Google Scholar] [CrossRef]
  32. Bin Kamilin, M.H.; Yamaguchi, S.; Bin Ahmadon, M.A. Radian Scaling and Its Application to Enhance Electricity Load Forecasting in Smart Cities Against Concept Drift. Smart Cities 2024, 7, 3412–3436. [Google Scholar] [CrossRef]
  33. Richter, L.; Lenk, S.; Bretschneider, P. Advancing Electric Load Forecasting: Leveraging Federated Learning for Distributed, Non-Stationary, and Discontinuous Time Series. Smart Cities 2024, 7, 2065–2093. [Google Scholar] [CrossRef]
  34. Malashin, I.; Tynchenko, V.; Gantimurov, A.; Nelyub, V.; Borodulin, A. Applications of Long Short-Term Memory (LSTM) Networks in Polymeric Sciences: A Review. Polymers 2024, 16, 2607. [Google Scholar] [CrossRef] [PubMed]
  35. Islam, M.S.; Alam, M.M.; Ahamed, A.; Meerza, S.I.A. Prediction of Diabetes at Early Stage using Interpretable Machine Learning. In Proceedings of the SoutheastCon 2023, Orlando, FL, USA, 13–16 April 2023; pp. 261–265. [Google Scholar] [CrossRef]
  36. Alharbi, F.R.; Csala, D. A Seasonal Autoregressive Integrated Moving Average with Exogenous Factors (SARIMAX) Forecasting Model-Based Time Series Approach. Inventions 2022, 7, 94. [Google Scholar] [CrossRef]
  37. Jamali, H.; Karimi, A.; Haghighizadeh, M. A new method of cloud-based computation model for mobile devices: Energy consumption optimization in mobile-to-mobile computation offloading. In Proceedings of the 6th International Conference on Communications and Broadband Networking, Singapore, 24–26 February 2018. [Google Scholar] [CrossRef]
  38. Chetry, B.P.; Kar, B. Online Signature Verification Using Chi-Square (χ2) Feature Selection and SVM Classifier. In Proceedings of the 2024 6th International Conference on Energy, Power and Environment (ICEPE), Shillong, India, 20–22 June 2024; pp. 1–6. [Google Scholar] [CrossRef]
  39. Hermo, J.; Bolón-Canedo, V.; Ladra, S. Fed-mRMR: A lossless federated feature selection method. Inf. Sci. 2024, 669, 120609. [Google Scholar] [CrossRef]
  40. Forghani, R.; Savadjiev, P.; Chatterjee, A.; Muthukrishnan, N.; Reinhold, C.; Forghani, B. Radiomics and Artificial Intelligence for Biomarker and Prediction Model Development in Oncology. Comput. Struct. Biotechnol. J. 2019, 17, 995–1008. [Google Scholar] [CrossRef]
  41. Farrokhi, Z.; Baesmat, K.; Regentova, E. Enhancing Urban Intelligence Energy Management: Innovative Load Forecasting Techniques for Electrical Networks. J. Power Energy Eng. 2024, 12, 72–88. [Google Scholar] [CrossRef]
  42. Zhao, Q.; Xiang, W.; Huang, B.; Wang, J.; Fang, J. Optimised extreme gradient boosting model for short term electric load demand forecasting of regional grid system. Sci. Rep. 2022, 12, 19282. [Google Scholar] [CrossRef] [PubMed]
  43. Zhao, M.; Guo, G.; Fan, L.; Han, L.; Yu, Q.; Wang, Z. Short-term natural gas load forecasting based on EL-VMD-Transformer-ResLSTM. Sci. Rep. 2024, 14, 20343. [Google Scholar] [CrossRef]
  44. Joshi, B.; Singh, V.K.; Vishwakarma, D.K.; Ghorbani, M.A.; Kim, S.; Gupta, S.; Chandola, V.K.; Rajput, J.; Chung, I.-M.; Yadav, K.K.; et al. A comparative survey between cascade correlation neural network (CCNN) and feedforward neural network (FFNN) machine learning models for forecasting suspended sediment concentration. Sci. Rep. 2024, 14, 10638. [Google Scholar] [CrossRef] [PubMed]
  45. Sarker, B.; Chakraborty, S.; Čep, R.; Kalita, K. Development of optimized ensemble machine learning-based prediction models for wire electrical discharge machining processes. Sci. Rep. 2024, 14, 23299. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Architecture and procedures of proposed method.
Figure 1. Architecture and procedures of proposed method.
Smartcities 08 00199 g001
Figure 2. Standard Long Short-Term Memory layer, where ⊗ indicates element-wise multiplication and ⊕ indicates element-wise addition; σ denotes the sigmoid activation function, and tanh generates the candidate cell state. The forget gate decides which information from previous time steps should be removed, the input gate incorporates new, relevant data into the cell state, and the output gate regulates which information is propagated to the next time step.
Figure 2. Standard Long Short-Term Memory layer, where ⊗ indicates element-wise multiplication and ⊕ indicates element-wise addition; σ denotes the sigmoid activation function, and tanh generates the candidate cell state. The forget gate decides which information from previous time steps should be removed, the input gate incorporates new, relevant data into the cell state, and the output gate regulates which information is propagated to the next time step.
Smartcities 08 00199 g002
Figure 3. Random Forest: general architecture.
Figure 3. Random Forest: general architecture.
Smartcities 08 00199 g003
Figure 4. Method in detail.
Figure 4. Method in detail.
Smartcities 08 00199 g004
Figure 6. Household one-year power consumption profile—Summerlin, Las Vegas.
Figure 6. Household one-year power consumption profile—Summerlin, Las Vegas.
Smartcities 08 00199 g006
Figure 7. Weather data for location of case study household.
Figure 7. Weather data for location of case study household.
Smartcities 08 00199 g007
Figure 8. Features selected based on CM.
Figure 8. Features selected based on CM.
Smartcities 08 00199 g008
Figure 9. Comparison of models for overall consumption (12-step ahead).
Figure 9. Comparison of models for overall consumption (12-step ahead).
Smartcities 08 00199 g009
Figure 10. Comparison of models on FAU downstairs consumption (12-step ahead).
Figure 10. Comparison of models on FAU downstairs consumption (12-step ahead).
Smartcities 08 00199 g010
Figure 11. Comparison of models on AC downstairs consumption (12-step ahead).
Figure 11. Comparison of models on AC downstairs consumption (12-step ahead).
Smartcities 08 00199 g011
Figure 12. Comparison of models on AC upstairs consumption (12-step ahead).
Figure 12. Comparison of models on AC upstairs consumption (12-step ahead).
Smartcities 08 00199 g012
Figure 13. Comparison of models on refrigerator consumption (12-step ahead).
Figure 13. Comparison of models on refrigerator consumption (12-step ahead).
Smartcities 08 00199 g013
Figure 14. Comparison of models on laundry/dryer consumption (12-step ahead).
Figure 14. Comparison of models on laundry/dryer consumption (12-step ahead).
Smartcities 08 00199 g014
Figure 15. Comparison of models on dishwasher/garbage disposal consumption (12-step ahead).
Figure 15. Comparison of models on dishwasher/garbage disposal consumption (12-step ahead).
Smartcities 08 00199 g015
Figure 16. Comparison of models on microwave consumption (12-step ahead).
Figure 16. Comparison of models on microwave consumption (12-step ahead).
Smartcities 08 00199 g016
Figure 17. Comparison of models on FAU upstairs consumption (12-step ahead).
Figure 17. Comparison of models on FAU upstairs consumption (12-step ahead).
Smartcities 08 00199 g017
Figure 18. Comparison of models on kitchen GFI /toaster consumption (12-step ahead).
Figure 18. Comparison of models on kitchen GFI /toaster consumption (12-step ahead).
Smartcities 08 00199 g018
Figure 19. Comparison of models on miscellaneous consumption (12-step ahead).
Figure 19. Comparison of models on miscellaneous consumption (12-step ahead).
Smartcities 08 00199 g019
Figure 20. Comparison of models on summation of all appliance consumption (12-step ahead).
Figure 20. Comparison of models on summation of all appliance consumption (12-step ahead).
Smartcities 08 00199 g020
Table 1. Challenges of statistical load forecasting.
Table 1. Challenges of statistical load forecasting.
Parametric Statistical Methods Gap
Handling high variability and peaks
Non-stationarity of load data
Difficulty handling missing data
Inaccuracy at different time horizons
Difficulty incorporating external variables
Inefficiency in real-time forecasting
Computational constraints with large datasets
Autocorrelation and multicollinearity
Risk of overfitting in complex models
Capturing seasonality and periodicity
Table 2. Challenges of ML load forecasting.
Table 2. Challenges of ML load forecasting.
Machine Learning Techniques Gap
Generalization challenges across different regions or time periods
Need for large, high-quality datasets
Hyperparameter tuning complexity
Overfitting with complex models
Limited effectiveness in incorporating domain knowledge
Difficulty interpreting model decisions (black-box nature)
Model retraining requirements for real-time data
Difficulty handling rare events or outliers
High computational requirements
Sensitivity to feature selection and engineering
Table 3. Results of Chi-Square feature selection.
Table 3. Results of Chi-Square feature selection.
FeatureTemperatureFeels-LikeUV IndexSea Level PressureSolar EnergySolar RadiationVisibility
Score7162.3325907.644932.464775.7044176.7744151.80955.890
FeatureTemperatureFeels-LikeUV IndexSea Level PressureSolar EnergySolar Radiation
Score3776.2952270.17867.17676.616600.80606284.650
Table 5. Evaluation Metrics and Formulas.
Table 5. Evaluation Metrics and Formulas.
MetricEquation
MSE [41]MSE = 1 N i = 1 N ( A i F i ) 2
MAE [42]MAE = 1 N i = 1 N F i A i
MAPE [43]MAPE = 1 N i = 1 N A i F i A i × 100 %
RMSE [44]RMSE = 1 N × i = 1 N ( F i A i ) 2
RMSLE [45]RMSLE = 1 N × i = 1 N ( l o g ( F i + 1 ) 2 l o g ( A i + 1 ) ) 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hassanpouri Baesmat, K.; Regentova, E.E.; Baghzouz, Y. Cloud-Enabled Hybrid, Accurate and Robust Short-Term Electric Load Forecasting Framework for Smart Residential Buildings: Evaluation of Aggregate vs. Appliance-Level Forecasting. Smart Cities 2025, 8, 199. https://doi.org/10.3390/smartcities8060199

AMA Style

Hassanpouri Baesmat K, Regentova EE, Baghzouz Y. Cloud-Enabled Hybrid, Accurate and Robust Short-Term Electric Load Forecasting Framework for Smart Residential Buildings: Evaluation of Aggregate vs. Appliance-Level Forecasting. Smart Cities. 2025; 8(6):199. https://doi.org/10.3390/smartcities8060199

Chicago/Turabian Style

Hassanpouri Baesmat, Kamran, Emma E. Regentova, and Yahia Baghzouz. 2025. "Cloud-Enabled Hybrid, Accurate and Robust Short-Term Electric Load Forecasting Framework for Smart Residential Buildings: Evaluation of Aggregate vs. Appliance-Level Forecasting" Smart Cities 8, no. 6: 199. https://doi.org/10.3390/smartcities8060199

APA Style

Hassanpouri Baesmat, K., Regentova, E. E., & Baghzouz, Y. (2025). Cloud-Enabled Hybrid, Accurate and Robust Short-Term Electric Load Forecasting Framework for Smart Residential Buildings: Evaluation of Aggregate vs. Appliance-Level Forecasting. Smart Cities, 8(6), 199. https://doi.org/10.3390/smartcities8060199

Article Metrics

Back to TopTop