Next Article in Journal
Mechatronic and Robotic Systems Utilizing Pneumatic Artificial Muscles as Actuators
Previous Article in Journal
A Smart Hydration Device for Children: Leveraging TRIZ Methodology to Combat Dehydration and Enhance Cognitive Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of Traditional and Data-Driven Algorithms for Energy Disaggregation Under Sampling and Filtering Conditions

by
Carlos Rodriguez-Navarro
,
Francisco Portillo
*,
Isabel Robalo
and
Alfredo Alcayde
Department of Engineering, University of Almeria, ceiA3, 04120 Almeria, Spain
*
Author to whom correspondence should be addressed.
Inventions 2025, 10(3), 43; https://doi.org/10.3390/inventions10030043
Submission received: 12 May 2025 / Revised: 9 June 2025 / Accepted: 11 June 2025 / Published: 13 June 2025

Abstract

Non-intrusive load monitoring (NILM) enables the disaggregation of appliance-level energy consumption from aggregate electrical signals, offering a scalable solution for improving efficiency. This study compared the performance of traditional NILM algorithms (Mean, CO, Hart85, FHMM) and deep neural network-based approaches (DAE, RNN, Seq2Point, Seq2Seq, WindowGRU) under various experimental conditions. Factors such as sampling rate, harmonic content, and the application of power filters were analyzed. A key aspect of the evaluation was the difference in testing conditions: while traditional algorithms were evaluated under multiple experimental configurations, deep learning models, due to their extremely high computational cost, were analyzed exclusively under a specific configuration consisting of a 1-s sampling rate, with harmonic content present and without applying power filters. The results confirm that no universally superior algorithm exists, and performance varies depending on the type of appliance and signal conditions. Traditional algorithms are faster and more computationally efficient, making them more suitable for scenarios with limited resources or rapid response requirements. However, significantly more computationally expensive deep learning models showed higher average accuracy (MAE, RMSE, NDE) and event detection capability (F1-SCORE) in the specific configuration in which they were evaluated. These models excel in detailed signal reconstruction and handling harmonics without requiring filtering in this configuration. The selection of the optimal NILM algorithm for real-world applications must consider a balance between desired accuracy, load types, electrical signal characteristics, and crucially, the limitations of available computational resources.

1. Introduction

The global energy transition urgently requires innovative solutions to address the dual challenges of climate change mitigation. Households account for approximately 20–30% of global energy use [1], with inefficient monitoring systems exacerbating financial burdens and greenhouse gas emissions from fossil fuel dependence. Non-intrusive load monitoring (NILM) [2] has emerged as a transformative approach to enable granular energy management by disaggregating appliance-level consumption from aggregated electrical signals. While traditional methods rely on intrusive submetering, which is costly and impractical for large-scale deployment, NILM offers a scalable alternative. Yet, its widespread adoption remains hindered by unresolved technical limitations.
Since Hart’s seminal work in the 1980s [3], NILM algorithms [4,5] have evolved considerably, incorporating machine learning and advanced signal processing techniques. In recent years, research in NILM has progressed significantly with the application of deep learning architectures and innovative models that enhance the extraction of temporal patterns and complex features from electrical signals. For instance, a convolutional neural network (CNN)-based approach that leverages temporal patternization to capture more detailed usage patterns and improve appliance classification from power signals has been proposed, demonstrating notable improvements in both accuracy and robustness compared to traditional methods [6]. Similarly, a transformer-based model that employs attention mechanisms to extract global dependencies between the aggregate signal and individual device signals (overcoming the limitations of recurrent models and reducing data preprocessing requirements) was introduced in [7]. Furthermore, harmonic and generative modeling for NILM have been explored, showing that the explicit incorporation of harmonic information and generative techniques can enhance disaggregation performance, particularly in scenarios with high harmonic distortion [8].
However, critical barriers persist in inconsistent benchmarking due to heterogeneous datasets, sensitivity to sampling rates (varying from milliseconds to minutes), and inadequate handling of harmonic distortions or low-power noise. Current evaluations often overlook the interplay between algorithmic robustness and real-world variables, such as the electrical characteristics of appliances or computational constraints for real-time applications. This gap limits the practical implementation of NILM systems (see Figure 1), despite their proven potential to reduce household consumption by up to 20% through behavioral feedback.
This paper advances NILM research by tackling key gaps in standardization, algorithmic evaluation, and metric-driven assessment. It promotes standardization using open-source tools, such as evaluating classical algorithms and neural network-based methods across diverse sampling rates, harmonic conditions, and power filters. The analysis integrates performance metrics, event detection, and computational efficiency to guide real-world deployment. By linking algorithmic performance [9] to dataset characteristics and hardware constraints, this work provides a framework for tailoring NILM solutions to specific appliance profiles and operational requirements [10]. The open tools and datasets underpinning this analysis enhance transparency, inviting replication and extension in future research.

2. Materials and Methods

Evaluating the performance of energy disaggregation algorithms requires a robust infrastructure for accurate data acquisition and standardized tools for comparative analysis. This study leverages datasets collected using open-source metering hardware, processed and analyzed through a dedicated software toolkit, and evaluated using a defined set of performance metrics under various experimental conditions.

2.1. Measurement Hardware

The datasets used in this study were generated using the OpenZmeter (oZm) platform, an open-source energy meter and power quality analyzer. While previous works describe versions v1 and v2 of the oZm, the core datasets analyzed in this article were obtained using the updated and high-precision oZm v3 [11]. This meter is distinguished by its capability to perform high-frequency electrical measurements at a sampling rate of 15,625 Hz, recording a broad spectrum of electrical variables, including voltage, current, and harmonics up to the 50th order for voltage, current, and power.
The oZm has been utilized in other research papers for different purposes, and its use in this work demonstrates the device’s robustness and versatility, establishing a new benchmark in high-frequency electrical monitoring for NILM research. Long-term data acquisition was facilitated via the oZm API.

2.2. Datasets

Evaluating the performance of energy disaggregation algorithms requires reliable electrical datasets that capture consumption at the aggregate and individual levels. In the field of NILM, the evaluation of algorithmic performance is fundamentally based on using such datasets. However, the heterogeneity of existing datasets often leads to inconsistency in algorithm benchmarking. Research on NILM has used several prominent public datasets for evaluation (see Figure 2), including the following:
  • Reference Energy Disaggregation Dataset (REDD) [12], a public dataset used for energy disaggregation research.
  • Minutely Power Dataset Almanac (AMPds) [13], a public dataset for load disaggregation and eco-feedback that includes data on the electricity, water, and natural gas consumption of residential households in Canada from 2012 to 2014.
  • UK Domestic Appliance-Level Electricity (UK-DALE) dataset [14], obtained from a two-year longitudinal study of UK households [15].
  • Electricity Consumption and Occupancy (ECO) dataset [16], which is used to evaluate the performance of NILM algorithms.
  • Green Energy Consumption Dataset (GREEND) [17], which comprises data on household energy consumption in Italy and Austria.
This paper employed two publicly available datasets based on open hardware to address some of the challenges related to data variability and allow a focused analysis of the impact of sampling and filtering conditions. Both datasets were collected at the University of Almeria (Spain) using the oZm platform.
The first dataset was DSUALM10H [18], introduced in June 2023. This is a new high-resolution dataset for NILM and a multichannel dataset containing measurements of 10 common household appliances. It is distinguished by its ability to perform high-frequency electrical measurements at a sampling rate of 15,625 Hz. It captures a broad spectrum of electrical variables, including voltage, current, and harmonics up to the 50th order for voltage, current, and power. It includes 150 electrical variables, such as current, voltage, and power transients.
DSUALM10H was generated using three oZm v3, providing 12 measurement channels, 1 for aggregate consumption and 10 for individual appliances. The monitored appliances included an electric oven, a microwave, a kettle, a vacuum cleaner, a radiator, a television, an electric shower heater, a fan, a refrigerator, and a freezer. A feature of DSUALM10H is that it includes harmonic content for each appliance.
The second dataset was DSUALM10, which was derived from DSUALM10H. It was obtained by eliminating the harmonic components of voltage, current, and power, thus containing only the fundamental electrical measurements. Like DSUALM10H, it is based on measurements from the 10 appliances using the oZm v3. Both datasets are publicly available [18,19] and generated with open hardware, which promotes transparency and reproducibility in NILM research. They serve as a basis for exploring how signal characteristics and sampling conditions impact the performance of algorithms.

2.3. Software Tools and Evaluation Metrics

Energy disaggregation and algorithm evaluation were conducted using the Non-Intrusive Load Monitoring Toolkit (NILMTK) v0.4.0 [20], an open-source framework designed to streamline NILM research by providing standardized workflows, algorithms, data converters, and a suite of built-in evaluation metrics. Within this framework, custom converters were developed to process data. The evaluation methodology followed the full NILMTK pipeline (from raw data conversion to disaggregation assessment).
To extend its capabilities, the NILMTK-Contrib extension was employed [21], providing implementations of traditional algorithms and deep learning-based approaches. In addition, it offers a rapid experimentation API for cross-building and cross-dataset validation, as well as support for training on synthetic aggregates and transfer learning across different sampling frequencies [22].
Core NILMTK-supported metrics were used to evaluate algorithm performance, covering both power estimation (regression) and event detection (classification) aspects [23] as follows:
1.
Mean Absolute Error (MAE) (Equation (1)), which measures the average absolute difference between the predicted and actual power consumption of an appliance. Lower MAE values indicate better estimation performance,
M A E = 1 N p r e d i c t e d i a c t u a l i
where, predictedi and actuali represent the individual prediction values and actual energy consumption values of the appliances, respectively, and N represents the total number of observations.
2.
Root Mean Squared Error (RMSE) (Equation (2)), which is like MAE, but gives greater weight to larger errors, represents the standard deviation of estimation errors,
R M S E = 1 T ( y t n y ^ t n ) 2
where y ^ t n represents the estimated power of device n at each time interval t, y t n denotes the actual power of the same device, and T is the total number of observations or time intervals recorded during energy consumption.
3.
F1-SCORE, which combines Precision (Equation (3)) and Recall (Equation (4)), assesses the accuracy in detecting ON/OFF events. A higher F1-SCORE reflects a better ability to correctly identify appliance operation states:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
TP represents true positives, FP false positives, and FN false negatives.
The F1-SCORE synthesizes these aspects, offering a composite measure that signals robust accuracy in identifying and predicting appliance states, as expressed in Equation (5):
F 1 = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l
4.
Normalized Disaggregation Error (NDE), which quantifies the total energy estimation error for each appliance, is normalized by actual consumption to enable fair comparisons,
N D E i = t = 1 T ( y ^ t i y t i ) 2 t = 1 T ( y t i ) 2
where y ^ t i is the estimated power consumption of the device i at time t, and y t i reflects the actual power consumption of the device over the total time interval T.

2.4. Algorithms

This study evaluated traditional algorithms and deep neural network-based models [24] for NILM using NILMTK and NILMTK-Contrib [24]. The main characteristics of the traditional algorithms are described below:
  • Combinatorial Optimization (CO). This algorithm performs an exhaustive search of all possible combinations of appliance states to find the one that best explains the aggregate signal. While it can yield good results in simple scenarios, its complexity grows exponentially with the number of devices and possible states, which limits its scalability [25].
  • Hart85. Based on finite state machines, this method detects ON/OFF events using dynamic active/reactive power thresholds. It is susceptible to parameter configuration and signal quality, showing highly variable performance depending on the appliance type and experimental conditions [3].
  • Mean. This method uses a moving average of the aggregate consumption signal over a time window to estimate the consumption of each appliance. It is notable for its simplicity, robustness, and low computational demand, although its accuracy may be limited for devices with complex or variable consumption patterns.
  • Factorial Hidden Markov Model (FHMM) [26]. This probabilistic model uses the Viterbi algorithm to infer the most probable sequence of device states, considering temporal transitions and relationships between them, with complexity O(T·SN) [27]. Unlike the standard FHMM implementation in NILMTK, this version incorporates several optimizations to improve efficiency; it leverages parallel processing on multicore CPUs via multithreading, replaces Python 3.7.12 loops with vectorized NumPy operations, manages the state space more efficiently through probability precomputing and dynamic pruning, optimizes memory usage with adjustable data types and sparse array storage, and uses JIT compilation with Numba for critical routines. In contrast, the original NILMTK version relies on sequential, generic implementations and non-vectorized data structures, resulting in substantially lower computational performance and resource efficiency.
The main characteristics of deep neural network-based models are as follows:
  • WindowGRU. Implemented as a bidirectional GRU network that processes temporary windows of aggregate electrical consumption, WindowGRU uses recurring layers with ReLU activation to predict the state of appliances in the last temporal step, is integrated into the NILMTK experimentation API for cross-evaluation between datasets, and is typically trained with Adam (30 epochs, learning rate 1 × 10−3) under GPU requirements (TensorFlow-GPU + CUDA) and normalization preprocessing, although non-optimized implementations of the activation functions may limit its practical performance [21].
  • Seq2Seq. This model implements an encoder–decoder-based deep neural network architecture, where a time window of the aggregated signal (e.g., 99 samples) is processed by the model to predict the corresponding sequence of consumption of an appliance, using recurrent and convolutional layers; each device has its own trained model, training is performed with normalized and batch data, and integration with the NILMTK API allows performance to be evaluated in different buildings and datasets in a flexible and reproducible way [21,28].
  • Denoising Autoencoder (DAE). Fully convolutional in this implementation, it uses Conv1D layers in the encoder (e.g., 3 layers with 8/16/32 filters and kernel = 4) and a symmetric transposed decoder to capture temporal patterns, injecting Gaussian noise (σ = 0.1–0.3) into the aggregate input to strengthen the model, and optimizes a combined loss function that integrates the Mean Square Error (MSE) of reconstruction with L1 regularization over the weights (controlled by λ) to avoid overfitting. Training is with Adam (learning rate ~1 × 10−3) on time windows of 150–600 standardized samples, thus achieving disaggregation through sparsa and locally invariant latent representations [21].
  • Recurrent Neural Networks (RNN). This model implements standard recurrent neural networks for energy disaggregation, processing temporal sequences of aggregate consumption through predefined windows (e.g., 100–600 samples). Each appliance has an independent RNN model trained with recurrent dense layers and ReLU activation, using Keras as a backend. The data are normalized and screened before training, which is performed with the Adam optimizer (learning rate ~1 × 10−3) and MSE loss function, integrated into the NILMTK API for cross-evaluation between buildings/datasets. The current implementation faces practical limitations due to version-specific dependencies (TensorFlow/Keras) and problems reported in the train–test division after migrating to internal Keras methods [21].
  • Seq2Point. This model implements a CNN-based neural network that takes temporal windows of the aggregated signal (e.g., 99 samples) and predicts the consumption value of an appliance only at the center point of the window, using a sequential architecture of Conv1D, Dense, Dropout, and Flatten layers, trained with MSE loss function, data normalization, and hyperparameter configuration, such as window size, epoch number, and batch size, all integrated into NILMTK’s rapid experimentation API to facilitate training, evaluation, and comparison between devices and datasets [29].

2.5. Experimental Conditions

To explore how different factors influence algorithm performance, evaluations were conducted under multiple experimental conditions, focusing on the following:
  • Sampling rate. Tested at intervals ranging from 90 and 60 s to higher resolutions (1 s, 500 ms, 250 ms, and 125 ms).
  • Harmonic content. By comparing DSUALM10H (with harmonics) and DSUALM10 (without them), the impact of harmonic components was assessed [18].
  • Power filtering. The effect of applying aggregate power thresholds (10 W, 50 W, 100 W) was evaluated before disaggregation.
This multifactorial evaluation (employing DSUALM10, DSUALM10H, and NILMTK metrics) aimed to comprehensively understand energy disaggregation algorithms’ behavior.

3. Results and Discussion

This section presents the evaluation outcomes for various energy disaggregation algorithms applied to DSUALM10H and DSUALM10, using the NILMTK-Contrib tool. In addition, the influence of different experimental conditions was investigated, including the application of power filters, the type of appliance, the sampling rate, and the presence of harmonics. Regarding algorithm performance, the results of visual disaggregation and an analysis of the execution times for all the algorithms considered are also included.

3.1. Evaluation Metrics

Four key metrics (MAE, NDE, F1-SCORE, and RMSE) were used to evaluate algorithm performance under identical experimental conditions (sampling rate, harmonic presence, power filters) for household appliances, with an 80/20 train–test data split, ensuring consistent scenario comparisons across metrics [22]. The inability to use sub-second sampling intervals for deep learning algorithms in NILMTK-Contrib (e.g., CNNs, LSTMs), unlike classical methods (CO, FHMM), stems from architectural constraints: classical approaches rely on statistical/discrete-event models independent of temporal granularity, while neural networks require fixed-length input windows (e.g., 500 samples at 1 s = 500 s context), where shorter intervals (<1 s) reduce effective temporal context, eroding long-term pattern capture, which is critical for performance. Additionally, higher sampling frequencies exponentially increase data volume and computational demands, and current NILMTK-Contrib implementations lack mechanisms to adapt hyperparameters (e.g., kernel sizes, pooling layers) to sub-second scales, unlike signal-level filters applicable across frequencies. The detailed results of the MAE, NDE, F1-SCORE, and RMSE metrics for all algorithms and experimental conditions are presented in Appendix A.

3.1.1. F1-SCORE

Figure 3 shows that, in general, neural network-based algorithms achieve higher or comparable F1-SCOREs to traditional algorithms for most home appliances. This suggests that neural networks are more robust and effective at handling the complexity and variability of electrical signals, especially in the presence of harmonics and unfiltered. However, there are exceptions: for example, the traditional Mean algorithm obtains an exceptionally high F1-SCORE for the TV, and Hart85 achieves good results only for some appliances.
On the other hand, while neural networks often outperform traditional algorithms, they also present difficulties in specific applications, such as vacuum cleaners and electric space heaters, where F1-SCOREs are low even for these advanced models. This indicates that, despite their sophistication, neural networks cannot always correctly detect all events in appliances with more complex or less predictable consumption patterns. In summary, the analysis confirms that, under the condition of 1-s sampling with harmonic content, neural networks offer a better capacity for event detection than traditional algorithms, showing a more balanced and robust performance against different types of appliances. Nevertheless, this advantage is accompanied by a considerably higher computational cost, which can limit practical applications.

3.1.2. MAE

The MAE varies considerably for traditional algorithms depending on the algorithm, appliance, and conditions. Stable and straightforward appliances such as the TV or fan have lower MAEs, especially with Mean. High-power or complex appliances, such as the electric oven, consistently present high MAEs with all traditional algorithms. Hart85 significantly improves its MAE by applying power filters.
Neural network-based algorithms, evaluated in the configuration of 1 s with harmonic content, generally outperform classical algorithms in all average metrics, including MAE. They tend to achieve lower MAEs, even for complex or high-powered appliances. For example, for the electric furnace, the MAE of RNN was 174.46 W, while Seq2Point was 285.67 W, Seq2Seq was 389.78 W, and DAE was 480.55 W. These values are considerably lower than the MAE reported for Mean (910.21 W), CO (553.7 W), and FHMM (322.3 W) without filters, and in some cases, even better than Hart85 (423.29 W) without filters. This reinforces the idea that neural models can better model complex consumption patterns. Figure 4 shows the differences in MAE per appliance among all the algorithms evaluated (including DAE, RNN, Seq2Point, Seq2Seq, and WindowGRU).

3.1.3. NDE

In traditional algorithms, Mean performs better with low NDEs for regular, unfiltered, low-consumption appliances. At the same time, Hart85 presents poor performance under these conditions, improving with the application of filters (see Figure 5).
The fridge has the worst overall performance with traditional algorithms. Algorithms based on neural networks, under the conditions evaluated (1 s with harmonic content), also present generally lower NDE on average than traditional algorithms, reflecting a greater precision in estimating normalized energy consumption. For the electric oven, the NDEs of RNN (0.295), Seq2Point (0.438), Seq2Seq (0.500), and DAE (0.610) are substantially lower than those of traditional unfiltered algorithms, indicating a better ability to estimate the consumption of this high-power appliance. For the fridge, whose NDEs are high with some traditional algorithms, the neural networks obtained NDEs between 0.762 and 0.7945, representing an improvement. The NDE confirms that while Mean is effective for regular loads, neural networks offer more robust performance for various consumption patterns.

3.1.4. RMSE

Like MAE, Mean shows the lowest RMSEs for stable, low-power unfiltered appliances (see Figure 6). Hart85 has a significantly high RMSE without filters, but improves substantially with its application. The electric oven consistently presents the highest RMSEs among all traditional appliances and algorithms. CO shows a significant increase in RMSE for the fridge and freezer with filters.
By including the neural network algorithms evaluated in the condition of 1 s with harmonic content, it is observed that they generally achieve lower RMSEs on average than classical algorithms. This is particularly notable for high-powered appliances such as the electric oven, where the RMSEs of RNN (323.0 W), Seq2Point (479.9 W), Seq2Seq (547.8 W), and DAE (667.6 W) are lower than those of traditional unfiltered algorithms, indicating a better ability to handle significant power fluctuations. Although Hart85 can achieve low RMSEs for high-power loads with filters, neural networks accomplish this without relying on this specific preprocessing in the tested configuration. The RMSE analyses, such as the MAE and NDE, underline that neural networks offer greater overall accuracy in estimating power, even in the presence of harmonics, although at a considerably higher computational cost.
The inclusion of algorithms based on neural networks in the analysis of metrics such as F1-SCORE, MAE, NDE, and RMSE allows a more complete evaluation, since these models usually outperform traditional algorithms on average and demonstrate a greater ability to handle the complexity and variability of appliances, especially without the need for filters; however, this better performance implies a much higher computational cost, so the choice of the most appropriate algorithm must balance the desired accuracy, the characteristics of the devices to be monitored, and the resource limitations of the environment where it will be implemented.

3.2. Influence of Experimental Factors

The following section discusses the impact of specific experimental conditions (filters, appliance type, sampling rate) on evaluation metrics.

3.2.1. Effect of Power Filtering

The comparative analysis of the influence of power filters on the disaggregation of electricity consumption metrics shows that, when evaluating the average of all appliances, the classical algorithms experience a slight improvement in absolute errors (MAE, RMSE) when applying a power filter (e.g., 100 W). However, this improvement is marginal and is accompanied by a degradation in the ability to detect events, especially in small and cyclic loads, as evidenced by the drop in the average F1-SCORE in the Hart85 algorithm. In contrast, the Mean algorithm demonstrates greater robustness, maintaining stable performance across all metrics despite filtering. Nonetheless, the improvements remained relatively limited (see Figure 7).
The results show that combining harmonics with deep learning techniques such as Seq2Point and RNN, along with low filter values, optimizes the F1-SCORE in the experimental framework analyzed. In contrast, the use of high filter values improves the RMSE in RNN, and the inclusion of harmonics benefits the performance of Seq2Seq, although it is detrimental to FHMM. Neural networks, specifically RNN and Seq2Point, exhibit up to 54% less error compared to traditional methods such as FHMM. To minimize NDE, neural networks—particularly Seq2Point and Seq2Seq—with medium to high filter values (50–100 W) and the use of harmonics, are the most efficient option, which contrasts with the results observed for F1-SCORE, where the Mean method stands out. Finally, to reduce MAE, neural networks, especially RNN with a filter value of 50 W and Seq2Point, prove to be optimal, while statistical methods such as Mean show significant limitations in numerical accuracy. This analysis highlights the superiority of neural architectures over conventional statistical approaches, both in terms of accuracy and adaptability to different evaluation metrics.
Neural network-based algorithms, evaluated with and without filtering despite their computational demands, demonstrate clear superiority over traditional methods across all average metrics, achieving superior MAE, RMSE, F1-SCORE, and NDE values. This indicates greater accuracy in estimating the energy consumed and detecting events in all appliances, not just high-power ones. Although neural models require more computational resources, their overall performance is superior, and they do not depend on filtering to obtain good results.
The figure summarizes these findings, highlighting that filtering only partially benefits classical algorithms and can distort the overall evaluation by impairing the detection of minor loads. At the same time, neural networks offer more balanced and accurate performance across the complete set of devices.

3.2.2. Effect of Type of Application

The performance of the algorithms used for energy disaggregation depends mainly on the type of appliance analyzed. In the case of appliances with stable and straightforward consumption patterns, such as televisions, fans, or incandescent lamps, classical algorithms, especially the Mean method, usually offer the best results in terms of accuracy and stability, showing excellent robustness to variations in the electrical signal. However, in high-powered devices or devices with more complex consumption patterns, such as electric ovens, heaters, or vacuum cleaners, errors in metrics increase for all algorithms. In these cases, models based on deep neural networks achieve better results in error metrics and event detection, outperforming classical methods, especially when power filters are not applied (see Figure 8a–d).
Traditional algorithms exhibit reduced performance for appliances that exhibit cyclical or intermittent patterns, such as refrigerators, freezers, or microwaves, especially if filters are applied. Neural models maintain competitive values and can better model these complex patterns. In general, no algorithm is superior in all cases: the Mean method is efficient for simple loads but loses competitiveness in more complicated scenarios. Algorithms such as Hart85 can improve in specific contexts. However, their performance is inconsistent, while neural methods offer a more balanced and robust performance against a greater variety of consumption patterns (but with more computational resources).
The choice of the most appropriate algorithm for energy disaggregation must consider both the type of appliance and the particularities of its consumption. Classical methods are recommended for stable and straightforward loads. At the same time, neural models are more appropriate for devices with complex consumption, high power, or cyclic patterns, provided that the necessary computational capacity is available.

3.2.3. Effect of the Sampling Rate

Sampling frequency is a key factor in energy disaggregation, as it determines the ability of algorithms to detect rapid events and complex patterns in appliance consumption. In classical algorithms, a higher sampling rate (1 s) significantly improves error metrics (MAE, RMSE, NDE) and event detection capability (F1-SCORE). As the interval between samples increases, errors increase and the F1-SCORE decreases, as brief events and rapid changes in consumption are lost, especially at small or intermittent loads. For example, the average MAE of the Mean algorithm drops from about 400 to 280 W, and the F1-SCORE improves from 0.48 to 0.51, when the sample rate increases from 90 s to 1 s, reflecting the importance of high-frequency measurement for accurate disaggregation (see Table 1). The higher the sampling frequency, the classical algorithms show fewer errors and higher F1-SCORE, especially in the Mean and CO algorithms, while decreasing the frequency (60 s or 90 s) results in increased errors and a decreased F1-SCORE; Hart85 and FHMM are less sensitive in F1-SCORE but their errors also grow less frequently, with Mean standing out as the most robust and stable algorithm.
In the case of algorithms based on deep neural networks, this study only evaluates the 1 s sampling condition due to its high computational cost. Although there is no direct data on the impact of reducing the frequency in these models, it is inferred that, as in the classical models, a lower frequency would degrade their performance by losing relevant temporal information. The neural models, evaluated at 1 s, present average metrics higher than the classical ones at any frequency, suggesting they make the most of the available temporal detail.
In summary, a high sampling rate is preferable to achieve maximum accuracy in energy disaggregation. However, it must be balanced with computational cost, especially in real-time or resource-constrained applications.

3.2.4. Effect of Harmonic Content

The presence of harmonics in the electrical signal variably affects the performance of energy disaggregation algorithms. In deep neural network-based models, the metrics were obtained with harmonics. Under these conditions, these algorithms show a robust and competitive performance, with an average F1-SCORE higher than most classical algorithms and errors generally lower for most appliances, especially in 1-s sampling setups and without a power filter. Therefore, harmonics do not prevent good results from detecting events and estimating power, surpassing the classics in most cases under the same conditions.
Classical algorithms were evaluated with and without harmonic content. Direct comparison of average metrics shows that the presence of harmonics does not produce drastic changes in the overall performance of these methods. However, it may cause slight variations in error metrics and event detection capability, depending on the type of appliance and experimental configuration. In general, Mean maintains its robustness and stability with or without harmonics, while Hart85, CO, and FHMM show small fluctuations in their metrics, but without a clear pattern of systematic deterioration. Harmonics can slightly affect the accuracy of event detection and slightly increase estimation errors in some devices. Still, the effect is usually less than that caused by other factors, such as the sampling rate or the application of filters.
In summary, the effect of the harmonic content on energy disaggregation metrics is limited compared to other experimental factors. Neural network-based algorithms maintain competitive performance even with harmonic content, and classical algorithms exhibit only slight variations in their metrics under these conditions. Therefore, the presence of harmonics does not represent a significant obstacle to the effectiveness of deep neural models in energy disaggregation tasks.

3.3. Disaggregated Data and Execution Times

The disparity in execution times between algorithms directly impacts their practical applicability. Traditional methods are considerably faster than deep learning-based approaches. Visual fidelity in reconstructing individual signals generally correlates with longer execution time. Traditional algorithms are distinguished by their computational efficiency, being generally much faster than neural networks. However, their performance and visual detail vary.
Mean is the fastest algorithm, with extremely low execution times (0.91 s for the visualization). It offers a considerably simplified disaggregation, as shown in Figure 9a. In terms of metrics, Mean is extremely efficient and accurate for stable and straightforward loads such as TVs or fans, showing low MAEs for these (e.g., TV 20.89, Fan 15.89, under 1 s, harmonics, no filter). However, its accuracy limits its usefulness for appliances with complex or high-power patterns, which usually have the highest MAE (e.g., 910.21 W in the electric oven). Despite this limitation, the Mean algorithm remains robust to variations, maintaining stable metrics even with filtering.
The CO algorithm is fast and maintains low and stable execution times (2.39 s for visualization with harmonics). It offers a remarkable level of disaggregation detail for its low computational cost (Figure 9b), which is often higher than other traditional methods. However, its behavior can be more erratic, being sensitive to the sampling rate and filters, and it can deteriorate significantly for cyclical appliances such as the refrigerator or freezer The 1 s configuration with harmonic content, without a filter, features high MAEs for the fridge (155.15 W) and the freezer (103.80 W), which can align with the observation that harmonics have a negative effect.
The FHMM algorithm has a considerably longer execution time than Mean and CO (25.67 s). Despite the increased computational effort, the level of visual detail in the disaggregation shown is surprisingly modest compared to CO (Figure 10a). It is noticeably slower than Mean and CO, which limits its viability in rapid response scenarios. Regarding the MAE values in the 1 s configuration with harmonic content, without filter, the values range from relatively low values for appliances such as the TV (18.56 W) to substantially higher values for more complex devices, such as the electric oven (322.33 W).
Hart85 exhibits slower execution than Mean and CO but is still manageable for non-strict environments (Figure 10b). Without filters, the MAE tends to be high in several appliances, such as the TV, fan, refrigerator, and freezer (TV 33.72 W, fan 23.88 W, fridge 26.61 W, freezer 24.34 W). Its performance improved with the application of power filters. Its dependence on the sampling rate is extreme, with near-zero performance in event detection at high frequencies without filters.
Algorithms based on deep neural networks offer greater fidelity in reconstructing individual signals at the cost of much higher computational demand. They were evaluated without a filter due to their high cost and their metrics were obtained with harmonics in 1 s sampling. They outperform classical algorithms in all average metrics when evaluated without a filter and with harmonic content.
DAE has a considerable execution time for visualization (555.63 s) with a window size of 99. It offers a much richer level of disaggregation, identifying specific consumption patterns. The MAEs in the 1 s configuration with harmonic content (Figure 11a) are generally low (e.g., microwave 108.51 W, TV 19.61 W, fan 14.59 W).
The scale of the Y-axis varies between algorithms because each NILMTK disaggregation method (such as Mean, CO, FHMM, or DNN) estimates appliance power from the total signal in a different way: simpler algorithms such as Mean tend to smooth the values and underestimate peaks, resulting in lower maximum power values. In contrast, others, such as CO, are more sensitive to abrupt changes and can estimate significantly higher maximum values. These differences reflect the unique behavior of each algorithm when interpreting and separating the power signal, so the variation in scale does not indicate an error, but rather the diversity in how each model processes and represents energy consumption.
RNN, while one of the most computationally expensive methods (11,892.43 s), offers a remarkable improvement in visual accuracy (Figure 11b). However, this benefit must be weighed against the high time cost, making it a suitable choice for applications where visual accuracy is a priority and computational resources are not a constraint.
Seq2Point has a runtime of 2012.42 s for visualization. The consistent disaggregated profile shows a good balance between complexity and detail (Figure 12a). The MAE is very low for several appliances (e.g., microwave 55.76, vacuum cleaner 87.95, electric shower heater 273.68 W, freezer 12.61 W). Seq2Seq completed the task in 857.34 s for visualization using a sequence length of 99, an epochs number of 10, and a batch size of 512. Their visual performance aligns with other neural models, evidencing good signal reconstruction, especially for complex loads (Figure 12b). It has the lowest MAE for the microwave (51.14 W) and the electric shower heater (226.71 W).
WindowGRU is the algorithm that requires the longest processing time for visualization (16,283.86 s). Despite its high computational cost, the level of detail obtained is comparable to that of other deep learning models (Figure 13), suggesting that cost does not always translate into a commensurate improvement in quality. The MAEs are generally higher than other neural models for various appliances.
Although WindowGRU exhibits the longest execution, the level of disaggregation detail it produced was comparable to that of neural network-based algorithms, offering no clear advantage in output quality despite its significantly higher computational cost.
In summary, there is a clear correlation between the execution time and the level of visual detail of the disaggregation results. Traditional methods offer fast times, although with less precision. At the same time, algorithms based on neural networks provide greater fidelity in reconstructing individual signals at the cost of a greater demand on computational resources. This information is key when choosing an algorithm according to the constraints and objectives of the practical application.
Figure 14 illustrates the difference in logarithmic-scale execution times between traditional and deep learning methods of speed, multiplying by several orders of magnitude the Mean and CO times, which limits its viability in scenarios where response time is a critical factor.
The analysis of average metrics (Figure 15) reveals a clear superiority of deep learning algorithms (Seq2Point, Seq2Seq, RNN, and DAE) over traditional methods, although with significant differences in execution times. Seq2Point stands out as the most balanced—with the best F1-SCORE (0.56), lowest MAE (111.63 W), and second-best RMSE (189.78 W)—but it requires 2012.42 s of processing. Seq2Seq achieves the best NDE (0.62) and RMSE (188.77 W) in 857.34 s, although with a slightly lower F1-SCORE (0.53). Traditional methods such as CO (1.22 s) and FHMM (66.31 s) show average performance (F1-SOCRE ≤ 0.464, MAE > 189 W), and Hart85 (8.67 s) presents a contradiction—good MAE/RMSE (144.05 W/282.7 W) but very poor F1-SCORE/NDE (0.114/0.957)—suggesting numerical accuracy at the expense of detection reliability. Mean shows limitations, with the worst MAE (241.03 W) despite an acceptable F1-SCORE (0.5) and minimal runtime (1.31 s), while WindowGRU (F1-SCORE = 0.5, MAE = 150.05 W) positions itself as an intermediate option but with the highest computational time (16,283.86 s). This hierarchy confirms that sequential models (Seq2Point/Seq2Seq) simultaneously optimize precision and consistency, but require up to 34 m of execution. In comparison, traditional methods excel in speed (≤1 m) at the cost of compromising key metrics, establishing a decisive trade-off between efficiency and performance.

4. Conclusions

The systematic analysis of the performance of energy disaggregation algorithms under different experimental conditions, considering both execution times and quantitative metrics such as MAE, RMSE, F1-SCORE, and NDE, reveals that there is no universally superior algorithm. The performance of each approach intrinsically depends on the type of appliance, the experimental configuration (sampling rate, harmonic content, and application of power filters), and the characteristics of the model. Traditional algorithms, such as Mean, CO, Hart85, and FHMM, excel at computational efficiency and speed, making them particularly suitable for real-time applications or resource-constrained environments. Mean offers remarkable robustness against experimental variations and maintains low errors for simple loads, although its simplicity limits its accuracy with complex consumption patterns. CO and Hart85 have specific advantages, such as low run times and significant improvements through preprocessing. However, their performance may be affected by the cyclical nature of certain appliances or by sensitivity to the sample rate and filters applied.
On the other hand, algorithms based on deep neural networks (DAE, RNN, Seq2Point, Seq2Seq, WindowGRU) demonstrate a superior ability to reconstruct individual signals and achieve better average metrics in most scenarios, especially in harmonic-based and unfiltered configurations. These models excel at event detection and accurate power estimation and are better suited to the complexity and variability of modern appliances. However, its main limitation lies in the high computational cost, with execution times extending to several hours, which restricts its applicability in contexts where speed and efficiency are a priority.
The influence of experimental factors such as the sampling rate, the presence of harmonics, and the application of filters is also decisive, affecting each algorithm differently. While Mean is relatively insensitive to these factors, Hart85 and CO experience more noticeable variations, and neural networks maintain competitive performance even with harmonic content.
The selection of the most appropriate disaggregation algorithm must be carried out considering, in an integrated way, the context of use, the type of loads to be monitored, the signal conditions, and especially the computational constraints of the environment. Traditional methods are preferable for applications that demand rapid response and efficiency. At the same time, deep neural network-based approaches are ideal when maximum accuracy is a priority, and the necessary resources are available. This holistic view is critical to developing NILM solutions that are accurate, efficient, and adaptable to real-world demands.

Author Contributions

Conceptualization, C.R.-N. and A.A.; methodology, C.R.-N., F.P. and A.A.; software, C.R.-N. and A.A.; validation, C.R.-N., F.P., I.R. and A.A.; formal analysis, C.R.-N. and A.A.; investigation, C.R.-N., F.P., I.R. and A.A.; resources, C.R.-N. and A.A.; data curation, C.R.-N. and A.A.; writing—original draft preparation, C.R.-N., F.P., I.R. and A.A.; writing— review and editing, C.R.-N., F.P., I.R. and A.A.; visualization, C.R.-N., F.P. and I.R.; supervision, A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data will be made available upon request to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Sample rate = 60 s, no harmonics.
Table A1. Sample rate = 60 s, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM354.42185.3625.3452.20238.14221.24226.2115.1717.4122.22
CO646.49239.2818.1244.82223.13397.39566.3816.5534.9827.12
Mean864.16186.8820.4639.05238.39380.81457.3114.4429.4423.86
Hart85597.70143.10125.59117.37190.18281.99397.07121.80111.55118.88
RMSE
(W)
FHMM737.06257.7229.5060.51246.90609.25449.7119.5229.8234.01
CO1000.95294.4623.6854.70337.63884.72729.6020.9543.9934.48
Mean896.58235.3120.7539.64247.11591.49507.1415.1831.2225.51
Hart851062.44244.08206.36195.05307.78648.09598.32211.81199.71209.23
F1FHMM0.690.570.360.630.260.440.710.470.670.18
CO0.390.510.770.440.360.170.330.750.410.50
Mean0.510.630.950.630.260.180.540.810.510.54
Hart850.000.480.430.550.000.000.240.370.580.40
NDEFHMM0.690.910.841.101.141.000.730.690.730.97
CO0.941.040.670.991.571.451.180.741.080.98
Mean0.840.830.590.721.150.970.820.540.770.73
Hart851.000.865.863.531.431.060.977.494.905.95
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A2. Sample rate = 30 s, no harmonics.
Table A2. Sample rate = 30 s, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM540.98229.1220.2140.98212.69397.22197.958.1226.5025.00
CO588.44212.4217.2141.30238.70501.21444.4120.0837.1634.05
Mean874.51192.2020.4339.23241.93394.17471.7114.6330.7325.39
Hart85597.70184.9433.7239.60201.16174.74397.0023.8826.6124.34
RMSE
(W)
FHMM975.05307.5320.5755.44358.10967.18435.4514.8841.1137.55
CO924.22283.0522.7450.43364.84936.05634.7024.2345.4140.59
Mean908.80238.1820.8140.22251.83622.33517.8315.5532.0926.86
Hart851072.45285.5135.3055.59319.34639.58600.5528.4841.4536.17
F1FHMM0.590.430.960.450.240.000.670.810.500.49
CO0.500.420.810.650.310.290.430.540.440.46
Mean0.480.600.960.610.240.140.500.780.480.50
Hart850.000.000.000.000.000.000.290.000.000.00
NDEFHMM0.911.080.581.001.621.510.690.520.991.04
CO0.860.990.640.911.651.461.010.851.101.12
Mean0.850.830.590.721.140.970.830.550.770.74
Hart851.001.001.001.001.451.000.961.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A3. Sample rate = 15 s, no harmonics.
Table A3. Sample rate = 15 s, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM270.94158.9919.2530.22220.82275.37158.9225.2537.7238.40
CO428.47228.6617.5940.18222.26354.78420.9217.3341.5340.24
Mean905.79196.0820.8041.08249.24394.17491.8215.6831.6926.38
Hart85588.75184.9433.7239.6083.13424.81490.2423.8826.6124.34
RMSE
(W)
FHMM700.01262.2625.7148.21377.24805.02412.8328.6951.3047.48
CO755.60319.0022.3952.48351.14746.42634.2321.6553.7447.91
Mean931.39252.9321.0041.20258.85639.07536.6216.1833.7528.53
Hart85906.99297.9335.4256.31228.76735.33671.0928.8342.7537.43
F1FHMM0.740.580.630.640.410.490.730.250.300.31
CO0.680.510.800.500.270.190.450.650.440.41
Mean0.470.580.950.590.230.140.470.750.470.47
Hart850.600.000.000.000.000.000.280.000.000.00
NDEFHMM0.640.880.730.861.651.230.641.001.201.27
CO0.691.070.630.931.541.140.990.751.261.28
Mean0.850.850.590.731.130.970.830.560.790.76
Hart850.831.001.001.001.001.121.041.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A4. Sample rate = 1 s, no harmonics.
Table A4. Sample rate = 1 s, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM443.89342.7020.2535.90247.73208.33522.4018.5831.7124.26
CO560.63292.5720.7241.56259.81391.28428.6816.53131.05109.20
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM944.11588.1826.4153.59407.12733.94787.8024.4936.5143.26
CO935.99555.1725.0455.60415.57816.30656.1023.21238.47189.35
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.490.270.600.580.310.600.260.530.460.54
CO0.490.430.780.420.290.270.360.650.320.33
Mean0.460.560.950.580.210.120.460.740.460.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.861.320.750.951.771.101.220.850.811.00
CO0.861.250.710.991.811.231.010.805.294.38
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A5. Sample rate = 500 ms, no harmonics.
Table A5. Sample rate = 500 ms, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM566.98418.8820.6844.82152.62468.49496.9014.3729.0331.24
CO550.79299.3514.9939.49222.44476.96392.8420.52151.17109.34
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM1068.42611.3120.8360.42318.941103.74541.3021.5147.4747.56
CO939.45559.7222.4752.56341.42935.18634.2426.39301.47193.86
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.420.380.950.230.140.000.460.590.400.15
CO0.560.390.770.510.200.160.450.540.370.39
Mean0.460.560.950.580.210.120.460.740.450.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.981.370.591.071.391.660.840.741.051.10
CO0.861.260.630.931.491.400.980.916.694.48
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A6. Sample rate = 250 ms, no harmonics.
Table A6. Sample rate = 250 ms, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM443.89342.7020.3335.87247.73208.33522.4018.6031.6724.34
CO518.94268.7120.9245.51188.73444.67406.7616.34179.36100.08
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM944.11588.1826.4653.57407.12733.94787.8024.5136.5143.31
CO895.57526.8025.3655.85312.39886.49650.1422.98341.74180.05
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.490.270.600.580.310.600.260.520.450.54
CO0.540.460.760.480.230.210.450.670.380.34
Mean0.460.560.950.580.210.120.460.740.450.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.861.320.750.951.771.101.220.850.811.00
CO0.821.180.720.991.361.331.000.797.584.16
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A7. Sample rate = 125 ms, no harmonics.
Table A7. Sample rate = 125 ms, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM322.33353.9719.2326.42199.3682.46463.4218.3126.9734.53
CO560.15261.4319.8045.39260.42390.16428.0616.14138.9198.95
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM801.09594.4425.7145.84364.99458.77741.9524.4245.5049.34
CO922.21517.6124.1455.30415.71839.33655.6022.83275.43183.00
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.660.290.610.640.380.810.380.480.560.39
CO0.520.370.770.500.300.210.430.670.370.40
Mean0.460.560.950.580.210.120.460.740.450.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.731.330.730.811.590.691.150.841.011.14
CO0.841.160.680.981.811.261.010.796.114.23
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A8. Sample rate = 90 s, with harmonics.
Table A8. Sample rate = 90 s, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM463.18233.0115.3732.27196.84559.81227.0718.0635.8826.64
CO456.30249.3717.3330.14214.61170.71425.7818.6236.3928.26
Mean853.69174.5920.4138.08245.06391.05426.0613.5327.9722.55
Hart85583.62159.2657.7651.44134.76208.15358.4453.2943.5048.99
RMSE
(W)
FHMM862.95298.6122.0747.27333.081052.38392.3422.7745.0535.36
CO797.28305.2821.9941.20319.84621.12604.5622.3545.0035.14
Mean881.07226.5620.7439.06254.08605.13480.3114.6630.3224.31
Hart851038.53238.0275.3271.34251.77626.67577.4679.2969.2376.12
F1FHMM0.530.320.720.590.500.000.600.630.320.62
CO0.700.340.830.670.320.000.530.760.530.65
Mean0.490.700.940.630.300.130.600.810.560.60
Hart850.000.480.390.530.000.000.220.320.590.44
NDEFHMM0.831.090.630.851.401.690.660.831.131.04
CO0.771.110.620.741.351.001.020.811.131.03
Mean0.850.830.590.701.070.970.810.530.760.72
Hart851.000.872.131.281.061.010.972.871.742.24
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A9. Sample rate = 60 s, with harmonics.
Table A9. Sample rate = 60 s, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM684.08153.6925.4243.44137.02476.60195.7119.0227.5826.01
CO646.49239.2818.1244.82223.13397.39566.3816.5534.9827.12
Mean864.16186.8820.4639.05238.39380.81457.3114.4429.4423.86
Hart85597.70143.10125.59117.37190.18281.99397.07121.80111.55118.88
RMSE
(W)
FHMM1018.62234.5428.7356.13284.431005.85413.9722.6242.1636.80
CO1000.95294.4623.6854.70337.63884.72729.6020.9543.9934.48
Mean896.58235.3120.7539.64247.11591.49507.1415.1831.2225.51
Hart851062.44244.08206.36195.05307.78648.09598.32211.81199.71209.23
F1FHMM0.510.670.280.370.430.000.720.320.100.16
CO0.380.510.770.430.360.170.330.750.410.50
Mean0.510.630.950.630.260.180.540.810.510.54
Hart850.000.480.430.550.000.000.240.370.580.40
NDEFHMM0.960.830.821.021.321.650.670.801.031.05
CO0.941.040.670.991.561.451.180.741.080.98
Mean0.840.830.590.721.150.970.820.540.770.73
Hart851.000.865.853.531.431.060.977.494.905.95
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A10. Sample rate = 30 s, with harmonics.
Table A10. Sample rate = 30 s, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM537.34249.8114.5138.16225.38397.22184.5917.0525.5635.89
CO588.44212.4217.2141.30238.70501.21444.4120.0837.1634.05
Mean874.51192.2020.4339.23241.93394.17471.7114.6330.7325.39
Hart85597.70184.9433.7239.60201.16174.74397.0023.8826.6124.34
RMSE
(W)
FHMM971.60324.6722.0353.36375.21967.18418.2822.4840.7144.45
CO924.22283.0522.7450.43364.84936.05634.7024.2345.4140.59
Mean908.80238.1820.8140.22251.83622.33517.8315.5532.0926.86
Hart851072.45285.5135.3055.59319.34639.58600.5528.4841.4536.17
F1FHMM0.620.320.720.490.120.000.680.460.530.38
CO0.500.420.810.650.310.290.430.530.440.45
Mean0.480.600.960.610.240.140.500.780.480.50
Hart850.000.000.000.000.000.000.290.000.000.00
NDEFHMM0.911.140.620.961.701.510.670.790.981.23
CO0.860.990.640.911.651.461.010.851.101.12
Mean0.850.830.590.721.140.970.820.550.770.74
Hart851.001.001.001.001.451.000.961.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A11. Sample rate = 15 s, with harmonics.
Table A11. Sample rate = 15 s, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM412.52293.6219.7051.50181.65395.56491.7715.6629.5426.51
CO491.85264.2317.8442.04211.51421.32434.9017.8642.8840.10
Mean905.79196.0820.8041.08249.24394.17491.8215.6831.6926.38
Hart85588.75184.9433.7239.6083.13424.81490.2423.8826.6124.34
RMSE
(W)
FHMM870.97353.9426.2064.09344.96991.87536.5516.1945.5228.55
CO860.52349.7821.9754.01325.13847.40638.8022.0255.3248.29
Mean931.39252.9321.0041.20258.85639.07536.6216.1833.7528.53
Hart85906.99297.9335.4256.31228.76735.33671.0928.8342.7537.43
F1FHMM0.670.320.590.190.040.000.470.750.450.47
CO0.560.400.810.480.340.090.410.660.430.35
Mean0.470.580.950.590.230.140.470.750.470.47
Hart850.600.000.000.000.000.000.280.000.000.00
NDEFHMM0.801.190.741.141.511.510.830.561.060.76
CO0.791.170.620.961.421.290.990.761.291.29
Mean0.850.850.590.731.130.970.830.560.790.76
Hart850.831.001.001.001.001.121.041.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A12. Sample rate = 1 s, with harmonics.
Table A12. Sample rate = 1 s, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM322.33353.9718.5626.45199.3682.46463.4218.9128.2034.55
CO553.66275.8116.3335.77264.40355.04411.2418.14155.15103.80
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM801.09594.4425.2145.87364.99458.77741.9524.8246.2349.64
CO914.17534.7323.7749.07419.65785.20647.2324.61296.09184.01
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.660.290.620.640.380.810.380.460.550.38
CO0.540.390.740.530.260.240.430.580.430.39
Mean0.460.560.950.580.210.120.460.740.450.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.731.330.710.811.590.691.150.861.031.15
CO0.841.200.670.871.831.181.000.856.574.25
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A13. Sample rate = 1 s, 10 W filter, no harmonics.
Table A13. Sample rate = 1 s, 10 W filter, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM512.31339.5420.6841.27274.60321.51584.1116.4537.1412.75
CO552.22284.7814.5444.67209.34381.89366.9219.28140.38100.55
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM1012.74587.0320.8341.33428.77913.30833.0823.0552.6234.53
CO926.43549.8322.1655.96332.86846.11611.5223.88277.65169.94
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.390.260.950.580.350.450.060.540.450.74
CO0.550.500.780.460.250.090.520.610.420.35
Mean0.460.560.950.580.210.120.460.740.460.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.931.320.590.731.871.371.290.801.170.80
CO0.851.230.630.991.451.270.940.836.163.93
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A14. Sample rate = 1 s, 10 W filter, with harmonics.
Table A14. Sample rate = 1 s, 10 W filter, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM440.61340.9720.3835.84248.01209.39519.9918.9131.6724.66
CO544.15298.6615.6038.32235.47421.72446.4519.17173.72116.68
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM940.44586.9226.5053.54407.34735.84785.9824.7136.4743.54
CO929.99561.1823.0251.34359.32872.96680.6525.37320.92210.87
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.500.270.600.580.310.600.260.510.460.54
CO0.530.430.760.540.260.170.420.590.350.33
Mean0.460.560.950.580.210.120.460.740.460.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.861.320.750.951.771.101.210.850.811.01
CO0.851.260.650.911.561.311.050.887.124.87
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A15. Sample rate = 1 s, 50 W filter, no harmonics.
Table A15. Sample rate = 1 s, 50 W filter, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM228.21288.4724.5552.84249.93415.60272.6415.8730.0329.22
CO535.97283.3815.0942.71175.60441.17433.9520.75151.55115.86
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM669.25527.9529.2365.59259.681038.44564.4016.3847.8546.53
CO944.00543.4622.6453.63297.29896.11653.5726.45293.14199.79
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.780.190.410.100.210.310.580.740.250.16
CO0.570.420.770.560.220.250.410.540.440.33
Mean0.460.560.950.580.210.120.460.740.460.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.611.190.821.161.131.560.870.571.061.08
CO0.861.220.640.951.291.351.010.916.504.62
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A16. Sample rate = 1 s, 50 W filter, with harmonics.
Table A16. Sample rate = 1 s, 50 W filter, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM910.22181.6920.6841.34247.16424.07611.5814.4824.0827.05
CO591.90303.0316.2645.97215.34457.24402.5216.07142.39111.57
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM935.23442.5020.8341.43406.931050.19852.4421.5843.5736.06
CO995.55558.8521.5855.38333.84881.98649.3922.79286.75198.89
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.460.010.950.580.350.000.140.550.400.46
CO0.470.360.810.490.240.170.410.670.340.39
Mean0.460.560.950.580.210.120.460.740.460.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.850.990.590.731.771.581.320.750.970.83
CO0.911.250.610.981.451.321.000.796.364.60
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A17. Sample rate = 1 s, 100 W filter, no harmonics.
Table A17. Sample rate = 1 s, 100 W filter, no harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM602.10258.8316.2448.84269.14333.10542.6015.87120.9024.05
CO506.12268.6922.6245.12217.97414.86408.6814.90156.64101.75
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM1099.57469.5723.3562.88424.46929.29802.2316.38219.3543.54
CO905.62522.5826.6556.44330.19837.05630.2121.65303.74195.80
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.240.300.700.330.350.390.210.740.460.56
CO0.560.440.760.470.230.220.480.700.340.39
Mean0.460.560.950.580.210.120.460.740.460.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM1.001.050.661.111.851.401.240.574.871.01
CO0.831.170.751.001.441.260.970.756.744.52
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A18. Sample rate = 1 s, 100 W filter, with harmonics.
Table A18. Sample rate = 1 s, 100 W filter, with harmonics.
MetricAlgorithmEF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE
(W)
FHMM568.62418.6220.7242.27152.89466.38488.1316.9431.7827.09
CO525.44279.5522.4438.35221.46454.92382.1416.60173.32120.58
Mean910.21217.5920.8941.34250.08398.60496.9915.8931.9126.84
Hart85423.29184.9433.7239.6083.13174.74426.2123.8826.6124.34
RMSE
(W)
FHMM1069.97610.8026.6058.66319.241101.25529.0223.3936.6536.05
CO917.93539.8626.5853.29346.34903.06619.7823.16328.51214.68
Mean935.22416.9121.0641.42259.79649.81541.3916.3836.6635.86
Hart85641.18445.6635.4656.46229.72666.32634.9028.9445.0843.28
F1FHMM0.410.380.540.280.140.000.460.450.450.46
CO0.560.430.730.510.220.180.500.650.340.32
Mean0.460.560.950.580.210.120.460.740.450.46
Hart850.810.000.000.000.000.000.330.000.000.00
NDEFHMM0.981.370.751.041.391.650.820.810.810.83
CO0.841.210.750.941.511.360.960.807.294.96
Mean0.850.940.590.731.130.980.840.570.810.83
Hart850.591.001.001.001.001.000.981.001.001.00
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A19. Metrics for DAE. Sample rate = 1 s.
Table A19. Metrics for DAE. Sample rate = 1 s.
EF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE (W)480.54108.5119.6141.87217.37280.54320.8914.5930.1927.55
RMSE (W)667.64218.2020.5442.10270.47615.92426.4115.7435.6136.18
F1-SCORE0.540.600.890.580.220.150.570.740.470.46
NDE 0.610.490.580.751.180.920.660.540.790.84
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A20. Metrics for RNN. Sample rate = 1 s.
Table A20. Metrics for RNN. Sample rate = 1 s.
EF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE (W)174.4497.5919.9841.78107.12318.48391.9113.2830.6614.93
RMSE (W)322.96202.9220.7842.00155.26597.30483.5815.6335.7929.69
F1-SCORE0.540.780.870.580.210.200.460.750.450.52
NDE0.290.460.590.740.680.900.750.540.790.69
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A21. Metrics for Seq2Point. Sample rate = 1 s.
Table A21. Metrics for Seq2Point. Sample rate = 1 s.
EF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE (W)285.6255.7622.5728.8987.95305.07273.6814.4529.6912.61
RMSE (W)479.89156.1223.2539.02147.50610.65359.9218.4235.0527.94
F1-SCORE0.570.850.800.660.310.160.540.690.450.52
NDE0.440.350.660.690.640.920.560.640.780.65
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A22. Metrics for Seq2Seq. Sample rate = 1 s.
Table A22. Metrics for Seq2Seq. Sample rate = 1 s.
EF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE (W)389.7451.1420.5334.26102.06331.77226.7113.8129.2114.64
RMSE (W)547.76123.4320.9638.82153.72618.95302.4017.8134.3529.54
F1-SCORE0.490.800.890.580.230.120.500.690.450.51
NDE0.500.280.590.690.670.930.470.620.760.68
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A23. Metrics for WindowGRU. Sample rate = 1 s.
Table A23. Metrics for WindowGRU. Sample rate = 1 s.
EF 1Mw 2TVIL 3VC 4ESH 5ESHw 6FanFridgeFreezer
MAE (W)495.63106.9822.2739.01124.24320.05324.5916.3428.7222.68
RMSE (W)682.75242.3322.6040.19179.78595.84467.9816.7136.6536.73
F1-SCORE0.460.520.950.580.210.170.460.740.450.46
NDE0.620.540.640.710.780.890.720.580.810.85
1 Electric furnace. 2 Microwave. 3 Incandescent lamp. 4 Vacuum cleaner. 5 Electric space heater. 6 Electric shower heater.
Table A24. Run time of the main algorithms executed in DSUALM10H.
Table A24. Run time of the main algorithms executed in DSUALM10H.
AlgorithmConfigurationTime (s)
COWithout harmonics2.61
With harmonics3.28
Graph (without harmonics)4.42
Graph (with harmonics)2.39
10 W filter without harmonics2.57
100 W filter without harmonicsN/A
10 W filter with harmonics3.31
100 W filter with harmonics3.11
MEANWithout harmonics1.31
Without harmonics + 10 W filter0.89
Without harmonics + 100 W filter0.93
With harmonics1.22
Mean graphic0.91
HART85Without harmonics8.54
With harmonics8.67
FHMMWithout harmonics66.31
With harmonics61.76
Graphic25.67
DAEWith harmonics (window size 99)555.63
RNNWith harmonics11,892.43
Seq2PointWith harmonics2012.42
Seq2SeqWith harmonics857.34
WindowGRUWith harmonics16,283.86

References

  1. Eurosatat Energy Consumption in Househols. Available online: https://ec.europa.eu/eurostat/statistics-explained/index.php?title=Energy_consumption_in_households (accessed on 10 May 2025).
  2. Liu, H. Non-Intrusive Load Monitoring, 1st ed.; Science Press Beijing: Beijing, China, 2020; Volume 1, ISBN 978-9811518591. [Google Scholar]
  3. Hart, G.W. Nonintrusive Appliance Load Monitoring. Proc. IEEE 1992, 80, 1870–1891. [Google Scholar] [CrossRef]
  4. Lakshmi Ravi, D. Non-Intrusive Load Monitoring (NILM): A Comprehensive Review. Int. J. Innov. Res. Technol. 2024, 11, 853–858. [Google Scholar]
  5. Souza, J. Non-Intrusive Load Monitoring (NILM): An Introduction and Practical Example. Available online: https://www.linkedin.com/pulse/non-intrusive-load-monitoring-nilm-introduction-practical-souza-rpy4f/ (accessed on 11 March 2025).
  6. Kim, H.; Lim, S. Temporal Patternization of Power Signatures for Appliance Classification in Nilm. Energies 2021, 14, 2931. [Google Scholar] [CrossRef]
  7. Sykiotis, S.; Kaselimi, M.; Doulamis, A.; Doulamis, N. Electricity: An Efficient Transformer for Non-Intrusive Load Monitoring. Sensors 2022, 22, 2926. [Google Scholar] [CrossRef] [PubMed]
  8. Sun, R.; Dong, K.; Zhao, J. DiffNILM: A Novel Framework for Non-Intrusive Load Monitoring Based on the Conditional Diffusion Model. Sensors 2023, 23, 3540. [Google Scholar] [CrossRef] [PubMed]
  9. Batra, N.; Parson, O.; Berges, M.; Singh, A.; Rogers, A. A Comparison of Non-Intrusive Load Monitoring Methods for Commercial and Residential Buildings. arXiv 2014, arXiv:1408.6595. [Google Scholar]
  10. Makonin, S.; Popowich, F. Nonintrusive Load Monitoring (NILM) Performance Evaluation. Energy Effic. 2015, 8, 809–814. [Google Scholar] [CrossRef]
  11. Viciana, E.; Arrabal-Campos, F.M.; Alcayde, A.; Baños, R.; Montoya, F.G. All-in-One Three-Phase Smart Meter and Power Quality Analyzer with Extended IoT Capabilities. Measurement 2023, 206, 112309. [Google Scholar] [CrossRef]
  12. Kolter, J.Z.; Johnson, M.J. REDD: A Public Data Set for Energy Disaggregation Research. In Proceedings of the 1st KDD Workshop on Data Mining Applications in Sustainability, San Diego, CA, USA, 21 August 2011; Association for Computing Machinery: New York, NY, USA, 2021. [Google Scholar]
  13. Makonin, S.; Popowich, F.; Bartram, L.; Gill, B.; Bajic, I.V. AMPds: A Public Dataset for Load Disaggregation and Eco-Feedback Research. In Proceedings of the 2013 IEEE Electrical Power & Energy Conference, Halifax, NS, Canada, 21–23 August 2013; 2013; pp. 1–6. [Google Scholar]
  14. Kelly, J.; Knottenbelt, W. The UK-DALE Dataset, Domestic Appliance-Level Electricity Demand and Whole-House Demand from Five UK Homes. Sci. Data 2015, 2, 150007. [Google Scholar] [CrossRef] [PubMed]
  15. Lucas, A.; Jansen, L.; Andreadou, N.; Kotsakis, E.; Masera, M. Load Flexibility Forecast for DR Using Non-Intrusive Load Monitoring in the Residential Sector. Energies 2019, 12, 2725. [Google Scholar] [CrossRef]
  16. Beckel, C.; Kleiminger, W.; Cicchetti, R.; Staake, T.; Santini, S. The ECO Data Set and the Performance of Non-Intrusive Load Monitoring Algorithms. In Proceedings of the BuildSys 2014—1st ACM Conference on Embedded Systems for Energy-Efficient Buildings, Memphis, TN, USA, 3 November 2014; Association for Computing Machinery: New York, NY, USA, 2014; pp. 80–89. [Google Scholar]
  17. Monacchi, A.; Egarter, D.; Elmenreich, W.; D’Alessandro, S.; Tonello, A.M. GREEND: An Energy Consumption Dataset of Households in Italy and Austria. In Proceedings of the 2014 IEEE International Conference on Smart Grid Communications (SmartGridComm), Venice, Italy, 3–6 November 2014; IEEE: New York, NY, USA, 2015. [Google Scholar]
  18. Rodriguez-Navarro, C. DSUALM10H. Available online: https://zenodo.org/records/13739915 (accessed on 26 March 2025).
  19. Rodriguez-Navarro, C.; Portillo, F.; Castro-Santos, L.; Filgueira-Vizoso, A.; Montoya, F.G.; Alcayde, A. Optimising Energy Efficiency Enhancing NILM through High—Resolution Data Analytics. REPQJ 2024, 22, 85–91. [Google Scholar] [CrossRef] [PubMed]
  20. Kelly, J.; Batra, N.; Parson, O.; Dutta, H.; Knottenbelt, W.; Rogers, A.; Sigh, A.; Srivastava, M. NILMTK v0.2: A Non-Intrusive Load Monitoring Toolkit for Large Scale Data Sets. In Proceedings of the BuildSys 2014—1st ACM Conference on Embedded Systems for Energy-Efficient Buildings, Memphis, TN, USA, 3 November 2014; Association for Computing Machinery: New York, NY, USA, 2014; pp. 182–183. [Google Scholar]
  21. Kukunuri, R.; Batra, N.; Pandey, A.; Malakar, R.; Kumar, R.; Krystalakos, O.; Zhong, M.; Meira, P.; Parson, O. NILMTK-Contrib: Towards Reproducible State-of-the-Art Energy Disaggregation. In BuildSys ’19: Proceedings of the 6th ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation, New York, NY, USA, 13–14 November 2019; Association for Computing Machinery: New York, NY, USA, 2019. [Google Scholar]
  22. Rodriguez-Navarro, C. Performance Analysis of Nine Disaggregation Algorithms on the DSUALM10 Dataset: A Notebook Study. Available online: https://zenodo.org/records/15575404 (accessed on 11 May 2025).
  23. Rodríguez-Navarro, C.; Portillo, F.; Martínez, F.; Manzano-Agugliaro, F.; Alcayde, A. Development and Application of an Open Power Meter Suitable for NILM. Inventions 2023, 9, 2. [Google Scholar] [CrossRef]
  24. Kelly, J.; Knottenbelt, W. Neural NILM: Deep Neural Networks Applied to Energy Disaggregation. In BuildSys 2015—Proceedings of the 2nd ACM International Conference on Embedded Systems for Energy-Efficient Built, Seoul, Republic of Korea, 4–5 November 2015; Association for Computing Machinery: New York, NY, USA, 2015; pp. 55–64. [Google Scholar]
  25. Yu, K.; Vasios, N.; Perol, T. Energy Disaggregation from Non-Intrusive Load Monitoring Combinatorial Optimization, Factorial Hidden Markov Model, Convolutional Neural Network; Harvard John A. Paulson School of Engineering and Applied Sciences: Boston, MA, USA, 2016; AM 207. [Google Scholar]
  26. Bonfigli, R.; Principi, E.; Fagiani, M.; Severini, M.; Squartini, S.; Piazza, F. Non-Intrusive Load Monitoring by Using Active and Reactive Power in Additive Factorial Hidden Markov Models. Appl. Energy 2017, 208, 1590–1607. [Google Scholar] [CrossRef]
  27. Ghahramani, Z.; Jordan, M.I. Factorial Hidden Markov Models. Mach. Learn. 1997, 29, 245–273. [Google Scholar] [CrossRef]
  28. Adewole, K.S.; Torra, V. Energy Disaggregation Risk Resilience through Microaggregation and Discrete Fourier Transform. Inf. Sci. 2024, 662, 120211. [Google Scholar] [CrossRef]
  29. Shastri, H.; Batra, N. Neural Network Approaches and Dataset Parser for NILM Toolkit. In Proceedings of the BuildSys 2021—2021 ACM International Conference on Systems for Energy-Efficient Built Environments, Coimbra, Portugal, 17–18 November 2021; Association for Computing Machinery: New York, NY, USA, 2021. [Google Scholar]
Figure 1. Advancing NILM for energy efficiency.
Figure 1. Advancing NILM for energy efficiency.
Inventions 10 00043 g001
Figure 2. NILM dataset characteristics.
Figure 2. NILM dataset characteristics.
Inventions 10 00043 g002
Figure 3. F1-SCORE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Figure 3. F1-SCORE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Inventions 10 00043 g003
Figure 4. MAE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Figure 4. MAE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Inventions 10 00043 g004
Figure 5. NDE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Figure 5. NDE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Inventions 10 00043 g005
Figure 6. RMSE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Figure 6. RMSE per algorithm and appliance (electric furnace (EF), microwave (MW), television (TV), incandescent lamp (Lamp), vacuum cleaner (VAC), electric space heater (ESH), electric shower heater (ESHw), fan, fridge, and freezer.).
Inventions 10 00043 g006
Figure 7. Metrics, including harmonics and filters. (a) MAE, (b) RMSE, (c) F1-SCORE, (d) NDE.
Figure 7. Metrics, including harmonics and filters. (a) MAE, (b) RMSE, (c) F1-SCORE, (d) NDE.
Inventions 10 00043 g007
Figure 8. Heat map for appliances and algorithms. (a) F1-SCORE, (b) RMSE, (c) NDE, (d) MAE.
Figure 8. Heat map for appliances and algorithms. (a) F1-SCORE, (b) RMSE, (c) NDE, (d) MAE.
Inventions 10 00043 g008
Figure 9. Disaggregated data for DSUALM10H. (a) Using Mean, (b) using CO. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11)).
Figure 9. Disaggregated data for DSUALM10H. (a) Using Mean, (b) using CO. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11)).
Inventions 10 00043 g009
Figure 10. Disaggregated data for DSUALM10H using FHMM (a) and Hart85 (b). The appliances are electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).
Figure 10. Disaggregated data for DSUALM10H using FHMM (a) and Hart85 (b). The appliances are electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).
Inventions 10 00043 g010
Figure 11. Disaggregated data for DSUALM10H. (a) Using DAE, (b) using RNN. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).)
Figure 11. Disaggregated data for DSUALM10H. (a) Using DAE, (b) using RNN. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).)
Inventions 10 00043 g011
Figure 12. Disaggregated data for DSUALM10H. (a) Using Seq2Point, (b) using Seq2Seq. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).)
Figure 12. Disaggregated data for DSUALM10H. (a) Using Seq2Point, (b) using Seq2Seq. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).)
Inventions 10 00043 g012
Figure 13. Disaggregated data for DSUALM10H using WindowGRU. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).)
Figure 13. Disaggregated data for DSUALM10H using WindowGRU. (Appliances: electric furnace (2), microwave (3), television (4), incandescent lamp (5), vacuum cleaner (6), electric space heater (7), electric shower heater (8), fan (9), fridge (10), and freezer (11).)
Inventions 10 00043 g013
Figure 14. Execution times by algorithm using a dedicated machine running Ubuntu 25.04, Intel Core i5-4210U CPU @ 1.70 GHz with 16 GB RAM and 500 GB SSD.
Figure 14. Execution times by algorithm using a dedicated machine running Ubuntu 25.04, Intel Core i5-4210U CPU @ 1.70 GHz with 16 GB RAM and 500 GB SSD.
Inventions 10 00043 g014
Figure 15. Standardized comparison of NILMTK-Contrib algorithm metrics.
Figure 15. Standardized comparison of NILMTK-Contrib algorithm metrics.
Inventions 10 00043 g015
Table 1. Metric averages based on sample rate, unfiltered, and with harmonic content.
Table 1. Metric averages based on sample rate, unfiltered, and with harmonic content.
Sampling RateAlgorithmMAE (W)RMSE (W)F1-SCORENDE
250 msMean2622980.520.82
CO2203900.441.09
FHMM2003600.380.98
Hart852903200.121
500 msMean2703100.510.83
CO2304100.431.11
FHMM2103700.390.99
Hart852953300.121
1 sMean2803230.510.85
CO2454800.431.17
FHMM2183730.41
Hart853033400.111
15 sMean2623070.480.87
CO2263870.411.19
FHMM2713250.121.01
Hart852103400.361.02
30 sMean2773280.480.88
CO2394200.411.21
FHMM2953530.131.02
Hart852253570.321.04
60 sMean3053550.460.83
CO2413930.391.16
FHMM3193880.111.04
Hart852573780.31.05
90 sMean3233740.440.86
CO2534100.371.18
FHMM3374060.11.06
Hart852713970.281.07
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rodriguez-Navarro, C.; Portillo, F.; Robalo, I.; Alcayde, A. Evaluation of Traditional and Data-Driven Algorithms for Energy Disaggregation Under Sampling and Filtering Conditions. Inventions 2025, 10, 43. https://doi.org/10.3390/inventions10030043

AMA Style

Rodriguez-Navarro C, Portillo F, Robalo I, Alcayde A. Evaluation of Traditional and Data-Driven Algorithms for Energy Disaggregation Under Sampling and Filtering Conditions. Inventions. 2025; 10(3):43. https://doi.org/10.3390/inventions10030043

Chicago/Turabian Style

Rodriguez-Navarro, Carlos, Francisco Portillo, Isabel Robalo, and Alfredo Alcayde. 2025. "Evaluation of Traditional and Data-Driven Algorithms for Energy Disaggregation Under Sampling and Filtering Conditions" Inventions 10, no. 3: 43. https://doi.org/10.3390/inventions10030043

APA Style

Rodriguez-Navarro, C., Portillo, F., Robalo, I., & Alcayde, A. (2025). Evaluation of Traditional and Data-Driven Algorithms for Energy Disaggregation Under Sampling and Filtering Conditions. Inventions, 10(3), 43. https://doi.org/10.3390/inventions10030043

Article Metrics

Back to TopTop