Next Article in Journal
Case Study: Improving the Quality of Dairy Cow Reconstruction with a Deep Learning-Based Framework
Next Article in Special Issue
Design and Implementation of a Cloud-IoT-Based Home Energy Management System
Previous Article in Journal
Deep Learning in the Detection of Disinformation about COVID-19 in Online Space
Previous Article in Special Issue
Toward an Intelligent Campus: IoT Platform for Remote Monitoring and Control of Smart Buildings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning-Based Ensemble Classifiers for Anomaly Handling in Smart Home Energy Consumption Data

by
Purna Prakash Kasaraneni
1,
Yellapragada Venkata Pavan Kumar
2,
Ganesh Lakshmana Kumar Moganti
2,* and
Ramani Kannan
3
1
School of Computer Science and Engineering, VIT-AP University, Amaravati 522237, Andhra Pradesh, India
2
School of Electronics Engineering, VIT-AP University, Amaravati 522237, Andhra Pradesh, India
3
Department of Electrical and Electronics Engineering, Universiti Teknologi Petronas (UTP), Seri Iskandar 32610, Malaysia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(23), 9323; https://doi.org/10.3390/s22239323
Submission received: 3 November 2022 / Revised: 22 November 2022 / Accepted: 25 November 2022 / Published: 30 November 2022
(This article belongs to the Special Issue Smart IoT System for Renewable Energy Resource)

Abstract

:
Addressing data anomalies (e.g., garbage data, outliers, redundant data, and missing data) plays a vital role in performing accurate analytics (billing, forecasting, load profiling, etc.) on smart homes’ energy consumption data. From the literature, it has been identified that the data imputation with machine learning (ML)-based single-classifier approaches are used to address data quality issues. However, these approaches are not effective to address the hidden issues of smart home energy consumption data due to the presence of a variety of anomalies. Hence, this paper proposes ML-based ensemble classifiers using random forest (RF), support vector machine (SVM), decision tree (DT), naive Bayes, K-nearest neighbor, and neural networks to handle all the possible anomalies in smart home energy consumption data. The proposed approach initially identifies all anomalies and removes them, and then imputes this removed/missing information. The entire implementation consists of four parts. Part 1 presents anomaly detection and removal, part 2 presents data imputation, part 3 presents single-classifier approaches, and part 4 presents ensemble classifiers approaches. To assess the classifiers’ performance, various metrics, namely, accuracy, precision, recall/sensitivity, specificity, and F1 score are computed. From these metrics, it is identified that the ensemble classifier “RF+SVM+DT” has shown superior performance over the conventional single classifiers as well the other ensemble classifiers for anomaly handling.

1. Introduction

Considering the global thrust towards the development of grid-independent and green energy systems for addressing the unrelenting growth of loads as well as environmental pollution, smart home and renewable energy-based microgrid culture has been increasing worldwide. Smart cities are new-era establishments where all the smart homes are jointly operated to consolidate and optimize electricity utilization. As these establishments are realized with a combination of electrical, communication, and information technology, the gathering of quality data is a challenging task. Smart homes connected to the power network continuously generate huge volumes of energy consumption data, which is normally a combination of timestamps and readings. The reading information in this data is a key value that helps in understanding the energy consumption behavior, billing generation, load profiling, forecasting, contingency analysis, device health condition analysis, etc. All these operations rely upon the quality of the data being captured. However, this data often may consist of different anomalies, viz., garbage data, outliers, redundant data, and missing data due to malfunctioning of advanced metering infrastructure, failure of communication channels, unanticipated issues in power networks, etc. If these anomalies are left unhandled in the dataset, there will be an adverse effect on the system operations and further delude the analytics of the energy consumption data. So, handling these anomalies is highly essential to enable analysts to perform accurate energy data analytics. Thus, the multifaceted nature of the smart home data when compared to other datasets gains importance in the data analysis field. So, this becomes an important research focus for data analysts when compared to the datasets of other applications. This is the prime motivation for the proposed work of this paper.
Big data refers to a huge quantity of data. However, the data quality is a more complex and significant aspect than the quantity in the direction of research [1]. Moreover, the issues related to data quality have gained much importance and attention in energy big data analytics [2]. The increased use of several intelligent devices in power system applications has become the major source of big data, which reflects on data storage, data processing, and data quality [3,4,5]. The failure of these intelligent devices makes data incomplete during the acquisition of energy consumption data. This incomplete data is commonly referred to as missing data [6,7,8]. Handling rather than ignoring this missing data drives toward better data analytics on energy consumption [9]. Hence, it is essential to analyze and impute the missing data in the smart home energy consumption data. Following this, several state-of-the-art works on missing data imputation and ensemble methods are discussed as follows.
The researchers suggested several ML-based imputation methods as well as thorough benchmarks for the comparison between conventional and modern methods [10]. An imputation algorithm “opt.impute” was introduced in [11] to achieve the finest solutions to the missing data. Further, an extensive review was conducted on the imputation of missing data using ML which helps in understanding the limitations of ML imputation methods [12]. A framework was implemented to improvise the multivariate imputation by chained equations (MICE) in imputing the missing sensor data [13]. A graph-based method was discussed in [14] to impute the missing sensor data. A copy-paste imputation method was introduced in [15] to impute the time-series data of energy. A mixture factor analysis method was discussed to estimate the missing data in the building’s energy load [16]. Different imputation methods, viz., MICE, KNN, and RF-based imputation were implemented to impute the missing data in the sensor data of the internet of things [17]. A data splitting-based imputation method named “nullify the missing values before the imputation” was proposed to impute the missing data [18].
A new statistical and ML-based imputation method was implemented in [19] to impute missing data in the applications of power grids. A fuzzy inductive reasoning method was discussed to deal with the missing data during the forecasting process in smart grids [20]. A six-stage particle swarm optimization imputation method was implemented for smart meter data collected from an Indian institution [21]. An imputation method based on a denoising autoencoder was presented in [22]. An imputation model named “bagged averaging of multiple linear regression” was discussed in [23] for imputing missing data in phasor measurement units. A two-stage deep autoencoder-based data imputation method was discussed in [24] for imputing missing data in wind farms. A bagging algorithm was implemented to impute the missing data in time-series data [25]. An autoencoder neural network was presented to impute missing data for classification [26]. The appropriate selection of the best imputation method and classification was discussed in [27]. An extensive study on the packages available in “R” for data imputation was presented in [28]. Electricity theft detection in smart grids using various ML algorithms and deep learning techniques was discussed in [29,30]. An AdaBoost ensemble model was implemented to detect electricity theft [31]. An improvised ensemble model of a general regression neural network and successive geometric transformations model was presented in [32] to recover the partial or fully missed data.
In summary, the abovementioned literature discusses the concepts of big data, sources of big data, and energy data analytics. In addition, the importance of handling anomalies in big data was discussed. To handle the anomalies in energy consumption datasets, a few imputation methods such as data splitting, fuzzy inductive reasoning, denoising autoencoder, and bagging are used. Further, to evaluate their performance, various single classifiers, namely, SVM, neural networks, etc., are used. However, these approaches are found ineffective to address the hidden issues of smart home energy consumption data due to the presence of a variety of anomalies such as garbage data, outlier data, redundant data, missing data, etc.
On the other hand, in recent days, the ensemble classification approach is supporting effective classification in data imputation in different applications, which was not tried for the smart home energy consumption data. With this motivation, this paper proposes ML-based ensemble classifiers to handle all the possible anomalies in smart home energy consumption data. The major contributions of this paper are summarized as follows:
The proposed approach initially identifies all anomalies and removes them, and then imputes this information. The entire implementation consists of four parts.
-
Part 1 (anomaly detection and removal) considers the original dataset and refines it by removing all the identified anomalies.
-
Part 2 (data imputation) considers this refined dataset and performs the missing data imputation using median, KNN, and bagging imputation methods, thereby producing an anomaly-free dataset.
-
Part 3 (single-classifier approaches) performs the classification of the dataset using the conventional single-classifier approaches such as RF, SVM, DT, NB, KNN, and NNET.
-
Part 4 (ensemble classifiers approaches) performs the classification of the dataset using the proposed ensemble classifier approaches such as RF+SVM+DT, RF+SVM+NB, RF+SVM+KNN, RF+SVM+NNET, RF+DT+NB, RF+DT+KNN, RF+DT+NNET, RF+NB+KNN, RF+NB+NNET, and RF+KNN+NNET.
To assess the classifiers’ performance, various metrics, namely, accuracy, precision, recall/sensitivity, specificity, and F1 score are computed. From these metrics, it is identified that the ensemble classifier “RF+SVM+DT” has shown superior performance over the conventional single classifiers as well the other ensemble classifiers for anomaly handling in smart home energy consumption data.
All these contributions are structured in the paper as follows. Section 2 presents the description of the dataset. Section 3 presents the description and implementation of the proposed approach. Section 4 presents simulation results and their discussion. Finally, Section 5 concludes the outcomes of the paper in a synopsized way.

2. Description of Dataset

To implement the proposed approach, the data of an appliance (refrigerator) from the Tracebase dataset [33] is considered. This dataset consists of 43 different appliances with 158 device IDs that are connected to various smart homes/buildings. Each appliance consists of CSV files that represent the energy consumption data of a day. A detailed description of this dataset can be obtained from [34]. Further, this dataset was considered and used in various literary works. The Tracebase dataset was used in the extensive study of different non-intrusive load monitoring (NILM) power consumption datasets described in [35,36,37]. The present and the future directions for energy management techniques using NILM datasets are discussed in [38].
The CSV file (dev_98C08A_2011.09.17.csv) data of the refrigerator appliance is prepared with the columns such as CAPTURED_DATE, CAPTURED_HOUR, CAPTURED_MINUTE, CAPTURED_SECOND, and CAPTURED_READING for implementing the proposed ensemble classifier approach.

3. Description and Implementation of the Proposed Approach

The conceptual model of the proposed approach is shown in Figure 1. It consists of four parts, viz., Part 1, Part 2, Part 3, and Part 4. The smart home energy consumption dataset will be given as input to Part 1. In Part 1, an analysis of the missing data will be carried out for understanding the missingness in the original dataset. Further, the identification and removal of different anomalies (viz., garbage data, outliers in the data, and redundant data) will be performed. From this, a dataset with the abovementioned anomalies removed will be produced and given as input to Part 2. In Part 2, the imputation of missing data will be completed. In Part 3, a single-classifier approach will be applied. This will provide a recommendation of the best single-classifier approach as the output. By taking this best single-classifier as the basis, the ensemble classifiers approach will be applied in Part 4. This will provide a recommendation of the best ensemble classifier to perform the imputation.
The implementation flow of the proposed ensemble classifiers approach through all the proposed parts is shown in Figure 2. The detailed description and implementation processes are discussed in Section 3.1, Section 3.2, Section 3.3 and Section 3.4 respectively for Part 1, Part 2, Part 3, and Part 4.

3.1. Implementation of Part 1 (Anomaly Detection and Removal)

The process starts from Part 1 by reading the smart home energy consumption dataset and saving it in an object “shec_dat”. Initially, the missing data information in this “shec_dat” is analyzed [6]. Further, the process is continued with the identification of garbage data in the dataset. To identify the garbage data i.e., the data other than the numerical data, a function grepl(“[[:digit:]] is used on each column of the dataset. If garbage data exists, those records are removed and the remaining data are given as input to identify the outliers data. If there is no garbage data, the existing dataset is used as it is and is given as input to identify the outliers data. The outliers data is the data that does not exist within the expected range. To identify outliers data, a boxplot analysis is applied to the data obtained after removing the garbage data. The boxplot analysis is a standardized approach to showing data distribution in a five-number summary (i.e., minimum, first quartile, median, third quartile, and maximum). The data that lies in between the “minimum” and “maximum” values are considered as the data within the range and useful for the analysis. The data that lies below the “minimum” and above the “maximum” values are considered outliers data and needs to be removed to achieve better analytics. The function boxplot() is applied to the readings column by using boxplot(shec_dat$CAPTURED_READING, plot=F)$out. If the outliers exist in the readings column, those records are removed and the remaining data are given as input to identify the redundant data. If there are no outliers, the existing dataset is used as it is to identify the redundant data. In general, redundant data refers to the duplication of the entire record in the dataset. However, in this case, there exist two types of redundant data in the dataset. They are the records with the same timestamp and same reading information, and records with the same timestamp and different reading information. The detailed process of identifying these types of redundant data is discussed in [39]. If the abovementioned types of redundant data exist, those records are removed. If there are no redundant data, the existing data is used as it is to perform the next step. At the end of Part 1, a dataset is obtained after removing all the anomalies (garbage data, outliers data, and redundant data). As several records in the dataset are removed due to the existence of different anomalies, this dataset consists of missing timestamps. Hence, these missing timestamps are filled, and the respective reading information is set to “NA (Not Available)” [8] before proceeding to the implementation of Part 2.

3.2. Implementation of Part 2 (Data Imputation)

Once all the missing records are finalized in the dataset obtained after removing all the anomalies, the imputation methods such as median imputation, KNN imputation, and bagging imputation are applied. The implementation of these imputation methods produces datasets with imputed reading values. Further, the single-classifier approach is applied to these imputed datasets. The implementation of the median, KNN, and bagging imputation methods is discussed in Section 3.2.1, Section 3.2.2 and Section 3.2.3 respectively.

3.2.1. Implementation of the Median Imputation Method

In the median imputation method, the median value of the reading information in the CAPTURED_READING column is calculated, and that value is used for imputing the missing reading information. This imputation method is simple and fast. The process of calculating the median value starts with the ordering of readings information in ascending order. Once the ordering of readings information is done, then the number of values (odd or even) in the CAPTURED_READING is taken into consideration. Here, the number of values plays a major role in calculating the median value of the readings information. The formula for calculating the median value is given in Equation (1).
M e d i a n ( D ) = { D ( s + 1 2 ) i f   s   i s   o d d D ( s 2 ) + D ( s 2 + 1 ) 2 i f   s   i s   e v e n
where D = list of values ordered in the CAPTURED_READING column, and s = number of values in the CAPTURED_READING column.
If the number of values in the CAPTURED_READING column is odd, then the middle value is considered as the median. If the number of values is even in the CAPTURED_READING column, then the average of the middle two values is considered as the median.

3.2.2. Implementation of the KNN Imputation Method

In the KNN imputation method, the distance between the k-nearest neighbor values is calculated by using the Euclidean distance metric. In the CAPTURED_READING column, the distance between the k-closest samples of the readings is calculated and that distance value is used to impute the missing reading information. The formula for calculating Euclidean distance is given in Equation (2).
d i s t ( p , q ) = i = 1 m ( p i q i ) 2
where dist = Euclidean distance, m = number of points, and pi & qi are the points.

3.2.3. Implementation of the Bagging Imputation Method

In the bagging imputation method, the term ‘bagging’ refers to bootstrap aggregation. The bootstrap is a statistical technique of iteratively resampling the data with replacement in the dataset. To perform this, initially, the number of bootstrap samples is to be fixed, and then the sample size. For each sample of bootstrap the following steps are performed: draw the sample with replacement, fit the model, anticipate the performance of the model based on the out-of-bag sample, and calculate the average of the sample of the model. The multiple iterations of sampling improve the prediction performance of the model. The bagging method fits a bagged tree. This method is simple, powerful, and accurate to impute the missing values in the readings information. However, it is computationally high-cost.

3.3. Implementation of Part 3 (Single-Classifier Approach)

In this section, the single-classifier approach is performed using various classifiers, viz., RF, SVM, DT, NB, KNN, and NNET, for the classification. All these classifiers are implemented individually on the dataset. To implement these, the dataset is divided into train_set and test_set. These classifiers are trained on the train_set using k-fold cross-validation. Here, the k-value considered is 10. Further, these classifiers are applied to the test_set to predict the classes Yes (Y) or No (N). Here, class ‘Y’ represents missing data, and class ‘N’ represents non-missing data. After the implementation, the performance metrics such as accuracy, precision, recall/sensitivity, specificity, and F1 score are computed using a confusion matrix to evaluate each classifier’s performance. The confusion matrix is shown in Figure 3 and the formulae for computing the performance metrics are given in Equations (3)–(7).
A c c u r a c y = T . P o s . + T . N e g . T . P o s . + T . N e g . + F . P o s . + F . N e g .
P r e c i s i o n = T . P o s . T . P o s . + F . P o s .
R e c a l l / S e n s i t i v i t y = T . P o s . T . P o s . + F . N e g .
S p e c i f i c i t y = T . N e g . T . N e g . + F . P o s .
F 1 S c o r e = 2 ( P r e c i s i o n R e c a l l ) ( P r e c i s i o n + R e c a l l )
If all the single classifiers are implemented and their performance is verified, then, the best single classifier is recommended. Otherwise, the performance metrics are re-verified.

3.4. Implementation of Part 4 (Ensemble Classifiers Approach)

This section uses the best single classifier recommended in Part3 as the input to develop ensemble classifiers. The ensemble of classifiers is performed using the “stacking” method. In stacking, there are two layers called the top layer and the bottom layer. The top layer consists of a classifier, which is referred to as a base classifier and the bottom layer consists of other classifiers. The output of the bottom layer is given as input to the top layer. The classifier used in the top layer is an ensemble with the output of the bottom layer classifiers, which produces an ensemble classifier. The stacking of classifiers is shown in Figure 4. From this figure, it is seen that the single classifiers used in the bottom layer are an ensemble with the recommended best classifier used in the top layer. For example, the single classifiers SVM and DT are part of the ensemble with the recommended best classifier RF. Similarly, all the other single classifiers form an ensemble with RF and produce ensemble classifiers. To implement these ensemble classifiers, the imputed datasets are given as input. Further, each imputed dataset is divided into train_set and test_set. The ensemble classifiers are trained on the train_set using k-fold cross-validation. Here, the k-value considered is 10. Further, these ensemble classifiers are applied to test_set to predict the classes Y or N. After the implementation, the performance metrics such as accuracy, precision, recall/sensitivity, specificity, and F1 score are computed using a confusion matrix to evaluate each ensemble classifier’s performance. If all the ensemble classifiers are implemented and their performance is verified then the best ensemble classifier for the imputation is recommended, otherwise the performance metrics are re-verified.

4. Simulation Results and Discussion

In keeping with the aims of the paper, the simulation results of the implementation are presented in three subsections. Section 4.1, Section 4.2 and Section 4.3 present the results corresponding to anomaly detection and removal, single-classifier approach, and ensemble classifiers approach, respectively.

4.1. Results Corresponding to Anomaly Detection and Removal

This section presents the details of the missing data in the original CSV file (original dataset) and the missing data in the dataset after eliminating the anomalies. The number of records in this original dataset is 155,374. During the analysis of missing data, 700 records are missed in the original dataset [7]. During the identification of garbage data, no garbage data (other than numerical data) are identified in the original CSV file. Hence, no records are removed and the same number of records (155,374) are available. During the identification of outliers data, there are 25 readings identified as outliers and the respective records are removed from the dataset. The removal of records with outliers left the dataset with 155,349 records. During the identification of redundant data, the records with the same timestamp and same reading are identified and those records are removed from the dataset. This removal left the dataset with 98,779 records. Further, the records with the same timestamp and different readings are identified and those records are removed from the dataset. This removal left the dataset with 72,597 records. Once the redundant data are removed, the missing data are filled with the respective timestamps and the respective reading with NA value, as shown in Figure 5 (all the highlighted rows). After this filling, there are 86,400 records in the dataset, out of these, 13,803 records contain missing readings.
The proportions of the available data and missing data in the original dataset and the dataset available after removing anomalies are shown in Figure 6a–c. These figures show the proportion of the missing data and available data in the considered dataset in three different scenarios, namely, (i) consideration of the original dataset, (ii) consideration of the dataset that is obtained after removing the anomalies, and (iii) consideration of the dataset after filling the missing timestamps, ready for the imputation.
From Figure 6a, it is understood that the proportion of available data is 99.55% and missing data is 0.45% in the original dataset. From Figure 6b, it is seen that the proportion of available data is 84% and missing data is 16% in all columns of the dataset obtained after removing anomalies. From Figure 6c, it is evident that there are no missing data in the columns CAPTURED_DATE, CAPTURED_HOUR, CAPTURED_MINUTE, CAPTURED_SECOND and the proportion of data availability is 84%. Further, there are missing readings in the column CAPTURED_READING with a proportion of 16%.

4.2. Results Corresponding to the Single-Classifier Approach

This section presents the performance of a single-classifier approach on the imputed datasets. The performance of classifiers in the median, KNN, and bagging imputation methods are discussed in Section 4.2.1, Section 4.2.2 and Section 4.2.3, respectively.

4.2.1. Performance of the Single-Classifier Approach in the Median Imputation Method

The performance metrics of each classifier are shown in Figure 7, where the red colored bar(s) indicate the highest value achieved corresponding to that particular metric. From this, the highest accuracy value of 98.1% is observed in RF, while the lowest accuracy value of 76.3% is observed in KNN, as shown in Figure 7a. The highest precision value of 99% is observed in RF, while the lowest precision value of 80.5% is observed in SVM and NB, as shown in Figure 7b. The highest recall value of 100% is observed in SVM and NB, while the lowest recall value of 87.9% is observed in KNN, as shown in Figure 7c. The highest specificity value of 95.9% is observed in RF, while the lowest specificity value of 0% is observed in SVM and NB, as shown in Figure 7d. The highest F1 Score value of 98.8% is observed in RF, while the lowest F1 Score value of 85.7% is observed in KNN, as shown in Figure 7e. From the subplots in Figure 7a–e, it is understood that the classifier RF has outperformed the others. Further, the performance summary of all the single classifiers is given in Table 1.

4.2.2. Performance of the Single-Classifier Approach in the KNN Imputation Method

The performance metrics of each classifier are shown in Figure 8, where the red colored bar(s) indicate the highest value achieved corresponding to that particular metric. From this, the highest accuracy value of 87.7% is observed in RF, while the lowest accuracy value of 68% is observed in NNET, as shown in Figure 8a. The highest precision value of 86.6% is observed in RF, while the lowest precision value of 80.3% is observed in NNET, as shown in Figure 8b. The highest recall value of 100% is observed in RF, SVM, DT, NB, and KNN, while the lowest recall value of 79.9% is observed in NNET, as shown in Figure 8c. The highest specificity value of 36.3% is observed in RF, while the lowest specificity value of 0% is observed in SVM and KNN, as shown in Figure 8d. The highest F1 Score value of 92.8% is observed in RF, while the lowest F1 Score value of 80.1% is observed in NNET, as shown in Figure 8e. From the subplots in Figure 8a–e, it is understood that the classifier RF has outperformed the others. Further, the percentage summary of all classifiers is given in Table 2.

4.2.3. Performance of the Single-Classifier Approach in the Bagging Imputation Method

The performance metrics of each classifier are shown in Figure 9, where the red colored bar(s) indicate the highest value achieved corresponding to that particular metric. From this, the highest accuracy value of 95.2% is observed in RF, while the lowest accuracy value of 75.7% is observed in NNET, as shown in Figure 9a. The highest precision value of 100% is observed in RF and DT, while the lowest precision value of 79.5% is observed in NNET, as shown in Figure 9b. The highest recall value of 100% is observed in SVM, and NB, while the lowest recall value of 84.3% is observed in DT, as shown in Figure 9c. The highest specificity value of 100% is observed in RF and DT, while the lowest specificity value of 0% is observed in SVM, NB, and NNET, as shown in Figure 9d. The highest F1 Score value of 96.9% is observed in RF, while the lowest F1 Score value 86.1% is observed in NNET, as shown in Figure 9e. From the subplots Figure 9a–e, it is understood that the classifier RF has outperformed the others. Further, the percentage summary of all classifiers is given in Table 3.

4.3. Results Corresponding to the Ensemble Classifiers Approach

This section presents the performance of the ensemble classifiers approaches on the imputed datasets. The performance of ensemble classifiers in the median, KNN, and bagging imputation methods are discussed in Section 4.3.1, Section 4.3.2 and Section 4.3.3 respectively.

4.3.1. Performance of the Ensemble Classifiers Approach in the Median Imputation Method

The performance metrics of each ensemble classifier are shown in Figure 10, where the red colored bar(s) indicate the highest value achieved corresponding to that particular metric. From this, the highest accuracy value of 98.9% is observed in RF+SVM+DT and RF+DT+NNET, while the lowest accuracy value of 72.5% is observed in RF+NB+KNN, as shown in Figure 10a. The highest precision value of 99.4% is observed in RF+SVM+NB, while the lowest precision value of 77.5% is observed in RF+SVM+KNN, as shown in Figure 10b.
The highest recall value of 100% is observed in RF+SVM+DT, RF+DT+NB, RF+DT+NNET, RF+NB+NNET, and RF+KNN+NNET, while the lowest recall value of 81.6% is observed in RF+SVM+NB, as shown in Figure 10c. The highest specificity value of 94.5% is observed in RF+SVM+DT, RF+DT+NB, RF+DT+NNET, and RF+KNN+NNET, while the lowest specificity value of 34.7% is observed in RF+NB+KNN, as shown in Figure 10d.
The highest F1 Score value of 99.3% is observed in RF+SVM+DT, RF+DT+NB, RF+DT+NNET, and RF+KNN+NNET, while the lowest F1 Score value of 82.2% is observed in RF+SVM+KNN and RF+NB+KNN, as shown in Figure 10e. From the subplots in Figure 10a–e, it is understood that the ensemble classifiers RF+SVM+DT and RF+DT+NNET have outperformed the others.
Further, the performance summary of all ensemble classifiers with respect to various parameters is given in Table 4.

4.3.2. Performance of the Ensemble Classifiers Approach in the KNN Imputation Method

The performance metrics of each ensemble classifier are shown in Figure 11, where the red colored bar(s) indicate the highest value achieved corresponding to that particular metric. From this, the highest accuracy value of 80.2% is observed in RF+DT+KNN, while the lowest accuracy value of 70.9% is observed in RF+SVM+KNN, as shown in Figure 11a. The highest precision value of 99.3% is observed in RF+DT+KNN, while the lowest precision value of 81.3% is observed in RF+NB+NNET, as shown in Figure 11b.
The highest recall value of 82.6% is observed in RF+NB+NNET, while the lowest recall value of 80.5% is observed in RF+SVM+DT, RF+SVM+NNET, as shown in Figure 11c. The highest specificity value of 43.6% is observed in RF+DT+NNET, while the lowest specificity value of 19.2% is observed in RF+SVM+NNET, as shown in Figure 11d. The highest F1 Score value of 89% is observed in RF+DT+KNN, while the lowest F1 Score value of 81.9% is observed in RF+NB+NNET, as shown in Figure 11e. From the subplots in Figure 11a–e, it is understood that the ensemble classifier RF+DT+KNN has outperformed the others. Further, the performance summary of all ensemble classifiers with respect to various parameters is given in Table 5.

4.3.3. Performance of the Ensemble Classifiers Approach in the Bagging Imputation Method

The performance metrics of each ensemble classifier are shown in Figure 12, where the red colored bar(s) indicate the highest value achieved corresponding to that particular metric. From this, the highest accuracy value of 89.6% is observed in RF+SVM+DT, while the lowest accuracy value of 71.2% is observed in RF+SVM+KNN, as shown in Figure 12a. The highest precision value of 98.8% is observed in RF+SVM+DT, while the lowest precision value of 86.5% is observed in RF+SVM+KNN, as shown in Figure 12b. The highest recall value of 89.4% is observed in RF+SVM+DT, while the lowest recall value of 78.8% is observed in RF+SVM+NNET, as shown in Figure 12c.
The highest specificity value of 91.1% is observed in RF+SVM+DT, while the lowest specificity value of 0.2% is observed in RF+SVM+NNET, as shown in Figure 12d. The highest F1 Score value of 93.9% is observed in RF+SVM+DT, while the lowest F1 Score value 82.8% is observed in RF+SVM+KNN, as shown in Figure 12e. From the subplots in Figure 12a–e, it is understood that the ensemble classifier RF+SVM+DT has outperformed the others.
Further, the performance summary of all ensemble classifiers with respect to various parameters is given in Table 6.

5. Conclusions

This paper proposes a machine learning-based ensemble classifiers approach to address the anomalies present in smart homes’ energy consumption data. This proposed approach has proven to be more effective than the conventional single-classifier approach that is presented in the literature. The salient observations from this work are summarized as follows:
All the possible anomalies are successfully identified and removed from the dataset. The number of records in the original dataset is 155,374 and the number of records available in the refined dataset after removing anomalies is 86,400, which is the actual expected number of records as per the dataset description.
Out of 86,400 records, 13,803 records are identified as records with missing data. This missing data has been successfully imputed by using various imputation methods (median, KNN, and bagging).
To assess the process of imputation, various conventional single-classifier approaches, as well as the proposed ensemble classifiers approaches, are implemented. From the computation of the performance metrics (accuracy, precision, recall/sensitivity, specificity, and F1 score), the RF classifier is identified as the superior single-classifier to all other single classifiers.
Out of the proposed ensemble classifiers, “RF+SVM+DT” has shown superior performance over the conventionally best single classifier (RF) as well the other ensemble classifiers for imputing the missing reading information.
Thus, the proposed ensemble classifiers approach has successfully handled anomalies that exist in the smart home energy consumption data.

Impacts and Implications of the Work

The proposed work in this paper helps in data preprocessing by the cleansing of data, which is typically essential to carry out precise analytics, and thereby, take superior decisions for energy management in smart buildings. Furthermore, the outcome of this work helps as a ready reference to understand the irregularities of the live data captured in a smart building/home/grid application for better data analytics. This impacts one of the important objectives of “United Nations Sustainable Development Goals (UN SDGs)—SDG 7: Energy” in producing an anomaly-free dataset for providing several customer services.
In addition, the identification of different data anomalies, viz., missing data, outliers data, garbage data, and redundant data in the energy consumption dataset, may be applied to the malfunctioning of metering infrastructure, failure/glitches of communication channels, cyber-attacks, energy thefts, unanticipated situations in power networks, etc.

Author Contributions

Conceptualization, P.P.K.; formal analysis, G.L.K.M.; funding acquisition, G.L.K.M.; investigation, P.P.K.; project administration, R.K.; software, P.P.K. and G.L.K.M.; supervision, Y.V.P.K.; Validation, Y.V.P.K.; writing—original draft, P.P.K.; writing—review & editing, Y.V.P.K. and R.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank VIT-AP University, Amaravati, Andhra Pradesh, India for funding the open-access publication fee for this research work.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

CSVcomma-separated values
DTdecision tree
KNNK-nearest neighbour
MLmachine learning
MICEmultivariate imputation by chained equations
NAnot available
NBnaive Bayes
NILMnon-intrusive load monitoring
NNETneural networks
RFrandom forest
SVMsupport vector machine

References

  1. Firmani, D.; Mecella, M.; Scannapieco, M.; Batini, C. On the Meaningfulness of “Big Data Quality”. Data Sci. Eng. 2016, 1, 6–20. [Google Scholar] [CrossRef] [Green Version]
  2. Chen, W.; Zhou, K.; Yang, S.; Wu, C. Data Quality of Electricity Consumption Data in a Smart Grid Environment. Renew. Sustain. Energy Rev. 2017, 75, 98–105. [Google Scholar] [CrossRef]
  3. Tu, C.; He, X.; Shuai, Z.; Jiang, F. Big Data Issues in Smart Grid—A Review. Renew. Sustain. Energy Rev. 2017, 79, 1099–1107. [Google Scholar] [CrossRef]
  4. Ghorbanian, M.; Dolatabadi, S.H.; Siano, P. Big Data Issues in Smart Grids: A Survey. IEEE Syst. J. 2019, 13, 4158–4168. [Google Scholar] [CrossRef]
  5. Bhattarai, B.P.; Paudyal, S.; Luo, Y.; Mohanpurkar, M.; Cheung, K.; Tonkoski, R.; Hovsapian, R.; Myers, K.S.; Zhang, R.; Zhao, P.; et al. Big Data Analytics in Smart Grids: State-of-the-art, Challenges, Opportunities, and Future Directions. IET Smart Grid 2019, 2, 141–154. [Google Scholar] [CrossRef]
  6. Kasaraneni, P.P.; Yellapragada, V.P.K. Simple and Effective Descriptive Analysis of Missing Data Anomalies in Smart Home Energy Consumption Readings. J. Energy Syst. 2021, 5, 199–220. [Google Scholar] [CrossRef]
  7. Kasaraneni, P.P.; Yellapragada Venkata, P.K. Analytical Approach to Exploring the Missing Data Behavior in Smart Home Energy Consumption Dataset. J. Renew. Energy Environ. 2022, 9, 37–48. [Google Scholar] [CrossRef]
  8. Kasaraneni, P.P.; Yellapragada, V.P.K. Systematic Statistical Analysis to Ascertain the Missing Data Patterns in Energy Consumption Data of Smart Homes. Int. J. Renew. Energy Res. 2022, 12, 1560–1573. [Google Scholar] [CrossRef]
  9. Emmanuel, T.; Maupong, T.; Mpoeleng, D.; Semong, T.; Mphago, B.; Tabona, O. A Survey on Missing Data in Machine Learning. J. Big Data 2021, 8, 140. [Google Scholar] [CrossRef]
  10. Jäger, S.; Allhorn, A.; Bießmann, F. A Benchmark for Data Imputation Methods. Front. Big Data 2021, 4, 693674. [Google Scholar] [CrossRef]
  11. Dimitris, B.; Pawlowski, C.; Ying, D.Z. From Predictive Methods to Missing Data Imputation: An Optimization Approach. J. Mach. Learn. Res. 2018, 18, 1–39. Available online: https://jmlr.csail.mit.edu/papers/volume18/17-073/17-073.pdf (accessed on 2 November 2022).
  12. Alabadla, M.; Sidi, F.; Ishak, I.; Ibrahim, H.; Affendey, L.S.; Che Ani, Z.; Jabar, M.A.; Bukar, U.A.; Devaraj, N.K.; Muda, A.S.; et al. Systematic Review of Using Machine Learning in Imputing Missing Values. IEEE Access 2022, 10, 44483–44502. [Google Scholar] [CrossRef]
  13. Wu, R.; Hamshaw, S.D.; Yang, L.; Kincaid, D.W.; Etheridge, R.; Ghasemkhani, A. Data Imputation for Multivariate Time Series Sensor Data with Large Gaps of Missing Data. IEEE Sens. J. 2022, 22, 10671–10683. [Google Scholar] [CrossRef]
  14. Jiang, X.; Tian, Z.; Li, K. A Graph-Based Approach for Missing Sensor Data Imputation. IEEE Sens. J. 2021, 21, 23133–23144. [Google Scholar] [CrossRef]
  15. Weber, M.; Turowski, M.; Cakmak, H.K.; Mikut, R.; Kuhnapfel, U.; Hagenmeyer, V. Data-Driven Copy-Paste Imputation for Energy Time Series. IEEE Trans. Smart Grid 2021, 12, 5409–5419. [Google Scholar] [CrossRef]
  16. Jeong, D.; Park, C.; Ko, Y.M. Missing Data Imputation Using Mixture Factor Analysis for Building Electric Load Data. Appl. Energy 2021, 304, 117655. [Google Scholar] [CrossRef]
  17. Okafor, N.U.; Delaney, D.T. Missing Data Imputation on IoT Sensor Networks: Implications for on-Site Sensor Calibration. IEEE Sens. J. 2021, 21, 22833–22845. [Google Scholar] [CrossRef]
  18. Bhagat, H.V.; Singh, M. NMVI: A Data-Splitting Based Imputation Technique for Distinct Types of Missing Data. Chemom. Intell. Lab. Syst. 2022, 223, 104518. [Google Scholar] [CrossRef]
  19. Su, T.; Shi, Y.; Yu, J.; Yue, C.; Zhou, F. Nonlinear Compensation Algorithm for Multidimensional Temporal Data: A Missing Value Imputation for the Power Grid Applications. Knowl.-Based Syst. 2021, 215, 106743. [Google Scholar] [CrossRef]
  20. Jurado, S.; Nebot, À.; Mugica, F.; Mihaylov, M. Fuzzy Inductive Reasoning Forecasting Strategies Able to Cope with Missing Data: A Smart Grid Application. Appl. Soft Comput. 2017, 51, 225–238. [Google Scholar] [CrossRef] [Green Version]
  21. Hemanth, G.R.; Charles Raja, S. Proposing Suitable Data Imputation Methods by Adopting a Stage Wise Approach for Various Classes of Smart Meters Missing Data—Practical Approach. Expert Syst. Appl. 2022, 187, 115911. [Google Scholar] [CrossRef]
  22. Ryu, S.; Kim, M.; Kim, H. Denoising Autoencoder-Based Missing Value Imputation for Smart Meters. IEEE Access 2020, 8, 40656–40666. [Google Scholar] [CrossRef]
  23. Le, N.T.; Benjapolakul, W. A Data Imputation Model in Phasor Measurement Units Based on Bagged Averaging of Multiple Linear Regression. IEEE Access 2018, 6, 39324–39333. [Google Scholar] [CrossRef]
  24. Liu, X.; Zhang, Z. A Two-Stage Deep Autoencoder-Based Missing Data Imputation Method for Wind Farm SCADA Data. IEEE Sens. J. 2021, 21, 10933–10945. [Google Scholar] [CrossRef]
  25. Andiojaya, A.; Demirhan, H. A Bagging Algorithm for the Imputation of Missing Values in Time Series. Expert Syst. Appl. 2019, 129, 10–26. [Google Scholar] [CrossRef]
  26. Choudhury, S.J.; Pal, N.R. Imputation of Missing Data with Neural Networks for Classification. Knowl. Based Syst. 2019, 182, 104838. [Google Scholar] [CrossRef]
  27. Sim, J.; Lee, J.S.; Kwon, O. Missing Values and Optimal Selection of an Imputation Method and Classification Algorithm to Improve the Accuracy of Ubiquitous Computing Applications. Math. Probl. Eng. 2015, 2015, 538613. [Google Scholar] [CrossRef] [Green Version]
  28. Yadav, M.L.; Roychoudhury, B. Handling Missing Values: A Study of Popular Imputation Packages in R. Knowl. Based Syst. 2018, 160, 104–118. [Google Scholar] [CrossRef]
  29. Banga, A.; Ahuja, R.; Sharma, S.C. Accurate Detection of Electricity Theft Using Classification Algorithms and Internet of Things in Smart Grid. Arab. J. Sci. Eng. 2022, 47, 9583–9599. [Google Scholar] [CrossRef]
  30. Khan, I.U.; Javeid, N.; Taylor, C.J.; Gamage, K.A.A.; Ma, X. A Stacked Machine and Deep Learning-Based Approach for Analysing Electricity Theft in Smart Grids. IEEE Trans. Smart Grid 2022, 13, 1633–1644. [Google Scholar] [CrossRef]
  31. Qu, Z.; Liu, H.; Wang, Z.; Xu, J.; Zhang, P.; Zeng, H. A Combined Genetic Optimization with AdaBoost Ensemble Model for Anomaly Detection in Buildings Electricity Consumption. Energy Build. 2021, 248, 111193. [Google Scholar] [CrossRef]
  32. Izonin, I.; Tkachenko, R.; Verhun, V.; Zub, K. An Approach towards Missing Data Management Using Improved GRNN-SGTM Ensemble Method. Eng. Sci. Technol. Int. J. 2021, 24, 749–759. [Google Scholar] [CrossRef]
  33. The Tracebase Data Set. Available online: http://www.tracebase.org (accessed on 30 September 2022).
  34. Purna Prakash, K.; Pavan Kumar, Y.V.; Reddy, C.P.; Pradeep, D.J.; Flah, A.; Alzaed, A.N.; Al Ahamdi, A.A.; Ghoneim, S.S.M. A Comprehensive Analytical Exploration and Customer Behaviour Analysis of Smart Home Energy Consumption Data with a Practical Case Study. Energy Rep. 2022, 8, 9081–9093. [Google Scholar] [CrossRef]
  35. Himeur, Y.; Alsalemi, A.; Bensaali, F.; Amira, A. Building Power Consumption Datasets: Survey, Taxonomy and Future Directions. Energy Build. 2020, 227, 110404. [Google Scholar] [CrossRef]
  36. Iqbal, H.K.; Malik, F.H.; Muhammad, A.; Qureshi, M.A.; Abbasi, M.N.; Chishti, A.R. A Critical Review of State-of-the-Art Non-Intrusive Load Monitoring Datasets. Electr. Power Syst. Res. 2021, 192, 106921. [Google Scholar] [CrossRef]
  37. Pipattanasomporn, M.; Chitalia, G.; Songsiri, J.; Aswakul, C.; Pora, W.; Suwankawin, S.; Audomvongseree, K.; Hoonchareon, N. CU-BEMS, Smart Building Electricity Consumption and Indoor Environmental Sensor Datasets. Sci. Data 2020, 7, 241. [Google Scholar] [CrossRef]
  38. Gopinath, R.; Kumar, M.; Prakash Chandra Joshua, C.; Srinivas, K. Energy Management Using Non-Intrusive Load Monitoring Techniques–State-of-the-Art and Future Research Directions. Sustain. Cities Soc. 2020, 62, 102411. [Google Scholar] [CrossRef]
  39. Kasaraneni, P.P.; Yellapragada, V.P.K.; Moganti, G.L.K.; Flah, A. Analytical Enumeration of Redundant Data Anomalies in Energy Consumption Readings of Smart Buildings with a Case Study of Darmstadt Smart City in Germany. Sustainability 2022, 14, 10842. [Google Scholar] [CrossRef]
Figure 1. Conceptual model for the proposed approach.
Figure 1. Conceptual model for the proposed approach.
Sensors 22 09323 g001
Figure 2. Implementation flow of the proposed approach.
Figure 2. Implementation flow of the proposed approach.
Sensors 22 09323 g002
Figure 3. Confusion matrix.
Figure 3. Confusion matrix.
Sensors 22 09323 g003
Figure 4. Stacking of classifiers.
Figure 4. Stacking of classifiers.
Sensors 22 09323 g004
Figure 5. Data after filling missing timestamps and placing NA.
Figure 5. Data after filling missing timestamps and placing NA.
Sensors 22 09323 g005
Figure 6. Proportion of missing data in the dataset. (a) Original dataset. (b) After removing the anomalies. (c) After filling missing timestamps.
Figure 6. Proportion of missing data in the dataset. (a) Original dataset. (b) After removing the anomalies. (c) After filling missing timestamps.
Sensors 22 09323 g006
Figure 7. Performance metrics for the single-classifier approach on the median imputed dataset. (a) Accuracy (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Figure 7. Performance metrics for the single-classifier approach on the median imputed dataset. (a) Accuracy (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Sensors 22 09323 g007
Figure 8. Performance metrics for the single-classifier approach on the KNN imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Figure 8. Performance metrics for the single-classifier approach on the KNN imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Sensors 22 09323 g008
Figure 9. Performance metrics for the single-classifier approach on the bagging imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Figure 9. Performance metrics for the single-classifier approach on the bagging imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Sensors 22 09323 g009
Figure 10. Performance metrics for the ensemble classifiers approach on the median imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Figure 10. Performance metrics for the ensemble classifiers approach on the median imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Sensors 22 09323 g010aSensors 22 09323 g010b
Figure 11. Performance metrics for the ensemble classifiers approach on the KNN imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Figure 11. Performance metrics for the ensemble classifiers approach on the KNN imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Sensors 22 09323 g011
Figure 12. Performance metrics for ensemble classifiers approach on bagging imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Figure 12. Performance metrics for ensemble classifiers approach on bagging imputed dataset. (a) Accuracy. (b) Precision. (c) Recall. (d) Specificity. (e) F1 Score.
Sensors 22 09323 g012
Table 1. Performance comparison of single-classifier approach on the median imputed dataset.
Table 1. Performance comparison of single-classifier approach on the median imputed dataset.
ClassifierAccuracy (%)Precision (%)Recall (%)Specificity (%)F1 Score (%)
RF98.19998.695.998.8
SVM80.580.5100089.2
DT9696.598.685.397.5
NB80.580.5100089.2
KNN76.383.687.928.685.7
NNET88.893.292.971.993
Superior ClassifierRFRFSVM, NBRFRF
Table 2. Performance comparison of the single-classifier approach on the KNN imputed dataset.
Table 2. Performance comparison of the single-classifier approach on the KNN imputed dataset.
ClassifierAccuracy (%)Precision (%)Recall (%)Specificity (%)F1 Score (%)
RF87.786.610036.392.8
SVM80.580.5100089.2
DT87.486.410035.192.7
NB81.1811003.389.5
KNN80.580.5100089.2
NNET6880.379.918.880.1
Superior ClassifierRFRFAll except NNETRFRF
Table 3. Performance comparison of the single-classifier approach on the bagging imputed dataset.
Table 3. Performance comparison of the single-classifier approach on the bagging imputed dataset.
ClassifierAccuracy (%)Precision (%)Recall (%)Specificity (%)F1 Score (%)
RF95.21009410096.9
SVM80.580.5100089.2
DT87.410084.310091.5
NB80.580.5100089.2
KNN76.279.794.50.486.5
NNET75.779.594086.1
Superior ClassifierRFRF, DTSVM, NBRF, DTRF
Table 4. Performance comparison of the ensemble classifiers approach on the median imputed dataset.
Table 4. Performance comparison of the ensemble classifiers approach on the median imputed dataset.
ClassifierAccuracy (%)Precision (%)Recall (%)Specificity (%)F1 Score (%)
RF+SVM+DT98.998.610094.599.3
RF+SVM+NB81.599.481.676.589.7
RF+SVM+KNN7377.587.536.882.2
RF+SVM+NNET91.79991.493.695
RF+DT+NB98.898.610094.599.3
RF+DT+KNN92.598.992.493.795.5
RF+DT+NNET98.998.610094.599.3
RF+NB+KNN72.578.985.934.782.2
RF+NB+NNET88.385.510062.692.2
RF+KNN+NNET98.898.610094.599.3
Superior ClassifierRF+SVM+DT
RF+DT+NNET
RF+SVM+NBRF+SVM+DT
RF+DT+NB
RF+DT+NNET
RF+NB+NNET
RF+KNN+NNET
RF+SVM+DT
RF+DT+NB
RF+DT+NNET
RF+KNN+NNET
RF+SVM+DT
RF+DT+NB
RF+DT+NNET
RF+KNN+NNET
Table 5. Performance comparison of the ensemble classifiers approaches on the KNN imputed dataset.
Table 5. Performance comparison of the ensemble classifiers approaches on the KNN imputed dataset.
ClassifierAccuracy (%)Precision (%)Recall (%)Specificity (%)F1 Score (%)
RF+SVM+DT78.897.280.519.788.1
RF+SVM+NB75.788.982.431.785.5
RF+SVM+KNN70.982.981.423.382.1
RF+SVM+NNET72.887.580.519.283.8
RF+DT+NB79.997.881.140.188.7
RF+DT+KNN80.299.380.629.589
RF+DT+NNET79.795.482.243.688.3
RF+NB+KNN76.191.381.32786
RF+NB+NNET71.181.382.627.481.9
RF+KNN+NNET77.593.181.531.286.9
Superior ClassifierRF+DT+KNNRF+DT+KNNRF+NB+NNETRF+DT+NNETRF+DT+KNN
Table 6. Performance comparison of the ensemble classifiers approaches on the bagging imputed dataset.
Table 6. Performance comparison of the ensemble classifiers approaches on the bagging imputed dataset.
ClassifierAccuracy (%)Precision (%)Recall (%)Specificity (%)F1 Score (%)
RF+SVM+DT89.698.889.491.193.9
RF+SVM+NB75.793.379.98.986.1
RF+SVM+KNN71.286.579.512.582.8
RF+SVM+NNET72.489.978.80.284
RF+DT+NB77.996.8800.787.6
RF+DT+KNN75.993.280.214.586.2
RF+DT+NNET75.993.280.214.586.2
RF+NB+KNN7794.580.416.986.8
RF+NB+NNET78.993.982.34087.7
RF+KNN+NNET75.292.679.89.385.7
Superior ClassifierRF+SVM+DTRF+SVM+DTRF+SVM+DTRF+SVM+DTRF+SVM+DT
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kasaraneni, P.P.; Venkata Pavan Kumar, Y.; Moganti, G.L.K.; Kannan, R. Machine Learning-Based Ensemble Classifiers for Anomaly Handling in Smart Home Energy Consumption Data. Sensors 2022, 22, 9323. https://doi.org/10.3390/s22239323

AMA Style

Kasaraneni PP, Venkata Pavan Kumar Y, Moganti GLK, Kannan R. Machine Learning-Based Ensemble Classifiers for Anomaly Handling in Smart Home Energy Consumption Data. Sensors. 2022; 22(23):9323. https://doi.org/10.3390/s22239323

Chicago/Turabian Style

Kasaraneni, Purna Prakash, Yellapragada Venkata Pavan Kumar, Ganesh Lakshmana Kumar Moganti, and Ramani Kannan. 2022. "Machine Learning-Based Ensemble Classifiers for Anomaly Handling in Smart Home Energy Consumption Data" Sensors 22, no. 23: 9323. https://doi.org/10.3390/s22239323

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop