Next Article in Journal
Economic, Environmental and Energetic Analysis of a Distributed Generation System Composed by Waste Gasification and Photovoltaic Panels
Previous Article in Journal
Burnout Investigation of Small Diameter Tubes Immersed in Nanofluids
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bi-Level Poisoning Attack Model and Countermeasure for Appliance Consumption Data of Smart Homes

1
Department of CSE, Jashore University of Science and Technology (JUST), Jashore 7400, Bangladesh
2
Centre for Cyber Security Research and Innovation, Deakin University, Geelong 3217, Australia
3
School of Computing Technologies, RMIT University, Melbourne 3001, Australia
*
Author to whom correspondence should be addressed.
Energies 2021, 14(13), 3887; https://doi.org/10.3390/en14133887
Submission received: 19 May 2021 / Revised: 10 June 2021 / Accepted: 17 June 2021 / Published: 28 June 2021
(This article belongs to the Special Issue Selected Papers on Power, Energy and Environmental Engineering)

Abstract

:
Accurate building energy prediction is useful in various applications starting from building energy automation and management to optimal storage control. However, vulnerabilities should be considered when designing building energy prediction models, as intelligent attackers can deliberately influence the model performance using sophisticated attack models. These may consequently degrade the prediction accuracy, which may affect the efficiency and performance of the building energy management systems. In this paper, we investigate the impact of bi-level poisoning attacks on regression models of energy usage obtained from household appliances. Furthermore, an effective countermeasure against the poisoning attacks on the prediction model is proposed in this paper. Attacks and defenses are evaluated on a benchmark dataset. Experimental results show that an intelligent cyber-attacker can poison the prediction model to manipulate the decision. However, our proposed solution successfully ensures defense against such poisoning attacks effectively compared to other benchmark techniques.

Graphical Abstract

1. Introduction

Home appliances consume a large portion of electrical energy, thus attracting researcher’s attention to understand the appliance energy usage patterns in buildings [1,2]. Even appliances in standby mode can be responsible for a significant increase in electricity consumption due to their continuous consumption of low electricity demand. To determine the correlation among different factors and to assess their impact on energy management systems, prediction models can be helpful. Many applications such as estimation of energy usage patterns, energy management, load control, demand-side management and demand-side response, simulation of building performance analysis and many more will be beneficiaries of electrical energy consumption models using regression-based predictive analyses [3,4].
While these predictive models have much potential for improved energy management and efficiency, measurement information and available internet traffic communication are heavily relied upon. In reality, home and building automation systems are communication sensitive and heavily rely on information exchange. This depends on the system structure and various data access points. Any of these access points can be controlled to infiltrate a network and change load and measurement information to destabilize the framework unpredictably. Although the smart-grid paradigm has started a new era with advanced communication and control for improved reliability and efficiency, it has created new challenges as well related to cybersecurity.
In the age of the Internet of Things (IoT), cybersecurity is not only a major concern to information technology but also to critical infrastructures like energy industries [5,6]. Power and energy systems benefit from advanced probabilistic modeling and have a large impact towards improved smart-grid reliability and energy efficiency. Thus they are more vulnerable to attacks by intelligent attackers that target these computational modules. The US Department of Energy (DoE) revealed 150 effective attacks that focused on data integrity and availability attacks within power networks [7]. The European Network of Transmission System Operators for Electricity, which addresses 42 European transmission framework administrators in 35 nations, discovered proof of a fruitful digital interruption in its office network [7]. In the year 2016, cyber-attackers conducted a blackout in the Ukraine power grid for 60 min [8]. Intruders gained access to the operation center of a western Ukrainian power company in 2015 and disconnected power from 225,000 families. A US report reasoned that the attack was launched through ‘spear phishing’ emails [9]. A detailed investigation of cyber-attack threat models is presented in [10] while the countermeasures are summarized in [11].
Accurate predictions can help decision makers to determine patterns of future consumption to better plan and optimize energy consumption, which will reduce power loss and increase energy efficiency of the energy systems. Similarly, building energy prediction will help to improve the energy efficiency of the building energy management system. However, corrupted prediction can lead to safety hazards, damage of power system assets and financial losses. For example, in poisoning attacks, intruders try to manipulate the training dataset to influence the outcome of a predictive model [12]. In these types of attacks, the attacker corrupts the learning model such that the model fails to correctly predict on new data during the testing phase. Thus, indirect access to the model is established, which can be used by the attacker to handle the model even in future. In another case, misclassification of unseen data may happen in evasion attacks during the testing phase. Here, an intruder may bypass a data security mechanism to convey an attack or other type of malware to a target model [13]. In privacy attacks, private information in training data can be stolen and utilized to perceive the private behavior of people [14].
However, among many cybersecurity issues, data poisoning attacks pose a great threat to energy consumption prediction. These types of attacks have been analyzed in many application domains such as worm signature generation [15], denial-of-service attack detection [16], PDF malware classification [17], etc. Although attacks on machine learning models, known as adverserial machine learning, has been an emerging research area, not many research works have been reported on manipulating machine learning models of energy consumption data. Such models should be updated regularly because data are generated continuously from different energy sources. In such cases, attacks become easier to mount on the models not only in the machine learning modules but also during data transfer. For instance, in a national energy/load management system, data are collected from different regions of a country through an online system, in which intruders can also put forward data of their choice by generating a data integrity attack. By controlling a couple of devices, intruders can submit counterfeit data, which are then utilized for preparing models applied to an enormous set of choices. Again, it is a challenging task to prevent poisoning attacks as current techniques against adversarial poisoned data perform poorly. In this work, a bi-level poisoning attack is carried out against an aggregated energy usage prediction model considering three popular regression models: (a) the ordinary least squares (OLS) model, (b) the ridge regression model and (c) the lasso regression model. Analyses are carried out based on a real-world dataset that contains different data sources and environmental parameters. Regression is extensively used for prediction purposes in many domains where a model tries to minimize a loss function and predict a numerical variable based on some predictor variables.
This paper aims to address the vulnerability due to a poisoning attack on household energy data and deal with potential countermeasures. Here, we consider bi-level poisoning attacks, where attackers inject poisoned data in two steps: during data propagation from devices of the smart home to the control center (during data transit) and during training of the machine learning model. Specifically, the contributions can be summarized as follows:
  • We assess a bi-level data poisoning strategy based on a sparse attack vector and optimization-based attack, which successfully corrupts the energy prediction model of home appliances (See Section 3);
  • An effective solution for the poisoned energy prediction model is also implemented. The proposed defense strategy is evaluated on various benchmark regression models (See Section 3);
  • Apparently, to the best of our knowledge, this is one of the earliest works on the attack and defense of poisoning attacks on ‘household energy prediction models’. Proposed methods are tested on a benchmark dataset from the UCI data repository (Section 5).

2. Related Work

This article deals with the cybersecurity vulnerabilities and countermeasures of the predictive models for household energy usage. Hence, we have divided the review into two broad categories.

2.1. Related Work on the Household Prediction Models

Diverse information and techniques are utilized in the existing literature to comprehend forecast models of energy utilization of home appliances. A methodology to appraise building energy utilization is assessed from the standard datasets in [18]. Hourly energy consumption is predicted from service bills using predetermined coefficients. Various residential accessory loads, including a fridge, PC, TV and clothes washer, are modeled using the Norton equivalent technique in [19]. Day-by-day energy use profiles for significant home devices are investigated in [20], which claims that refrigerators show better uniformity than user-dependent accessories such as clothes washers. The authors in [21] develop a model that can detect and estimate different home appliance loads using an explicit duration hidden Markov model. The review paper in [22] finds crucial appliances and their parameters affecting electricity consumption in domestic buildings. Another study focuses on electrical appliance modeling for accurate energy simulations for buildings [23]. However, most of the models presented here are concerned with building simulation studies. However, a number of studies are also found in the literature dealing with electricity load prediction in operational phases.
A large variety of parameters are considered in the prediction models of electricity demand. The most important parameters for such models are rainfall index, time of day, outdoor temperature, global solar radiation, wind speed, etc. [24]. The impact of climate-related factors on monthly power demand are studied in [25]. Another study [26] considers brief-stretch power estimation for different houses and finds temporal distribution as a significant factor. The investigation in [27] uncovers that climate, area, and floor zone are the main factors, and the numbers of fridges and entertainment devices are the most impacting determinants, of day-by-day least utilization. To predict an individual appliance’s energy consumption, a system was developed by [28] where different data, for example, past utilization, season, month, etc., were utilized to supervise the model.

2.2. Related Work on the Security Vulnerabilities of the Predictive Models

It is important to note that prediction models have a large probability of security vulnerabilities to potential attackers. Recently, the safety aspects of machine learning models have attracted the attention of researchers. A large variety of attacks on different domains have been designed and analyzed, which are especially in three categories: poisoning attacks, evasion attacks and privacy attacks. Such attacks have been practically demonstrated in many application domains.
A family of poisoning attacks has been investigated against a support vector machine (SVM)-based handwritten digit recognition system in [12]. Such attack infuses uniquely created training data that expand the SVM’s test data. Another work on malware detection in [17] reveals that feature selection techniques can be seriously compromised under poisoning attacks. A poisoning attack and a corresponding defense mechanism are proposed in [16], with regard to a specific anomaly detector for identifying peculiarities in background networks. A noise injection attack on a syntactic-based programmed worm signature generator is introduced in [15].
A group of evasion attacks is constructed in [29] to evaluate the robustness of a neural network. Three standard datasets (MNIST, CIFAR-10, ImageNet) are used to evaluate proposed attacks. In the paper [30], the authors tentatively research the viability of classifier evasion attack utilizing a genuine, established framework called PDFRATE. Another work [31] shows how an attacker can abuse AI, as utilized in the SpamBayes spam filter, to render it pointless regardless of whether the attacker’s entrance is restricted to just 1% of the training messages. In the work [13], the authors present a basic yet compelling gradient-based methodology that can be abused to deliberately evaluate the security of a few, generally utilized classification tasks against evasion assaults. The methodology is assessed on the security task of malware recognition in PDF files, and shows that such frameworks can be effortlessly sidestepped.
A large amount of research can be found on privacy attacks in the literature. Privacy attacks in pharmacogenetics are investigated in [32], wherein AI models are utilized to control clinical medicines dependent on a patient’s background and genotype. It justifies that, given the model and some segment data about a patient, attackers can foresee the patient’s hereditary markers. Ref. [33] quantitatively explores how AI models spill data about the individual records on which they were trained. Utilizing realistic datasets, including a clinic release dataset whose membership is sensitive from the data protection point of view, they show that these models can be defenseless against membership inference attacks.
While a number of research works have focused on household energy consumption prediction [18,19,20,21,22,23,24,25,26,27,28], the security issues of these models are not well addressed in the literature. However, though there are many works on poisoning attacks in machine learning models for different domains [12,13,15,16,17,29,30,31,32,33], the impact of a false data injection attack during communication along with a poisoning attack on machine learning models specifically for energy systems (to the best of our knowledge) cannot be found in the literature. In this paper, we perform bi-level poisoning attacks on regression models for predicting the energy usage of household appliances.

3. Proposed Bi-Level-Poisoning-Based Adversarial Model on Energy Data

A household prediction model heavily relies on the sensing and decision making as shown in Figure 1. Information is sensed from smart home appliances and sent to a server through an advanced communication technique where decision making occurs. However, an intelligent attacker can perform one specific type of attack known as a data integrity attack in two different ways. One is during the communication of the information that they can manipulate, which is similar to the man-in-the-middle attack or a spoofing attack. Another involves poisoning the machine learning model. Based on this framework and model, we have proposed a bi-level poisoning attack model.

3.1. Poisoning Attack during Communication

Different devices in the smart home send signals and streaming data periodically to the central machine learning module. An attacker may capture and possibly alter the data between smart home devices and the server. The attacker may also be able to intercept all relevant signals passing between them and corrupt a fraction of the entire dataset. In this section, we construct a false data injection attack during communication from smart home appliances to the server.
False information infusion assaults require the attackers to know the current design of the smart home framework. In the event that the attacker can determine the current arrangement of the automated smart building management system, attackers can inject poisonous measurements that will mislead the decision making process of the machine learning module. As this system configuration does not change frequently in a smart home, it is trivial for the assailants to acquire such design data to launch these attacks. Another requirement for the attackers is the manipulation of the sensor measurements. The attackers need to truly mess with the sensor, or manipulate the sensor measurements before they are used for training in the machine learning module. Strong protection against unauthorized physical access to these sensors will make it nontrivial to control the sensor estimations. Be that as it may, it is the beginning stage of our exploration, and the current outcomes can act as a basis for future examination of more complicated models. We consider the scenario where the assailant is obliged to get to some particular sensors and perform arbitrary bogus information infusion assaults, in which the assailant expects to discover any assault vector as long as it can prompt an off-base assessment of state factors.
We expect that there are p sensors that give p estimations ( s 1 , …, s p ) and these sensors send a total of q observations ( o 1 , …, o q ) periodically. The relationship between these sensor measurements and q observations can be characterized by a q p matrix M.
Allow s a to address the vector of noticed estimations that may contain pernicious information. s a can be addressed as
s a = s + i
where s = ( s 1 , …, s p ) is the vector of original measurements and i = ( i 1 , …, i p ) is the pernicious information added to the first estimations. We allude to i as an assault vector. The kth component i k being nonzero implies that the assailant bargains the kth sensor and afterward replaces its unique estimation s a with a fake estimation s k + i k . The attacker can pick any nonzero self-assertive vector as the assault vector i and afterward build the pernicious estimations s i = s + i .
We accept that the attacker approaches x explicit sensors. Instinctively, the assailant can just adjust the estimations of these x sensors. Therefore, the assailant cannot basically pick any v = [ v 1 , . . . , v p ] T and use i = M v as the assault vector. For those sensors that cannot be accessed by the attacker, the infused errors should stay 0.
Formally, we let L s e n s o r s = l 1 , …, l x be the arrangement of lists of the x meters that the aggressor approaches. The aggressor can adjust estimations s l z , where l z   ϵ   L s e n s o r s . To dispatch a bogus information infusion assault without being distinguished, the aggressor needs to discover a nonzero assault vector i = ( i 1 , …, i p ) such that i l = 0 for l L s e n s o r and i is a direct mix of the segment vectors of M (i.e., i = M v ).
As we consider an arbitrary bogus information infusion assault that causes incorrect estimation of the observations, the errors that infuse into some unacceptable assessment could be useful.
Consequently, the assault vector i satisfies the condition:
i = ( i 1 , . . . , i p ) = M v
with i l = 0 for l L s e n s o r , where L s e n s o r is the arrangement of files of the meters that can be accessed by the aggressor.

3.2. Poisoning Attack on the Predictive ML Module

The second level of attack performs a poisoning attack on a linear regression model. We consider the optimization-based poisoning attack proposed by [34]. In an ideal case, a linear regression model generates a function f ( x , θ ) = w T x + b after the training stage. This linear function predicts the estimation of y at x. A regularization parameter preserves the generalization capability of the algorithm on unseen data. Based on the types of regularization term, different linear regression methods are used: ordinary least squares (OLS), ridge regression, LASSO etc. Information, s, from the smart home appliances is sent through an advanced communication system to the central server (machine learning module), which is corrupted by the man-in-the-middle attack. Thus, poisoned data, s a , are stored in the server that is used for training the ML module. Let us assume the data stored in the server are denoted by S = s a .
A proposed second-level poisoning attack tries to corrupt the learning model during the training phase. This corrupted model modifies the prediction results on new data in the testing phase. Both the white-box and black-box attacks are considered in this attack. In a white-box attack, the intruder has knowledge about the training data S t r , list of features x, learning algorithm L A and prepared boundary parameters γ . For white-box attacks, this optimization problem can be written as:
a r g m a x S p L F ( S , γ * p )
where γ * p a r g m i n γ L ( S t r U S p , γ ) and L F is the loss function; S is the untainted dataset.
However, in a black-box attack, it has no information on the training set S t r yet can gather a substitute informational collection S t r . The list of features x and learning algorithm L A are known, while the prepared boundary parameters γ are unknown. Optimization of L A on the substitute data set S t r can be helpful in estimating these parameters. The attack’s capability in the optimization-based attack is usually confined to the upper-bounding of the number p of poisoning points that can be infused into the training information. The feature values and response variables are randomly selected within a range of [0, 1]. However, poisoning rates less than 25% are well advised, because normally the attacker can handle just a small segment of the training information. More details on the optimization-based poisoning attack (Algorithm 1) can be obtained from [34].
Algorithm 1: Poisoning Attack on the Predictive ML Module.
Energies 14 03887 i001

4. Proposed Defense Mechanism against Bi-Level-Poisoning-Based Adversarial Model

Proposed bi-level poisoning attack deploys two levels of attack: one is during communication from smart home appliances to the ML module, another is while training the ML module. Hence, we have proposed two levels of security mechanism with two distinct algorithms, which will work collaboratively and provide robust defense against such kinds of poisoning attacks.

4.1. Defense Mechanism against Poisoning Attack during Communication

The first level of poisoning attack manipulates measurement sensors. As discussed earlier, from a set of measurement sensors with some observations a 2-D matrix will eventually form. Hence, this kind of attack is similar to alteration of a 2-D matrix where the attack vector represents a sparse data set. Defending this attack is similar to recovering a low-rank matrix with a small corrupted portion. If it is viewed from the point of a robust classical principal component analysis (PCA) problem, a large number of application domains are found to face such a problem. Principal component analysis (PCA) assumes that high-dimensional information lies almost in a much lower-dimensional subspace. The correct assumption of this subspace is important to reduce the dimension of the data and to process, analyze, compress and visualize the data. If s a = s + i , where s a is a large matrix of data arranged as columns, s is a rank-r matrix and i is a matrix of Gaussian random variables, the goal of PCA is to estimate s.
Though PCA finds the optimal estimate of the subspace, it breaks down even under small corruptions. Robust PCA (RPCA) [35] can recover the low-rank matrix s from s a with gross but sparse errors i. However, although it provides superior convergence rates, the complexity of robust PCA is too high. An accelerated proximal gradient algorithm is faster and more scalable for estimating s from s a in the presence of sparse error i. To defend the first-level poisoning attack on the dataset, we apply nuclear norm minimization and a 1-norm-minimization-based convex programming surrogate named the accelerated proximal gradient algorithm (Algorithm 2) [36].
The problem of the robust PCA can be formulated as follows:
p * = m i n s , i | | s | | * + λ | | i | | 1 s a = s + i
A relaxation of Equation (4) is considered in [36] as follows:
m i n s , i μ | | s | | * + μ λ | | i | | 1 + 1 2 | | s + i s a | | F 2
Instead of fixing μ to any small value, convergence can be achieved in the accelerated proximal gradient algorithm by repeatedly decreasing the value of μ .
Algorithm 2: Defense Mechanism Using Accelerated Proximal Gradient Algorithm in First-Level Attack.
Energies 14 03887 i002

4.2. Defense Mechanism against Poisoning Attack on the Predictive ML Module

The second-level attack is on the linear regression model that poisons the training dataset so that the prediction model gives the wrong decision while testing. Existing protection propositions can be grouped into two classes: noise versatile safeguards and adversarially versatile safeguards. The main idea behind these noise-resilient regression algorithms is to distinguish and eliminate anomalies from a dataset. While these techniques ensure robustness against commotion and anomalies, an attacker can in any case produce harmful information that influences the prepared model. Specifically, an assailant can produce poisoning points that are basically the same as the valid information dissemination (called inliers); however, these can in any case deceive the model. Incidentally, these current regression techniques are not strong against inlier attack points picked to maximally deceive the assessed regression model.
However, recently proposed adversarially-tough regression algorithms regularly give guarantees under solid presumptions about information and clamor dissemination. These algorithms are based on some assumptions such as information, commotion fulfilling the sub-Gaussian assumption, the component matrix having a low rank and it being possible to project the matrix to a lower dimensional space. Each of these strategies have provable robustness guarantees; however, the suppositions on which they depend are not typically fulfilled.
To defend such an attack, the TRIM defense algorithm [34] is used, which takes a principled approach instead of just removing outliers from the training set. It applies an iterative approach and in each iteration, it estimates the regression parameters, γ , and trains on a subset of points with the lowest residuals. Moreover, a loss function L F is also used, calculated iteratively on a disparate subset of the residuals. The size of the original training set s t r is n and the attacker injects poisoned samples S p , where p = β . n . We have to ensure that β is less than 1 so that the majority of training data remain unpoisoned. The linear regression model can be trained with a subset of legitimate training points of size n, if all p poisoning points can be identified. However, separating the legitimate and attacked points is difficult as the true distribution of the unpoisoned training data is clearly unknown. The TRIM defense mechanism tries to determine a subset of training points that have the lowest residuals. In this work, we applied trimmed optimization techniques (Algorithm 3) for the adversarial linear regression model, which has been effectively assessed for the energy dataset.
Algorithm 3: Defense Mechanism Against Poisoning Attack on the Predictive ML Module.
Energies 14 03887 i003

5. Results and Discussion

We carried out our attack and protection mechanisms in Python, utilizing the numpy and sklearn bundles. We utilized the standard cross-validation strategy to divide the whole dataset into training, testing and validation sets. We utilized two primary measurements for assessing our calculations: mean square error (MSE) for the adequacy of the assaults and protections, and running time for their expense.

5.1. Datasets

The dataset we used in this work comes from the UCI AI storehouse [37]. It contains the temperature and dampness of various rooms in a low-energy house from a remote sensor network, information from a climate station and one sub-metered electrical fuel utilization source (lights). The low-energy house followed passive house certification design rules, thus having a yearly heating and cooling load of about 15 kWh/m 2 each year. The data were gathered each 10 min from various electric energy meters including a heat recuperation ventilation unit, homegrown high-temperature water heat siphon and electric baseboard radiators. Another sub-metered load (lights) was used for the investigation since it has been demonstrated to be a decent indicator of room inhabitance when joined with relative stickiness estimations. Weather information from the closest air terminal climate station was converged by date and time in this investigation to assess its effect on the forecast of the energy utilization of appliances.

5.2. Effects of Poisoning Attacks on Energy Consumption Data

In this section, we analyze the effects of sparse-error-based and optimization-based bi-level poisoning attacks on energy datasets for three popular regression models: ordinary least squares (OlS), LASSO and ridge regression. Figure 2a shows the mean square error (MSE) of the proposed bi-level poisoning attack for a 5% poisoning rate. Overall, the ordinary least squares (OLS) model shows the highest error and the ridge model shows the lowest error of 0.07 and 0.04, respectively. For the attacks without poisoning, all the models show similar errors, although OLS has less error than the other models. Figure 2b shows the mean square error (MSE) of the proposed bi-level poisoning attack for a 10% poisoning rate. It is seen that the proposed bi-level poisoning attack affects the OLS model most, whereas the ridge model is least affected. The lasso model shows a significant MSE value of 0.07 compared to a 5% poisoning error.
Effects of bi-level poisoning attacks on different models for various poisoning rates can be well-understood from Figure 3. Various poisoning rates starting from 5% to 25% are assessed for three models. The ordinary least squares model for energy consumption prediction is mostly affected by a data poisoning attack with varying MSE for different poisoning rates. A change in poisoning rate from 10% to 15% results in a sharp change in error for the OLS model. Though it is assumed that an increase in poisoning rate will increase MSE, for the case of OLS, MSE increased up to the 15% poisoning rate, but after this point, there was a fall in MSE to 20%, from which point it started to increase again. These behaviors of OLS can be revealed as we focus on the working principles of the OLS method. The ordinary least squares (OLS) linear regression model is not penalized for its selection of weights. During the training stage, the model may place a large weight on the features that seem to be important. As a large amount of predictor variables are affected by poisoning attacks and such manipulation can make these variables correlated, OLS parameter estimates face a large variance that makes the model unreliable.
However, both the lasso and ridge regression models show exponential curves for increasing poisoning rates. The ridge model showed a gradual rise in MSE as the poisoning rate increased from 5% to 25%. The lasso model faced an irregular increase in MSE as poisoning rate increased. Unlike OLS, the lasso model is penalized for the sum of absolute values of the weights. Therefore, the absolute values of weight will not only be reduced but also many will tend to be zeros. However, ridge penalizes the model for the sum of squared values of the weights. Here, the weights have smaller absolute values and tend to penalize the extremes of the weights, thus weights are more evenly distributed. The proposed bi-level poisoning attack injects poisonous data in such a way that a lot of predictor variables are manipulated. The relevancy of these affected features to the prediction results in different effects on the lasso and ridge models. For the case of the lasso model, relevancy with all predictor variables generates more errors in prediction tasks, whereas the ridge model faces fewer errors than the lasso model due to strong relevancy with all the feature variables.
Table 1 shows the real effect of proposed bi-level poisoning attacks on the energy dataset. MSE results obtained with the proposed attack were translated into application -specific parameters. In the prediction model of electrical energy consumption, the goal is to predict probable energy use (in watt-hour (WH)) for home appliances. Actual energy consumption at any time instant is 580 WH, which is predicted as 579.96% by the model without poisoning. However, every regression model is conquerable by poisoning attacks with a significant change by a factor of 2.87 or more for a 10% poisoning rate. The same scenarios are also found in the case of smaller poisoning rates. For a 5% poisoning rate, the change in energy consumption is 128% for the ridge model, 139% for OLS and 145% for lasso regression.
The attack additionally needs to meet a time prerequisite to expand its probability of staying stealthy. If the time needed for attack development is too high, the working conditions may change and this will affect the likelihood of the assault being detected. Henceforth, in this part we assess the performance of the proposed bi-level poisoning attack in terms of effectiveness, e.g., how quickly it can create an assault. We will compare the performance sparse-error-based attack against the optimization-based attack. The time requirements for both attack constructions for two different poisoning rates are plotted in Figure 4.
Overall, poisoning more data points requires more time for both types of attack. In addition, the sparse-error-based attack during communication from smart home appliances to the ML module requires more time than the optimization-based attack in the ML module. For instance, it requires 0.35 s for the sparse-error-based attack while it needs only 0.25 s for the optimization-based attack for a 5% poisoning rate. However, a significant increase in time requirements is noted for the sparse-error-based attack than the optimization-based attack for a 10% poisoning rate. Therefore, the optimization-based attack construction method requires the least computational time.

5.3. Defense Algorithms

In this part, we assess the accelerated proximal gradient algorithm (APG) and TRIM defense mechanism separately against the bi-level poisoning attack. Figure 5a–c show MSEs for OLS, LASSO and ridge regression, respectively, for the model with no defense mechanism and the model with the TRIM defense algorithm.
The optimization-based attack generates inlier points with comparative conveyance like the training dataset, for which the TRIM technique is much more effective. For OLS, LASSO and ridge regression, the mean square error (MSE) of the TRIM defense algorithm is within 1% of the original models. This demonstrates that the TRIM technique is a significant defense mechanism against a poisoning attack in the ML module. In addition, the defense we evaluated ran very fast, taking an average time of only 0.02 s.
However, the sparse errors introduced during the first-level attack are minimized using the accelerated proximal gradient (APG) algorithm (Figure 6). To inject sparse errors, the intruder should have sound knowledge on the running configuration of a smart-home system. The non-changing behavior of the architecture of smart home systems adds an extra benefit to the attacker. However, it is assumed that only a few sensors in a smart home can be accessed by the attacker. In Figure 6, only 50 observations are illustrated, where among 28 sensors and data sources, only the temperature sensors are assumed to be accessed by the intruder. We apply poisonous data in the temperature sensors. As the APG security mechanism is deployed, it successfully captures the poisonous data s a and separates the original measurement s and injected sparse error i.
However, no defense mechanism is one hundred percent capable of detecting and recovering poisonous data. Especially for APG (Figure 7), it is seen that although it reduced the error rate to a great extent, an increasing graph in error rate is also observed in accordance with an increase in poisoning rate. For example, in the case of 25% poisoning, more than 0.04% error exists in spite of the deployment of the APG algorithm.

6. Conclusions

In this paper, bi-level poisoning attacks on linear regression models for forecasting energy utilization of appliances in a smart home were performed. These reveal that poisoning attacks have an adverse effect on building energy consumption prediction models. In particular, bi-level poisoning attacks during communication and training of the ML module may worsen the situation. Even if the attacker can poison at a rate of 10%, it can change the predicted value up to 287% in the proposed model. The prediction will become worse if the poisoning rate is increased. To tackle this emerging challenge, a combined multiple linear regression attacking solution was proposed against the poisoning attacks on the prediction model. This combined defense mechanism is time efficient and can reduce the mean square error to a great extent. Attacks and defense algorithms were systematically evaluated on a standard data set containing data from a wireless network, climate information from an air terminal station and energy consumption data from home appliances. The real implications of poisoning attacks in energy consumption prediction models of home appliances have been well assessed in this work. In future, we will work to develop more scalable poisoning attacks on different machine learning models and their defense mechanisms for the power industry.

Author Contributions

Conceptualization, A.A., Z.R. and M.B.; methodology, A.A.; software, M.B. and S.M.G.; validation, A.A. and Z.R.; formal analysis, A.A. and Z.R.; investigation, A.A., Z.R. and M.B.; resources, A.A. and M.B.; data curation, A.A. and M.B.; writing—original draft preparation, M.B.; writing—review and editing, A.A.; visualization, M.B. and S.M.G.; supervision, A.A.; project administration, A.A.; funding acquisition, A.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in the UCI Machine Learning Repository at https://archive.ics.uci.edu/ml/datasets/Appliances+energy+prediction.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cetin, K.S. Characterizing large residential appliance peak load reduction potential utilizing a probabilistic approach. Sci. Technol. Built Environ. 2016, 22, 720–732. [Google Scholar] [CrossRef]
  2. Jones, R.V.; Lomas, K.J. Determinants of high electrical energy demand in UK homes: Socio-economic and dwelling characteristics. Energy Build. 2015, 101, 24–34. [Google Scholar] [CrossRef] [Green Version]
  3. Seem, J.E. Using intelligent data analysis to detect abnormal energy consumption in buildings. Energy Build. 2007, 39, 52–58. [Google Scholar] [CrossRef]
  4. Zhao, P.; Suryanarayanan, S.; Simoes, M.G. An energy management system for building structures using a multi-agent decision-making control methodology. IEEE Trans. Ind. Appl. 2012, 49, 322–330. [Google Scholar] [CrossRef]
  5. Sadeghi, A.R.; Wachsmann, C.; Waidner, M. Security and privacy challenges in industrial internet of things. In Proceedings of the 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC), San Francisco, CA, USA, 8–12 June 2015; pp. 1–6. [Google Scholar]
  6. Abir, S.M.A.A.; Anwar, A.; Choi, J.; Kayes, A.S.M. IoT-Enabled Smart Energy Grid: Applications and Challenges. IEEE Access 2021, 9, 50961–50981. [Google Scholar] [CrossRef]
  7. Macola, I.G. The Five Worst Cyberattacks against the Power Industry Since 2014. Available online: https://www.power-technology.com/features/the-five-worst-cyberattacks-against-the-power-industry-since2014/ (accessed on 9 April 2021).
  8. Polityuk, P.; Oleg Vukmanovic, S.J. Ukraine’s Power Outage was a Cyber Attack: Ukrenergo. 17 December 2016. Available online: https://www.reuters.com/article/us-ukraine-cyber-attack-energy/ukraines-power-outage-was-a-cyber-attack-ukrenergo-idUSKBN1521BA (accessed on 9 April 2021).
  9. Volz, D. U.S. Government Concludes Cyber Attack Caused Ukraine Power Outage. December 2015. Available online: https://www.reuters.com/article/us-ukraine-cybersecurity-idUSKCN0VY30K (accessed on 9 April 2021).
  10. Reda, H.T.; Anwar, A.; Mahmood, A. Comprehensive Survey and Taxonomies of False Injection Attacks in Smart Grid: Attack Models, Targets, and Impacts. arXiv 2021, arXiv:cs.CR/2103.10594. [Google Scholar]
  11. Reda, H.T.; Anwar, A.; Mahmood, A.N.; Tari, Z. A Taxonomy of Cyber Defence Strategies Against False Data Attacks in Smart Grid. arXiv 2021, arXiv:cs.CR/2103.16085. [Google Scholar]
  12. Biggio, B.; Nelson, B.; Laskov, P. Poisoning attacks against support vector machines. arXiv 2012, arXiv:1206.6389. [Google Scholar]
  13. Biggio, B.; Corona, I.; Maiorca, D.; Nelson, B.; Šrndić, N.; Laskov, P.; Giacinto, G.; Roli, F. Evasion attacks against machine learning at test time. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Berlin/Heidelberg, Germany, 2013; pp. 387–402. [Google Scholar]
  14. Fredrikson, M.; Jha, S.; Ristenpart, T. Model inversion attacks that exploit confidence information and basic countermeasures. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, 12–16 October 2015; pp. 1322–1333. [Google Scholar]
  15. Perdisci, R.; Dagon, D.; Lee, W.; Fogla, P.; Sharif, M. Misleading worm signature generators using deliberate noise injection. In Proceedings of the 2006 IEEE Symposium on Security and Privacy (S&P’06), Berkeley/Oakland, CA, USA, 21–24 May 2006; p. 15. [Google Scholar]
  16. Rubinstein, B.I.; Nelson, B.; Huang, L.; Joseph, A.D.; Lau, S.H.; Rao, S.; Taft, N.; Tygar, J.D. Antidote: Understanding and defending against poisoning of anomaly detectors. In Proceedings of the 9th ACM SIGCOMM Conference on Internet Measurement, Chicago, IL, USA, 4–6 November 2009; pp. 1–14. [Google Scholar]
  17. Xiao, H.; Biggio, B.; Brown, G.; Fumera, G.; Eckert, C.; Roli, F. Is feature selection secure against training data poisoning? In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 1689–1698. [Google Scholar]
  18. Fumo, N.; Mago, P.; Luck, R. Methodology to estimate building energy consumption using EnergyPlus Benchmark Models. Energy Build. 2010, 42, 2331–2337. [Google Scholar] [CrossRef]
  19. Ghorbani, M.; Rad, M.S.; Mokhtari, H.; Honarmand, M.; Youhannaie, M. Residential loads modeling by norton equivalent model of household loads. In Proceedings of the 2011 Asia-Pacific Power and Energy Engineering Conference, Wuhan, China, 25–28 March 2011; pp. 1–4. [Google Scholar]
  20. Cetin, K.; Tabares-Velasco, P.; Novoselac, A. Appliance daily energy use in new residential buildings: Use profiles and variation in time-of-use. Energy Build. 2014, 84, 716–726. [Google Scholar] [CrossRef]
  21. Guo, Z.; Wang, Z.J.; Kashani, A. Home appliance load modeling from aggregated smart meter data. IEEE Trans. Power Syst. 2014, 30, 254–262. [Google Scholar] [CrossRef] [Green Version]
  22. Jones, R.V.; Fuertes, A.; Lomas, K.J. The socio-economic, dwelling and appliance related factors affecting electricity consumption in domestic buildings. Renew. Sustain. Energy Rev. 2015, 43, 901–917. [Google Scholar] [CrossRef] [Green Version]
  23. Ruellan, M.; Park, H.; Bennacer, R. Residential building energy demand and thermal comfort: Thermal dynamics of electrical appliances and their impact. Energy Build. 2016, 130, 46–54. [Google Scholar] [CrossRef]
  24. Bacher, P.; Madsen, H.; Nielsen, H.A.; Perers, B. Short-term heat load forecasting for single family houses. Energy Build. 2013, 65, 101–112. [Google Scholar] [CrossRef] [Green Version]
  25. Hor, C.L.; Watson, S.J.; Majithia, S. Analyzing the impact of weather variables on monthly electricity demand. IEEE Trans. Power Syst. 2005, 20, 2078–2085. [Google Scholar] [CrossRef]
  26. Saldanha, N.; Beausoleil-Morrison, I. Measured end-use electric load profiles for 12 Canadian houses at high temporal resolution. Energy Build. 2012, 49, 519–530. [Google Scholar] [CrossRef]
  27. Kavousian, A.; Rajagopal, R.; Fischer, M. Determinants of residential electricity consumption: Using smart meter data to examine the effect of climate, building characteristics, appliance stock, and occupants’ behavior. Energy 2013, 55, 184–194. [Google Scholar] [CrossRef]
  28. Basu, K.; Hawarah, L.; Arghira, N.; Joumaa, H.; Ploix, S. A prediction system for home appliance usage. Energy Build. 2013, 67, 668–679. [Google Scholar] [CrossRef]
  29. Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (sp), San Jose, CA, USA, 22–24 May 2017; pp. 39–57. [Google Scholar]
  30. Rndic, N.; Laskov, P. Practical evasion of a learning-based classifier: A case study. In Proceedings of the 2014 IEEE Symposium on Security and Privacy, Berkeley, CA, USA, 18–21 May 2014; pp. 197–211. [Google Scholar]
  31. Nelson, B.; Barreno, M.; Chi, F.J.; Joseph, A.D.; Rubinstein, B.I.; Saini, U.; Sutton, C.A.; Tygar, J.D.; Xia, K. Exploiting Machine Learning to Subvert Your Spam Filter. LEET 2008, 8, 1–9. [Google Scholar]
  32. Fredrikson, M.; Lantz, E.; Jha, S.; Lin, S.; Page, D.; Ristenpart, T. Privacy in pharmacogenetics: An end-to-end case study of personalized warfarin dosing. In Proceedings of the 23rd {USENIX} Security Symposium ({USENIX} Security 14), San Diego, CA, USA, 20–22 August 2014; pp. 17–32. [Google Scholar]
  33. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership inference attacks against machine learning models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–24 May 2017; pp. 3–18. [Google Scholar]
  34. Jagielski, M.; Oprea, A.; Biggio, B.; Liu, C.; Nita-Rotaru, C.; Li, B. Manipulating machine learning: Poisoning attacks and countermeasures for regression learning. In Proceedings of the 2018 IEEE Symposium on Security and Privacy (SP), Francisco, CA, USA, 20–24 May 2018; pp. 19–35. [Google Scholar]
  35. Wright, J.; Ganesh, A.; Rao, S.; Ma, Y. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization. In Coordinated Science Laboratory Report no. UILU-ENG-09-2210, DC-243; University of Illinois at Urbana: Champaign, IL, USA, 2009. [Google Scholar]
  36. Lin, Z.; Ganesh, A.; Wright, J.; Wu, L.; Chen, M.; Ma, Y. Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix. In Coordinated Science Laboratory Report no. UILU-ENG-09-2214, DC-246; University of Illinois at Urbana: Champaign, IL, USA, 2009. [Google Scholar]
  37. Candanedo, L.M.; Feldheim, V.; Deramaix, D. Data driven prediction models of energy use of appliances in a low-energy house. Energy Build. 2017, 140, 81–97. [Google Scholar] [CrossRef]
Figure 1. Proposed system architecture. Energy consumption data for each home appliance are sent over a wireless network to the central machine learning server (ML module). During communication from smart home to ML module, attacker may inject false data. However, during training of the ML models, training data can be manipulated by attackers.
Figure 1. Proposed system architecture. Energy consumption data for each home appliance are sent over a wireless network to the central machine learning server (ML module). During communication from smart home to ML module, attacker may inject false data. However, during training of the ML models, training data can be manipulated by attackers.
Energies 14 03887 g001
Figure 2. Mean square error (MSE) of attacks on different regression models.
Figure 2. Mean square error (MSE) of attacks on different regression models.
Energies 14 03887 g002
Figure 3. MSE for different poisoning rates.
Figure 3. MSE for different poisoning rates.
Energies 14 03887 g003
Figure 4. Elapsed time (in seconds) for attack construction for optimization-based attack and sparse-error-based attack.
Figure 4. Elapsed time (in seconds) for attack construction for optimization-based attack and sparse-error-based attack.
Energies 14 03887 g004
Figure 5. Effects of TRIM defense mechanism on different regression models are expressed by mean square error (MSE). MSE without defense mechanism and after applying TRIM defense are compared in this figure.
Figure 5. Effects of TRIM defense mechanism on different regression models are expressed by mean square error (MSE). MSE without defense mechanism and after applying TRIM defense are compared in this figure.
Energies 14 03887 g005
Figure 6. Accelerated proximal gradient algorithm for handling sparse-error-based attacks during communication from smart-home appliances to the ML module.
Figure 6. Accelerated proximal gradient algorithm for handling sparse-error-based attacks during communication from smart-home appliances to the ML module.
Energies 14 03887 g006
Figure 7. Effect of accelerated proximal gradient algorithm.
Figure 7. Effect of accelerated proximal gradient algorithm.
Energies 14 03887 g007
Table 1. Initial energy consumption (WH) and prediction (WH) after poisoning attack for different poisoning rates.
Table 1. Initial energy consumption (WH) and prediction (WH) after poisoning attack for different poisoning rates.
Poisoning
Rate
Appliance’s
Energy Use
(in WH)
Prediction
without
Poisoning
Change in Predicted Value
OLSRidgeLasso
1%580579.9644.07%23.25%27.46%
2%91.10%53.11%51.62%
3%70.61%70.64%103.65%
4%95.17%105.28%118.17%
5%139.25%128.53%145.64%
6%186.28%158.39%169.79%
7%165.79%175.92%221.83%
8%190.35%210.56%236.34%
9%234.43%233.82%263.81%
10%281.46%263.67%287.96%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Billah, M.; Anwar, A.; Rahman, Z.; Galib, S.M. Bi-Level Poisoning Attack Model and Countermeasure for Appliance Consumption Data of Smart Homes. Energies 2021, 14, 3887. https://doi.org/10.3390/en14133887

AMA Style

Billah M, Anwar A, Rahman Z, Galib SM. Bi-Level Poisoning Attack Model and Countermeasure for Appliance Consumption Data of Smart Homes. Energies. 2021; 14(13):3887. https://doi.org/10.3390/en14133887

Chicago/Turabian Style

Billah, Mustain, Adnan Anwar, Ziaur Rahman, and Syed Md. Galib. 2021. "Bi-Level Poisoning Attack Model and Countermeasure for Appliance Consumption Data of Smart Homes" Energies 14, no. 13: 3887. https://doi.org/10.3390/en14133887

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop