Next Article in Journal
The Possibility of Achieving Zero CO2 Emission in the Indonesian Cement Industry by 2050: A Stakeholder System Dynamic Perspective
Previous Article in Journal
The Alien Invader, Rhus typhina L., Outperforms Its Native Competitor in the Scenario of Nitrogen Deposition and Arbuscular Mycorrhizal Fungi (AMF) Inoculation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Time Series Data Preparation for Failure Prediction in Smart Water Taps (SWT)

1
Centre for Water Systems, University of Exeter, Exeter EX4 4QF, UK
2
Department of Computer Science, EMPS, University of Exeter, Exeter EX4 4QF, UK
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(7), 6083; https://doi.org/10.3390/su15076083
Submission received: 13 January 2023 / Revised: 17 February 2023 / Accepted: 22 February 2023 / Published: 31 March 2023

Abstract

:
Smart water tap (SWT) time series model development for failure prediction requires acquiring data on the variables of interest to researchers, planners, engineers and decision makers. Thus, the data are expected to be ‘noiseless’ (i.e., without discrepancies such as missing data, data redundancy and data duplication) raw inputs for modelling and forecasting tasks. However, historical datasets acquired from the SWTs contain data discrepancies that require preparation before applying the dataset to develop a failure prediction model. This paper presents a combination of the generative adversarial network (GAN) and the bidirectional gated recurrent unit (BiGRU) techniques for missing data imputation. The GAN aids in training the SWT data trend and distribution, enabling the imputed data to be closely similar to the historical dataset. On the other hand, the BiGRU was adopted to save computational time by combining the model’s cell state and hidden state during data imputation. After data imputation there were outliers, and the exponential smoothing method was used to balance the data. The result shows that this method can be applied in time series systems to correct missing values in a dataset, thereby mitigating data noise that can lead to a biased failure prediction model. Furthermore, when evaluated using different sets of historical SWT data, the method proved reliable for missing data imputation and achieved better training time than the traditional data imputation method.

1. Introduction

A sustainable solution for rural water delivery requires accurate water infrastructure assessment and efficient data processing techniques. These techniques need data, which should come from regular usage of the water infrastructure. However, most rural water installations lack accurate data from the available repository [1]. Therefore, with inadequate, partial, or missing data regarding the smart water taps, it is difficult to develop a comprehensive failure prediction model or an early warning system. Furthermore, investment in extensive inspection and data-gathering programmes on smart rural taps to overcome data gaps may not be financially feasible for rural water management agencies [2]. So, to achieve failure prediction for rural water taps, the available time series data generated from the system usage and data manipulation is sufficient for critical analysis irrespective of the discrepancies.
Part of the aim of this paper is to develop a failure prediction model for smart water taps to support proactive maintenance, which can help provide a sustainable water supply to rural communities in sub-Saharan Africa and similar contexts. Solar-powered smart water taps (SWT) deployed to rural areas in some parts of Africa are perceived as low-cost and reliable water supply sources for domestic use in the region. These SWTs, often referred to as e-taps, dispense water when a pre-paid token comes in contact with them. During their functional time, the smart taps generate time series datasets that can be analysed for the purpose of developing a failure prediction model that will ensure a functional sustainability of the water taps. This study also shows the importance of appropriately preparing a case study dataset for time series failure prediction problems. The study also ensures that existing water taps in rural communities in Africa are sustained and managed in a way that satisfies the current water and future socioeconomical ecological demand of sub-Saharan Africa. Furthermore, it shows the procedures for preparing time series datasets for efficient analysis with machine-learning (ML) techniques. Finally, because of the inconsistencies in the case study dataset, the research presents a method for data imputation to correct data discrepancy in the dataset.
Most time series can be defined as a set of generated observations x t acquired at specified time steps t . On the other hand, a discrete time series can be described as one in which a discrete set T 0 (where T 0 = 0 ,   1 ) of times at which observations were made, for example, recording water usage observations at fixed time intervals [3].
For most data-driven applications, the ML learning paradigm—a time series data representation and generalisation of the learning structure used on datasets—is mostly adopted. However, the data representation impacts the machine-learning performance based on the quality of the available dataset. Therefore, poor data representation may reduce the machine-learning model’s performance, while better data representation can result in a better-performing ML model [4]. Therefore, selecting appropriate feature values for feature construction and data representations from an input dataset is a core element of the ML principle. In addition, an efficient feature engineering task requires some time, which takes a larger part of the time required to build an ML model. This task is also domain-specific as it requires human input. Therefore, the dataset should follow the standard time series data collection method [5].
When handling missing data in a dataset, two helpful sides of a case study can identify the impact on the time series prediction result. Firstly, the quantity of missing data is beneficial for our case study. Therefore, the quantity of missing data directly impacts the prediction result and conclusion. The second practical aspect that impacts the prediction result is the cause of missing data in a case study. Therefore, the most important step in analysing a time series is describing the dataset and selecting appropriate computational or mathematical models for the interest data.
Furthermore, in order to accommodate future observations (new data), this study assumes that each observation x t is a subset of a given random variable X t . Finally, the proposed research assumes that a time series can be described as a collection of random variables plotted according to the order obtained over time. For example, X t = x 1 ,   x 2 ,   x 3 ,   .   .   . ,   x n (where x 1 ,   x 2 ,   x 3 ,   .   .   . ,   x n denote values recorded at each time step) is a time series with a sequence of random variables at different timestamps [6].
This research’s main contribution involves preparing data for time series failure prediction, emphasising the need for data cleaning and procedures for missing data imputation. The proposed method of the data preparation has the following features:
(1)
Data preparation through missing data imputation using the generative adversarial network (GAN);
(2)
BiGRU saves computational time by combining the model’s cell state and hidden state during data imputation;
(3)
Analysis of the missing data proportion in the dataset;
(4)
Identification of the dataset patterns.

2. Classification of Missing Data

Missingness in datasets can be described as missing data values from a sample dataset [2,7]. The source of missingness in a dataset is very important in data analysis because it affects the technique required to address the problem [8]. Several underlying factors can cause missing data in a case study:
  • lost or forgotten data value;
  • missing value due to non-applicability to the instances of the case study event;
  • missing value due to irrelevance to the instances of the study.
For example, SWT data variables can be measured; however, the record of such variable values may be missing for unknown reasons. Additionally, the failure of the SWT sensors (the case study), database communication error, human error, electrical component failure or related factors can cause a dataset to be missing.
On the other hand, the SWT variable may not be recorded over a given period for unknown reasons. For instance, missing records over a long period or at random. Additionally, data may not be captured because it is recorded but not useful to a study. In dealing with missing data from the SWT case study, a distinction is made between missing data due to identifiable reasons and missing data due to unidentifiable reasons. In the former case, data imputation for missing data due to identifiable factors can cause a bias in the result of a study [9]. Therefore, imputing data for unidentifiable reasons can be done based on the assumption that the data are missing randomly. Thus, while data missing for identifiable reasons are non-recoverable, data can be recovered for unidentifiable reasons.
Data imputation is a challenging task that consumes time. Selecting relevant variables from a dataset involves the identification of useful potential predictor subsets from a large candidate set [10]. Therefore, the general steps involved in handling missing data include:
  • Analysis of the missing data proportion in the dataset;
  • Identification of the dataset patterns;
  • Detecting the cause of missing data;
  • Selecting the appropriate data imputation method.
The mathematical definition of missing data is as shown in Equations (1)–(3) [8] (Salgado, Azevedo, Proen, & Vieira, 2016):
X = X a ,   X b
where:
X —dataset
X a —observed data
X b —missing data
A binary response for each observation is designed to determine the missing observation, as shown in Equation (2):
B = 1   i f   X a 0   i f   X b
where:
B —binary response
However, a probability P r   B can be used to describe the missing value mechanisms to show that observation may be missing given X a and   X b :
P r   ( B | x a ,   x b )
The three missing data mechanisms are subject to the dependence of the response probability (i.e., whether or not the response probability depends on the observed or missing data values).

2.1. Missing Completely at Random (MCAR)

This phenomenon occurs when a dataset’s missing values are dependent on the measurement of the observed and unobserved data values. Given X with some missing values, the missing values are MCAR if the likelihood of X missing data is unrelated to X values or other data variables. For example, if the data value of every sixth event in the e-tap is not recorded or omitted, the mechanism for that missing data is MCAR [11,12]. Therefore, the probability of a missing observation depends on itself alone and thus reduces to:
P r   B | x a ,   x b = P r   B

2.2. Missing at Random (MAR)

The MAR mechanism gives a weaker assumption than the previous mechanism (MCAR). The MAR mechanism is a probability of a missing data value relating to the observable data only (i.e., statistically, the observed data are related to the missing data) [13]. Therefore, estimating the missing data value from the observed data is possible. Furthermore, the probability of X missing data values is not related to the values of X . The mathematical equation for this is shown in Equation (5):
P r   B | x a ,   x b = P r   ( B | x a )

2.3. Missing Not at Random (MNAR)

The MNAR mechanism allows missing data to depend on unobserved data values, a missing mechanism where both MCAR and MAR do not apply. Therefore, determining the missing mechanism may be impossible if the data are unseen. In MNAR, the whole data distribution is only identifiable with further assumptions [14]. The next section of this paper shows the characteristics of the case study dataset, which aids in identifying the missingness mechanism to adopt in handling the challenge of missing data and data inconsistencies found in the dataset.

3. Materials and Methods

  • The Data and Training Environment
This research implementation was in collaboration with an industry partner. This industry partner was the data provider, including real historical data collected from the SWT setups’ daily usage over one year, from 2 August 2017 to 11 July 2018, from 27 water points at Jarreng village in the Gambia, West Africa. The data represented the water system’s activity, which was collected in real time through the NFC component of the SWT via a remote server. The data also contained 1,054,009 samples with varying discrepancies and tap performances. A missing rate p was used to randomly generate missing samples, while the data distribution and imputation were achieved using the GAN and BiGRU methods. The data observation had an interval of 1 h, which was used to accommodate all activities of the SWT.
The computing environment was an HP workstation with 16 GB of RAM and one terabyte of memory. The training environment was the Anaconda distribution of Python 3.0 and TensorFlow version 1.7.1 installed on the workstation. The evaluation metric was the root mean squared error (RMSE).

3.1. Problem Formulation

Data imputation is a challenging task that consumes time. Moreover, selecting relevant variables from a dataset involves the identification of useful potential predictor subsets from a large candidate set [10]. Therefore, we defined missing data as presented in Equation (1).
Given the SWT setups with l taps, the SWT time series X recorded at times T = t 0 ,   .   .   . ,   t k 1 with samples X = x 0 ,   .   .   .   ,   x i ,   .   .   . , x k 1   k × d (where x i n -dimensional sample; k —original space dimensionality; d —number of data points).
In this formulation, the missing data pattern and location are deemed important. Therefore, we intend to show how to recover some of the unknown data values D   k × d from the case study noisy dataset X . We also assumed that D   k × d with values in 0 ,   1 is a mask matrix where D which indicates the existence or non-existence of a dataset value. For example, D i j = 1 if x i j exist and 0 otherwise. The aim is to impute missing data values as closely as possible to the existing data value. Therefore, we impute the SWT dataset to fill in the missing values based on their spatiotemporal relation.
In order to present a three-dimensional SWT multivariate example, four SWTs with three observations are shown in the matrix in Equations (6)–(9).
x = x 0 1 m i s s i n g x 0 3 x 0 4 x 1 1 x 1 2 m i s s i n g x 1   4 x 2 1 m i s s i n g x 2 3 m i s s i n g
D = 1 0 1 1 1 1 0 0 0 1 1 0
t = t 0 t 1 t 2 t 3
Next, a new matrix is defined thus:
ρ   k × d
This matrix records the time between the last value and the current value.

3.2. Characteristics of Data in the Investigated SWTs

Following daily use, the SWT generates a large quantity of time series datasets, which contain much information about the system’s behaviour. Therefore, it is important to have a time series dataset with continuous intervals without any breaks between them. For example, it is acceptable to have a dataset that consists of only Year–Month–Day or other time elements alone, such as Hour–Minute–Second (or even both). However, the time intervals must be provided in a continuous time step. For example, Figure 1a shows some data inconsistencies. Although the data structure looks fine, the time element contains inconsistent time steps. For example, 04:22:00 precedes 04:22:09 (a 9 s time step), after which the minutes and seconds change to 3 and 4 s intervals. Figure 1b is another sample of incoherent time steps found in the DateTime column of the dataset.
It is scientifically advisable that the time series dataset or sample be collected in a manner (i.e., stochastically) that reflects the daily, monthly or strictly adheres to hour/minute/second intervals [3]. Another technique is to roll up the time steps and take the statistical mean value of the columns [15]. Figure 2 also shows examples of missing day and month values. For example, from 2018-07-11, the next date is shown as 2019-12-04. Thus, a few months were missing from the list of months in some of the datasets.
To further emphasise the noise in the dataset due to a poor data collection method, Figure 3 shows an example of missing years, months and days (inclusive of a missing timestamp as well). For example, the year 2006 ends in February, and the next time step begins in January 2007. Similarly, only two time steps exist in the real data of 2007.

3.3. Missingness in a Dataset

In order to effectively handle missingness in a dataset, the method used should be peculiar to the case study datasets, the cause of the missingness and the percentage of missing data in the case study. However, in selecting the model for our study, we considered the simplicity of the missing data handling model and its ability to minimise the bias it introduces to the dataset. We did this to ensure that high-quality datasets could be used for modelling. For example, the water tap dataset contained many missing values in some variables (e.g., Figure 2 and Figure 3). Furthermore, in certain variables (e.g., Figure 3), incorrect values were also found.
An initial exploratory data analysis (EDA) was performed on the dataset [16]. The EDA aids in iteratively interacting with the dataset to extract relevant information from the case study. For example, the information could be the quality, the shape and relevant insights about the dataset. Figure 4 shows an initial EDA performed on the historical dataset acquired from the study location.
Figure 4 depicts the data variable and their corresponding missing values. From the plot, the variables; Litre, Voltage, TotalPrice and the amount spent all have significant missing dataset values, which can affect the result of any data analysis. Finally, after ascertaining the missingness in the case study dataset, we then formulate the missing data as described in Section 3.4.

3.4. Missing Data Formulation

The missing data pattern and location are deemed important in the missing data formulation. Therefore, we show how to recover some of the unknown data values D   k × d from the case study noisy dataset X . It is assumed that D   k × d with values in 0 ,   1 is a mask matrix where D indicates the existence or non-existence of a dataset value. For example, D i j = 1 if x i j exist and 0 otherwise. The aim is to impute missing data values as closely as possible to the existing data value. Therefore, the imputation of the SWT dataset is to fill in the missing values based on their spatiotemporal relation.

3.5. Implementation of Missing Data Imputation

This section gives a detailed description of the proposed SWT data imputation methods. The data imputation approach used in this work involves using GAN and BiGRU. Each approach plays a significant role in achieving a clean dataset for time series failure prediction in SWTs.

3.5.1. Generative Adversarial Network (GAN)

Goodfellow [17] proposed the generative adversarial network, which constitutes two parts, namely: the generator network ( G ) and the discriminator network ( D ). We applied the generator network component to map random SWT data noise (i.e., the low-dimensional vectors) to the data samples. Secondly, we used the discriminator network to receive low- and high-dimensional data values and distinguish the two sets [18]. Then, we simultaneously trained the generator and discriminator net to compete during training. Therefore, the GAN method generated a new synthetic dataset in line with the training set distribution.
On the other hand, the discriminator network (𝔇) was used to distinguish between the forged time series and the real SWT dataset because it had access to the real SWT and forged data samples. Finally, we used the value function V   G ,   D to evaluate the model learning cost. The learning process involves the computation shown in Equation (10):
m a x   D m i n G   V   G ,   D
where:
V   G ,   D = E p   d a t a x l o g   D X + E p g x l o g   1 D X
Therefore, the generator parameter stays fixed while the discriminator is updated or vice versa. However, in practice, the traditional GAN suffers instability (mode collapse). We further introduced the training stability by introducing Wasserstein GAN (WGAN), which is formulated as shown in:
m a x   θ D m i n θ G E p   d a t a x   D X E p   g x ( D G X + Z )
Therefore, to efficiently impute missing SWT data values, we used the bidirectional GRU to compress the missing e-tap data value into the low-dimensional vector. Then, the low-dimensional vector was applied to the reconstruction of the forged SWT time series ( x ) in order to confuse the discriminator network. The discriminator generator, on the other hand, harnessed the recursive function to create a distinction between x and x .

3.5.2. Bidirectional Gated Recurrent Unit (BiGRU)

It is noteworthy that the SWT data value contained rough time lags that were handy in the method development. In cases where past observations were missing over a significant period, we applied the BiGRU to introduce a decay vector β , which decreased the GRU memory by combining the cell state and the input gate, as shown in Equation (12).
β t i = 1 e m a x 0 ,   W β δ t i + b β
where:
W β and b β —training parameters
Therefore, based on Equation (12), the decay vector β is a value between 0–1, but it is not zero. Following this, the GRU hidden state h t i 1 is updated through a multiplication (through a listwise manner) of the decay vector β as shown in Equation (13):
h t i 1 = β t i     h t i 1
where:
h t i 1 —the hidden state
However, because the method uses a BiGRU, we designed the imputation process to correct missing data values and simultaneously save implementation time. Thus, two hidden states are added up, as shown in Equation (14):
h t i 1 = ( h t i + h t i )
Updating the BiGRU for SWT missing data imputation is as follows:
μ t i 1 = σ   W μ 1 h t i 1   1   ,   x t i + b μ 1
r t i 1 = σ   W r 1 h t i 1   1   ,   x t i + b r 1
h ˜ t i 1 = tanh W h ˜ 1 r t i 1 h t i 1   1   ,   x t i + b h ˜ 1
h t i 1 = 1 μ t i     h t i 1   1 + μ t i 1   h ˜ t i 1
μ t i 2 = σ   W μ 2 h t i 1   2   ,   x t i + b μ 2
r t i 2 = σ   W r 2 h t i 1   2   ,   x t i + b r 2
h ˜ t i 2 = tanh W h ˜ 2 r t i 2 h t i 1   2   ,   x t i + b h ˜ 2
h t i 2 = 1 μ t i     h t i 1   2 + μ t i 1   h ˜ t i 2
where
σ —activation function
μ t i 1 and μ t i 2 —reset gates
r t i 1 and r t i 2 —reset gates
h ˜ t i 1 , h ˜ t i 2 , h t i 1 and h t i 2 —hidden states
W μ , W r , W h ˜ , b μ , b r and b h ˜ —learning parameters
Based on the preceding equations, the decay vector β computation ensures that the value of the decay vector decreases as the time lag increases.

3.6. The Proposed GAN for SWT Dataset Imputation

Generally, random noises are passed to the GAN model’s generator network to generate a synthesised set of data values. However, in the proposed method, the SWT dataset already contains highly noisy sets of time series, which makes the addition of new noises appropriate. Usually, the incomplete SWT dataset would need additional noise before the model training begins. Figure 5 shows the proposed method and how it tries to reconstruct the required synthesised time series x .
From Figure 5, the incomplete SWT time series is harnessed by the generator network to impute samples from which the synthesised time series x was generated, while the BiGRUI is the component we used to construct the generator and discriminator networks. The discriminator network, on the other hand, compares the synthesise time series x with the actual time series x values and attempts to distinguish between the x and the x values. Convergence of the discriminator network occurs where a distinction between the two variables’ values cannot be achieved. The training of the denoiser model is as shown in Equation (23):
G η + x   = x
The generator network G Z produces new time series x similar to the actual SWT time series x , by adding to the generator’s loss function a squared error metric. This process makes it easy to impute the values of the actual SWT time series x with the synthesise time series x . Therefore, the BiGRUI accepts x   as input for processing. After the time series processing, the recurrent neural network’s hidden state connects to the fully connected layer of the network. In turn, the fully connected layer of the network then compresses the low-dimensional vector Z as output. The low-dimensional vector becomes the initial input of the second fully connected layer. Progressively, this becomes the input of the next BiGRUI layer of the recurrent neural network. This process continues until the final stage of the network generates the sample x by combining all the previous outputs.
The discriminator network also consists of the BiGRUI layer and a fully connected neural network layer. The discriminator distinguishes the synthesised samples from the actual data samples. By doing this, the discriminator outputs a probability showing the extent of the authenticity of the data samples. In order to achieve this, the discriminator’s loss function is defined as below:
L D = D x + D x
The complete SWT time series x and the incomplete SWT data with their time lags becomes the input to the discriminator network and are processed with the aid of the BiGRUI. After successfully processing the SWT time series, the fully connected layer of the BiGRUI accepts the last neural network’s hidden layer as input and gives out a probability. The generator was updated more than once, while the discriminator network was only enhanced once through a single iteration. After the data imputation, the triple exponential method for data smoothing was applied to correct any remaining data imbalance in the dataset.

Triple Exponential Smoothing

In order to achieve data imbalance correction, we used the triple exponential smoothing (ES) method to smooth out data imbalances left in the dataset. The ES accepts the imputed dataset from the GAN network, deseasonalises it and creates an adaptive normalisation for the new imputed dataset. The exponential method can also capture global dependencies on the dataset and extract relevant data variables based on the global information. Technically, the ES aids in assigning weights of previous values to predict the future weight of observed values [19].
The triple exponential smoothing method smooths out datasets with seasonality and trends in the SWT dataset by applying three constants: the smoothing constant, the trend smoothing constant β and the seasonal change constant γ . The multiplicative triple ES method (which is the focus of this section) is shown in Equations (25).
y t = α   x t s t L + 1 α y t 1 + b t 1 b t = β y t y t 1 + 1 β b t 1 s t = γ   x t y t + 1 γ s t L
where
y t —exponentially smoothed outcome
x t —the observed value of the time series
y t 1 —forecasted value at time t 1
α —smoothing constant, 0   α 1
β —trend smoothing constant, 0 β 1
γ —seasonal change constant 0 β 1
s t —seasonal indices
Applying the triple ES method enabled the model to handle the SWT data imbalance after data imputation. Data imbalance can be described as a classification task where the observations are not evenly distributed.
Other failure imputation methods exist, and these methods were chosen as the baseline to compare with our proposed GAN and BiGRU data imputation method. Therefore, the following baseline methods were selected: (i) K—nearest neighbour (KNN) [20], mean-value-filled (MEAN) [21], matrix factorisation (MF) [22], multiple imputations by chained equation (MICE) [23], generative adversarial network (GAN) [17] and GAIN [24]. These methods are contemporary benchmarks for data imputation and served as baseline techniques for the proposed failure imputation model.

4. Results and Discussion

This section discusses the results of the data preparation method proposed in this paper. Every SWT dataset with incomplete values x was mapped to all the low-dimensional vectors in the network. After the mapping, a time series is reconstructed from the low-dimensional vector. This reconstruction synthesised data values close to the real SWT dataset. The synthesised data values become values for the missing data values. The formula for the imputation is given as follows:
x = x   D + 1 m x
The proposed data imputation method was tested on a historical e-tap dataset, and the results are analysed and presented in this section. The proposed data imputation method results corrected discrepancies in our dataset before we applied the ML techniques to build a failure prediction model. Table 1 shows the model’s result compared to some selected baseline models.
Table 1 shows that the proposed BiGAN for data imputation in SWT performed better than other baseline models, and as such, it was the best fit for data imputation. The RMSE of a better model is expected to be smaller. Therefore, it can be seen that the result of the BiGRU-enabled GAN (BiGAN) has a better evaluation result for the SWT data preparation model. Compared to the regular GAN, it can also be seen from Table 1 that a significant difference makes the BiGAN a better choice for the e-tap than the other data imputation methods.
Furthermore, Figure 6 shows the value of RMSE for different data points at different timestamps. It can be noticed that the BiGRU enabled GAN has the lowest RMSE values, showing that the method is a good choice for the study. Compared with the complete data (the synthesised dataset), which has a higher RMSE value, the BiGRU GAN used for this study shows better performance. Therefore, it is a good fit for the SWT missing data imputation approach.
After data imputation, the new dataset still contained trends and seasonality. Therefore, this required adopting a method that can cater to those parameters in the dataset. One of the exponential smoothing methods was adopted against other moving average methods. This adoption of the triple exponential smoothing method is because the SWT dataset is large and still contains some outliers that led to overfitting at some point in developing the failure prediction model. The other data smoothing methods (like the moving average) are suitable for smaller datasets, and there is currently no scholarly article on which method to choose for a particular study. However, the choice of what data imputation method to adopt is dependent on the dataset available and the target result of the analysis.
The triple exponential smoothing method was chosen as the most appropriate for the SWT time series data imputation because the dataset shows trends. However, the single and double exponential smoothing methods cannot work in this case because they do not contain the seasonality parameter to solve the problem. Figure 7, Figure 8 and Figure 9 show the results of the three methods on the same dataset.
Figure 7 and Figure 8 show the single and double exponential smoothing methods that cannot solve seasonality in the dataset. The double exponential smoothing method can solve the trend challenge by using the trend constant. However, the seasonality problem cannot be solved because the method does not contain the seasonality parameter.
Compared with the single and double exponential smoothing methods, the triple exponential smoothing method, as shown in Figure 9, solved the seasonality problem, making it a better choice for the data smoothing task.
Sub-Saharan African rural villages depend on solar-powered water withdrawal taps for a clean domestic water supply [25]. However, these inadequate solar-powered taps sometimes failure and typically cause a shortage in the water supply. The result obtained from this research aid in ensuring a sustainable regime that will guarantees consistent water supply for the growing rural community.

5. Conclusions

The interest of this paper is to develop a missing data imputation method to correct missing data in a time series dataset. The aim was to prepare our dataset for failure prediction. Therefore, the degree of missingness, the time (i.e., the computational time) required to prepare (preprocess) the dataset and the quantity of dataset under investigation must be considered in deciding an appropriate approach for the data preparation task.
In order to achieve our aim, we used the BiGRU-enabled GAN generator network to generate newly synthesised samples similar to our dataset. At the same time, our model enabled the discriminator network to distinguish between the synthesised and original datasets. This distinction normally comes with some errors, but the RMSE helped the BiGRU by minimising the error. The results show that the proposed BiGAN achieved the lowest RMSE score, making it a better choice for missing data imputation.
In this paper, we considered time a crucial factor and that consideration influenced the data preparation technique proposed. The e-tap dataset is a large dataset with many missing values. As such, it requires a method that can salvage some of the implementation time to develop an efficient failure prediction model that works in real time. In order to achieve time efficiency for the model, the bidirectional GRU was adopted and fused with the GAN techniques.
After data imputation, the dataset still exhibited some seasonality and trends, which needed to be corrected before the dataset could be used for effective and unbiased failure prediction. Therefore, the triple exponential smoothing method was chosen to solve the seasonality and trends in the dataset. This method helped to further prepare the dataset for efficient failure prediction model development. With this proposed method, we can conclude that the BiGAN is a better choice for missing data imputation involving large datasets.

Author Contributions

Conceptualization, N.M.O.; Methodology, N.M.O.; Formal analysis, N.M.O.; Data curation, N.M.O.; Supervision, F.A.M. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be made available on request as it has a non disclosure agreement attached to it.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jones, S.A.; Sanford Bernhardt, K.L.; Kennedy, M.; Lantz, K.; Holden, T. Collecting critical data to assess the sustainability of rural infrastructure in low-income countries. Sustainability 2013, 5, 4870–4888. [Google Scholar] [CrossRef] [Green Version]
  2. Kabir, G.; Tesfamariam, S.; Hemsing, J.; Sadiq, R. Handling incomplete and missing data in water network database using imputation methods. Sustain. Resilient Infrastruct. 2020, 5, 365–377. [Google Scholar] [CrossRef]
  3. Peter, J.; Brockwell, R.A.D. Time Series: Theory and Methods, 2nd ed.; Springger Science+Business, Media LLC: New York, NY, USA, 2006. [Google Scholar]
  4. Najafabadi, M.M.; Villanustre, F.; Khoshgoftaar, T.M.; Seliya, N.; Wald, R.; Muharemagic, E. Deep learning applications and challenges in big data analytics J. Big Data 2015, 2, 1. [Google Scholar] [CrossRef] [Green Version]
  5. Domingos, P. A Few Useful Things to Know About Machine Learning. Commun. ACM 2012, 55, 78–87. [Google Scholar] [CrossRef] [Green Version]
  6. Robert, H.; Shumway, D.S.S. Time Series Analysis and Its Applications, 4th ed.; Springer Science+Business Media: New York, NY, USA, 2016. [Google Scholar]
  7. Valis, D.; Hasilova, K.; Forbelska, M.; Pietrucha-Urbanik, K. Modelling water distribution network failures and deterioration. In Proceedings of the IEEE International Conference on Industrial Engineering and Engineering Management, Singapore, 10–13 December 2017; pp. 924–928. [Google Scholar] [CrossRef]
  8. Salgado, C.M.; Azevedo, C.; Proen, H.; Vieira, S.M. Missing Data. In Secondary Analysis of Electronic Health Records; Springer Nature: Basingstoke, UK, 2016; pp. 1–427. [Google Scholar] [CrossRef] [Green Version]
  9. Ma, J.; Cheng JC, P.; Jiang, F.; Chen, W.; Wang, M.; Zhai, C. A bi-directional missing data imputation scheme based on LSTM and transfer learning for building energy data. Energy Build. 2020, 216, 109941. [Google Scholar] [CrossRef]
  10. Hadeed, S.J.; O’Rourke, M.K.; Burgess, J.L.; Harris, R.B.; Canales, R.A. Imputation methods for addressing missing data in short-term monitoring of air pollutants. Sci. Total Environ. 2020, 730, 139140. [Google Scholar] [CrossRef] [PubMed]
  11. Lee, K.J.; Tilling, K.M.; Cornish, R.P.; Little, R.J.; Bell, M.L.; Goetghebeur, E.; Hogan, J.W.; Carpenter, J.R. Framework for the treatment and reporting of missing data in observational studies: The Treatment And Reporting of Missing data in Observational Studies framework. J. Clin. Epidemiol. 2021, 134, 79–88. [Google Scholar] [CrossRef] [PubMed]
  12. Yang, Y.; Kim, J.; Cho, I.-H. Parallel Fractional Hot Deck Imputation and Variance Estimation for Big Incomplete Data Curing. IEEE Trans. Knowl. Data Eng. 2022, 34, 3912–3926. [Google Scholar] [CrossRef]
  13. Hamori, S.; Motegi, K.; Zhang, Z. Copula-based regression models with data missing at random. J. Multivar. Anal. 2020, 180, 104654. [Google Scholar] [CrossRef]
  14. Li, W.; Yang, S.; Han, P. Robust estimation for moment condition models with data missing not at random. J. Stat. Plan. Inference 2020, 207, 246–254. [Google Scholar] [CrossRef]
  15. Offiong, N.M.; Wu, Y.; Memon, F.A. Predicting failures in electronic water taps in rural sub-Saharan African communities: An LSTM-based approach. Water Sci. Technol. 2020, 82, 2776–2785. [Google Scholar] [CrossRef] [PubMed]
  16. Verbeeck, N.; Caprioli, R.M.; Van de Plas, R. Unsupervised machine learning for exploratory data analysis in imaging mass spectrometry. Mass Spectrom. Rev. 2020, 39, 245–291. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  18. Zhao, F.; Lu, Y.; Li, X.; Wang, L.; Song, Y.; Fan, D.; Zhang, C.; Chen, X. Multiple imputation method of missing credit risk assessment data based on generative adversarial networks. Appl. Soft Comput. 2022, 126, 109273. [Google Scholar] [CrossRef]
  19. Barrow, D.; Kourentzes, N.; Sandberg, R.; Niklewski, J. Automatic robust estimation for exponential smoothing: Perspectives from statistics and machine learning. Expert Syst. Appl. 2020, 160, 113637. [Google Scholar] [CrossRef]
  20. Sina, D.; Thomas, B. Anomaly Detection in Univariate Time Series: An Empirical Comparison of Machine Learning Algorithms. In Proceedings of the ICDM, Beijing, China, 8–11 November 2019; pp. 1–15. [Google Scholar]
  21. Khan, S.I.; Hoque, A.S.M.L. SICE: An improved missing data imputation technique. J. Big Data 2020, 7, 37. [Google Scholar] [CrossRef] [PubMed]
  22. Natarajan, S.; Vairavasundaram, S.; Natarajan, S.; Gandomi, A.H. Resolving data sparsity and cold start problem in collaborative filtering recommender system using Linked Open Data. Expert Syst. Appl. 2020, 149, 113248. [Google Scholar] [CrossRef]
  23. Li, L.; Prato, C.G.; Wang, Y. Ranking contributors to traffic crashes on mountainous freeways from an incomplete dataset: A sequential approach of multivariate imputation by chained equations and random forest classifier. Accid. Anal. Prev. 2020, 146, 105744. [Google Scholar] [CrossRef]
  24. Yoon, J.; Jordon, J.; Van Der Schaar, M. GAIN: Missing data imputation using generative adversarial nets. In Proceedings of the 35th International Conference on Machine Learning, ICML, Stockholm, Sweden, 10–15 July 2018; Volume 13, pp. 9042–9051. [Google Scholar]
  25. Foster, R.; Cota, A. Solar Water Pumping Advances and Comparative Economics. Energy Procedia 2014, 57, 1431–1436. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a): Data inconsistencies; (b): Data inconsistencies (missing times).
Figure 1. (a): Data inconsistencies; (b): Data inconsistencies (missing times).
Sustainability 15 06083 g001
Figure 2. Further data inconsistencies.
Figure 2. Further data inconsistencies.
Sustainability 15 06083 g002
Figure 3. Data inconsistencies (missing months, days and timestamps).
Figure 3. Data inconsistencies (missing months, days and timestamps).
Sustainability 15 06083 g003
Figure 4. An Initial exploratory data analysis.
Figure 4. An Initial exploratory data analysis.
Sustainability 15 06083 g004
Figure 5. The proposed missing data imputation method.
Figure 5. The proposed missing data imputation method.
Sustainability 15 06083 g005
Figure 6. The RMSE metric.
Figure 6. The RMSE metric.
Sustainability 15 06083 g006
Figure 7. Single exponential smoothing for SWT.
Figure 7. Single exponential smoothing for SWT.
Sustainability 15 06083 g007
Figure 8. Double exponential smoothing for SWT.
Figure 8. Double exponential smoothing for SWT.
Sustainability 15 06083 g008
Figure 9. Triple exponential smoothing for SWT.
Figure 9. Triple exponential smoothing for SWT.
Sustainability 15 06083 g009
Table 1. BiGAN RMSE Comparison with Baseline Models after Imputation.
Table 1. BiGAN RMSE Comparison with Baseline Models after Imputation.
Missingness (%)KNNMEANMFMICELASTGAINGANBiGAN
10–200.460.380.320.530.310.350.330.29
30–400.570.420.480.660.420.440.540.41
50–600.620.510.550.720.560.500.620.52
70–800.740.680.610.740.670.610.750.60
90–1000.810.730.780.810.770.780.770.68
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Offiong, N.M.; Memon, F.A.; Wu, Y. Time Series Data Preparation for Failure Prediction in Smart Water Taps (SWT). Sustainability 2023, 15, 6083. https://doi.org/10.3390/su15076083

AMA Style

Offiong NM, Memon FA, Wu Y. Time Series Data Preparation for Failure Prediction in Smart Water Taps (SWT). Sustainability. 2023; 15(7):6083. https://doi.org/10.3390/su15076083

Chicago/Turabian Style

Offiong, Nsikak Mitchel, Fayyaz Ali Memon, and Yulei Wu. 2023. "Time Series Data Preparation for Failure Prediction in Smart Water Taps (SWT)" Sustainability 15, no. 7: 6083. https://doi.org/10.3390/su15076083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop