Next Article in Journal
Eddy Current Transducer with Rotating Permanent Magnets to Test Planar Conducting Plates
Previous Article in Journal
Accurate Multiple Ocean Bottom Seismometer Positioning in Shallow Water Using GNSS/Acoustic Technique
Previous Article in Special Issue
Intelligent and Dynamic Ransomware Spread Detection and Mitigation in Integrated Clinical Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using the IBM SPSS SW Tool with Wavelet Transformation for CO2 Prediction within IoT in Smart Home Care

Department of Cybernetics and Biomedical Engineering, Faculty of Electrical Engineering and Computer Science, VSB—Technical University of Ostrava, Ostrava 70800, Czech Republic
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(6), 1407; https://doi.org/10.3390/s19061407
Submission received: 31 January 2019 / Revised: 7 March 2019 / Accepted: 13 March 2019 / Published: 21 March 2019

Abstract

:
Standard solutions for handling a large amount of measured data obtained from intelligent buildings are currently available as software tools in IoT platforms. These solutions optimize the operational and technical functions managing the quality of the indoor environment and factor in the real needs of residents. The paper examines the possibilities of increasing the accuracy of CO2 predictions in Smart Home Care (SHC) using the IBM SPSS software tools in the IoT to determine the occupancy times of a monitored SHC room. The processed data were compared at daily, weekly and monthly intervals for the spring and autumn periods. The Radial Basis Function (RBF) method was applied to predict CO2 levels from the measured indoor and outdoor temperatures and relative humidity. The most accurately predicted results were obtained from data processed at a daily interval. To increase the accuracy of CO2 predictions, a wavelet transform was applied to remove additive noise from the predicted signal. The prediction accuracy achieved in the selected experiments was greater than 95%.

1. Introduction

An intelligent building is one that is responsive to the requirements of occupants, organizations, and society. An intelligent building requires real-time information about its occupants so that it can continually adapt and respond [1]. Intelligent buildings respond to the needs of occupants and society, promoting the well-being of those living and working in them [2]. The researchers point out that the Adaptive House (the concept of a home which programs itself), Learning Homes and Attentive Homes must be programmed for a particular family and home and updated in line with changes in their lifestyle. The system monitors actions taken by the residents and looks for patterns in the environment which reliably predict these actions, where a neural network learns these patterns and the system then performs the learned actions automatically for improving the Quality of Life (QoL) [3]. Privacy, reliability and false alarms are the main challenges to be considered for the development of efficient systems to detect and classify the Activities of Daily Living (ADL) and Falls [4]. In order to provide a user-friendly environment for the management of the operational and technical functions along with providing support for the independent housing of senior citizens and disabled persons in buildings indicated as Smart House Care (SHC), it is necessary to make appropriate visualization of the technological process as required by the users with the possibility of the indirect monitoring of the seniors’ life activities based on the information obtained from the sensors used for the common management of the operational and technical functions in SHC. The properly devised visualization complements the final correct functionality of the intelligent building. Beaudin et al. described the use of computational and sensor technology in intelligent buildings with a focus on health monitoring in the households. This work comprises of a visualization for displaying health data and a proposal for improving the health and wellbeing of the users [5]. Booysen et al. explored machine-machine communication (M2M) to address the need for autonomous control of remote and distributed mobile systems in intelligent buildings [6]. Basu et al. used miniature wireless sensors in the wireless network to track and recognize the behavior of persons in the house. The visualization includes sensor data from the building to capture the duration of the activities [7]. Fleck et al. described a system based on intelligent cameras for 24-h monitoring and supervision of senior citizens. On this occasion, visualization is used to display relevant life information in this intelligent environment, which includes the evaluation of the seniors’ position and recognition of life activities [8]. The current trend for processing large volumes of measured quantities using Soft Computing Methods (SC) [9,10] is to use the available Big Data Analysis tools within the IoT platform [11,12].
The Internet of Things, shortly known as IoT, can be assumed as an integration layer which creates an interconnection of several physical devices, sensors, actuators, and controllers [13]. In simple words, the IoT allows objects other than computers or smartphones to use the Internet for sending and receiving information [14]. The number of connected IoT devices is rapidly growing. This growth could be the result of a very wide range of applications, ranging from basic home appliances and security systems to more sophisticated applications. For example, Xu et al. used IoT in order to construct a real-time system monitoring system for micro-environment parameters such as temperature, humidity, PM10 and PM2.5 [15]. Q. Min et al. suggested using IoT for monitoring of discrete manufacturing process based on IoT [16]. Wang et al. presented a feasible and reliable plantation monitoring system based on the Internet of Things, that combined the wireless sensor network, embedded development, GPRS communication technology, web service, and Android mobile platform [17]. Windarto et al. presented an application of IoT by implementing automation of lights and door in a room [18]. Coelho et al. used IoT to collect data from multiple heterogeneous sensors that were providing different types of information at a variety of locations in a smart home [19]. Data collection in this kind of examples usually results in big data. The Oxford dictionary defines big data as “extremely large data sets that may be analyzed computationally to reveal patterns, trends, and associations, especially relating to human behavior and interactions” [20]. One of the possibilities that come with big data processing is a predictive analysis which can provide predictions about the future or otherwise unknown events. Predictive analysis offers a wide range of applications such as social networking, healthcare, mobility, insurance, finance marketing, etc. Nyce suggested to use the predictive analysis for risk identifications and probabilities in order to provide an appropriate insurance rate, his method took advantage of marketing records, underwriting records, and claims records [21]. Predictive analysis includes many different statistical techniques ranging from data mining, predictive modeling to machine learning. Predictive modeling may be applied to many areas such as weather forecasting, business, Bayesian spam filters, advertising and marketing, fraud detection, etc. Predictive models are based on variables that are most likely to influence the outcome [22]. These variables are also known as predictors. There are many types of predictive models, such as neural networks and Decision trees. Ahmad et al. compared different methods of predictive modeling for solar thermal energy systems such as random forest, extra trees, and regression trees [23].
IBM offers a variety of services in terms of predictive analysis such as Watson analytics and SPSS [24,25,26,27,28,29,30,31,32,33,34,35,36]. AlFaris et al. reviewed the smart technologies; the interface and integration of the meters, sensors and monitoring systems with the home energy management system (HEMS) within the IoT with the outline that the smart home in practice provides the ability to the house to be net-zero energy building. Especially that it reduces the power demand and improve the energy performance by 37% better than ASHRAE standards for family villas sector [37]. Alirezaie et al. presents a framework called E-care@home, consisting of an IoT infrastructure, which provides information with an unambiguous, shared meaning across IoT devices, end-users, relatives, health and care professionals and organizations and demonstrates the proposed framework using an instantiation of a smart environment that is able to perform context recognition based on the activities and the events occurring in the home [38]. Bassoli et al. introduced a new system architecture suitable for human monitoring based on Wi-Fi connectivity, where the proposed solution lowers costs and the implementation burden by using the Internet connection that leans on standard home modem-routers, already present normally in the homes, and reducing the need for range extenders thanks to the long range of the Wi-Fi signal with energy savings of up to 91% [39]. Catherwood et al. presented an advanced Internet of Things point-of-care bio-fluid analyzer; a LoRa/Bluetooth-enabled electronic reader for biomedical strip-based diagnostics system for personalized monitoring, where practical hurdles in establishing an Internet of Medical Things network, assisting informed deployment of similar future systems are solved [40].
In this paper, the authors focused on designing a methodology for processing data obtained from sensors that measure non-electrical quantities in an SHC environment for the purpose of monitoring the presence of people in a room through the KNX (Konnex bus) and BACnet (Building Automation and Control Networks) technologies designed for SHC automation. Standard solutions using Software (SW) tools in IoT platforms are currently available and can process large amounts of measured data. These solutions monitor and optimize the quality of the indoor environment, taking into account the real needs of residents.
This paper examines the possibilities of increasing the accuracy in CO2 predictions in Smart Home Care (SHC) using IBM SPSS software tools in the IoT to determine occupancy times of a monitored SHC room. The accuracy of CO2 predictions from the processed data was compared and evaluated at daily, weekly and monthly intervals for the spring and autumn periods. To predict CO2 levels from the measured indoor and outdoor temperatures and relative humidity, the Radial Basis Function (RBF) method (feedforward neural network) was applied. To improve the accuracy of CO2 predictions, a wavelet transform was applied to remove additive noise from the predicted signal.
For the classification of prediction quality with Wavelet Transformation (WT) additive noise cancelation, a correlation analysis (correlation coefficient R), calculated MSE (Mean Squared Error), Mean Absolute Error (MAE), Euclidean distance (ED), City Block distance (CB) are used.

2. Description of Used Technologies in SHC

The Smart two-floor wooden house (hereafter, Smart Home; floor area of 12.1 m × 8.2 m; (Figure 1) was built as a training center of the Moravian-Silesian Wood Cluster (MSWC). The wooden house (SHC) was built to a passive standard in accordance with standards ČSN 75 0540-2 and ČSN 730540-2 (2002) [41]. The measured values for the internal climate monitoring in the living space were evaluated by selection and measurement of CO2, the temperature (T) and relative humidity (rH) in room 204 with the application of air-quality sensors Siemens QPA2062 implemented in the BACnet system [42]. The basic element of the BACnet control system is a sub-station DESIGO PX PXC100-E.D. The DESIGO PX sub-stations are controlled by a user-friendly control panel PXM 20E [43]. The BACnet technology is used for HVAC (heating, ventilating, and air conditioning) control in SHC. The process sub-station BACnet/IP DESIGO PX-PXC100 E.D. forms a foundation of the control system for the measurement of nonelectrical quantities and control of operating and technical functions in SH [44]. The communication between the individual modules is executed via a standard BACnet protocol over Ethernet. Mutual communication between the sub-stations (peer-to-peer) is supported there [45]. KNX technology is used for lighting and blinds control and the switching on/off of socket circuits [46]. Furthermore, information about the security status of the building is transferred from the Electronic Security System (ESS) using KNX technology [47]. The KNX technology is suitably integrated into the BACnet technology using an interface providing interoperability between the communication protocols [48]. The visualization of control and regulation of the operating and technical functions in the building (HVAC, blinds, lighting, etc.) and storage of the measured data in the database was created using DESIGO Insight software tools [49]. To unify the monitoring and control of various automation and electronic systems into a single environment, it is necessary to use a software tool that allows you to consolidate several types of bus systems, standards and communication, and data protocols into o single monitoring application. In our case, we took advantage of the opportunities offered by the PI System software tool (hereinafter referred to as PI) produced by OSIsoft (Figure 1) [50].
The PI System includes SW tools such as PI ProcessBook for user-friendly data readout with the ability to create an application for the visualization and monitoring of SHC resident activities (Figure 2 and Figure 3) [51,52,53].

Visualization for the SHC Created in the PI ProcessBook Tool

Visualization of the wooden house in the PI Process Book tool is divided into several screens, which can be continuously accessed from the main screen (Figure 2). The SHC control technology is integrated on each visualization screen in accordance with the Building Management System (BMS). These screens further comprise of buttons for entry into individual rooms. After clicking on the relevant room button, a new window, containing a detailed description of the technology used in the specific room shown in the individual charts, will appear. Each technological element is illustrated in the chart, wherein the individual charts are sorted in accordance with the groups of elements used (Figure 3). The individual technology units are then displayed on separate screens and can also be viewed from the main screen. This solution was chosen because it is not possible to place all the information about the technologies implemented on one screen so that the screen remained well-arranged. The additional distribution of the technologies into individual screens will allow the user to get a better insight into what elements belong to the individual technologies and which do not anymore. Thanks to this solution selected, orientation in the enclosed charts are easier as well as the analysis of the individual quantities and actions in the building.
The visualization, monitoring, and processing of the measured values of non-electric variables, such as the measurement of temperature, humidity, and CO2 for monitoring the quality of the indoor environment of the selected room in the building described, are implemented using the PI System software application and SPSS IBM SW Tool (Figure 1).

3. Proposed Method for Creating an Optimized Model of CO2 Concentration Prediction

3.1. Implementation of Predictive Analysis Using the IBM SPSS Modeler

The IBM SPSS Modeler allows users to build models using a simplified, easy-to-use, and object-oriented user interface. The user is provided with Modelling algorithms, such as prediction, classification, segmentation, and association detection. The model results can be easily deployed and read into databases, IBM SPSS Statistics and a wide variety of other applications. Working with IBM SPSS Modeler can be divided into three basic steps:
  • Importing the data into IBM SPSS Modeler
  • Performing a series of analyses on the imported data
  • Evaluation and exporting the data
This sequence is also known as a Datastream because the data is flowing from the source to each analysis node and then to the output. The IBM SPSS Modeler allows the users to work with multiple data streams at once. These data streams can be build and modified using the stream canvas area of the application. These streams are created by drawing diagrams of relevant data operations. IBM SPSS Modeler’s Node Palette area of the displays shows most of the available data and modeling tools. The user may perform a simple drag and drop on each item in the nodes palette to add them to the current stream. The node palette items are divided into a few main categories as follows [54]:
  • Source: contains nodes that allow importing data into IBM SPSS Modeler from external sources such as analytic servers, databases, XML files, Microsoft Excel etc.
  • Record Ops: includes Nodes performing operations on data records, such as selecting, merging, and appending.
  • Field Ops: these nodes can perform operations on data fields such as filtering, deriving new fields, and determining the measurement level for given fields.
  • Graphs: Provides nodes that can graphically represent the data from before and after modeling.
  • Modeling: contains available modeling algorithms such as neural networks, decision trees, clustering, and data sequencing.
  • Output: consists of nodes that can provide outputs, such as plots, charts, evaluation, etc.
  • Export: composed of the nodes that can export the output to other applications, such as Microsoft Excel.
  • IBM SPSS Statistics: dedicated to nodes for importing or exporting data to IBM SPSS Statistics.
Neural networks are one of the many ways to achieve predictive analysis. IBM SPSS Modeler offers multiple types of neural networks for predictive analysis. The text further describes the procedure for determining the appropriate method of predicting the course of CO2 concentration from the measured values taken by the indoor temperature sensor Ti (°C), (QPA 2062) in an SHC room (range 0 to 50 °C/−35 to 35 °C, accuracy ± 1K) and relative humidity rH (%), (QPA 2062) (range 0 and 100%, accuracy ±5%), outdoor temperature To (°C), (AP 257/22), (range: −30 … + 80 °C, resolution: 0.1 °C) using the RBF. The RBF was selected due to its higher speed of training [55]. The RBF network is a feed-forward network that requires supervised learning. Unlike multilayer perceptron’s (MLP), this network consists of only one hidden layer. Overall, there are three layers in the RBF network: the input layer, RBF layer, and the output layer. The IBM SPSS algorithm guide describes mathematical models of these layers [56] (Figure 4) as following:
Input layer: J 0 = P   u n i t s ,   a 0 : 1 ,   , a 0 : J 0 with a 0 : j = X j
RBF layer: j 1   u n i t s ,   u n i t s ,   a 1 : 1 ,   a 1 : j 1 ; with a 1 : j = j ( X )
j ( X ) = e ( p = 1 P 1 2 σ j p 2 ( x p μ j p ) 2 ) j = 1 J 1 e ( p = 1 P 1 2 σ j p 2 ( x p μ j p ) 2 )
Output layer: j 2 = R   u n i t s ,   a I : 1 ,   a I : j 2 with a I : r = ω I : r + j = 1 j 1 ω r j j ( X )
Where:
  • X ( m ) : Input vector
  • I: Number of layers (for RBM = 2)
  • Ji: Number of units in layer i
  • j ( X ( m ) ) : jth unit for input X ( m ) ,   j = 1 ,   ,   j 1 .
  • µj: Center of j
  • σj: Width of j
  • a i : j m : Unit j of layer i
  • ω r j : weight connecting rth output unit and jth hidden unit of RBF layer
The training of RBF can be divided into two stages. The first stage determines the basis function by clustering methods and the second stage determines the weights given to the basis function. SPSS measures the accuracy of neural networks by calculating the percentage of the records for which the predicted value matches the observed value. For the continues values, the accuracy is calculated by 1 minus the average of the absolute values of the predicted values minus the observed values over the maximum predicted value minus the minimum predicted value (the following formula) [56].
Accuracy = 1 n m = 1 M ( 1 | y r ( m ) y ^ r ( m ) | m a x m ( y r ( m ) ) m i n m ( y r ( m ) ) ) ,

3.2. Signal Trend Detection Based on Wavelet Transformation

Materials and Methods

In this section, we introduce a method for the CO2 concentration prediction optimization based on the Wavelet transformation additive noise canceling. Based on the experimental results, the predicted CO2 trend contains glitches representing the fast change part of the signal. Such signal segments may significantly deteriorate the quality of the prediction. We propose an optimized scheme of the neural network prediction based on the Wavelet filtration appearing as a robust method due to a wide variability of the filtration settings. Such a system significantly improves the prediction system based on the neural network.
In the signal processing, we assume that each signal y(t) is composed of two essential parts, namely, they are the signal trend T(t) and a component having a stochastic character X(t) which is perceived as the signal noise and details. Based on this definition, we can use the following signal formulation (3):
y ( t ) = T ( t ) + X ( t )
The major problem when the signal trend is being extracted is noise detection. There are many applications of the trend detection including the CO2 measurement. Such a signal may be influenced by the glitches which should be removed to obtain a smooth signal for further processing. The wavelet analysis represents a transformation of the signal y(t) to obtain two types of coefficients, particularly they are the wavelet and scaling coefficients. These coefficients are completely equivalent with the original CO2 signal. It is supposed that wavelet coefficients are related to changes along a specifically defined scale. The main idea of the signal trend detection is to perform an association of the scaling coefficients with the signal trend T(x). On the other hand, the wavelet coefficients are supposed to be associated with the signal noise, which is mainly represented by the glitches when processing the CO2 signal. In our analysis, we considered an uncorrelated noise, adapting the wavelet estimator to work as a kernel estimator. The advantage of such an approach is formulating an estimator based on the sampled data irregularity. In this method, we used the scaling coefficients as estimators of the signal trend. We supposed that the sampled CO2 observations are represented by Y ( t n ) , thus, the CO2 estimator is given by Equation (4):
T ^ ( t ) = n = 0 N 1 Y ( t n ) E J ( t , s ) d s
Integration is done over a set of the intervals ( A n ( s ) ), their union forms perform partitioning interval covering all the observations   t n , where t n A n . Consequently, E J is defined as Equation (5):
E J ( t , s ) = 2 J k θ ( 2 J t k ) θ ( 2 J s k )
In this expression, θ ( t ) represents the scaling function. This function is defined as follows (Equation (6)):
θ ( t ) = k c k θ ( 2 t k )
The wavelet function is defined by Equation (7):
ψ ( t ) = ( 1 ) k c 1 k θ ( 2 t k )
The first crucial task is an appropriate selection of the mother’s wavelet for the predicted CO2 signal filtration. Supposing the Daubechies wavelets can well reflect morphological structure therefore, this family was used for our model. Particularly, in our approach, we used the Daubechies wavelet (Db6), with the D6 scaling function utilizing the orthogonal Daubechies coefficients.

3.3. Validation Ratings Used

In order to carry out the objective comparison, the following parameters were considered:
Mean Absolute Error (MAE) represents the estimator measuring of the difference between two continuous variables. The MAE is given by the following expression:
M A E = 1 n i = 1 n | y i y ı ^ |
Mean squared error (MSE) represents the estimator measuring the average of the error squares between two signals. The MSE represents a risk function which corresponds with the expected value of the squared or quadratic error loss. The MSE is given by the following expression:
M S E ( x 1 , x ^ 2 ) = 1 n i = 1 n ( x ( i ) x ^ ( i ) ) 2
Euclidean distance (ED) represents an ordinary straight-line distance between two points lying in the Euclidean space. Based on this distance, the Euclidean space becomes a metric space. The lower the Euclidean distance we achieve, the more similar are two signal samples. In our analysis, we considered a mean of the ED. The Euclidean distance is given by the following expression:
d ( x 1 , x ^ 2 ) = ( x 1 x ^ 2 ) 2 + ( y 1 y ^ 2 ) 2
City Block distance (CB) represents a distance between two signals x 1 , x ^ 2 in the space with the Cartesian coordinate system. This parameter can be interpreted as a sum of the lengths of the projections of the line segments between the points onto the coordinate axes. CB distance is defined as follows:
d c b ( x 1 , x ^ 2 ) = x 1 x ^ 2 = i = 1 n | x 1 i x ^ 2 i |
The Correlation coefficient (R) measures a level of the linear dependency between two signals. The more the signals are considered linearly dependable, the higher the correlation coefficient is. In comparison with the previous parameters, the correlation coefficient represents a normative parameter. Zero correlation stands for the total dissimilarity between two signals, measured in a sense of their linear dependency. Contrarily, 1 and −1 stand for full positive and full negative correlation.
As we have already stated above, in our work, we analyze two-month CO2 predictions. In each measurement, we have a prediction from the neural network with 10, 50, 100, 150, 200, 250, 300, 350 and 400 neurons. Thus, we completely analyzed 9 predicted signals for each measurement. These signals are compared against the reference based on the evaluation parameters stated above. In terms of the Euclidean distance and MSE, lower values indicate a higher agreement between the signal and reference and thus, a better result. Contrarily, a higher correlation coefficient shows better results. In the following part of the analysis, we report the results of the quantification comparison. All the testing is done for the Wavelet Db6, with 6-level decomposition and the Wavelet settings as follows: threshold selection rule—Stein’s Unbiased Risk and soft thresholding for selection of the detailed coefficients.

4. The Practical Experimental Part

4.1. First Part of the ADL Information in SHC from the CO2 Concentration Course, Blinds, Slats and On/Off Control of Lights

The first experimental part in the study addressed the real needs of seniors who live in their own flats despite advanced age and mental and physical disabilities. These people strive to maintain maximum self-sufficiency and, thus, remove as much burden from their relatives, neighbors, friends or surroundings as possible. An example is a married couple, one of whom is mentally impaired, the other being the caregiver. They stay in touch with their family (their children) by SMS to keep them informed about how they are. In situations of acute need, the children are ready to come and help. In this example, the indirect ADL (Activities of Daily Living) in Room R203 (Figure 2) in SHC can be detected by monitoring operational and technical functions, such as

Discussion of the First Experimental Part

In the aforementioned results of Experimental Part 1 (Figure 5, Figure 6, Figure 7 and Figure 8), the presence of persons in the SHC monitored area is clearly time-localized according to the ADL. ADL information in SHC can be thus reliably forwarded to the close relatives. The information obtained at daily, weekly and monthly intervals can also be used alongside SHC automation technologies in a so-called “smart building”. This type of building records house activities and uses the accumulated data to automatically control technologies according to the predictable needs of users, such as controlling lights, blinds, heating, forced ventilation and cooling based on the usual patterns of use and facilitates cost savings for programming and configuring the intelligent house control system [2].

4.2. Second Experimental Part: ADL Monitoring Information from Prediction CO2 Concentration Course Within IoT SPSS SW Tool

As it was described earlier, the main goal is predicting CO2 concentration based on data collected by indoor humidity, indoor temperature, and outdoor temperature. The sample data were collected from the SHC. The procedure of this implementation can be divided into a few steps as following:
  • Pre-processing the data
  • Developing a data stream using IBM SPSS Modeler
  • Testing various training data from different times of the year
  • Analyzing the results and selecting the best model
  • Uploading the selected model method to Watson studio for IoT implementation

4.2.1. Pre-Processing

Data Normalization using the min-max method (often known as feature scaling) was used as a pre-processing method. This method scales the parameters in the range between 0 and 1. Since the experimental data were stored in data files, the pre-processing stage was performed. The implementation was performed by calculating the minimum and maximum values of each parameter. Then the normalized values were calculated using Equation (12).
N o r m a l i z e d   v a l u e = c u r r e n t   v a l u e min   v a l u e max   v a l u e min   v a l u e

4.2.2. Developing a Data Stream Using IBM SPSS Modeler

Figure 9 shows the data stream developed in IBM SPSS Modeler. In the first stage, the data were fed to IBM SPSS using Excel files by adding Excel node from the source category of the node palette. In the next stage, a data type selection operator was added from the field operator category. This node carried the task of setting the default target to CO2 concentration and default inputs (in this case inputs were used as predictors) to humidity, indoor temperature, and outdoor temperature values.
There are a few common validation methods used in IBM SPSS Modeler such as K-fold, V-fold, N-fold, and Partitioning. The IBM SPSS Modeler User manual recommends using the partitioning method for large datasets due to the faster processing time. The partitioning method randomly divides the data sets into three parts of training, testing, and validation. The ratio of this division can be selected in the software using percentages of the data set. The partition node form field operator category was added to the stream in order to divide it into three parts. The first part (40% of the data) for the training of the neural network, using target values (CO2) and predictors (humidity, indoor temperature, and output temperature). The second part for testing (30% of the data) the neural network by using target and predictor values. The testing partition is used for selecting the most suitable model and prevention of overfitting. The last partition was dedicated for validation (30% of the data) of the developed model using only predictors and comparing the prediction with the reference signal. In other words, the validation partition is used to determine how well the model truly performs [19].
In the next step, an automatic data preparation field operator was used in order to transform the data for better predictive accuracy. The transformed data were fed to a neural network for training a model. The neural network is using the RBF model with various numbers of neurons for multiple implementations (Figure 10). The resulting model from the trained neural network is represented by a nugget gem. Varies nodes were connected to this nugget gem for additional analysis of the model such as an Excel node for exporting reference data and predicted values to an Excel file (a filter node was used to select which values to store in the output Excel file), a Time plot node to display the time plot of reference versus predicted values, a Multiplot node for displaying the plots from portioned data and an Analysis node for displaying the details such as linear correlation, mean absolute error, etc.

4.2.3. Testing Various Training Data from Different Times of the Year

For this stage of the experiment, seven different data sets from the spring and fall of 2018 were selected. The data collection was performed at the rate of one sample per minute. The first selected data interval was the whole month of May (Table 1). The validation results indicate that model number 7 with 76.6% accuracy and relevantly small error (MAE = 0.006, MSE = 9.75 × 10−5) represents the best prediction. By repeating the experiment with data from November 2018 (Table 2), a slight improvement in the accuracy of all models can be observed. Additionally, by observing Table 3, it is apparent that model number 9 holds the best results in terms of accuracy (80.6%), linear correlation (0.898), MAE (0.011) and MSE (1.553 × 10−3).
The training process was repeated by replacing the intervals with a week in May 2018 (Table 3) and a week in November 2018 (Table 4). With a reduction in the data size, the chances of overfitting were reduced, resulting in better generalization and higher accuracy (Figure 11). These improvements can be observed in Table 3 and Table 4. In both cases, the model number 9 shows the finest accuracy (6th to 13th of May 2018: 98.2%; 6th to 13th of November 2018: 96.2%), highest linear correlation (6th to 13th of May 2018: 0.985; 6th to 13th of November 2018: 0.979) and the lowest error (6th to 13th of May 2018: MSE: 3.14 × 10−6 MAE: 0.001; 6th to 13th of November 2018: MSE: 4.12 × 10−5, MAE: 0.003). Additionally, a consistency between the accuracy results of the two experiments can be observed.
For the last few experiments, the size of the data sets was reduced to one day. The 15th and 12th of May (Table 5 and Table 6) and the 15th of November (Table 7), Figure 12 were selected for this experiment. Once again, a significant improvement of the accuracies due to the reduction in the interval lengths can be observed. The result from the 12th of May (Table 5) implies that model number 5 shows the maximum accuracy (98.1%). By considering the linear correlation and error values, it can be concluded that model number 3 shows better overall characteristics (Figure 13). In the case of 15th of May, model number 3 provides the most accurate result with an impressive 99.9% of accuracy, closely followed by model number 9 with 99.8% of accuracy. In the next step, the experiment was repeated with data from November 15. Similar to the few of the previous cases, model number 9 shows the highest accuracy (99.7%), relevantly high linear correlation value (0.895) and the lowest errors (MAE: 0.003, MSE: 6.45 × 10−5).

4.2.4. Analyzing the Results and Selecting the Best Model

Table 8 shows the average of the accuracy, linear correlation MAE and MSE in each experiment. By observing this table, it is apparent that the experiments with an interval length of one month, hold the lowest average accuracies (63.1% and 72%). It can also be observed that the experiment with the 15th of May as period holds the highest average accuracy (99.5%), highest linear correlation (0.995) and relevantly low error values (MAE: 1.78 × 10−3, MSE: 2.44 × 10−5). Additionally, the experiment with a interval length of a week in May shows a slightly smaller average accuracy (94.7%) and linear correlation (0.967) but it has overall lower error values (MAE: 1.78 × 10−3, MSE: 2.44 × 10−5) (Figure 14).
Table 9 contains the average results from all experiments for each model. The overall trend of this Table points toward the fact that with the increasing number of neurons, the accuracy of the model’s increases. This increase may be insignificant after a certain point. For example, models number 5, 6, 7, 8 and 9 share a similar average accuracy (91.2%, 91.1%, 90.2%, 90.2% and 92.1%), linear correlation (0.935, 0.931, 0.935, 0.936 and 0.936), MAE (8.15 × 10−3, 9.14 × 10−3, 7.71 × 10−3, 7.71 × 10−3 and 7.85 × 10−3) and MSE (1.02 × 10−3, 1.45 × 10−3, 1.15 × 10−3, 1.15 × 10−3 and 1.17 × 10−3).
Figure 11 clearly demonstrates the relationship between accuracy, the experiments interval length and the number of neurons. By summing up all of the obtained results, it is clear that model number 3 (Table 6) trained with the data from 15th of May holds the highest accuracy (99.8%), linear correlation (0.996) and relevantly small error values. Model number 9 shows the best overall average accuracy (Table 9) and, in case of the experiments with the interval of 12th (see Figure 12) and 15th May, the accuracy difference between this model and the most accurate model (model number 3) is negligible. Therefore, model number 9 trained with the May 15th interval was selected as the most suitable model for the next stages (Figure 12).

4.2.5. IoT Implementation with Watson Studio SW Tool

The data streams created in the IBM SPSS Modelers can be stored as a file (“.str” format). The IBM Watson Studio allows the user to import the developed data stream simply by uploading the stored files. This allows the data streams that were originally developed in the IBM SPSS Modeler to take advantage of cloud computing, cloud storage and the possibility of near-real-time streaming. As it was explained earlier, model number 9 (with 400 neurons) trained with the data from 15th of May was selected as the best overall result of this experiment. Specifically, this model showed high accuracy, high linear correlation, low MAE and MSE errors. Therefore, it was uploaded to Watson studio. Figure 15 shows the streamed developed in SPSS in Watson studio for near-real-time training (Excel files were replaced with assets on the cloud). Figure 16 shows a data flow stream in Watson that includes model 9 trained with data from the 15th of May for near real-time prediction.

4.2.6. Discussion of the Second Experimental Part

By evaluating the obtained results from the implementation with IBM SPSS Modeler, it is apparent that as it was expected that the experiments with interval length of one day showed better overall accuracy (average value up to 99.5%) and the experiments with sample periods of one month showed the least overall accuracy (average value up to 72.2%). Additionally, in four out of six experiments, model number 9 held the highest accuracy, and in the other two cases, it had the only insignificant difference with the most accurate models. Therefore, it was selected as the overall most accurate model. As it was mentioned earlier, in terms of the training interval, 15th of May showed the most accurate results. Therefore, model number 9 with the 15th of May training interval (Table 9) was selected and exported to an IBM cloud data stream. Furthermore, the results demanded additional filtering in order to reduce the noise and provide smoother results.

4.3. Third Part—Testing and Quantitative Comparison (WT Additive Noise Canceling)

4.3.1. Testing and Quantitative Comparison

In our research, we analyzed signals representing the CO2 signals. We had a set of the estimated (predicted) signals being compared against the real measured signal CO2, which is perceived as a reference. In our analysis, we are comparing two-month CO2 prediction. We compared one-day, one-week and one-month predictions for May and November 2018.
Based on the observations, it is apparent that the predicted CO2 signals do not have a smooth process. They are frequently influenced by rapid oscillations, so-called glitches and signal fluctuations (Figure 17 and Figure 18). Such signal variations represent the signal noise, impairing the real trend of the CO2 prediction, which should be reduced. In our analysis, we used wavelet filtration to eliminate such signals to obtain the signal trend for further processing.
As we have already stated above, we used the mother’s wavelet Db6 for the CO2 signal trend detection. Firstly, we take advantage of the fact that different level of the decomposition allows perceiving more or less signal details represented by the detailed coefficients. Since we need to perceive the signal trend by eliminating the steep fluctuations, we need to consider an appropriate level of the decomposition. An experimental comparison of individual wavelet settings is reported in Figure 17 and Figure 18. Based on the experimental results, we used the 6-level decomposition for the CO2 signal trend detection. The filtration procedure further utilizes the following settings: threshold selection rule—Stein’s Unbiased Risk and soft thresholding for selection of the detailed coefficients.
Wavelet filtration was used for the extraction of the CO2 signal trend, simultaneously rapid changes of the signal were removed. On Figure 19 and Figure 20, there is a comparison among the reference signal and predicted signals by wavelet transformation for day and month predictions from May and November 2018.
Based on the results, wavelet filtration is capable of filtering rapid signal changes whilst preserving the signal trend. To justify this situation, we report the selected situations showing the glitches deteriorating a smooth signal trend, and a respective wavelet approximation largely reducing such signal parts (Figure 21, Figure 22 and Figure 23). Among these cases, we mark the most significant glitches as green in the originally predicted signals to highlight the Wavelet smoothing effectivity.
As it is obvious, the CO2 prediction contains lots of significant occurrences represented by the glitches and spikes, significantly deteriorating the smoothness of the analyzed signal. Wavelet appears to be a reliable alternative for reduction of those parts of the signal. On the other hand, we are aware that trend detection, in some cases, reduces the peaks and thus, the original signal’s amplitude is reduced. Such situations are reported in Figure 21, Figure 22 and Figure 23.
In the last part of our analysis, the objective comparison is carried out. As we have already stated, we are comparing predicted CO2 signals with signals being filtered out by the wavelet transformation. All the signals are compared against the reference CO2 signals for day and week predictions.
As we have already stated above, in our work we analyzed the two-month CO2 prediction. In each measurement, we have a prediction from a neural network with 10, 50, 100, 150, 200, 250, 300, 350 and 400 neurons. Thus, we completely analyzed the predicted signals of 9 models for each measurement. These signals are compared against the reference based on the evaluation parameters stated above. In terms of the Euclidean distance and MSE, lower values indicate a higher agreement between the signal and reference and thus, better result. Contrarily, a higher correlation coefficient indicates better results. In the following part of the analysis, we report the results of the quantification comparison. All the testing is done for the Wavelet Db6, with 6-level decomposition and the following Wavelet settings: threshold selection rule—Stein’s Unbiased Risk and soft thresholding for selection of the detailed coefficients. Figure 24 shows the MSE evaluation for CO2 prediction in the following time intervals: (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018. Figure 25 shows the correlation coefficient evaluation for the CO2 prediction in the following time intervals: (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018. Figure 26 shows the Euclidean distance evaluation for the CO2 prediction in the following time intervals: (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Lastly, we summarize the achieved results of all the predicted signals. In Table 10 and Table 11, we compare the individual parameters—Mean Square Error (MSE), Correlation index (Corr) and Euclidean distance (ED)—for individual CO2 predictions from May and November 2018. Each parameter is averaged for all the predictions and the difference (diff) between the original prediction and Wavelet smoothing is evaluated (Table 12).

4.3.2. Discussion of the Third Experimental Part

As it is obvious, the predicted CO2 signals contain lots of significant occurrences represented by the glitches and spikes, significantly deteriorating the smoothness of the analyzed signals. Such steep fluctuations may have a significant impact on CO2 accuracy. Wavelet appears to be a reliable alternative for reduction of those parts of the signal. On the other hand, we are aware that trend detection, in some cases, reduces the peaks and thus, the original signal’s amplitude is reduced. In our work, we have studied the Daubechies wavelet family. These wavelets, as it is known, can well reflect the morphological structure of the signals. We are particularly using the Db6 wavelet for trend detection.
Alternatively, we mention the comparison in Reference [50] of the CO2 filtration based on the LMS algorithm. In this study, the authors employed adaptive filtration. The main limitation of this method is a necessity of the reference signal and a slow adaptation of the filtration procedure, as well as depending on the accuracy of the step size parameter μ calculation and inaccurate determination of the arrival and departure time of the person from the monitored area. Furthermore, the Wavelet filtration presented in this study achieves better results in a context of the objective comparison against the LMS filtration. Wavelet filtration has a much stronger potential for the CO2 filtration due to a possibility of the application of a variety wavelets allowing for the extraction of specific morphological signal features in various decomposition levels and, thus, better optimize the CO2 prediction. These facts predetermine wavelets to be a robust system for the CO2 prediction enhancement.
In the last part of our analysis, the objective comparison is carried out. As we have already stated, we compared originally measured CO2 signals with predicted signals being filtered out by the wavelets. To carry out the objective comparison, the following parameters are considered: when considering the MSE, we get better results for wavelet trend detection. This means that we have minimized the difference between the gold standard and filtered signals. The correlation coefficient gives higher values for the predicted CO2 signals. Regardless, we have achieved just slight differences. The reason might be caused by the fact that the trend detection largely omits higher peaks, therefore, the linear dependence for wavelet filtration is smaller when compared with the predicted signals. Using the Wavelet filtration leads to more accurate results against the predicted signals and signals are much more smoothed, not containing steep fluctuations. On the other hand, we are aware of a certain loss of the amplitude. Therefore, in the future, it would be worth investigating the frequency features of the CO2 signals to objectively determine frequency modifications while filtering by the wavelets.

5. Conclusions

The authors of the paper focused on designing a methodology that specifies a procedure for processing data measured by sensors in an SHC environment for the purpose of indirectly monitoring the presence of people in an SHC area through KNX and BACnet technologies commonly applied in building automation. This paper explores the possibilities of improving accuracy in CO2 predictions in SHC using IBM SPSS software tools in the IoT to determine the occupancy times of a monitored SHC room. The RBF method was applied to predict CO2 levels from the measured indoor and outdoor temperatures and relative humidity. The accuracy of CO2 predictions from the processed data was compared and evaluated at daily, weekly and monthly intervals for the spring and autumn periods. As it was expected the most accurate results were provided by experiments with the daily intervals (accuracy was about 99%) while the monthly intervals resulted in the least accurate results (accuracy was about 80%). Overall, the developed stream in IBM SPSS Modeler is capable of predicting the CO2 concentration values using the values of humidity and indoor and outdoor temperature. By providing a live data asset to the IBM Cloud, the uploaded model can achieve near real-time prediction of CO2 concentration values. Using a wavelet transform mathematical method to cancel additive noise led to more accurate results in predicted signals. The signals were also much smoother and did not contain sharp fluctuations, although there was a certain loss in amplitude, resulting in inaccuracies when the maximum achieved CO2 value was determined. Future work should, therefore, focus on finding an optimal method for canceling additive noise in real time, which would help increase the overall accuracy of CO2 predictions [57,58,59,60]. Additionally, the real-life and live performance of this implementation should be examined.

Author Contributions

J.V. methodology; J.V., J.K. (Jan Kubicek), O.M.G. software; J.V., J.K. (Jan Kubicek) validation; J.V., J.K. (Jan Kubicek), O.M.G. formal analysis; J.V. investigation; J.V., O.M.G. resources; J.V. data curation; J.V., J.K. (Jan Kubicek), O.M.G. writing—original draft preparation; J.V., O.M.G. writing—review and editing; J.V. visualization; J.V. supervision; J.K. (Jiri Koziorek) project administration; J.K. (Jiri Koziorek) funding acquisition.

Funding

This research was funded by the Ministry of Education of the Czech Republic (Project No. SP2018/170) and by the European Regional Development Fund in the Research Centre of Advanced Mechatronic Systems project, project number CZ.02.1.01/0.0/0.0/16_019/0000867 within the Operational Programme Research, Development and Education.

Acknowledgments

This work was supported by the Student Grant System of VSB Technical University of Ostrava, grant number SP2019/118. This work was supported by the European Regional Development Fund in the Research Centre of Advanced Mechatronic Systems project, project number CZ.02.1.01/0.0/0.0/16_019/0000867 within the Operational Programme Research, Development and Education. The work and the contributions were supported by the project SV4507741/2101, ‘Biomedicínské inženýrské systémy XIII’.

Conflicts of Interest

The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Vanus, J.; Martinek, R.; Kubicek, J.; Penhaker, M.; Nedoma, J.; Fajkus, M. Using the PI processbook software tool to monitor room occupancy in smart home care. In Proceedings of the 2018 IEEE 20th International Conference on e-Health Networking, Applications and Services (Healthcom), Ostrava, Czech Republic, 17–20 September 2018. [Google Scholar] [CrossRef]
  2. Clements-Croome, D.J. Intelligent Buildings; ICE Publishing: London, UK, 2013; p. 1. [Google Scholar]
  3. Harper, R. Inside the Smart Home; Springer: Berlin/Heidelberg, Germany, 2003; p. 24. ISBN 1852336889. [Google Scholar]
  4. Longhi, S.; Freddi, A.; Monteriù, A. Human Monitoring, Smart Health and Assisted Living: Techniques and Technologies (Healthcare Technologies); Institution of Engineering and Technology: Stevenage, UK, 2017. [Google Scholar]
  5. Beaudin, J.S.; Intille, S.S.; Morris, M.E. To Track or Not to Track: User Reactions to Concepts in Longitudinal Health Monitoring. J. Med. Internet Res. 2006, 8, e29. [Google Scholar] [CrossRef]
  6. Booysen, M.J. Machine-To-Machine (M2M) Communications in Vehicular Networks. KSII Trans. Internet Inf. Syst. 2012, 6. [Google Scholar] [CrossRef]
  7. Basu, D.; Moretti, G.; Gupta, G.S.; Marsland, S. Wireless sensor network based smart home: Sensor selection, deployment and monitoring. In Proceedings of the IEEE Sensors Applications Symposium Proceedings, Galveston, TX, USA, 19–21 February 2013; pp. 49–54. [Google Scholar] [CrossRef]
  8. Fleck, S.; Strasser, W. Smart Camera Based Monitoring System and Its Application to Assisted Living. Proc. IEEE 2008, 96, 1698–1714. [Google Scholar] [CrossRef]
  9. Pantazaras, A.; Lee, S.E.; Santamouris, M.; Yang, J. Predicting the CO2 levels in buildings using deterministic and identified models. Energy Build. 2016, 127, 774–785. [Google Scholar] [CrossRef]
  10. Ríos-Moreno, G.J.; Trejo-Perea, M.; Castañeda-Miranda, R.; Hernández-Guzmán, V.M.; Herrera-Ruiz, G. Modelling temperature in intelligent buildings by means of autoregressive models. Autom. Constr. 2007, 16, 713–722. [Google Scholar] [CrossRef]
  11. Aggarwal, M.; Madhukar, M. IBM’s Watson analytics for health care: A miracle made true. In Cloud Computing Systems and Applications in Healthcare; IGI Global: Hershey, PA, USA, 2017; pp. 117–134. [Google Scholar]
  12. Kaur, A.; Jasuja, A. Health monitoring based on IoT using Raspberry PI. In Proceedings of the 2017 International Conference on Computing, Communication and Automation (ICCCA), Greater Noida, India, 5–6 May 2017; pp. 1335–1340. [Google Scholar]
  13. Petnik, J.; Vanus, J. Design of Smart Home Implementation within IoT with Natural Language Interface. IFAC PapersOnline 2018, 51, 174–179. [Google Scholar] [CrossRef]
  14. McEwen, A.; Cassimally, H. Designing the Internet of Things, 1st ed.; John Wiley & Sons Ltd.: Chichester, UK, 2008; p. 9. ISBN 978-1-118-43062-0. [Google Scholar]
  15. Xu, B.; Zheng, J.; Wang, Q. Analysis and Design of Real-Time Micro-Environment Parameter Monitoring System Based on Internet of Things. In Proceedings of the 2016 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Chengdu, China, 15–18 December 2016. [Google Scholar]
  16. Min, Q.; Ding, Y.F.; Xiao, T.; Wang, S. Research of Visualization Monitoring Technology Based on Internet of Things in Discrete Manufacturing Process. In Proceedings of the 2015 2nd International Symposium on Dependable Computing and Internet of Things (Dcit), Wuhan, China, 16–18 November 2015; pp. 128–133. [Google Scholar]
  17. Wang, Y.; Song, J.; Liu, X.; Jiang, S.; Liu, Y. Plantation Monitoring System Based on Internet of Things. In Proceedings of the 2013 IEEE International Conference on Green Computing and Communications and IEEE Internet of Things and IEEE Cyber, Physical and Social Computing, Beijing, China, 20–23 August 2013; pp. 366–369. [Google Scholar]
  18. Windarto, Y.E.; Eridani, D. Door and Light Control Prototype Using Intel Galileo Based Internet of Things. In Proceedings of the 2017 4th International Conference on Information Technology, Computer, and Electrical Engineering (ICITACEE), Semarang, Indonesia, 18–19 October 2017; pp. 176–180. [Google Scholar]
  19. Coelho, C.; Coelho, D.; Wolf, M. An IoT Smart Home Architecture for Long-Term Care of People with Special Needs. In Proceedings of the 2015 IEEE 2nd World Forum on Internet of Things (WF-IOT), Milan, Italy, 14–16 December 2015; pp. 626–627. [Google Scholar]
  20. Oxford Dictionaries, Definition of Big Data in English. Available online: https://en.oxforddictionaries.com/definition/big_data (accessed on 25 November 2018).
  21. Nyce, C. Predictive Analytics White Paper. 2007. Available online: https://www.the-digital-insurer.com/wp-content/uploads/2013/12/78-Predictive-Modeling-White-Paper.pdf (accessed on 25 November 2018).
  22. Rouse, M. Predictive Modeling. Available online: https://searchenterpriseai.techtarget.com/definition/predictive-modeling (accessed on 25 November 2018).
  23. Ahmad, M.W.; Reynolds, J.; Rezgui, Y. Predictive modelling for solar thermal energy systems: A comparison of support vector regression, random forest, extra trees and regression trees. J. Clean. Prod. 2018, 203, 810–821. [Google Scholar] [CrossRef]
  24. IBM SPSS Software. Available online: https://www.ibm.com/analytics/spss-statistics-software (accessed on 25 November 2018).
  25. IBM Watson. Available online: https://www.ibm.com/watson/ (accessed on 25 November 2018).
  26. Nagwanshi, K.K.; Dubey, S. Statistical Feature Analysis of Human Footprint for Personal Identification Using BigML and IBM Watson Analytics. Arab. J. Sci. Eng. 2018, 43, 2703–2712. [Google Scholar] [CrossRef]
  27. Perumal, T.; Datta, S.K.; Bonnet, C. IoT Device Management Framework for Smart Home Scenarios. In Proceedings of the 2015 IEEE 4th Global Conference on Consumer Electronics (GCCE), Osaka, Japan, 27–30 October 2015; pp. 54–55. [Google Scholar]
  28. Arnold, O.; Kirsch, L.; Schulz, A. An Interactive Concierge for Independent Living. In Proceedings of the 2014 IEEE 3rd Global Conference on Consumer Electronics (GCCE), Tokyo, Japan, 7–10 October 2014; pp. 59–62. [Google Scholar]
  29. Carvalko, J.R. Law and policy in an era of cyborg-assisted-life the implications of interfacing in-the-body technologies to the outer world. In Proceedings of the 2013 IEEE International Symposium on Technology and Society, Toronto, ON, Canada, 27–29 June 2013; IEEE: New York, NY, USA, 2013; pp. 204–215. [Google Scholar]
  30. Cervenka, P.; Hlavaty, I.; Miklosik, A.; Lipianska, J. Using cognitive systems in marketing analysis. Econ. Ann. XXI 2016, 160, 56–61. [Google Scholar] [CrossRef]
  31. Chen, Y.; Argentinis, E.; Weber, G. IBM Watson: How Cognitive Computing Can Be Applied to Big Data Challenges in Life Sciences Research. Clin. Ther. 2016, 38, 688–701. [Google Scholar] [CrossRef] [PubMed]
  32. Coccoli, M.; Maresca, P.; Stanganelli, L.; Knowledge Systems Institute. Teaching Computer Programming Through Hands-on Labs on Cognitive Computing. In DMS 2016: The 22nd International Conference on Distributed Multimedia Systems; Knowledge Systems Institute: Skokie, IL, USA, 2016; pp. 158–164. [Google Scholar]
  33. Devarakonda, M.; Zhang, D.; Tsou, C.H.; Bornea, M. Problem-Oriented Patient Record Summary: An Early Report on a Watson Application. In Proceedings of the 2014 IEEE 16th International Conference on E-Health Networking, Applications and Services (Healthcom), Natal, Brazil, 15–18 October 2014; pp. 281–286. [Google Scholar]
  34. Guidi, G.; Miniati, R.; Mazzola, M.; Iadanza, E. Case Study: IBM Watson Analytics Cloud Platform as Analytics-as-a-Service System for Heart Failure Early Detection. Future Internet 2016, 8, 32. [Google Scholar] [CrossRef]
  35. Kolker, E.; Ozdemir, V. How Healthcare Can Refocus on Its Super-Customers (Patients, n = 1) and Customers (Doctors and Nurses) by Leveraging Lessons from Amazon, Uber, and Watson. OMICS J. Integr. Biol. 2016, 20, 329–333. [Google Scholar] [CrossRef] [PubMed]
  36. Murtaza, S.S.; Lak, P.; Bener, A.; Pischdotchian, A. How to Effectively Train IBM Watson: Classroom Experience. In Proceedings of the 2016 49th Hawaii International Conference on System Sciences, Koloa, HI, USA, 5–8 January 2016; Bui, T.X., Sprague, R.H., Eds.; IEEE Computer Society: Los Alamitos, CA, USA, 2016; pp. 1663–1670. [Google Scholar]
  37. AlFaris, F.; Juaidi, A.; Manzano-Agugliaro, F. Intelligent homes’ technologies to optimize the energy performance for the net zero energy home. Energy Build. 2017, 153, 262–274. [Google Scholar] [CrossRef]
  38. Alirezaie, M.; Renoux, J.; Köckemann, U.; Kristoffersson, A.; Karlsson, L.; Blomqvist, E.; Tsiftes, N.; Voigt, T.; Loutfi, A. An Ontology-based Context-aware System for Smart Homes: E-care@home. Sensors 2017, 17, 1586. [Google Scholar] [CrossRef]
  39. Bassoli, M.; Bianchi, V.; de Munari, I. A Plug and Play IoT Wi-Fi Smart Home System for Human Monitoring. Electronics 2018, 7, 200. [Google Scholar] [CrossRef]
  40. Catherwood, P.A.; Steele, D.; Little, M.; McComb, S.; McLaughlin, J. A Community-Based IoT Personalized Wireless Healthcare Solution Trial. IEEE J. Transl. Eng. Health Med. 2018, 6, 1–13. [Google Scholar] [CrossRef] [PubMed]
  41. Vanus, J.; Belesova, J.; Martinek, R.; Nedoma, J.; Fajkus, M.; Bilik, P.; Zidek, J. Monitoring of the daily living activities in smart home care. Hum. Cent. Comput. Inf. Sci. 2017, 7, 30. [Google Scholar] [CrossRef]
  42. Vanus, J.; Martinek, R.; Bilik, P.; Zidek, J.; Dohnalek, P.; Gajdos, P. New method for accurate prediction of CO2 in the Smart Home. In Proceedings of the Conference Record IEEE Instrumentation and Measurement Technology Conference, Taipei, Taiwan, 23–26 May 2016. [Google Scholar]
  43. Vanus, J.; Martinek, R.; Bilik, P.; Zidek, J.; Skotnicova, I. Evaluation of Thermal Comfort of the Internal Environment in Smart Home Using Objective and Subjective Factors. In Proceedings of the 2016 17th International Scientific Conference on Electric Power Engineering, Gothenburg, Sweden, 16–18 May 2016; IEEE: New York, NY, USA, 2016; pp. 524–528. [Google Scholar]
  44. Vanus, J.; Martinek, R.; Nedoma, J.; Fajkus, M.; Cvejn, D.; Valicek, P.; Novak, T. Utilization of the LMS Algorithm to Filter the Predicted Course by Means of Neural Networks for Monitoring the Occupancy of Rooms in an Intelligent Administrative Building. IFAC PapersOnline 2018, 51, 378–383. [Google Scholar] [CrossRef]
  45. Vanus, J.; Martinek, R.; Nedoma, J.; Fajkus, M.; Valicek, P.; Novak, T. Utilization of Interoperability between the BACnet and KNX Technologies for Monitoring of Operational-Technical Functions in Intelligent Buildings by Means of the PI System SW Tool. IFAC PapersOnline 2018, 51, 372–377. [Google Scholar] [CrossRef]
  46. Vanus, J.; Smolon, M.; Martinek, R.; Koziorek, J.; Zidek, J.; Bilik, P. Testing of the voice communication in smart home care. Hum. Cent. Comput. Inf. Sci. 2015, 5, 15. [Google Scholar] [CrossRef]
  47. Vanus, J.; Stratil, T.; Martinek, R.; Bilik, P.; Zidek, J. The Possibility of Using VLC Data Transfer in the Smart Home. IFAC PapersOnline 2016, 49, 176–181. [Google Scholar] [CrossRef]
  48. Vanus, J.; Valicek, P.; Novak, T.; Martinek, R.; Bilik, P.; Zidek, J. Utilization of regression analysis to increase the control accuracy of dimmer lighting systems in the Smart Home. IFAC PapersOnline 2016, 49, 517–522. [Google Scholar] [CrossRef]
  49. Vanus, J.; Vojcinak, P.; Martinek, R.; Kellnar, M.; Machacek, Z.; Bilik, P.; Koziorek, J.; Zidek, J. Building heating technology in Smart Home using PI System management tools. Perspect. Sci. 2016, 7, 114–121. [Google Scholar] [CrossRef]
  50. Vanus, J.; Machac, J.; Martinek, R.; Bilik, P.; Zidek, J.; Nedoma, J.; Fajkus, M. The design of an indirect method for the human presence monitoring in the intelligent building. Hum. Cent. Comput. Inf. Sci. 2018, 8, 28. [Google Scholar] [CrossRef]
  51. Skotnicova, I.; Lausova, L.; Michalcova, V.; Vanus, J. Temperatures and heat transfer beneath a ground floor slab in a passive house. In Nano, Bio and Green—Technologies for a Sustainable Future Conference Proceedings, SGEM 2016, Vol II (International Multidisciplinary Scientific GeoConference-SGEM, Vienna, Austria, 2–5 November 2016; Stef92 Technology Ltd.: Sofia, Bulgaria, 2016; pp. 269–276. [Google Scholar]
  52. Vanus, J.; Machac, J.; Martinek, R.; Bilik, P. Design of an application for the monitoring and visualization of technological processes with pi system in an intelligent building for mobile devices. In Proceedings of the 9th International Scientific Symposium on Electrical Power Engineering (ELEKTROENERGETIKA), Stara Lesna, Slovakia, 12–14 September 2017; pp. 518–522. [Google Scholar]
  53. Vanus, J.; Machacek, Z.; Koziorek, J.; Walendziuk, W.; Kolar, V.; Jaron, Z. Advanced energy management system in Smart Home Care. Int. J. Appl. Electromagn. Mech. 2016, 52, 517–524. [Google Scholar] [CrossRef]
  54. IBM SPSS Modeler 16 Algorithms Guide. Available online: ftp://public.dhe.ibm.com/software/analytics/spss/documentation/modeler/16.0/en/AlgorithmsGuide.pdf (accessed on 25 November 2018).
  55. IBM SPSS Modeler 17 Algorithms Guide. Available online: ftp://public.dhe.ibm.com/software/analytics/spss/documentation/modeler/17.0/en/AlgorithmsGuide.pdf (accessed on 25 November 2018).
  56. IBM SPSS Modeler 18 Algorithms Guide. Available online: ftp://public.dhe.ibm.com/software/analytics/spss/documentation/modeler/18.0/en/AlgorithmsGuide.pdf (accessed on 25 November 2018).
  57. PI System™ from Data to Knowledge to Transformation. Available online: https://www.osisoft.com/pi-system/#tab1 (accessed on 10 January 2019).
  58. Liu, Z.; Cheng, K.; Li, H.; Cao, G.; Wu, D.; Shi, Y. Exploring the potential relationship between indoor air quality and the concentration of airborne culturable fungi: A combined experimental and neural network modeling study. Environ. Sci. Pollut. Res. 2018, 25, 3510–3517. [Google Scholar] [CrossRef] [PubMed]
  59. Sterman, M.; Baglione, M. Simulating the use of CO2 concentration inputs for controlling temperature in a hydronic radiant system. In Proceedings of the ASME International Mechanical Engineering Congress and Exposition (IMECE), Tampa, FL, USA, 3–9 November 2017; Volume 6. [Google Scholar] [CrossRef]
  60. Zuraimi, M.S.; Pantazaras, A.; Chaturvedi, K.A.; Yang, J.J.; Tham, K.W.; Lee, S.E. Predicting occupancy counts using physical and statistical Co2-based modeling methodologies. Build. Environ. 2017, 123, 517–528. [Google Scholar] [CrossRef]
Figure 1. A block diagram of data transfer from the Smart Home Care technology through the PI OPC (Ole for Process Control) interface to the PI Server and PI ProcessBook.
Figure 1. A block diagram of data transfer from the Smart Home Care technology through the PI OPC (Ole for Process Control) interface to the PI Server and PI ProcessBook.
Sensors 19 01407 g001
Figure 2. The main visualization screen in SW tool PI ProcessBook, first floor.
Figure 2. The main visualization screen in SW tool PI ProcessBook, first floor.
Sensors 19 01407 g002
Figure 3. A screen for room 204 with a detailed description of individual technologies and individual charts in SW tool PI ProcessBook.
Figure 3. A screen for room 204 with a detailed description of individual technologies and individual charts in SW tool PI ProcessBook.
Sensors 19 01407 g003
Figure 4. The Radial Basis Function (RBF) neural network diagram.
Figure 4. The Radial Basis Function (RBF) neural network diagram.
Sensors 19 01407 g004
Figure 5. The reference (measured) CO2 concentration course and Illuminance E (lx) course (12 November 2018) in room R203 in Smart Home Care (SHC) for Activities of Daily Living (ADL) monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), Time of Person Presence (TPP) ∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45. Light on I. 8:15:54, Light off II. 17:43:47.
Figure 5. The reference (measured) CO2 concentration course and Illuminance E (lx) course (12 November 2018) in room R203 in Smart Home Care (SHC) for Activities of Daily Living (ADL) monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), Time of Person Presence (TPP) ∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45. Light on I. 8:15:54, Light off II. 17:43:47.
Sensors 19 01407 g005
Figure 6. The reference (measured) CO2 concentration course, indoor temperature and relative humidity course (12 November 2018), in room R203, outdoor temperature course (12 November 2018) within the measurement of nonelectrical values in SHC for ADL monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), (∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), (TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), (TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), (TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), (TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), (TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), (TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45. I. The window was opened at 8:22:44, II. The window was closed at 8:50:42, III. The window was opened partially at 10:02:19, IV. The entire window was opened at 10:42:34, V. The window was closed at 11:05:25, VI. The window was opened partially at 11:36:07, VII. The window was closed at 14:00:35.
Figure 6. The reference (measured) CO2 concentration course, indoor temperature and relative humidity course (12 November 2018), in room R203, outdoor temperature course (12 November 2018) within the measurement of nonelectrical values in SHC for ADL monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), (∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), (TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), (TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), (TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), (TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), (TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), (TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45. I. The window was opened at 8:22:44, II. The window was closed at 8:50:42, III. The window was opened partially at 10:02:19, IV. The entire window was opened at 10:42:34, V. The window was closed at 11:05:25, VI. The window was opened partially at 11:36:07, VII. The window was closed at 14:00:35.
Sensors 19 01407 g006
Figure 7. The reference (measured) CO2 concentration course and Blinds course (12 November 2018) in room R203 in SHC for ADL monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), ∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45. I. The Blind 43 (Figure 2) was lowered down in time 8:32:37, II. The Blind 48 (Figure 2) was lowered down in time 8:37:24 III. The Blind 33 (Figure 2) was lowered down in time 8:43:12, IV. The blinds were pulled partially in time 11:52:29.
Figure 7. The reference (measured) CO2 concentration course and Blinds course (12 November 2018) in room R203 in SHC for ADL monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), ∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45. I. The Blind 43 (Figure 2) was lowered down in time 8:32:37, II. The Blind 48 (Figure 2) was lowered down in time 8:37:24 III. The Blind 33 (Figure 2) was lowered down in time 8:43:12, IV. The blinds were pulled partially in time 11:52:29.
Sensors 19 01407 g007
Figure 8. The reference (measured) CO2 concentration course and Slats course (12 November 2018) in room R203 (Figure 2) in SHC for ADL monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), ∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), (TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), (TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45.I. Slats 47 (Figure 2) were turned at 8:32:37, II. Slats 42 were turned at 8:33:18. III. Slats 42 were turned at 8:38:05, IV. Slats 42 were turned at 8:39:47, V. Slats 32 were turned at 8:43:12, IV. Slats 42 and 47 were turned at 11:52:29.
Figure 8. The reference (measured) CO2 concentration course and Slats course (12 November 2018) in room R203 (Figure 2) in SHC for ADL monitoring. (1) arrival (12 November 2018 7:33:00), (2) departure (12 November 2018 8:28:11), ∆t1 = 0:55:11; (3) arrival (12 November 2018 09:12:52), (4) departure (12 November 2018 10:29:16), TPP ∆t2 = 1:16:24; (5) arrival (12 November 2018 11:41:55), (6) departure (12 November 2018 11:44:18), TPP ∆t3 = 0:02:23. (7) arrival (12 November 2018 12:18:04), (8) departure (12 November 2018 12:26:15), (TPP ∆t1 = 0:08:11; (9) arrival (12 November 2018 12:51:50), (10) departure (12 November 2018 12:53:12), (TPP ∆t2 = 0:01:22; (11) arrival (12 November 2018 13:51:52), (12) departure (12 November 2018 14:32:47), TPP ∆t3 = 0:40:55; (13) arrival (12 November 2018 14:52:34), (14) departure (12 November 2018 16:02:30), TPP ∆t2 = 1:09:56; (15) arrival (12 November 2018 16:23:18), (16) departure (12 November 2018 16:42:03), TPP ∆t3 = 0:18:45.I. Slats 47 (Figure 2) were turned at 8:32:37, II. Slats 42 were turned at 8:33:18. III. Slats 42 were turned at 8:38:05, IV. Slats 42 were turned at 8:39:47, V. Slats 32 were turned at 8:43:12, IV. Slats 42 and 47 were turned at 11:52:29.
Sensors 19 01407 g008
Figure 9. The developed stream using IBM SPSS Modeler.
Figure 9. The developed stream using IBM SPSS Modeler.
Sensors 19 01407 g009
Figure 10. The block diagram of the developed Stream using IBM SPSS Modeler.
Figure 10. The block diagram of the developed Stream using IBM SPSS Modeler.
Sensors 19 01407 g010
Figure 11. The accuracy for each interval of experiments.
Figure 11. The accuracy for each interval of experiments.
Sensors 19 01407 g011
Figure 12. Model number 9 trained and validated using data from yjr 15th of May 2018.
Figure 12. Model number 9 trained and validated using data from yjr 15th of May 2018.
Sensors 19 01407 g012
Figure 13. Model number 9 trained and validated using data from the interval of the 6th of May 2018 to 13th of May 2018.
Figure 13. Model number 9 trained and validated using data from the interval of the 6th of May 2018 to 13th of May 2018.
Sensors 19 01407 g013
Figure 14. Model number 3 trained and validated using data from the interval of 12th of May 2018.
Figure 14. Model number 3 trained and validated using data from the interval of 12th of May 2018.
Sensors 19 01407 g014
Figure 15. Watson Studio with the data stream developed in IBM SPSS Modeler.
Figure 15. Watson Studio with the data stream developed in IBM SPSS Modeler.
Sensors 19 01407 g015
Figure 16. Watson Studio with the data flow stream that uses model 9.
Figure 16. Watson Studio with the data flow stream that uses model 9.
Sensors 19 01407 g016
Figure 17. The comparison of the original predicted CO2 signal and Wavelet filtration from 15 May 2018 for 2-level decomposition (left), 6-level decomposition (middle) and 7-level decomposition (right).
Figure 17. The comparison of the original predicted CO2 signal and Wavelet filtration from 15 May 2018 for 2-level decomposition (left), 6-level decomposition (middle) and 7-level decomposition (right).
Sensors 19 01407 g017
Figure 18. The comparison of the original predicted CO2 signal and Wavelet filtration from 1–29 May 2018 for 2-level decomposition (left), 6-level decomposition (middle) and 7-level decomposition (right).
Figure 18. The comparison of the original predicted CO2 signal and Wavelet filtration from 1–29 May 2018 for 2-level decomposition (left), 6-level decomposition (middle) and 7-level decomposition (right).
Sensors 19 01407 g018
Figure 19. An example of one-day prediction from 15th May 2018 of the CO2 signal. Filtration is done by using the Db6 wavelet with 6-level decomposition.
Figure 19. An example of one-day prediction from 15th May 2018 of the CO2 signal. Filtration is done by using the Db6 wavelet with 6-level decomposition.
Sensors 19 01407 g019
Figure 20. An example of whole-month prediction from November 2018 of the CO2 signal. Filtration is done by using the Db6 wavelet with 6-level decomposition.
Figure 20. An example of whole-month prediction from November 2018 of the CO2 signal. Filtration is done by using the Db6 wavelet with 6-level decomposition.
Sensors 19 01407 g020
Figure 21. The 6–13 November 2018 prediction of the CO2 signal containing several glitches and signal spikes, marked as green rectangle.
Figure 21. The 6–13 November 2018 prediction of the CO2 signal containing several glitches and signal spikes, marked as green rectangle.
Sensors 19 01407 g021
Figure 22. The 15 November 2018 prediction of the CO2 signal containing several glitches and signal spikes, marked as green rectangle.
Figure 22. The 15 November 2018 prediction of the CO2 signal containing several glitches and signal spikes, marked as green rectangle.
Sensors 19 01407 g022
Figure 23. The 4 November–3 December 2018 prediction of the CO2 signal containing several glitches and signal spikes, marked as green rectangle.
Figure 23. The 4 November–3 December 2018 prediction of the CO2 signal containing several glitches and signal spikes, marked as green rectangle.
Sensors 19 01407 g023
Figure 24. The Mean Squared Error (MSE) evaluation for the CO2 prediction (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Figure 24. The Mean Squared Error (MSE) evaluation for the CO2 prediction (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Sensors 19 01407 g024
Figure 25. The correlation coefficient evaluation for the CO2 prediction (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Figure 25. The correlation coefficient evaluation for the CO2 prediction (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Sensors 19 01407 g025
Figure 26. The Euclidean distance evaluation for CO2 prediction (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Figure 26. The Euclidean distance evaluation for CO2 prediction (a) 6–13 May 2018, (b) 15 May 2018, (c) 1–29 May 2018, (d) 6–13 November 2018, (e) 15 November 2018, (f) 4 November–3 December 2018.
Sensors 19 01407 g026
Table 1. The trained models—Training period: May 2018.
Table 1. The trained models—Training period: May 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11018.60.4050.0120.4320.0123.31 × 10−4
25040.20.6280.0100.6380.0102.42 × 10−4
310061.20.7820.0080.7870.0081.56 × 10−4
415069.60.8280.0070.8310.0071.24 × 10−4
520075.30.8590.0060.8640.0061.02 × 10−4
625076.10.8610.0050.8670.0069.91 × 10−5
730076.60.8640.0050.8700.0069.75 × 10−5
835076.40.8640.0050.8700.0069.75 × 10−5
940074.10.8540.0060.8610.0061.06 × 10−4
Table 2. The trained models—Training period: November 2018.
Table 2. The trained models—Training period: November 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11035.50.6030.0240.6210.0241.862 × 10−3
25063.90.7960.0160.8040.0161.377 × 10−3
310072.00.8410.0130.8520.0131.295 × 10−3
415077.00.8720.0120.8780.0121.414 × 10−3
520079.00.8840.0100.8890.0111.393 × 10−3
625080.60.8940.0100.8980.0101.385 × 10−3
730080.60.8950.0100.8980.0101.385 × 10−3
835080.60.8950.0100.8980.0101.385 × 10−3
940080.60.8930.0110.8980.0111.553 × 10−3
Table 3. The trained models—6th of May 2018 to 13th of May 2018.
Table 3. The trained models—6th of May 2018 to 13th of May 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11075.70.8550.0040.8570.0043.32 × 10−5
25094.60.9690.0010.9690.0017.46 × 10−6
310096.30.9760.0010.9760.0015.40 × 10−6
415097.00.9790.0010.9810.0014.41 × 10−6
520097.60.9820.0010.9840.0013.67 × 10−6
625097.60.9830.0010.9840.0013.64 × 10−6
730097.60.9830.0010.9840.0013.64 × 10−6
835097.60.9830.0010.9840.0013.16 × 10−6
940098.20.9830.0010.9850.0013.14 × 10−6
Table 4. The trained models—6th of November 2018 to 13th of November 2018.
Table 4. The trained models—6th of November 2018 to 13th of November 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11052.00.7240.0160.7140.0175.33 × 10−4
25085.80.9150.0080.9210.0081.64 × 10−4
310091.80.9470.0050.9540.0059.68 × 10−5
415096.10.9690.0030.9730.0035.07 × 10−5
520096.40.9730.0030.9770.0034.61 × 10−5
625096.70.9780.0030.9780.0024.13 × 10−5
730096.70.9780.0030.9780.0024.13 × 10−5
835096.70.9780.0030.9780.0024.13 × 10−5
940096.80.9750.0030.9790.0034.12 × 10−5
Table 5. The trained models—Training period: 12th of May 2018.
Table 5. The trained models—Training period: 12th of May 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11088.90.9270.0640.9220.0679.81 × 10−3
25097.80.9720.0260.9770.0242.63 × 10−3
310098.00.9710.0220.9670.0213.11 × 10−3
415096.70.9400.0370.9340.0386.54 × 10−3
520098.10.9370.0350.9430.0325.43 × 10−3
625095.00.9210.0350.9010.0408.48 × 10−3
730095.70.9410.0310.9320.0306.42 × 10−3
835095.70.9410.0310.9350.0306.42 × 10−3
940095.70.9410.0310.9350.0306.42 × 10−3
Table 6. The trained models—Training period: 15th of May 2018.
Table 6. The trained models—Training period: 15th of May 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11097.80.9870.0060.9870.0057.30 × 10−5
25099.70.9960.0020.9980.0011.29 × 10−5
310099.90.9970.0010.9980.0018.98 × 10−6
415099.80.9960.0020.9980.0011.35 × 10−5
520099.70.9950.0020.9970.0011.68 × 10−5
625099.50.9930.0020.9930.0022.72 × 10−5
730099.50.9930.0020.9930.0022.72 × 10−5
835099.50.9930.0020.9930.0022.72 × 10−5
940099.80.9960.0010.9970.0011.24 × 10−5
Table 7. The trained models—Training period: 15th of November 2018.
Table 7. The trained models—Training period: 15th of November 2018.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)TestingValidation
Linear CorrelationMAELinear CorrelationMAEMSE
11079.00.8950.0060.8900.0071.61 × 10−4
25094.80.9870.0020.9870.0023.07 × 10−5
310095.40.9260.0020.9520.0024.94 × 10−5
415091.90.8610.0030.8970.0031.13 × 10−4
520091.90.8610.0030.8970.0031.13 × 10−4
625091.90.8610.0030.8970.0031.13 × 10−4
730091.90.8610.0030.8970.0031.13 × 10−4
835091.90.8610.0030.8970.0031.13 × 10−4
940099.70.9130.0020.8950.0036.45 × 10−5
Table 8. The average of the obtained results in each experiment.
Table 8. The average of the obtained results in each experiment.
Measurement PeriodAccuracy (%)Validation
Linear CorrelationMAEMSE
May63.10.787.44 × 10−31.51 × 10−4
November72.20.8481.30 × 10−21.45 × 10−3
May 6 to 1394.70.9671.33 × 10−37.52 × 10−6
November 6 to 1389.90.9395.00 × 10−31.17 × 10−4
May 1295.70.9403.12 × 10−26.14 × 10−3
May 1599.50.9951.78 × 10−32.44 × 10−5
November 1592.00.9123.22 × 10−39.67 × 10−5
Table 9. The average of the obtained results in the six performed experiments.
Table 9. The average of the obtained results in the six performed experiments.
Order of Measurement (Model Number)Number of Neurons (-)Accuracy (%)Validation
Linear CorrelationMAEMSE
11064.00.7741.94 × 10−21.83 × 10−3
25087.80.8998.88 × 10−36.38 × 10−4
310082.40.9257.29 × 10−36.75 × 10−4
415089.80.9279.29 × 10−31.18 × 10−3
520091.20.9358.15 × 10−31.02 × 10−3
625091.10.9319.14 × 10−31.45 × 10−3
730091.20.9357.71 × 10−31.15 × 10−3
835090.20.9367.71 × 10−31.15 × 10−3
940092.10.9367.85 × 10−31.17 × 10−3
Table 10. The summarization of the evaluation parameters of the original CO2 prediction.
Table 10. The summarization of the evaluation parameters of the original CO2 prediction.
Date of PredictionMSE [-]Corr [%]ED [-]
6–13 May 20187.577 × 10−697.00.266
15 May 20182.436 × 10−599.50.177
1–29 May 20181.505 × 10−477.92.439
6–13 November 20181.172 × 10−494.21.017
15 November 20189.761 × 10−593.30.366
4 November–3 December 20184.065 × 10−479.83.429
Table 11. The summarization of the evaluation parameters of the wavelet smoothing CO2 prediction.
Table 11. The summarization of the evaluation parameters of the wavelet smoothing CO2 prediction.
Date of PredictionMSE [-]Corr [%]ED [-]
6–13 May 20186.442 × 10−698.10.238
15 May 20181.473 × 10−699.60.157
1–29 May 20181.360 × 10−480.62.292
6–13 November 20181.047 × 10−495.60.925
15 November 20187.061 × 10−595.70.315
4 November–3 December 20183.781 × 10−481.33.281
Table 12. The difference parameters of CO2 prediction.
Table 12. The difference parameters of CO2 prediction.
Date of PredictionDiff MSE [-]Diff Corr [%]Diff ED [-]
6–13 May 20181.134 × 10−60.00490.0277
15 May 20184.886 × 10−68.524 × 10−40.0196
1–29 May 20181.443 × 10−50.0270.147
6–13 November 20181.258 × 10−50.00790.092
15 November 20182.699 × 10−50.0220.051
4 November–3 December 20182.842 × 10−50.0150.147

Share and Cite

MDPI and ACS Style

Vanus, J.; Kubicek, J.; Gorjani, O.M.; Koziorek, J. Using the IBM SPSS SW Tool with Wavelet Transformation for CO2 Prediction within IoT in Smart Home Care. Sensors 2019, 19, 1407. https://doi.org/10.3390/s19061407

AMA Style

Vanus J, Kubicek J, Gorjani OM, Koziorek J. Using the IBM SPSS SW Tool with Wavelet Transformation for CO2 Prediction within IoT in Smart Home Care. Sensors. 2019; 19(6):1407. https://doi.org/10.3390/s19061407

Chicago/Turabian Style

Vanus, Jan, Jan Kubicek, Ojan M. Gorjani, and Jiri Koziorek. 2019. "Using the IBM SPSS SW Tool with Wavelet Transformation for CO2 Prediction within IoT in Smart Home Care" Sensors 19, no. 6: 1407. https://doi.org/10.3390/s19061407

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop