You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

19 January 2017

Significant Change Spotting for Periodic Human Motion Segmentation of Cleaning Tasks Using Wearable Sensors

and
Department of Biomedical Engineering, National Yang-Ming University, 155, Li-Nong Street, Section 2, Peitou, Taipei 11221, Taiwan
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Sensors for Ambient Assisted Living, Ubiquitous and Mobile Health

Abstract

The proportion of the aging population is rapidly increasing around the world, which will cause stress on society and healthcare systems. In recent years, advances in technology have created new opportunities for automatic activities of daily living (ADL) monitoring to improve the quality of life and provide adequate medical service for the elderly. Such automatic ADL monitoring requires reliable ADL information on a fine-grained level, especially for the status of interaction between body gestures and the environment in the real-world. In this work, we propose a significant change spotting mechanism for periodic human motion segmentation during cleaning task performance. A novel approach is proposed based on the search for a significant change of gestures, which can manage critical technical issues in activity recognition, such as continuous data segmentation, individual variance, and category ambiguity. Three typical machine learning classification algorithms are utilized for the identification of the significant change candidate, including a Support Vector Machine (SVM), k-Nearest Neighbors (kNN), and Naive Bayesian (NB) algorithm. Overall, the proposed approach achieves 96.41% in the F1-score by using the SVM classifier. The results show that the proposed approach can fulfill the requirement of fine-grained human motion segmentation for automatic ADL monitoring.

1. Introduction

During the past decades, there has been extraordinary improvements in public health and medicine, allowing people to live longer and healthier. The proportion of the aging population is rapidly increasing around the world [,]. Aging brings many challenges to the elderly due to their cognitive decline, chronic age-related diseases, as well as limitations in physical activity, vision, and hearing. Moreover, multiple chronic diseases can limit daily functioning and hinder independent daily living []. Therefore, the early detection of meaningful change in aging is fundamentally important for timely prevention and treatment. The measurement of daily functionality through the activities of daily living (ADL) monitoring has become very important work in the assessment of functional status of the elderly. Traditionally, approaches for measuring ADL performance and assessment of clinically meaningful change are based on questionnaires, interviews, and in-person visits. However, it often has poor demarcation as to onset and transition to new states. The decline of ADL performance progress is slow over time, making the changeover to a new state difficult to recognize. These approaches are also hard to continuously identify the full range of potential events or trends of interest.
In recent years, advances in microelectromechanical systems (MEMS) and information and communications technology (ICT) have created new opportunities for automatic ADL monitoring to improve the quality of life and provide adequate medical service for the elderly. Several well-known approaches have been applied to the development of automatic ADL monitoring systems, such as ambient-based and wearable-based approaches. Ambient-based ADL monitoring systems refer to the practice that diversity of multimodal sensors, actuators, and devices are installed in the environment, which can recognize and inference individual’s ADLs by the detection of interactions between the individual and the environment. Most of the ambient-based approaches detect the environmental change based on sensor events. For example, using infrared sensors placed on doors to detect room transitions [], or using RFID reader sensors to trace object usage []. Some works install large sensors in the home environment to monitor ADL performance []. However, some applications include, e.g., the monitoring of specific motion in fluid intake [], or the classification of dietary intake gestures [,] for an automated ADL monitoring system, where reliable human motion on a more fine-grained level is needed, especially for hand gesture spotting. Wearable-based ADL monitoring systems are considered one of the most prevalent approaches to capture and track human motion during ADL tasks. The attributes of wearable sensors, such as high computational power, small size, and low cost, allow people to wear them over extended periods of time. Wearable sensors can be embedded into clothes, eyeglasses, belts, shoes, watches, and mobile devices to collect various information about ADLs, such as human motion, vital signs, and context information. However, many technical challenges arise from the fact that there is a considerable amount of variability inherent in individual ADL patterns if there is no manual intervention or prior knowledge.
The continuous monitoring of particular human motion during daily living can provide a measurement of daily functionality for remote health monitoring applications, especially in early detection of a change in aging. Analysis of the specified gestures or motions can provide the clinical professional with specific ADL performance metrics, such as the number of repetitions performed and the duration of each repetition, many of which are hard to monitor by current assessment strategies. Automatic ADL monitoring requires reliable ADL information on a fine-grained level, especially for the assessment of daily functionality in remote health monitoring applications, especially in early detection of a change in aging. To enable continuous monitoring and measurement, the ADL monitoring system has to segment and classify specific gestures and motions from continuous data. The accuracy of the segmentation approach affects the ability and granularity of the monitoring system. Previous work also shows that the performance of the best classifier sharply decreases when the segmented data are non-differentiable []. Therefore, some critical issues in data segmentation techniques need to be dealt with. Two common technical challenges are identified as follows:
(1)
Since the data stream obtained from the sensor is continuous and unsegmented, there is no idea about the numbers of human motions have actually been performed, while ADLs last for a period of time;
(2)
Human motions can be performed slowly or quickly, which causes all motions during ADL performance to be full of variance even for the individual.
In this work, we propose a significant change spotting mechanism for periodic human motion segmentation during cleaning task performance. A novel approach is proposed based on the search for significant change of human motion, which can manage critical technical issues in segmentation techniques, such as continuous data segmentation, individual variance, and category ambiguity. Three different classification algorithms are utilized for the identification of the significant change candidate, including a Support Vector Machine (SVM), k-Nearest Neighbors (kNN), and Naïve Bayesian (NB) algorithm. The proposed approach for cleaning task monitoring can provide an example of the feasibility in daily living based on a recording from a healthy adult. Additionally, the proposed approach is suitable for dealing with the high dimensionality signals of hand motion, and explicitly models both spatial and temporal variability to provide accurate segmentation results. The rest of the paper is organized as follows: we briefly introduce related work in Section 2; in Section 3, the novel significant change spotting approach is proposed to provide robust and adaptive human motion segmentation for cleaning task monitoring, which is based on search and spotting of significant changes during cleaning task performance; the experimental results validate the capabilities of our proposed mechanism are addressed in Section 4; in Section 5, we discuss the limitations of the current work and potentiality of future work; and, finally, the conclusion is presented in Section 6.

3. Proposed Segmentation Approach

In the performance of daily cleaning tasks, an activity scenario is an event or series of actions and gestures, including spontaneous and periodic human actions. The periodic human actions in this work are composed of motions, but with different directions, forces, and time intervals. For example, the participant performed, 10 times, cleaning motions during a window cleaning task, where each cleaning activity can be performed in up, down, forward, and backward directions according to the cleaning tasks. Obviously, the change of direction can be a significant event to gather the human motion segment from continuous sensing data. In order to detect the significant event for accurate segmentation from continuous cleaning data, there are two main considerations of the proposed approach for human motion segmentation. Firstly, rather than segmenting cleaning motion based on detection of a trigger point, the proposed significant change spotting is to observe a significant series from continuous cleaning motion. A significant series can be defined as an interval of series sensing data that indicates the human applied force to change the direction. Secondly, to reduce computation complexity in the presence of high dimensionality data, the segmentation and spotting mechanism should be as computationally efficient as possible. A series of processes for significant change spotting has been employed, such as dominant axis identification, event detection, and feature analysis.
Figure 1 shows the architecture of the proposed system. It consists of four processes: data acquisition, pre-processing, a hierarchical candidate search for significant change, and significant change spotting. In the data acquisition process, the data collects sensing data from acceleration, and the band pass filters are utilized for all sensing data. The choice of activity of daily living, experimental setup, and pre-processing are described in respective subsections. As shown in Figure 2, a hierarchical candidate search for significant change is the two-stage process that searches all possible candidates for significant change spotting based on the observation of a significant series during cleaning motion performance. Finally, feature analysis and machine learning-based spotting are executed for the process of significant change spotting.
Figure 1. The architecture of the proposed approach for the segmentation mechanism, including data acquisition, pre-processing, hierarchical candidate search for significant change, and significant change spotting.
Figure 2. Framework of the hierarchical candidate search for significant change from raw data.

3.1. Data Acquisition and Pre-Processing

3.1.1. Choice of Activity of Daily Living

There are two main types of ADLs, namely basic ADLs (BADLs) and instrumental ADLs (IADLs). The BADLs are activities necessary for self-care, while IADLs are not. Examples of BADLs are feeding, bathing, dressing, grooming, etc. IADLs, such as using telephone, housework, doing laundry, etc. In previous works, ADL monitoring systems have been developed for five common ADLs from the Barthel Index, including feeding, grooming, dressing, mobility, and stair-climbing []. Some works focus on the specific measure of a particular interest in mobility performance []. However, IADL assessments are more suitable than BADL assessments for the early detection of daily functional decline []. This work aims to segment cleaning motions (e.g., wiping, sweeping, and mopping) during cleaning tasks, which is one of the general IADL categories to assess the self-care and independent living capability [].

3.1.2. Experiments Setup and Pre-Processing

We conducted an experiment on six subjects involving four males and two females, aged from 23 to 29 years, who performed four cleaning tasks, included cleaning a table, cleaning a window, sweeping a floor, and mopping a floor, the details of which are described in Table 1 and Figure 3. All cleaning tasks are constrained to an area of 5.5 m × 6.2 m for the purposes of capturing video of the session. The targets of cleaning tasks, such as the window and table are 1.5 m × 1 m and 1.2 m × 0.7 m. Each subject is asked to perform a series of cleaning tasks for five days over two weeks. The cleaning tasks was repeated or sustained for a period and no time constraints, but marked with the start and stop times, and labeled with the name of each task. The overall dataset covers 112.4 min.
Table 1. The description of the cleaning tasks.
Figure 3. (a) Clean Window; (b) Clean Table; (c) Sweep Floor; (d) Mop Floor.
The wearable sensor named Opal (APDM, Inc., Portland, OR, USA) is utilized to record sensing data of hand movements in this work. The dimension and weight of the sensor node are 48.5 × 36.5 × 13.5 mm and weigh 22 g. The wearable sensor includes a triaxial accelerometer and transmits sensing data wirelessly through an RF module. The sampling rate is 40 Hz, the range is ±6 g, and battery life of the sensor allows eight hours of continuous data-logging. The sensor placement usually depends on the target application and motions as different sensors provide different information which may reflect more important information in specific applications than others. The sensor of this work is attached to the right wrist since the execution of cleaning tasks has a close relationship to the performance of the upper limb. Furthermore, wrist-worn sensors have the attribute of low-intrusiveness, making them suitable for continuous collection of motion data in daily living. Then, an access point receives the acceleration wirelessly from the sensor nodes and transmits the sensing data to the terminal through a USB connection, as shown in the Figure 4. A third-order Butterworth filter, having cut-off frequencies of 0.1 Hz and 12 Hz is applied to the raw data in the process of pre-processing, in order to remove the artefacts and noise [,,].
Figure 4. Illustration of sensor placement and the hardware platform.

3.2. Hierarchical Candidate Search for Significant Change

As shown in Figure 2, firstly, the sliding window approach with 50% overlapping is applied to divide the sensor data into chunks, where the window size is four seconds. The setting provided a sufficient buffer to cover at least one complete human motion as the longest duration of complete human motion does not exceed two seconds in this work. Each chunk is a sliding buffer and possibly contains several gestures, which provides a global view of the continuous sensor data. Then dominant axis identification is executed for each of the sliding frames. Finally, a candidate search algorithm for a significant series is proposed to detect the events indicating direction changes of gestures. The detail of the dominant axis selection and candidate search algorithm for a significant series are described in the following subsections.
Formally, let a sequence of raw sensing data R = { ( x i , y i , z i ) | i = 1 , 2 , , N R } be divided into sliding frames F = { ( f i ) | i = 1 , 2 , , N F } with a defined window size. Secondly, dominant axis identification is applied to each sliding frame and obtains a vector of the dominant axis for a significant series search D = { ( d f i k ) | k = 1 , 2 , , N D } . Finally, the process of the candidate search for significant change is applied to any sliding frame and gathers a vector of candidates for significant change spotting S ( f i ) = { ( S f i j ) | j = 1 , 2 , , N F } , f i F .

3.2.1. Dominant Axis Classifier

In order to improve the proposed algorithm’s robustness and computational complexity in the presence of high-dimensionality data, a dominant selection is designed for each sliding window. The force of motion tends to focus on a particular direction during cleaning task performance, thus, it can be assumed that the cleaning motion undergoing the significant change of axis is the dominant one. The axis with the highest difference among all axes is assumed to be dominant in the sliding frame. The dominant axis can be identified by a threshold-based identifier. Given that an existing sliding frame f α belonging to F and with a total of three axes from triaxial acceleration x ,   y , z , the dominant axis can be identified by Equation (1):
argmax x ,   y ,   z   D A I ( x ,   y , z ) = { d i = A v g m a x i A v g m i n i | i = x , y , z }
where A v g m i n i and ( A v g m a x i ) are the average minimum (maximum) value, and defined as the average of the first smallest (largest) ten percent of values in the axis i of the sliding frame. The value of the axis in the sliding window with the maximum is identified to be dominant for the current sliding frame.

3.2.2. Hierarchical Candidate Search for Significant Series

Since hand motion in cleaning tasks is more complicated than ambulatory movements, the improved search algorithm for a significant series is proposed in this work. The pseudocode of the search algorithm is described in Algorithm 1, and the illustration of the proposed candidate search for a significant series is shown in Figure 5. Traditionally, the exemplar motions are scanned firstly for key features or events, such as peaks, zero crossing points (ZCPs), or mean crossing points (MCPs) in velocity or acceleration as a way to approximate a potential segmentation region in the sliding frame. Previous works also show that using velocity features allows the segmentation algorithm to spot the general shape of the interested motion more robustly than approaches that rely on distance measures []. As shown in Figure 6a, the detection of MCPs is adopted in the proposed segmentation approach. Firstly, the threshold is defined as the average of sensing data in the dominant axis (for Line 1). Then detection of the MCP during human motion is determined by events composed of two successive data, which cross the threshold in ascending order (for Lines 2–7). These detected events are assumed as potentially part of the significant series. Instead of using a trigger point for the segmentation, this work focuses on spotting a significant series as the transition region for the human motion segmentation. However, the significant series are not always performed as the monotonic series since motions can be performed in various ways. Much of the noise, such as local peaks and noise spikes, that occurs in the signal might be caused by the vibration of muscle, individual habit, and the variability of movement velocity. In order to tackle the issues to accurately locate the starting and ending points of a complete candidate, the extension function is applied to all events (for Lines 8 and 13).
Figure 5. Tri-axial acceleration of a subject performing floor mopping. (a) The threshold is determined by the dominant axis (acceleration of the X-axis). Each circle denotes the detected event, which contains starting and ending points, and crossing the threshold in ascending order. (b) The covered region by the starting and end points denote the searched candidate for significant change spotting.
Figure 6. From bottom to top, the spotting process includes a candidate search for significant change and significant change spotting. Finally, the spotted candidates are compared against to the ground truth in terms of the classification result. If there are no searched candidates, the output of cleaning gesture must be negative.
As shown in Figure 5a, for any detected event, one of the data points with the smaller value is considered as the initial starting point of candidate, and the other one with the larger value is considered as the initial ending point of candidate. In the following, the searching end point for the candidate is introduce firstly. If the significant point with the maximum value in the current extension is equal to the initial ending point or searched significant point in the previous extension, the initial ending point or searched significant point is identified as the ending point of a complete candidate, and the algorithm stops searching for the ending point of candidate. Suppose d i n i t i a l _ e n d is the initial ending point of detected event before extension, l is the number of times of extension during the ending point search with the initial value equal to 1, α u p p e r is the upper boundary of each time extension, and the ending point d e n d of the candidate can be found in N t o t a l _ e n d times of extension. The extension function E e n d is applied to search for the ending point of the candidate, which can be defined as followed:
d e n d = E e n d ( d i n i t a i l _ e n d , α u p p e r ,   l )   = { d i n i t i a l e n d , x l = d i n i t i a l e n d E e n d ( x l , α u p p e r ,   l + 1 ) , x l d i n i t i a l _ e n d ,
x l = a r g m a x ( d i n i t a i l _ e n d , d i n i t a i l _ e n d + 1 , , d i n i t a i l _ e n d + α u p p e r ) ,
where 1 l N t o t a l _ e n d , d e n d is the ending point of the candidate, and x l is the significant point with the local maximum in the l t h extension, which can be defined in Equation (3).
In the following, the searching starting point for a candidate is introduced. If the significant point with the minimum value in current extension is equal to the initial starting point or searched significant point in the previous extension, the initial ending point or searched significant point is identified as the starting point of the complete candidate, and the algorithm stops searching for the starting point of the candidate. Suppose d i n i t i a l _ s t a r t is the initial starting point of a detected event before extension, p is the number of times of extension during the starting point search with the initial value equal to 1, α l o w e r is the lower boundary of each time extension, and the starting point d s t a r t of the candidate can be found in N t o t a l _ s t a r t times of extension. The extension function E s t a r t is applied to search starting point of the candidate, which can be defined as followed:
d s t a r t = E s t r a t ( d i n i t a i l _ s t a r t , α l o w e r ,   p ) = { d i n i t a i l _ s t a r t , x p = d i n i t i a l _ s t a r t E s t r a t ( x p , α l o w e r , p + 1 ) , x p d i n i t i a l _ s t a r t ,
x p = a r g m i n ( d i n i t a i l _ s t a r t α l o w e r , d i n i t a i l _ s t a r t + 1 , , d i n i t a i l _ s t a r t 1 , d i n i t a i l _ s t a r t )
where 1 l N t o t a l _ s t a r t , d s t a r t is the starting point of the candidate, and x p is the significant point with the local minimum in the p t h extension, which can be defined in Equation (5).

3.3. Significant Change Spotting

In this subsection, the applied features extraction and selected machine learning-based classifier are introduced. A machine learning-based classifier is used to identify whether the candidate is really a significant or non-significant series. Three typical machine learning algorithms are selected as the core technique of a machine learning-based classifier for a performance comparison of the proposed segmentation mechanism, including Naïve Bayes (NB), k-nearest neighbor (kNN), and support vector machine (SVM).

3.3.1. Feature Analysis and Selection

We investigate two feature sets for significant change spotting. The first feature set contains traditionally used statistical features. Since the total number of samples within each significant series candidate is much smaller, complex statistical features, such as skewness, kurtosis, and spectral entropy may not be reliably calculated. Therefore, we only consider statistical features that can be reliably calculated at a primitive level. The statistical features are listed in Table 2.
Table 2. Statistical features calculation and description.
The second set of features are called physical features, which are derived based on the physical parameters of an upper limb functionality assessment [,]. Most of these physical features can provide accurate and objective information about the movement quality and smoothness for human m. Table 3 lists the physical features including in this work.
Table 3. Physical features calculation and description.
Feature selection for a machine learning-based classifier is the critical process during significant change spotting. Selecting the suitable features can reduce the computational effort of the classification process, and improve the classifier performance. Furthermore, the selected features can provide causal relationships between features and classes for researchers. The sequential forward selection approach is utilized to select relevant features during feature selection. The method is employed by repeatedly adding features to the feature vector with the currently best quality.
Algorithm 1: Candidate search algorithm for significant series
Input:A sequence of data D = { d i | i = 1 , 2 , , N D } , a set of events E = { e i | i = N E } ,
the numbers of events N E = 0
Output:A series of candidates for significant series S = { s i | i = N S } , N S is the total
number of candidates for significant series
1: θ t h r e s ← calculateMean( X ) /* Threshold definition for event detection
2:for i from 1 to N D do  /* Detect the event for the set of candidates
3: if d i   θ t h r e s 0   and   d i + 1   θ t h r e s 0   t h e n
4:   N E = N E + 1
5:   e N E ( d i , d i + 1 )
   E e N E
6: end if
7:end for
8:for i = 1   to   N E   d o    /* Candidate search for the significant series
9: ( d j , d k ) getCurrentSet ( e i )
10: d c m a x extendLocalMaximum ( d j )
11: d c m i n extendLocalMaximum ( d k )
12: S getCandidate ( d c m i n , d c m a x )
13:end for
14:return S

3.3.2. Spotting and Classification

There are various classification approaches that have been developed from simple threshold-based, to more advanced, machine learning algorithms. Considering the computational complexity, classification performance, target activity, and type of application, the selection of appropriate machine learning-based approaches is a critical component [,]. While this work focuses on the segmentation process belonging to the early process of activity monitoring system, some powerful machine learning-classifiers with high computational complexity are not included, such as neural networks. Therefore, the following three typical types of machine learning-based classifiers are utilized and introduced.
Naive Bayes (NB) classifier: NB is a statistical approach to estimate the possibility of a given instance. The Naïve Bayes probabilistic model makes assumptions that instances have possibilities relating to an independent class. The NB classifier selects a class label that has the greatest probability based on the observed data, using Bayes rule and assuming conditional independence. This work uses a Bayesian classifier where Gaussian distributions were used to model the priors of the classes.
K-nearest neighbor (kNN): The kNN algorithm is one of the most popular algorithms for machine learning. It is a supervised learning algorithm where the result of new instance query is classified based on the majority of a kNN category. The kNN algorithm is amongst the most basic of all machine learning algorithms: a target is assigned by a majority vote of its k-nearest neighbors (k is a positive integer, typically small). The kNN classifier uses the kNN algorithm as the prediction value of the new query instance. The classifier does not use any model to fit and is only based on memory. Many researchers have proved that the kNN classifier accomplishes results with good performance in different applications. The kNN classifier utilized in this work uses a Euclidean formula to calculate the distance between two points in multidimensional space, and classifies test data with k = 5 .
Support vector machine (SVM): SVM is one of the standard tools for machine learning and data mining. SVM aims to define the hyperplane that maximizes the distance between the two parallel hyperplanes. Intuitively, the larger margin means a lower generalization error of the classifier. SVM identifies a function that maximizes the distance to data points in both classes. SVM classifiers are applied with a RBF (radial basis function) kernel function in this work.

4. Results and Data Analysis

4.1. Evaluation Methodology

The leave-one-out cross-validation (LOOCV) is adopted in this work. LOOCV uses one segment as the testing set and the remaining set as the training set. There are a total of k segments created according to the total days performed by the subject. For example, there are five segments while he subject performed five cleaning tasks on five different days. One segment is used as the testing dataset while the rest of the ( k   1 ) segments are used as the training dataset, and repeated k times to evaluate the proposed segmentation approach.
As shown in the Figure 6, several tags are designed to validate the results of the proposed approach. While spotting is the main target, there are several metrics to measure performance of the proposed significant change spotting for motion segmentation. The recognition performance is measured in precision, recall, and F1-score commonly used for information retrieval assessments. These evaluation metrics are defined in Equations (6)–(8). Precision indicates the percentage of times that a recognition result made by the system is correct. Recall means the percentage of times that an activity performed by a user is detected by the system. F1-score is an integrated measure that combines both.
Recall = TP TP + FN
Precision = TP TP + FP
F 1 - score = 2   ·   Recall   ·   Precision Recall + Precision

4.2. Search and Spotting Results

Figure 7 shows the overall performance metrics of the kNN, SVM, and NB classifiers, however, which is not easy to express and expound the proposed significant change spotting approach of human motion segmentation for cleaning task monitoring. Therefore, each performance metric is split into nine areas in row major order, and each area can be defined as the average performance of 25 pairwise combinations of different upper and lower boundaries ( α upper , α lower ). For example, area A4 contains the pairwise combinations of the upper boundary from 5 to 10, the lower boundary from 1 to 5, and a total of 25 pairwise combinations.
Figure 7. Overall performance comparison of different classifiers. (a) An example of performance divided into nine areas. Each area can be defined as the average performance of 25 pairwise combinations of different upper and lower boundaries. (b) Pairwise combinations of different upper and lower boundaries (samples) are shown in each performance metric.
Table 4 shows nine areas of performance comparisons of different classifiers. The average performance of the three utilized classifiers with the highest recall (99.27 ± 0.25) is located in the A1 area, the highest precision (93.1 ± 0.76) is located in A9, and the highest F1-score (95.26 ± 0.25) is located in A9.
Table 4. The overall performance comparison of different classifiers for four cleaning tasks.
Since wearable sensors are sensitive to motion performance, this work explores whether it is feasible to build a general segmentation model for four different cleaning tasks in a laboratory. Figure 8 shows the proposed significant change spotting approach for four cleaning tasks in recall, precision, and F1-score, respectively. Obviously, the proposed approach for each cleaning tasks has its suitable areas of upper and lower boundary. Firstly, for the performance of floor sweeping, the areas with the same upper boundary show a downward trend in recall as the lower boundary increases, but the performance in precision is the inverse. A similar situation also happened in the floor mopping task, where the areas with the same lower boundary show a downward trend in recall as the upper boundary increases, but the performance in precision is the inverse. Secondly, for the performance of cleaning the table, the difference between areas with the highest recall (99.03% in A9) and with the lowest recall (98.49% in A7) is only 0.54%, but the difference between areas with the highest precision (90.19% in A8) and with the lowest precision (84.69% in A3) is 5.51%. Thirdly, the difference between areas with the highest recall (99.03% in A9) and with the lowest recall (98.49% in A7) in four cleaning tasks are not above 2.26%. Therefore, the performance of the proposed segmentation in F1-score really depends on the performance in precision.
Figure 8. Average performance of three classifiers for four cleaning tasks in (a) recall; (b) precision; and (c) F1-score.
Table 5 shows the pairwise combination with the best performance in recall, precision, and F1-score for four cleaning tasks. Firstly, for the pairwise combination with the best performance in recall, the pairwise combination with low α lower has the best performance. It also shows that most of the pairwise comparisons for the four cleaning tasks have the best performance with high α upper and α lower precision and recall. As shown, the NB classifier has the best performance in recall for the four cleaning tasks. However, the performance of each cleaning task in precision and F1-score has its appropriate classifier. Overall, the best performance in the F1-score of the pairwise combinations is (9, 19), which achieves 96.41% in the F1-score by using the SVM classifier.
Table 5. The best performance of the proposed segmentation approach for four cleaning tasks.
Table 6 shows the selected features of each dominant axis in descending order. The top four selected features in all axes are the same, but with different order. RMS and Movement Time are only selected once, and Maximum is selected when the dominant axis is the X or Y axis. The features STD and Mean, Minimum, and Maximum are selected amongst the three axes.
Table 6. Selected features of each dominant axis in ranked order for the significant change spotting.

5. Discussion

Overall, the proposed segmentation approach using kNN and SVM have better performance than using NB. Three metric performance differences between the two classifiers are only 0.03% to 0.12%. However, kNN requires storage and access to all of the training data. Thus, kNN is less scalable for a large dataset and less practical for wearable sensors given their limited storage. Therefore, SVM is more suitable for implementation on the wearable sensors.
Based on our results, the most challenging cleaning tasks for the proposed segmentation approach is “mop floor”. Since the mopping motions are composed of fragmental parts, the significant changes during mop floor is invisible to the classifiers. This outcome varies in intensity amongst individuals, which leads the classifiers to have poor performance in recall across four cleaning tasks. However, the F1-score is still adequate across all participants, and even using the NB classifier achieves an average of 91.27% in the A1 area, at least.
As shown in Figure 8, the proposed segmentation with the higher α upper and α lower shows the non-obvious downward trend in recall. The significant improvement in precision is shown as α upper and α lower increase, which leads the performance in precision as the main effect on the F1-score. One reason for these phenomena is that the effect of muscle vibration causing significant changes are not monotonous. Therefore, the low α upper and α lower of the extension function is unable to find the true significant changes in the significant series search. Only parts of the significant change are found, which easily confuses the classifier and finally leads to misclassification. Therefore, the ability of the proposed approach to include true significant changes is considered as the critical factor for the results, even with the powerful classifier.
As shown in the Table 6, there is a strong commonality in the top four selected features, including Maximum, Minimum, Mean, and STD, when the dominant axes are the X and Y axes. Such strong commonality reflects and fulfills the attribute of the proposed significant change approach. The different ranked order also reveals the different consideration of each axis. Specifically, the features peak number and jerk are not selected in this work, which have similar results with previous studies []. This is because target activities are performed smoothly by the healthy subject. Further study should be carried out to deal with different situations, especially for a subject with a functional disability of an upper limb.
We observed a large amount of individual variability in cleaning styles. Various forms of the gesture of holding the utensil during cleaning tasks were observed, even though the subjects were asked to perform on the same target. Additionally, some subjects perform cleaning tasks continuously without any temporary stop, while some others take a temporary stop during the cleaning tasks to check the state of the cleaned target. These could be attributed to an individual’s own cleaning style.
In this study we focus on building an effective segmentation approach for each observed cleaning style; all cleaning gestures are given different labels. However, more cleaning tasks and other ADLs are not considered in this work, such as sink cleaning in the kitchen and bathroom, and toilet cleaning. Without any question, this posed an additional challenge to the classification task. In future work, ambient sensors are planned to be employed in order to reflect greater interaction information between body motion and context, such as room transition and the use of utensils, which might contribute significantly to the motion segmentation of cleaning tasks.
Despite the importance of high recall, precision, and F1-score, the practical problem that still needs to be addressed depends largely on the application domain. In contrast to physical activity monitoring used for energy consumption estimation, gesture performance measurement during ADLs are particular challenging. This is because the requirements of activity information are different. In order to detect early changes in aging, which reflects a decline in the ability of independently perform ADLs, the requirement of continuously and fine-grained activities information collection is essential. The experimental results show that the proposed approach can tackle many technical challenges, such as continuous signal segmentation, individual variability, and activity ambiguity. The proposed significant spotting-based segmentation approach can play an important role in activity monitoring, and provide an efficient approach for continuous cleaning task monitoring.

6. Conclusions

In order to develop an automatic ADL monitor for the assessment of daily functionality in remote health monitoring applications, especially in early detection of a change in aging, the automatic ADL monitoring requires reliable ADL information on a fine-grained level. In this work, we proposed a significant change spotting mechanism for periodic human motion segmentation during performance of cleaning tasks. A novel approach is proposed based on the search for significant change of gestures, which can deal with critical technical issues, such as continuous data segmentation, individual variance, and category ambiguity. Three different classification algorithms are utilized to validate the significant change candidate through machine learning-based spotting, including a Support Vector Machine (SVM), k-Nearest Neighbors (kNN), and Naïve Bayesian (NB) algorithm. The experimental results have demonstrated that the proposed segmentation approach achieves a high F1-score. The proposed approach can be a part of automatic ADL monitoring activity recognition for the fine-grained information of cleaning tasks, such as the number of cleaning gestures performed, and the duration of each cleaning task. Such fine-grained information of cleaning tasks has the potential to assist caregivers and clinicians to identify a decline in the ability of independently performing ADLs in a quantitative and objective way. In future work, cleaning task monitoring based on the proposed segmentation approach will be presented and designed in the real world. The study will collect the data from the elderly to assess the ability of the proposed approach. Furthermore, a discussion about, and a comparison of, filters, machine learning algorithms, and segmentation approaches are planned to be investigated.

Acknowledgments

This work was supported in part by grants from the Ministry of Science and Technology (MOST 105-2221-E-010-008).

Author Contributions

Kai-Chun Liu and Chia Tai Chan contributed equally in this work. Both of them conceived the idea and design the methodology. Kai-Chun Liu conducts the experiments and evaluated experimental results. Chia Tai Chan supervised the overall work and provided suggestions in each section of this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nations University. World Population Ageing 2015; Nations University: New Orleans, LA, USA, 2015. [Google Scholar]
  2. Rashidi, P.; Mihailidis, A. A survey on ambient-assisted living tools for older adults. IEEE J. Biomed. Health Inform. 2013, 17, 579–590. [Google Scholar] [CrossRef] [PubMed]
  3. Pol, M.; van Nes, F.; van Hartingsveldt, M.; Buurman, B.; de Rooij, S.; Kröse, B. Older people’s perspectives regarding the use of sensor monitoring in their home. Gerontologist 2016, 5, 131–143. [Google Scholar] [CrossRef] [PubMed]
  4. Wren, C.R.; Tapia, E.M. Toward scalable activity recognition for sensor networks. In Location-and Context-Awareness; Springer: Berlin, Germany, 2006; pp. 168–185. [Google Scholar]
  5. Patterson, D.J.; Fox, D.; Kautz, H.; Philipose, M. Fine-Grained Activity Recognition by Aggregating Abstract Object Usage. In Proceedings of the Ninth IEEE International Symposium on Wearable Computers, Osaka, Japan, 18–21 October 2005.
  6. Nef, T.; Urwyler, P.; Büchler, M.; Tarnanas, I.; Stucki, R.; Cazzoli, D.; Müri, R.; Mosimann, U. Evaluation of three state-of-the-art classifiers for recognition of activities of daily living from smart home ambient data. Sensors 2015, 15, 11725–11740. [Google Scholar] [CrossRef] [PubMed]
  7. Amft, O.; Bannach, D.; Pirkl, G.; Kreil, M.; Lukowicz, P. Towards Wearable Sensing-Based Assessment of Fluid Intake. In Proceedings of the 8th IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops), Mannheim, Germany, 29 March–2 April 2010; pp. 298–303.
  8. Junker, H.; Amft, O.; Lukowicz, P.; Tröster, G. Gesture spotting with body-worn inertial sensors to detect user activities. Pattern Recognit. 2008, 41, 2010–2024. [Google Scholar] [CrossRef]
  9. Krishnan, N.C.; Juillard, C.; Colbry, D.; Panchanathan, S. Recognition of hand movements using wearable accelerometers. J. Ambient Intell. Smart Environ. 2009, 1, 143–155. [Google Scholar]
  10. Pietka, E. Expert systems in parameter extraction of the ECG signal. In proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 4–7 November 1988.
  11. Shoaib, M.; Bosch, S.; Incel, O.D.; Scholten, H.; Havinga, P.J. Complex human activity recognition using smartphone and wrist-worn motion sensors. Sensors 2016, 16, 426. [Google Scholar] [CrossRef] [PubMed]
  12. Fleury, A.; Vacher, M.; Noury, N. Svm-based multimodal classification of activities of daily living in health smart homes: Sensors, algorithms, and first experimental results. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 274–283. [Google Scholar] [CrossRef] [PubMed]
  13. Chernbumroong, S.; Cang, S.; Atkins, A.; Yu, H. Elderly activities recognition and classification for applications in assisted living. Expert Syst. Appl. 2013, 40, 1662–1674. [Google Scholar] [CrossRef]
  14. Mathie, M.; Coster, A.; Lovell, N.; Celler, B. Detection of daily physical activities using a triaxial accelerometer. Med. Biol. Eng. Comput. 2003, 41, 296–301. [Google Scholar] [CrossRef] [PubMed]
  15. Bidargaddi, N.; Sarela, A.; Klingbeil, L.; Karunanithi, M. Detecting walking activity in cardiac rehabilitation by using accelerometer. In Proceedings of the 3rd International Conference on Intelligent Sensors, Sensor Networks and Information, Cairns, Australia, 3–6 December 2007.
  16. Foerster, F.; Smeja, M.; Fahrenberg, J. Detection of posture and motion by accelerometry: A validation study in ambulatory monitoring. Comput. Hum. Behav. 1999, 15, 571–583. [Google Scholar] [CrossRef]
  17. Karantonis, D.M.; Narayanan, M.R.; Mathie, M.; Lovell, N.H.; Celler, B.G. Implementation of a real-time human movement classifier using a triaxial accelerometer for ambulatory monitoring. IEEE Trans. Inf. Technol. Biomed. 2006, 10, 156–167. [Google Scholar] [CrossRef] [PubMed]
  18. Preece, S.J.; Goulermas, J.Y.; Kenney, L.P.; Howard, D.; Meijer, K.; Crompton, R. Activity identification using body-mounted sensors—A review of classification techniques. Physiol. Meas. 2009, 30, R1–R33. [Google Scholar] [CrossRef] [PubMed]
  19. Preece, S.J.; Goulermas, J.Y.; Kenney, L.P.; Howard, D. A comparison of feature extraction methods for the classification of dynamic activities from accelerometer data. IEEE Trans. Biomed. Eng. 2009, 56, 871–879. [Google Scholar] [CrossRef] [PubMed]
  20. Brezmes, T.; Gorricho, J.-L.; Cotrina, J. Activity recognition from accelerometer data on a mobile phone. In Distributed Computing, Artificial Intelligence, Bioinformatics, Soft Computing, and Ambient Assisted Living; Springer: Berlin, Germany, 2009; pp. 796–799. [Google Scholar]
  21. Tapia, E.M.; Intille, S.S.; Haskell, W.; Larson, K.; Wright, J.; King, A.; Friedman, R. Real-time recognition of physical activities and their intensities using wireless accelerometers and a heart rate monitor. In Proceedings of the 11th IEEE International Symposium on Wearable Computers, Boston, MA, USA, 11–13 October 2007.
  22. Nam, Y.; Park, J.W. Child activity recognition based on cooperative fusion model of a triaxial accelerometer and a barometric pressure sensor. IEEE J. Biomed. Health Inform. 2013, 17, 420–426. [Google Scholar] [PubMed]
  23. San-Segundo, R.; Lorenzo-Trueba, J.; Martínez-González, B.; Pardo, J. Segmenting human activities based on hmms using smartphone inertial sensors. Pervasive Mob. Comput. 2016, 30, 84–96. [Google Scholar] [CrossRef]
  24. Stiefmeier, T.; Roggen, D.; Tröster, G. Gestures are strings: Efficient online gesture spotting and classification using string matching. In Proceedings of the ICST 2nd international conference on Body area networks, Florence, Italy, 11–13 June 2007.
  25. Keogh, E.; Chu, S.; Hart, D.; Pazzani, M. Segmenting time series: A survey and novel approach. Data Min. Time Ser. Databases 2004, 57, 1–22. [Google Scholar]
  26. Banos, O.; Galvez, J.-M.; Damas, M.; Pomares, H.; Rojas, I. Window size impact in human activity recognition. Sensors 2014, 14, 6474–6499. [Google Scholar] [CrossRef] [PubMed]
  27. Amft, O.; Tröster, G. Recognition of dietary activity events using on-body sensors. Artif. Intell. Med. 2008, 42, 121–136. [Google Scholar] [CrossRef] [PubMed]
  28. Sekine, M.; Tamura, T.; Akay, M.; Fujimoto, T.; Togawa, T.; Fukui, Y. Discrimination of walking patterns using wavelet-based fractal analysis. IEEE Trans. Neural Syst. Rehabil. Eng. 2002, 10, 188–196. [Google Scholar] [CrossRef] [PubMed]
  29. Kiani, K.; Snijders, C.; Gelsema, E. Computerized analysis of daily life motor activity for ambulatory monitoring. Technol. Health Care 1997, 5, 307–318. [Google Scholar] [PubMed]
  30. Koenig, N.; Mataric, M.J. Behavior-based segmentation of demonstrated tasks. In Proceedings of the International Conference on Development and Learning, Kerkrade, The Netherlands, 5–7 July 2006.
  31. Kohlmorgen, J.; Lemm, S. A dynamic hmm for on-line segmentation of sequential data. In Proceedings of the Advances in Neural Information Processing Systems, Las Vegas, NV, USA, 25–28 June 2001.
  32. Kulic, D.; Takano, W.; Nakamura, Y. Online segmentation and clustering from continuous observation of whole body motions. IEEE Trans. Robot. 2009, 25, 1158–1166. [Google Scholar] [CrossRef]
  33. Chiappa, S.; Peters, J.R. In Movement extraction by detecting dynamics switches and repetitions. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, USA, 6–9 December 2010.
  34. Reyes-Ortiz, J.-L.; Oneto, L.; Samà, A.; Parra, X.; Anguita, D. Transition-aware human activity recognition using smartphones. Neurocomputing 2016, 171, 754–767. [Google Scholar] [CrossRef]
  35. Field, M.; Stirling, D.; Pan, Z.; Ros, M.; Naghdy, F. Recognizing human motions through mixture modeling of inertial data. Pattern Recognit. 2015, 48, 2394–2406. [Google Scholar] [CrossRef]
  36. Mathie, M. Monitoring and Interpreting Human Movement Patterns Using a Triaxial Accelerometer; The University of New South Wales: Sydney, Australia, 2003. [Google Scholar]
  37. Selles, R.W.; Formanoy, M.A.; Bussmann, J.B.; Janssens, P.J.; Stam, H.J. Automated estimation of initial and terminal contact timing using accelerometers; development and validation in transtibial amputees and controls. IEEE Trans. Neural Syst. Rehabil. Eng. 2005, 13, 81–88. [Google Scholar] [CrossRef] [PubMed]
  38. Lin, J.F.-S.; Kulić, D. Online segmentation of human motion for automated rehabilitation exercise analysis. IEEE Trans. Neural Syst. Rehabil. Eng. 2014, 22, 168–180. [Google Scholar] [CrossRef] [PubMed]
  39. Fod, A.; Matarić, M.J.; Jenkins, O.C. Automated derivation of primitives for movement classification. Auton. Robots 2002, 12, 39–54. [Google Scholar] [CrossRef]
  40. Mahoney, F.I. Functional evaluation: The barthel index. Md. State Med. J. 1965, 14, 61–65. [Google Scholar] [PubMed]
  41. Yang, S.; Li, Q. Inertial sensor-based methods in walking speed estimation: A systematic review. Sensors 2012, 12, 6102–6116. [Google Scholar] [CrossRef] [PubMed]
  42. Stucki, G.; Stucki, S.; Brühlmann, P.; Michel, B.A. Ceiling effects of the health assessment questionnaire and its modified version in some ambulatory rheumatoid arthritis patients. Ann. Rheum. Dis. 1995, 54, 461–465. [Google Scholar] [CrossRef] [PubMed]
  43. Lawton, M.P.; Brody, E.M. Assessment of older people: Self-maintaining and instrumental activities of daily living. The Gerontologist 1969, 9, 179–186. [Google Scholar] [CrossRef] [PubMed]
  44. Banos, O.; Damas, M.; Pomares, H.; Prieto, A.; Rojas, I. Daily living activity recognition based on statistical feature quality group selection. Expert Syst. Appl. 2012, 39, 8013–8021. [Google Scholar] [CrossRef]
  45. Lemmens, R.J.; Janssen-Potten, Y.J.; Timmermans, A.A.; Smeets, R.J.; Seelen, H.A. Recognizing complex upper extremity activities using body worn sensors. PLoS ONE 2015, 10, e0118642. [Google Scholar] [CrossRef] [PubMed]
  46. Biswas, D.; Cranny, A.; Gupta, N.; Maharatna, K.; Achner, J.; Klemke, J.; Jöbges, M.; Ortmann, S. Recognizing upper limb movements with wrist worn inertial sensors using k-means clustering classification. Hum. Mov. Sci. 2015, 40, 59–76. [Google Scholar] [CrossRef] [PubMed]
  47. De los Reyes-Guzmán, A.; Dimbwadyo-Terrer, I.; Trincado-Alonso, F.; Monasterio-Huelin, F.; Torricelli, D.; Gil-Agudo, A. Quantitative assessment based on kinematic measures of functional impairments during upper extremity movements: A review. Clin. Biomech. 2014, 29, 719–727. [Google Scholar] [CrossRef] [PubMed]
  48. Cook, D.J.; Schmitter-Edgecombe, M.; Dawadi, P. Analyzing activity behavior and movement in a naturalistic environment using smart home techniques. IEEE J. Biomed. Health Inform. 2015, 19, 1882–1892. [Google Scholar] [CrossRef] [PubMed]
  49. Bulling, A.; Blanke, U.; Schiele, B. A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. (CSUR) 2014, 46, 33. [Google Scholar] [CrossRef]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.