Next Article in Journal
A Biomimetic Miniaturized Microphone Array for Sound Direction Finding Applications Based on a Phase-Enhanced Electrical Coupling Network
Previous Article in Journal
Noninvasive Assessment of Aortic Pulse Wave Velocity by the Brachial Occlusion-Cuff Technique: Comparative Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Selective Ensemble Based on Extreme Learning Machine for Sensor-Based Human Activity Recognition

1
School of Artificial Intelligence, Hebei University of Technology, Tianjin 300130, China
2
School of Engineering, Merz Court, Newcastle University, Newcastle upon Tyne NE1 7RU, UK
3
National Research Center for Rehabilitation Technical Aids, Beijing Key Laboratory of Rehabilitation Technical Aids for Old-Age Disability, Key Laboratory of Human Motion Analysis and Rehabilitation Technology of the Ministry of Civil Affairs, Beijing 100176, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(16), 3468; https://doi.org/10.3390/s19163468
Submission received: 5 June 2019 / Revised: 27 July 2019 / Accepted: 6 August 2019 / Published: 8 August 2019
(This article belongs to the Section Physical Sensors)

Abstract

:
Sensor-based human activity recognition (HAR) has attracted interest both in academic and applied fields, and can be utilized in health-related areas, fitness, sports training, etc. With a view to improving the performance of sensor-based HAR and optimizing the generalizability and diversity of the base classifier of the ensemble system, a novel HAR approach (pairwise diversity measure and glowworm swarm optimization-based selective ensemble learning, DMGSOSEN) that utilizes ensemble learning with differentiated extreme learning machines (ELMs) is proposed in this paper. Firstly, the bootstrap sampling method is utilized to independently train multiple base ELMs which make up the initial base classifier pool. Secondly, the initial pool is pre-pruned by calculating the pairwise diversity measure of each base ELM, which can eliminate similar base ELMs and enhance the performance of HAR system by balancing diversity and accuracy. Then, glowworm swarm optimization (GSO) is utilized to search for the optimal sub-ensemble from the base ELMs after pre-pruning. Finally, majority voting is utilized to combine the results of the selected base ELMs. For the evaluation of our proposed method, we collected a dataset from different locations on the body, including chest, waist, left wrist, left ankle and right arm. The experimental results show that, compared with traditional ensemble algorithms such as Bagging, Adaboost, and other state-of-the-art pruning algorithms, the proposed approach is able to achieve better performance (96.7% accuracy and F1 from wrist) with fewer base classifiers.

1. Introduction

In recent years, many works [1,2] have shown that human activity recognition (HAR) has enabled various applications. For instance, daily activities may provide information for health conditions of human beings, and some diseases, such as cerebral small vessel disease [3] and stroke [4], have been proved to be related to the mobility of the human body. Therefore, HAR has been utilized to detect some diseases. In addition, the HAR system can obtain the users’ daily energy expenditure, which can be utilized as a reference for their exercise advice. Moreover, sports training such as swimming [5] and badminton [6] also benefits from HAR. According to the types of data acquisition devices employed, HAR can be divided into vision-based and sensor-based approaches. Vision-based approaches recognize different activities by using video or image sequences. Although vision-based approaches have experienced great breakthroughs in recent years, they still suffer from some drawbacks, including privacy, pervasiveness and complexity [7]. With the development of microelectronics, sensor-based approaches that make use of sensor readings from accelerometers, gyroscopes and magnetomers have attracted more attention around the world. These three kinds of sensor have been utilized in a lot of studies [1,2,5,6], demonstrating their advantages and superior performance in HAR.
A lot of machine learning algorithms have been explored for sensor-based activity recognition. In [8], a neural network was utilized for recognizing three states of activity, including static, transition and dynamic state and 15 kinds of activities. The neural network was applied to recognize eight different activities of construction workers, and showed the best recognition accuracy when compared with five other machine learning algorithms [9]. KNN was also utilized to recognize everyday activities in [10], and a 99.01% overall accuracy was reported in their experiments. In [11], a decision tree (DT) classifier was applied to the detection of activity intensity in youth with cerebral palsy. The computationally efficient support vector machine (SVM) classifier has also been applied in HAR. Wu et al. [12] utilized KNN and SVM as classifiers to demonstrate the proposed features and feature selection method in HAR. By using coordinate transformation and principal component analysis, an online-independent support vector machine (OISVM) [13] has showed that it is effective in improving the robustness of HAR system. Since experimental conditions such as the datasets and extracted features are different, it is difficult to compare the performances of the above classifiers.
The recently proposed extreme learning machine (ELM) [14] is an effective efficient learning algorithm based on single-layer feedforward network (SLFN). It has many advantages, including a simple structure, faster learning rate, and better generalization ability. Therefore, ELM has been widely used in HAR in recent years. In [15], ELM was applied to realize location-adaptive activity recognition; due to the advantages of ELM, experiments showed that the proposed model could adapt the classifier to new device locations quickly. Xiao et al. [16] proposed kernel Fisher discriminant analysis (KDA)-based ELM classifier to recognize six kinds of activity, the experiments showed that it could achieve higher accuracy and faster learning speed than the BP and SVM. An ELM ensemble learning algorithm called average combining extreme learning machine (ACELM) was proposed by [17] to construct a more stable classifier. Moreover, several different variants of ELM have also been proposed and applied in problems of HAR, such as the imbalanced datasets problem [18,19], class incremental learning [20], and cross-person activity recognition [21,22]. However, due to its simple structure and the randomly generated hidden layer parameters, including input weights and hidden layer bias values, a single ELM classifier usually produces unstable outputs, especially when the testing data and the training data are very different in distribution [23].
Ensemble learning has primarily been considered for improving the generalization performance and recognition accuracy of a single classifier. The ensemble learning algorithm was first proposed by Hansen et al. [24]. Their research shows that the ensemble of multiple neural networks can improve the generalization performance of neural networks. Currently, Bagging and Boosting are the two most popular ensemble algorithms. Despite the significant progress of ensemble learning, the accuracy improvement is not proportional to the number of base classifiers. Furthermore, an ensemble learning algorithm that produces too many base classifiers may lead to large computational complexity and low efficiency. Selective ensemble, which is also known as ensemble pruning, is an approach for addressing these issues. In general, the set of base classifiers determined by ensemble pruning tries to meet the performance criterion of maximizing the recognition accuracy and minimizing computation time. If a classifier pool contains M base classifiers, 2M-1 nonempty base classifier subsets can be generated. This makes selecting a subset of classifier with the optimal performance to be an NP-complete problem [25].
To improve the performance of the system, many ensemble pruning approaches have been proposed, and these methods can be categorized into three main groups: ordering-based, optimization-based, and clustering-based pruning approaches [26]. Ordering-based pruning is the most widely used algorithm. For example, two selective techniques for multiple neural networks: forward selection and backward elimination were proposed by Ahmad and Zhang [27,28] to improve model generalization. Li et al. [29] proposed a maximum relevance and minimum redundancy-based ensemble pruning (MRMREP) method for ensemble learning-based facial expression recognition. The proposed method utilized two important factors (the correlation between target labels and predictions, the redundancy between classifiers) to order all base classifiers. Through the experiment, the proposed MRMREP can achieve superior results compared with other ensemble pruning methods. Cao et al. [30] designed a multi-sensor fusion with ensemble pruning system (MSF-EP) for activity recognition and presented four ordering-based ensemble pruning methods to optimize the multi-sensor deployment. A novel ordering-based metric named the margin and diversity-based measure (MDM) was proposed by [31] to explicitly evaluate the importance of base classifiers. Comparative experiments with the other state-of-the-art ensemble pruning methods proved the effectiveness of the algorithm.
Optimization-based pruning has also attracted tremendous attention from scholars. Zhou [32] proposed a genetic algorithm-based selective ENsemble (GASEN) approach that utilized the genetic algorithm (GA) to evolve the weights of base neural networks. According to the evolved weights of base neural networks, it selects some neural networks with higher prediction accuracy and a large diversity between each other and to make up the ensemble. The experiments showed that it has stronger generalization ability compared with some popular ensemble approaches such as Bagging and Boosting. Zhu et al. [33] proposed an optimization-based pruning method based on improved discrete artificial fish swarm algorithm (IDAFSA), which utilized an artificial fish swarm algorithm as an optimization strategy to find the optimal classifier subset instead of the GA. Experimental studies on 29 datasets from the UCI provide the effectiveness of the algorithm. In [34], a bee algorithm (BA) was utilized to select the optimal ensemble subset from a pool of different base classifiers including support vector machine, k-nearest neighbor and linear discriminant analysis classifiers. The proposed method can achieve 83% of accuracy, 93% of specificity and 60% of sensitivity in the mammogram.
The clustering-based pruning approaches are derived from clustering techniques. This method mainly includes two steps: Firstly, the base classifiers in the ensemble are divided into different clusters. The classifiers from the same cluster have similar classification results, while the classifiers from different clusters perform in a more diverse manner. Nowadays, several clustering techniques are utilized in ensemble pruning, including k-means [35], hierarchical agglomerative clustering [36], and deterministic annealing [37]. Finally, in order to increase the diversity of the ensemble, we obtain the base classifier in different clusters. For example, Bakker et al. [38] utilized the classifiers at the centroid of each cluster to constitute the final ensemble.
Although there are many HAR studies based on ensemble learning technology [39,40,41,42,43,44], to our best knowledge, there is still no work attempting to improve the performance of HAR through a selective ensemble approach. Most of the ensemble learning-based HAR studies [17,30,39] combined all the trained base classifiers for recognition. However, some base classifiers may be redundant and have poor performance, which may affect the performance of the recognition system. Therefore, a selective ensemble-based approach may be a good choice for improving the performance of ensemble-based HAR. As a traditional kind of ordering-based pruning method, pairwise diversity can be utilized to measure the diversity among base classifiers and shows good performance in many research works when utilized as a strategy for pre-pruning base classifiers [45,46]. Additionally, glowworm swarm optimization (GSO) is a biomimetic optimization algorithm [47] that has advantages of fast convergence speed and good global convergence. It has been utilized in multiple-objective environmental economic dispatch [48], sensor deployment [49], and vehicle routing problems [50]. Compared with GSO, other heuristic algorithms, such as the genetic algorithm, can also successfully solve the ensemble pruning problem. However, when the number of base classifiers increases, other heuristic algorithms will encounters problems when solving the ensemble pruning problem, including poor solution quality, large time consumption, and low convergence. Based on these considerations, this paper proposes a novel selective ensemble method, DMGSOSEN, which combines pairwise diversity and the GSO algorithm for HAR. Firstly, considering the diversity of base classifiers in the initial pool of ensemble, bootstrap sampling is utilized to train base ELMs. Secondly, we utilize pairwise diversity measures for each base classifier to pre-prune the base ELMs. This step can preserve the base classifier with large diversity, eliminate the redundant base classifier and reduce the complexity of the GSO-based pruning stage. Finally, further pruning is carried out by using the GSO method and the remaining base ELMs are integrated by majority voting.
The contributions of this paper can be described as follows:
(1) We propose a novel sensor-based HAR approach based on ELM and DMGSOSEN for improving the recognition performance and reducing the size of ensemble. The DMGSOSEN is a novel ensemble pruning approach that combines existing algorithms, it has good capacity of selecting the generated base classifiers to show its desirable performance for HAR.
(2) We find that the double-fault measure has better performance when compared with four other pairwise diversity measures. Based on the double-fault measure pre-pruning, we utilize discrete glowworm swarm optimization algorithm to further search the optimal sub-ensemble.
(3) The DMGSOSEN-based approach could select superior base classifiers adaptively through optimization algorithm, which makes it more practicable to deal with the various styles of activity.
(4) We demonstrate the efficiency of the proposed DMGSOSEN-based HAR approach with dataset acquired from different body positions.
The remainder of this paper is organized as follows: In Section 2, we present details of the proposed HAR approach based on ELM and DMGSOSEN. The DMGSOSEN is a novel combination of existing algorithms for ensemble pruning in ensemble learning-based HAR. Section 3 and Section 4 describe the experimental dataset and experimental setup, respectively. Following that, comparative experiments are carried out to validate the effectiveness of the proposed approach in Section 5. Finally, we draw conclusions in Section 6.

2. The Proposed HAR Approach Based on ELM and DMGSOSEN

Figure 1 shows the architecture of the proposed HAR approach. As shown in Figure 1, the proposed selective ensemble learning method for HAR contains three modules: base classifier generation, base classifier selection by DMGSOSEN, and classifier fusion. First, the initial pool of base classifier is constructed through bootstrap sampling. Because there may be poor performance and redundant base classifiers in the initial pool, we will not utilize all the base classifiers to establish an ensemble recognition system. Second, the DMGSOSEN is proposed to select superior individual classifiers from the initial pool of base classifier and optimize the ensemble. The DMGSOSEN combines double-fault measure and GSO algorithm uniquely for base classifier selection in HAR. Third, majority voting is utilized to integrate the selected base classifiers. In the following subsections, we describe details of these three modules.

2.1. Base Classifier Generation

In this paper, ELM is selected as base classifier of the ensemble system due to its simple structure and good generalization performance. The basic structure of ELM has input, hidden and output layer nodes, which is shown in Figure 2. The only parameter that needs to be set is the number of hidden layer nodes. For any N different samples (xj, tj), j = 1, 2, …, N, where x j = [ x j 1 , x j 2 , , x jn ] T is the jth sample, each sample contains n-dimensional features, and t j = [ t j 1 , t j 2 , , t jm ] T is the encoded class label. All samples belong to m different classes and the ELM mathematical model with L hidden neurons can be expressed as:
i = 1 L β i g ( w i x j + b i ) = t j ,   j = 1 , 2 , , N
where g(x) is the excitation function, wi, bi, and βi are the input weight, hidden layer bias and output weights of the ith hidden neuron node respectively. Equation (1) can be written in matrix form:
H β = T
where β represents the output weight, T is the corresponding coding class label, and H is the hidden layer output matrix:
H = g ( w 1 x 1 + b 1 ) g ( w L x 1 + b L ) g ( w 1 x N + b 1 ) g ( w L x N + b L ) N × L
The output weight β can be calculated by Equation (4):
β = H T
where H is the generalized inverse matrix of H.
HAR requires the recognition system to not only have good recognition accuracy, but also to have good generalization ability. Although ELM can handle general recognition problems, different subject-related features and various styles of a certain activity usually degrade the performance of the ELM. Furthermore, the generalization ability of an algorithm is usually influenced by training samples with category representations that can determine the decision boundaries of different activity categories. The ensemble learning techniques can be utilized to improve the generalization ability of a single classifier. Diversity is an important principle for base classifier generation, it is of great significance for improving the generalization ability of ensemble learning. Bootstrap and bagging have been utilized in several studies [27,28,33] to improve the diversity of base classifiers for enhancing the generalization ability of ensemble. In this paper, the bootstrap sampling method is utilized to obtain the training dataset for each ELM.

2.2. Pairwise Diversity Measures

The diversity between the base classifiers is a key factor in determining the performance of an ensemble system. The diversity measure between the base classifiers is not simple, despite a lot of theories have been proposed to measure the diversity among base classifier, there is currently no uniform definition of diversity among classifiers. Considering the small computational complexity of pair-based diversity measures and their good performances on ensemble pruning, five pairwise diversity measures which belongs to ordering-based method will be compared with respect to their pre-pruning performances. We will choose the pairwise diversity measure with the best performance for DMGSOSEN. These five methods will be described as follows:
Disagreement [51] was proposed by Skalak based on the concept of diversity. The larger the disagreement measure, the greater the diversity between the base classifiers. The disagreement measure for the two base classifiers Ci and Cj can be calculated by the following formula:
Dis ij = b + c a + b + c + d
where d represents the number of samples when classifier Ci and Cj recognize errors, a represents the number of samples when classifier Ci and Cj recognize correctly, b represents the number of samples when classifier Ci recognizes errors while classifier Cj recognizes correctly, c represents the number of samples when classifier Ci recognizes correctly while classifier Cj recognizes errors.
Correlation coefficient [52] is derived from statistics, and the correlation coefficient of two classifiers can be calculated by Equation (6).
ρ ij = a d b c ( a + b ) ( c + d ) ( a + c ) ( b + d )
The Q-statistic was proposed by Yule [53], and can be regarded as a simplified operation of the correlation coefficient. It is defined by Equation (7).
Q ij = a d b c a d + b c
The Kappa-statistic is widely used in statistics and it was used to analyze the diversity between classifiers for the first time by Margineantu and Dietterich [54]. The amount of computation using paired Kappa is less than the Q-statistic measure. The smaller the paired Kappa measure, the smaller the correlation of the base classifier. The formula is as shown in Equation (8).
K ij = 2 ( a d b c ) ( a + b ) ( b + d ) + ( a + c ) ( c + d )
Giacinto and Roli proposed a double-fault measure in 2001 [55], which can be utilized to calculate the proportion of samples misclassified by both classifiers. It can be expressed as Equation (9).
DF ij = d a + b + c + d

2.3. Discrete Glowworm Swarm Optimization

After pre-pruning the base classifier based on the pairwise diversity measures, this paper utilizes the GSO to select base classifiers with better performance to optimize the sub-ensemble. As the traditional GSO is proposed for continuous optimization problems, it is not suitable for selective ensemble which belongs to a discrete combinatorial optimization problem. To make GSO suitable for dealing with discrete problems in binary space, the discrete glowworm swarm optimization (DGSO) is detailed in this section. First, the GSO algorithm is briefly described.

2.3.1. GSO

GSO is a heuristic algorithm inspired by mimicking the luminescent behavior of glowworms in nature. In GSO, glowworms are randomly distributed throughout the entire search space with a certain amount of fluorescein. In the range of the field of view, glowworms constantly move closer to those that are brighter than themselves, thus achieving group optimization and finally converging on the global optimal solution. The basic steps are as follows:
  • Step 1: Initialization of algorithm parameters.
  • Step 2: Convert the fitness value J(xi (t)) corresponding to the position xi(t) of glowworm i at time t to the fluorescein value li(t) by Equation (10).
    l i ( t ) = ( 1 ρ ) l i ( t 1 ) + γ J ( x i ( t ) )
    where ρ is the fluorescein decay constant belonging to (0, 1) and the γ is the fluorescein enhancement constant.
  • Step 3: Each glowworm selects a neighborhood set Ni(t) whose individual brightness is higher than itself in its dynamic decision domain radius r d i ( t ) ( 0 < r d i ( t ) < r s ) .
  • Step 4: Calculate the probability pij(t) of the movement of individual i to the individual j (jNi (t)) in its neighborhood set Ni(t) by Equation (11).
    p ij ( t ) = l j ( t ) l i ( t ) k N i ( t ) l k ( t ) l i ( t )
  • Step 5: Select the moving object and update the glowworm position according to Equation (12).
    x i ( t + 1 ) = x i ( t ) + s ( x j ( t ) x i ( t ) x j ( t ) x i ( t ) )
    where s > 0 is the step that one glowworm is moving towards the other.
  • Step 6: Update the dynamic decision radius of glowworm by Equation (13).
    r d i ( t + 1 ) = min ( r s , max ( 0 , r d t ( t ) + β ( n t N i ( t ) ) ) )
    where β is a constant and nt is a parameter used to control the number of neighbors.

2.3.2. DGSO

Based on the traditional GSO algorithm, the modifications of DGSO mainly include the following aspects: the encoding method of the solution, the position update method of the glowworm and the construction of the fitness function. Through these improvements, DGSO is able to search in a binary discrete space. These modifications will be detailed in this section.
(a) Encoding method
When using DGSO to solve the selective selection problems, the structure of the solution can be expressed by:
x i ( t ) = [ x i 1 ( t ) , x i 2 ( t ) , , x ih h ( t ) , , x iD ( t ) ]
The above formula indicates the position of the ith glowworm in the tth iteration and D′ represents the dimension of each glowworm in the population, that is, the number of base classifiers after pre-pruning by diversity measures. In selective ensemble, xih can only be 0 or 1, xih(t) = 1 indicates that the ith glowworm selects the hth base classifier in the tth iteration and xih(t) = 0 means that the ith glowworm does not select the hth base classifier in the tth iteration.
The initial position of the glowworm is obtained by:
x ih ( 0 ) = 1 r a n d 0.5 0 r a n d < 0.5
where rand is a randomly number generated from (0, 1).
(b) Glowworm position update
The fixed step search method is not suitable for DGSO in binary discrete space. To make the search process of the discrete GSO algorithm simple and efficient, this paper selects the position update formula according to probability. In the tth iteration of the DGSO algorithm, current glowworm position can be expressed as x i ( t ) = [ x i 1 ( t ) , x i 2 ( t ) , , x iD ( t ) ] and target glowworm position can be expressed as x j ( t ) = [ x j 1 ( t ) , x j 2 ( t ) , , x jD ( t ) ] . When the position update is performed, each dimension variable in the individual position row vector is updated with a certain probability, thereby realizing the update of the entire vector which is the position update. The specific position update formula is as follows:
x ik ( t + 1 ) = x ik ( t ) ,   r ( k ) p 1 x jk ( t ) ,   p 1 < r ( k ) p 2   x ik ( t ) 1 ,   r ( k ) > p 2
where r is a randomly generated D′-dimensional vector r = { r 1 , r 2 , , r k , , r D } and rk ∈ [0,1], p1 and p2 ∈ [0,1] are both selected parameters for the update formula.
(c) Infeasible solution
When the elements in the solution vector appear to all be 0 or all be 1, these two cases correspond to the selective ensemble system containing no base classifiers and all base classifiers, respectively. Therefore, both of the above cases are considered to be infeasible solutions. For both cases, Equation (15) is utilized in this paper to randomly generate feasible solutions to improve search efficiency.
(d) Distance between glowworms
Since selective ensemble learning is a discrete combination optimization problem and the solution vector of discrete GSO only contains two values of 0 and 1, the traditional Euclidean distance is not suitable for calculating the distance between glowworm. Therefore, the Hamming distance metric is utilized in this paper instead, it is the number of different characters in the same position corresponding to two equal length vectors. If the positions of the individual glowworm i and j in the tth iteration are:
x i ( t ) = ( x i 1 , x i 2 , , x iD ) , x j ( t ) = ( x j 1 , x j 1 , , x iD )
Then the distance between the individual glowworm i and j at the tth iteration is recorded as:
h m _ d ij ( t ) = h a m m i n g _ d i s t a n c e ( x i ( t ) , x j ( t ) )
(e) Fitness function
The fitness function of the selective ensemble problem can be defined as:
F n = 1 m j = 1 m A c c ( f ( x j ) , y j )
where Fn is the recognition accuracy between the recognized category and the actual category, A c c ( f ( x j ) , y j ) = 1 ,   if   f ( x j ) = y j 0 ,   if   f ( x j ) y j , m represents the number of test samples, f(xj) and yj represent the recognized category and actual category on the jth test sample, respectively. The higher the fitness value, the higher the selective ensemble accuracy.

2.4. The Proposed DMGSOSEN-Based Classifier Selection

Step 1:
Establish an initial pool of base ELMs. In this paper, bootstrap sampling is utilized to generate D training subsets Si, i = 1, 2, ..., D. The base ELM is trained on each subset Si, so a base classifier pool with D ELMs can be obtained.
Step 2:
Pre-prune the base classifier pool based on pairwise diversity measures. For ensemble selection, it is not only computationally expensive but also difficult to search for the optimal sub-ensemble when using an optimization-based pruning method, especially when the initial pool of base classifiers is large in size. To tackle this problem, we pre-prune the initial base ELMs pool in order to reduce the number of base classifiers before using the GSO method. The base classifiers that make up the ensemble system should not only have good performance, but also have great diversity, in order to ensure good generalization ability of ensemble system. Thus, the five kinds of pairwise diversity measures mentioned in Section 2.2 are respectively utilized to calculate the diversity of each base ELM and eliminate the base ELMs with small diversity in the base classifier pool. The performances of the five pairwise diversity measures will be compared, and we will choose the best one as the evaluative criteria. The pairwise diversity measure of each base ELM can be obtained by:
Div i = 1 D j = 1 D div ij
where Divi means the pairwise diversity measure of the ith base ELM, divij represents the pairwise diversity measure between the ith base ELM and the jth base ELM, 1 < ijD.
Step 3:
DGSO pruning. After pre-pruning by pairwise diversity measure, the D base classifiers in the base classifier pool retain the D′ base classifiers. Next, the GSO algorithm is used to continue pruning the D′ base classifiers.
Step 3.1:
Initialize the basic parameters of the DGSO. These parameters include population size g, maximum iteration number iter_max, fluorescein volatilization factor ρ, fluorescein update rate γ, dynamic decision domain update rate β, threshold nt of glowworm contained in the neighborhood set Ni(t), initial fluorescein value l0, initial dynamic decision radius r d i ( 0 ) , perceived radius rs, initial solution x(0), the parameters of position update formula: p1, p2.
Step 3.2:
A set of glowworms in initial positions can been obtained. Calculate the fitness value of the glowworm according to Equation (19) and the corresponding fluorescein value by Equation (10). The fitness value J(xi(t)) corresponding to the position xi(t) of the glowworm i at the tth iteration is converted to fluorescein value li(t). Save the glowworm’s position with the maximum fitness function Fmax.
Step 3.3:
Calculate the Hamming distance between glowworm individuals by Equation (18). Each glowworm selects a neighbor set Ni(t) whose fluorescein values are larger than itself in its dynamic decision domain radius r d i ( t ) ( 0 < r d i ( t ) r s ) .
Step 3.4:
Calculate the probability pij(t) of the glowworm i moving to the individual j(jNi(t)) in the neighborhood set Ni(t) by Equation (11) and select the moving object by the roulette method according to the probability.
Step 3.5:
Randomly generate a D′-dimensional vector r between 0 and 1, and r ( k ) { r 1 , r 2 , , r k , , r D } . According to the value of r(k), update the position of each candidate glowworm by Equation (16).
Step 3.6:
In the set of glowworms in new positions, the glowworm’s position with the maximum fitness function is Fmax. If Fmax > Fmax, set Fmax = Fmax, Otherwise, Fmax = Fmax. Update the dynamic decision domain radius of the glowworm individual according to Equation (13).
Step 3.7:
Check termination criteria. If the maximum number of iterations is not reached, return to Step 3.2. Otherwise, go to Step 4.
Step 4:
Lastly, the glowworm with the best fitness value is considered for majority voting. Then, the base ELMs participating in the final ensemble can be acquired, which corresponds to the coding combination with the fitness value. The final recognition result is obtained by majority voting: f(x) = arg max Ni, Ni is the number of base classifiers that the sample x is recognized as the ith activity.
The flowchart of the proposed DMGSOSEN approach is illustrated in Figure 3.

3. Experimental Dataset and Feature Extraction

3.1. Dataset

In the experiment, we acquired the dataset in our laboratory by using the TRIGNOTM wireless system from Delsys Company, as shown in Figure 4. The TRIGNOTM wireless system contains a data acquisition platform and a collection node, which are shown in Figure 4a,b, respectively. The collection node integrates a triaxial accelerometer which has a sampling frequency of 150 Hz and an acceleration range of ±6 G with a resolution of 0.016 (G is the gravitational constant). Figure 4c presents the fixed position of the collection node and the workflow of the system implementation. Since the experimental platform has wireless transmission function, the acceleration signal can be transmitted to the data acquisition platform from collection node. The ZigBee protocol is utilized in the study. Once received by the acquisition platform, the data are transmitted and stored in the computer. Five healthy students, including 3 males and 2 females, participated in the data collection. Their ages ranged from 20 to 34, and their average age was 26. Each participant was asked to fix the collection node to five different body parts: chest, waist, left wrist, left ankle and right arm. Before the start of each experiment, we utilized straps to fix the sensors on the body and checked the sensors were in the same position as the previous subject. The activities performed by each subject included walking, running, going upstairs, going downstairs, jumping and standing. These activities were separated and there were no transitions. Therefore, a dataset with five sensor locations could be obtained. Figure 5 shows the activity data of “walking” from the selected five positions. The preprocessing of the acceleration signal includes removing abnormal data and signal denoising. Data points with numerical anomalies in the acceleration signal sequence were removed. Discrete wavelet transform was adopted to filter out noise signals in this paper and the wavelet function Coif5 was utilized to filter out noise signals from acceleration signals. Then, the sliding window was utilized to divide the acceleration signal after preprocessing; 300 samples were chosen as the window length, and a 50% overlap between adjacent windows was adopted.

3.2. Feature Extraction

After using sliding window to divide the triaxial acceleration data, we extracted features from these windows. These features include the maximum, the minimum, the mean value, standard deviation σ, skewness S, kurtosis K, correlation coefficient C between three axes, signal magnitude area (SMA), and number of zero crossings which is number of zero crossings in a window after subtracting the window mean value from every window sample. Various research works have proven the effectiveness of these features on HAR [8,13,16,30]. These features can be expressed as follows:
m e a n = 1 N i = 1 N a i
σ = 1 N i = 1 N ( a i m e a n ) 2
S = 1 N i = 1 N ( a i m e a n σ ) 3
K = 1 N i = 1 N ( a i m e a n ) 4 / σ 4 3
C xy = cov ( x , y ) / ( σ x σ y )
SMA = i = 1 N ( x ( i ) + y ( i ) + z ( i ) )
where ai is the acceleration data I = 1, 2, …, N. N is the number of samples, cov(x, y) is the covariance of the x- and y-axis acceleration. x(i), y(i) and z(i) respectively indicate the values of x-axis, y-axis and z-axis acceleration signals at the ith sampling point. After feature extraction, all features were normalized to the interval [0, 1]. Considering the balance of the data, the number of each activity sample of each dataset is as consistent as possible. The first column of Table 1 shows the activity performed by each subject and the right column shows the quantities of feature samples of different activities from the five body positions.

4. Experimental Setup

The experiments were implemented in Matlab 2014a using a computer with a 2.8 GHz processor and 6 GB memory. The parameters of DGSO are set as follows: population size g = 20, maximum iteration number iter_max = 100, fluorescein volatilization factor ρ = 0.4, fluorescein update rate γ = 0.5, dynamic decision domain update rate β = 0.06, initial fluorescein value l0 = 2, threshold nt = 5, initial dynamic decision radius r d i ( 0 ) = 7, perceived radius rs = 12, p1 = 0.15, p2 = 0.75. All of these parameters were defined empirically. The leave-one-out (LOO) strategy was utilized to evaluate the proposed method. The data from four subjects was utilized as training data and half of the data from another remaining subject was utilized for selecting the ensemble with best performance. In addition, the other half of the data was utilized for testing the proposed approach. The verification was repeated 5 times until the data from all subjects had been utilized for pre-pruning and testing.

4.1. Pre-Pruning Based on Pairwise Diversity Measures

To select the most effective method for pre-pruning the initial base classifier pool, according to diversity of base ELMs (from highest to lowest), the recognition accuracies of five body positions for ordered bagging based on initial pool with 50 base ELMs are shown in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10. It can be seen from Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 that, as there are few base classifiers in the initial stage of the ensemble, the ensemble system lacks diversity, which affects the ensemble’s accuracy. The ensemble accuracy reaches the maximum at an intermediate number of base ELMs. Then, there are a large number of redundant base classifiers in the ensemble system, which will result in a decrease in ensemble accuracy. Therefore, the ensemble accuracy increases first and then decreases with the number of base ELMs. This demonstrates that the ensemble accuracy can be improved by pre-pruning base ELMs with lower diversity with other base ELMs. Additionally, we can also find that the double-fault measure can achieve better results than the other four diversity measures with five sensor locations. Hence, the double-fault measure will be utilized for pre-pruning base ELMs in this paper.
The parameter D′ of the pre-prune is important, which determines the number of base classifiers in the GSO pruning. As we can see from Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10, the number of base ELMs when the maximum ensemble accuracy obtained is different. It is unscientific to set fixed parameter D′ when data from different sensor positions and the number of base classifiers D are considered. Therefore, we set the parameter D′ according to statistical methods. Suppose Divi is the double-fault measure of the ith base ELM and [Div1, Div2, ……, DivD] represents the double-fault measure vector of D base classifiers, D i v ¯ = 1 D i = 1 D D i v i is the arithmetic mean of the double-fault measure of D base classifiers, we eliminate the base classifiers whose Divi is smaller than D iv ¯ , and the remaining D′ base ELMs are utilized for the GSO-based selective ensemble. For initial pool with 50 base ELMs, the D′ is 18, 25, 27, 30, 29 for waist, chest, right arm, left ankle and left wrist, respectively.

4.2. Performance Measures

The accuracy measure is used to evaluate the performance of the proposed method, which can be expressed as follows:
Accuracy = T P + T N T P + T N + F P + F N
where the variables TP, TN, FP, and FN, respectively, represent the number of true positive, true negative, false positive, and false negative outcomes in a given experiment.
In addition, F1 evaluation criteria are also considered. F1 is defined as the combination of precision and the recall, which are defined as follows:
precision = T P T P + F P
recall = T P T P + F N
The F1 is calculated as follows:
F 1 = 2 × recall × precision recall + precision

5. Experimental Results

To verify the proposed HAR approach, initial base classifier pools of different sizes (50, 100, 150, 200) were set up and utilized for the experiment. Table 2, Table 3, Table 4 and Table 5 show the comparative recognition performance of DMGSOSEN-based HAR approach with the best, average, and worst performance of the base ELMs in the initial pool. It can be observed from Table 2, Table 3, Table 4 and Table 5 that the performance obtained by the proposed approach is much better than the best and average performance of base ELMs in the initial base classifier pool. In addition, it can be found that no matter which position is considered, the proposed DMGSOSEN-based HAR approach does not achieve the best results when the size of initial base classifier pool is the largest. When the number of the initial base classifier is 100 and 150, the proposed approach is more likely to achieve better results. Furthermore, we also find that the recognition performances of the five positions are quite different, and the waist is more likely than the other four positions to achieve optimal recognition result.
Additionally, in order to gain a better insight into the activity recognition problem and the proposed DMGSOSEN base HAR method, the corresponding confusion matrix was constructed. Table 6, Table 7, Table 8, Table 9 and Table 10 show the results of using an initial pool of 100 base classifiers with the data from the five positions, respectively. We can observe that confusion occurs in most cases between activities such as (GU, GD), (W, R), (J, GD) and (J, GU), especially when the data from the right arm, left ankle and left wrist were utilized. Furthermore, we find that data from the wrist and chest are superior to other positions for recognizing similar activities, such as (GU, GD) and (W, R). In addition, we can also observe that no matter which position is utilized, the activity standing (S) is much easier to recognize than activities such as running (R), going upstairs (GU) and going downstairs (GD).

5.1. Compared to Traditional Ensemble Algorithm-Based HAR

In addition, several comparative experiments were carried out to evaluate the proposed approach in comparison with the traditional ensemble methods Bagging and Adaboost. Table 11, Table 12, Table 13 and Table 14 show the comparative results with initial pools of different sizes. In Table 11, Table 12, Table 13 and Table 14, n represents the number of base classifiers selected by the proposed method. It can be seen from Table 11, Table 12, Table 13 and Table 14 that although the ensembles derived using the traditional methods Bagging and Adaboost have more base classifiers, the proposed DMGSOSEN outperforms these two methods with fewer base classifiers, which shows that it may be better to derive ensembles with many base classifiers than with all. Furthermore, we find that the proposed method eliminates more than 60% of the base classifiers in the initial pool and achieves better recognition performance compared with Bagging and Adaboost, thus demonstrating the effectiveness of the proposed DMGSOSEN for HAR.

5.2. Compared to the State-of-the-Art Pruning Approach-Based HAR

To better assess the performance of the proposed DMGSOSEN-based HAR approach, we utilized an initial pool containing 100 base ELMs in order to compare it with other state-of-the-art pruning method-based HAR. These pruning methods included aggregation ordering in bagging (AGOB) [56], ordered bagging ensemble (POBE) [57], D-D-ELM [58], DF-D-ELM [59], GASEN [31], MOAG [60], RRE [61], and DivP [38]. Among these, AGOB, POBE and MOAG are all studies of ordering-based selective ensembles, and the basic classifiers are ordered by using their proposed metrics. GASEN utilizes GA to optimize the weights of the base classifier and the combination of base classifiers with the best performance constitutes the final ensemble. RRE attempts to make full use of the votes of the worst single model in the ensemble. DivP applies GA to combine five pairwise diversity matrices, utilizing a graph coloring method to generate candidate ensembles. DD-ELM and DF-D-ELM attempts to remove the base ELMs by using the disagreement measure and the double-fault measure, respectively.
Table 15, Table 16 and Table 17 show the comparative results of these methods at the five sensor locations. It can be seen from the Table 15, Table 16 and Table 17 that the proposed selective ensemble HAR method based on DMGSOSEN achieves the best recognition performance when compared with the other algorithms. Although the number of base classifiers selected by the proposed method is slightly more than the DivP, it can achieve better performance than DivP with data of five sensor locations. Overall, the recognition performance and the number of base classifiers demonstrate that the proposed DMGSOSEN-based HAR approach performs better than other selective ensemble approaches, which indicates that the DMGSOSEN method has stronger generalization ability and learning efficiency in HAR tasks.

5.3. Compared to the Previous Studies in HAR

To further evaluate the performance of this study, we compared it with some previous studies in HAR, including EEMD+FS+SVM [12], ACELM [17], CELearning [41], tFFT+Convnet [62] and KPCA+DBN [63]. These studies, conducted in recent years, include deep learning, ensemble learning and feature selection for HAR. The methods and results of all these studies are shown in Table 18. Although these studies are based on their different datasets and methods, we can know the relative performance of this research in the field of HAR. It is obvious that our study has the best performance compared with other previous studies. We can achieve 96.7% recognition accuracy and F1 score by using our proposed base ensemble ELM approach.

6. Conclusions

Traditional HAR systems based on a single classifier are likely to perform poorly due to the diversity of activity styles. Combining multiple classifiers appears to be a very effective approach for improving the performance and generalization ability of the HAR system. However, there would be some base classifiers that are redundant and perform poorly in multiple classifier systems, providing no contribution to the performance of the HAR system. To tackle this issue, a HAR approach based on ELM and DMGSOSEN is proposed in this paper. The DMGSOSEN is a novel ensemble pruning method using a combination of existing algorithms for ensemble learning-based HAR. Compared to the other four pairwise diversity measures, the double-fault measure shows better performance for pre-pruning the initial pool on five sensor locations. The experimental results on the dataset with five positions show that the DMGSOSEN-based HAR approach can achieve better recognition performance with fewer base ELMs compared with traditional ensemble HAR methods: Bagging, Adaboost and other state-of-the-art pruning-based HAR methods.
In future work, more complex activities will be added to test the proposed method, and we will optimize the module’s performance by considering other state-of-art machine learning methods, such as deep leaning. For example, when determining base classifiers, kernel extreme learning machine (KELM) is an improvement of ELM with characteristics of fast training and good generalization. In addition, more combinations of diversity measures and heuristic searching algorithms such as particle swarm optimization or fish swarm algorithm will be attempted to search for a sub-ensemble for constructing a selective ensemble-based HAR system.
The dataset utilized in this work only contains six daily activities from five subjects, who were all healthy with similar ages. This is a limitation of this work. In future works, we will attempt to collect data from more subjects with different living behaviors, ages, genders, etc., and more high-level activities (open door, cooking, etc.) will be considered in order to verify the proposed method. Furthermore, some public datasets should be utilized to test the performance of the proposed method and compare it with some state-of-the-art approaches. Moreover, this study is also limited due to the lack of a validation set completely different from the training set. We will utilize datasets with different ages or physical characteristics to test the applicability of the proposed method.

Author Contributions

Conceptualization, Y.T.; Data curation, L.C.; Funding acquisition, L.C., Y.G., X.W.; Investigation, L.C., Y.G.; Methodology, Y.T., J.Z.; Validation, L.C., Y.G.; Writing—original draft, Y.T.; Writing—review & editing, J.Z., X.W.

Funding

This work was supported by China Scholarship Council (201806700006), National Natural Science Foundation of China (NO. 61803143 and 61703135) and National Key Technology Research and Development Program of the Ministry of Science and Technology of China (NO. 2015BAI06B03).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Field, M.; Stirling, D.; Zengxi, P.; Ros, M.; Naghdy, F. Recognizing human motions through mixture modeling of inertial data. Pattern Recognit. 2015, 48, 2394–2406. [Google Scholar] [CrossRef]
  2. Yao, R.; Lin, G.; Shi, Q.; Ranasinghe, D.C. Efficient dense labelling of human activity sequences from wearables using fully convolutional networks. Pattern Recognit. 2018, 78, 252–266. [Google Scholar] [CrossRef]
  3. Chen, Y.Q.; Huang, M.Y.; Hu, C.Y.; Zhu, Y.C.; Han, F.; Miao, C.Y. A coarse to fine feature selection method for accurate detection of cerebral small vessel disease. In Proceedings of the International Joint Conference on Neural Networks, Vancouver, BC, Canada, 24–29 July 2016; pp. 2609–2616. [Google Scholar]
  4. Chen, Y.Q.; Yu, H.C.; Miao, C.Y.; Yang, X.D. Using motor patterns for stroke detection. Sci. Adv. Comput. Psychophysiol. 2015, 46, 12–14. [Google Scholar]
  5. Mooney, R.; Corley, G.; Godfrey, A.; Quinlan, L.R.; ÓLaighin, G. Inertial sensor technology for elite swimming performance analysis: A systematic review. Sensors 2016, 16, 18. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, Z.L.; Guo, M.; Zhao, C. Badminton stroke recognition based on body sensor networks. IEEE Trans. Hum. Mach. Syst. 2016, 46, 769–775. [Google Scholar] [CrossRef]
  7. Lara, O.D.; Labrador, M.A. A survey on human activity recognition using wearable sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
  8. Khan, A.M.; Lee, Y.K.; Lee, S.Y.; Kim, T.S. A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 1166–1172. [Google Scholar] [CrossRef]
  9. Akhavian, R.; Behzadan, A.H. Smartphone-based construction workers’ activity recognition and classification. Autom. Constr. 2016, 71, 198–209. [Google Scholar] [CrossRef]
  10. Wannenburg, J.; Malekian, R. Physical activity recognition from smartphone accelerometer data for user context awareness sensing. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 3142–3149. [Google Scholar] [CrossRef]
  11. Trost, S.G.; Pinkham, M.F.; Lennon, N.; O’Neil, M.E. Decision trees for detection of activity intensity in youth with cerebral palsy. Med. Sci. Sports Exerc. 2016, 48, 958–966. [Google Scholar] [CrossRef]
  12. Wang, Z.L.; Wu, D.H.; Chen, J.M.; Ghoneim, A.; Hossain, M.A. A triaxial accelerometer-based human activity recognition via EEMD-based features and game-theory-based feature selection. IEEE Sens. J. 2016, 16, 3198–3207. [Google Scholar] [CrossRef]
  13. Chen, Z.H.; Zhu, Q.C.; Soh, Y.C.; Zhang, L. Robust human activity recognition using smartphone sensors via CT-PCA and online SVM. IEEE Trans. Ind. Inform. 2017, 13, 3070–3080. [Google Scholar] [CrossRef]
  14. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the IEEE International Joint Conference Neural Network, Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar]
  15. Chen, Y.Q.; Zhao, Z.T.; Wang, S.Q.; Chen, Z.Y. Extreme learning machine-based device displacement free activity recognition model. Soft Comput. 2012, 16, 1617–1625. [Google Scholar] [CrossRef]
  16. Xiao, W.D.; Lu, Y.L. Daily human physical activity recognition based on kernel discriminant analysis and extreme learning machine. Math. Probl. Eng. 2015, 2015, 790412. [Google Scholar] [CrossRef]
  17. Yuan, Y.; Wang, C.H.; Zhang, J.Z.; Xu, J.D.; Li, M. An ensemble approach for activity recognition with accelerometer in mobile-phone. In Proceedings of the 2014 IEEE 17th International Conference on Computational Science and Engineering, Chengdu, China, 19–21 December 2014; pp. 1469–1474. [Google Scholar]
  18. Wu, D.H.; Wang, Z.L.; Chen, Y.; Zhao, H.Y. Mixed-kernel based weighted extreme learning machine for inertial sensor based human activity recognition with imbalanced dataset. Neurocomputing 2016, 190, 35–49. [Google Scholar] [CrossRef]
  19. Gao, X.Y.; Chen, Z.Y.; Tang, S.; Zhang, Y.D.; Li, J.T. Adaptive weighted imbalance learning with application to abnormal activity recognition. Neurocomputing 2016, 173, 1927–1935. [Google Scholar] [CrossRef]
  20. Zhao, Z.T.; Chen, Z.Y.; Chen, Y.Q.; Wang, S.Q.; Wang, H.G. A class incremental extreme learning machine for activity recognition. Cogn. Comput. 2014, 6, 423–431. [Google Scholar] [CrossRef]
  21. Deng, W.Y.; Zheng, Q.H.; Wang, Z.M. Cross-person activity recognition using reduced kernel extreme learning machine. Neural Netw. 2014, 53, 1–7. [Google Scholar] [CrossRef]
  22. Wang, Z.L.; Wu, D.H.; Gravinac, R.; Fortino, G.; Jiang, Y.M.; Tang, K. Kernel fusion based extreme learning machine for cross-location activity recognition. Inf. Fusion 2017, 37, 1–9. [Google Scholar] [CrossRef]
  23. Zhu, X.H.; Ni, Z.W.; Cheng, M.Y.; Jin, F.F.; Li, J.M.; Weckman, G. Selective ensemble based on extreme learning machine and improved discrete artificial fish swarm algorithm for haze forecast. Appl. Intell. 2018, 48, 1757–1775. [Google Scholar] [CrossRef]
  24. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef] [Green Version]
  25. Martínez-Muñoz, G.; Hernández-Lobato, D.; Suárez, A. An analysis of ensemble pruning techniques based on ordered aggregation. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 245–259. [Google Scholar] [CrossRef]
  26. Ykhlef, H.; Bouchaffra, D. An efficient ensemble pruning approach based on simple coalitional games. Inf. Fusion 2017, 34, 28–42. [Google Scholar] [CrossRef]
  27. Ahmad, Z.; Zhang, J. Bayesian selective combination of multiple neural networks for improving long-range predictions in nonlinear process modelling. Neural Comput. Appl. 2005, 14, 78–87. [Google Scholar] [CrossRef]
  28. Ahmad, Z.; Zhang, J. Selective combination of multiple neural networks for improving model prediction in nonlinear systems modelling through forward selection and backward elimination. Neurocomputing 2009, 72, 1198–1204. [Google Scholar] [CrossRef]
  29. Li, D.Y.; Wen, G.H. MRMR-based ensemble pruning for facial expression recognition. Multimed. Tools Appl. 2018, 77, 15251–15272. [Google Scholar] [CrossRef]
  30. Cao, J.J.; Li, W.F.; Ma, C.C.; Tao, Z.W. Optimizing multi-sensor deployment via ensemble pruning for wearable activity recognition. Inf. Fusion 2018, 41, 68–79. [Google Scholar] [CrossRef]
  31. Guo, H.P.; Liu, H.B.; Li, R.; Wu, C.A.; Guo, Y.B.; Xu, M.L. Margin & diversity based ordering ensemble pruning. Neurocomputing 2018, 275, 237–246. [Google Scholar]
  32. Zhou, Z.H.; Wu, J.; Tang, W. Ensembling neural networks: many could be better than all. Artif. Intell. 2002, 137, 239–263. [Google Scholar] [CrossRef]
  33. Zhu, X.H.; Ni, Z.W.; Ni, L.P.; Jin, F.F.; Cheng, M.Y.; Li, J.M. Improved discrete artificial fish swarm algorithm combined with margin distance minimization for ensemble pruning. Comput. Ind. Eng. 2019, 128, 32–46. [Google Scholar] [CrossRef]
  34. Ashwaq, Q.; Shahnorbanun, S.; Dheeb, A.; Siti, N.H.S.A.; Rizuana, I.H.; Shantini, A. Heterogeneous ensemble pruning based on bee algorithm for mammogram classification. Int. J. Adv. Compu. Sci. Appl. 2018, 9, 231–239. [Google Scholar]
  35. Xu, S.; Chan, K.; Gao, J.; Xu, X.; Li, X.; Hua, X.; An, J. An integrated K-means-Laplacian cluster ensemble approach for document datasets. Neurocomputing 2016, 214, 495–507. [Google Scholar] [CrossRef]
  36. Bouguettaya, A.; Yu, Q.; Liu, X.; Zhou, X.; Song, A. Efficient agglomerative hierarchical clustering. Expert Syst. Appl. 2015, 42, 2785–2797. [Google Scholar] [CrossRef]
  37. Yu, J.; Yang, M. Deterministic annealing Gustafson-Kessel fuzzy clustering algorithm. Inf. Sci. 2017, 417, 435–453. [Google Scholar]
  38. Bakker, B.; Heskes, T. Clustering ensembles of neural network models. Neural Netw. 2003, 16, 261–269. [Google Scholar] [CrossRef]
  39. Chen, Z.H.; Jiang, C.Y.; Xie, L.H. A novel ensemble ELM for human activity recognition using smartphone sensors. IEEE Trans. Ind. Inf. 2019, 15, 2691–2699. [Google Scholar] [CrossRef]
  40. Acharya, S.; Mongan, W.M.; Rasheed, I.; Liu, Y.Q.; Anday, E.; Dion, G.; Fontecchio, A.; Kurzweg, T.; Dandekar, K.R. Ensemble learning approach via kalman filtering for a passive wearable respiratory monitor. IEEE J. Biomed. Health Inf. 2019, 23, 1022–1031. [Google Scholar] [CrossRef] [PubMed]
  41. Xu, S.J.; Tang, Q.F.; Jin, L.P.; Pan, Z.G. A cascade ensemble learning model for human activity recognition with smartphones. Sensors 2019, 19, 2307. [Google Scholar] [CrossRef]
  42. Chelli, A.; Patzold, M. A machine learning approach for fall detection and daily living activity recognition. IEEE Access 2019, 7, 38670–38687. [Google Scholar] [CrossRef]
  43. Guan, Y.; Plötz, T. Ensembles of Deep LSTM Learners for Activity Recognition using Wearables. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2017, 1, 1–28. [Google Scholar] [CrossRef] [Green Version]
  44. Sena, J.; Santos, J.B.; Schwartz, W.R. Multiscale DCNN Ensemble Applied to Human Activity Recognition Based on Wearable Sensors. In Proceedings of the 2018 26th European Signal Processing Conference (EUSIPCO), Rome, Italy, 3–7 September 2018; pp. 1202–1206. [Google Scholar]
  45. Cavalcanti, G.D.C.; Oliveira, L.S.; Moura, T.J.M.; Carvalho, G.V. Combining diversity measures for ensemble pruning. Pattern Recognit. Lett. 2016, 74, 38–45. [Google Scholar] [CrossRef]
  46. Yang, C.; Yin, X.; Hao, H.; Yan, Y.; Wang, Z.B. Classifier ensemble with diversity: Effectiveness analysis and ensemble optimization. Acta Autom. Sin. 2014, 40, 660–674. [Google Scholar]
  47. Krishnanand, K.N.; Ghose, D. Detection of multiple source locations using a glowworm metaphor with applications to collective robotics. In Proceedings of the IEEE Swarm Intelligence Symposium, Pasadena, CA, USA, 8–10 June 2005; pp. 84–91. [Google Scholar]
  48. Nelson Jayakumar, D.; Venkatesh, P. Glowworm swarm optimization algorithm with topsis for solving multiple objective environmental economic dispatch problem. Appl. Soft Compu. 2014, 23, 375–386. [Google Scholar] [CrossRef]
  49. Liao, W.H.; Kao, Y.C.; Li, Y.S. A sensor deployment approach using glowworm swarm optimization algorithm in wireless sensor networks. Expert Syst. Appl. 2011, 38, 12180–12188. [Google Scholar] [CrossRef]
  50. Marinaki, M.; Marinakis, Y. A glowworm swarm optimization algorithm for the vehicle routing problem with stochastic demands. Expert Syst. Appl. 2016, 45, 145–163. [Google Scholar] [CrossRef]
  51. Skalak, D.B. The sources of increased accuracy for two proposed boosting algorithms. In Proceedings of the 13th American Association for Artificial Intelligence, Integrating Multiple Learned Models Workshop, Portland, OR, USA, 4–8 August 1996; pp. 120–125. [Google Scholar]
  52. Kuncheva, L.I.; Whitaker, C.J. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Mach. Learn. 2003, 51, 181–207. [Google Scholar] [CrossRef]
  53. Tang, E.K.; Suganthan, P.N.; Yao, X. An analysis of diversity measures. Mach. Learn. 2006, 65, 247–271. [Google Scholar] [CrossRef] [Green Version]
  54. Margineantu, D.D.; Dietterich, T.G. Pruning adaptive boosting. In Proceedings of the Fourteenth International Conference on Machine Learning, Nashville, TN, USA, 8–12 July 1997; Volume 97, pp. 211–218. [Google Scholar]
  55. Giacinto, G.; Roli, F. Design of effective neural network ensembles for image classification purposes. Image Vis. Comput. 2001, 19, 699–707. [Google Scholar] [CrossRef]
  56. Martínezmuñoz, G.; Suárez, A. Aggregation ordering in bagging. In Proceedings of the IASTED international conference on artificial intelligence and applications, Innsbruck, Austria, 16–18 February 2004; pp. 258–263. [Google Scholar]
  57. Martínezmuñoz, G.; Suárez, A. Pruning in ordered bagging ensembles. In Proceedings of the Twenty-third International Conference on Machine learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 609–616. [Google Scholar]
  58. Lu, H.J.; An, C.L.; Ma, X.P.; Zheng, E.H.; Yang, X.B. Disagreement measure based ensemble of extreme learning machine for gene expression data classification. Chin. J. Comput. 2013, 36, 341–348. [Google Scholar] [CrossRef]
  59. Lu, H.J.; An, C.L.; Zheng, E.H.; Lu, Y. Dissimilarity based ensemble of extreme learning machine for gene expression data classification. Neurocomputing 2014, 128, 22–30. [Google Scholar] [CrossRef]
  60. Guo, L.; Boukir, S. Margin-based ordered aggregation for ensemble pruning. Pattern Recogn. Lett. 2013, 34, 603–609. [Google Scholar] [CrossRef]
  61. Dai, Q.; Zhang, T.; Liu, N. A new reverse reduce-error ensemble pruning algorithm. Appl. Soft Comput. 2015, 28, 237–249. [Google Scholar] [CrossRef]
  62. Ronao, C.A.; Cho, S.B. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  63. Hassan, M.M.; Uddin, M.Z.; Mohamed, A.; Almogren, A. A robust human activity recognition system using smartphone sensors and deep learning. Future Gener. Comput. Syst. 2018, 81, 307–313. [Google Scholar] [CrossRef]
Figure 1. The framework of proposed selective ensemble-based HAR approach.
Figure 1. The framework of proposed selective ensemble-based HAR approach.
Sensors 19 03468 g001
Figure 2. The basic structure of ELM.
Figure 2. The basic structure of ELM.
Sensors 19 03468 g002
Figure 3. Workflow of the DMGSOSEN.
Figure 3. Workflow of the DMGSOSEN.
Sensors 19 03468 g003
Figure 4. Human activity data acquisition platform based on acceleration sensor: (a) the data acquisition platform, (b) data collection node containing a triaxial accelerometer, (c) experimental data acquisition process.
Figure 4. Human activity data acquisition platform based on acceleration sensor: (a) the data acquisition platform, (b) data collection node containing a triaxial accelerometer, (c) experimental data acquisition process.
Sensors 19 03468 g004
Figure 5. The triaxial accelerometer data of “walking” from the chest, waist, left wrist, left ankle and right arm.
Figure 5. The triaxial accelerometer data of “walking” from the chest, waist, left wrist, left ankle and right arm.
Sensors 19 03468 g005
Figure 6. Recognition accuracy from waist position for ordered bagging according to five pairwise diversity measures.
Figure 6. Recognition accuracy from waist position for ordered bagging according to five pairwise diversity measures.
Sensors 19 03468 g006
Figure 7. Recognition accuracy from chest position for ordered bagging according to five pairwise diversity measures.
Figure 7. Recognition accuracy from chest position for ordered bagging according to five pairwise diversity measures.
Sensors 19 03468 g007
Figure 8. Recognition accuracy from right arm position for ordered bagging according to five pairwise diversity measures.
Figure 8. Recognition accuracy from right arm position for ordered bagging according to five pairwise diversity measures.
Sensors 19 03468 g008
Figure 9. Recognition accuracy from left ankle position for ordered bagging according to five pairwise diversity measures.
Figure 9. Recognition accuracy from left ankle position for ordered bagging according to five pairwise diversity measures.
Sensors 19 03468 g009
Figure 10. Recognition accuracy from left wrist position for ordered bagging according to five pairwise diversity measures.
Figure 10. Recognition accuracy from left wrist position for ordered bagging according to five pairwise diversity measures.
Sensors 19 03468 g010
Table 1. Same acquisition feature samples of different activities from the five body positions.
Table 1. Same acquisition feature samples of different activities from the five body positions.
ActivityFeature Samples
Walking (W)6164
Running (R)6028
Going upstairs (GU)5772
Going downstairs (GD)5836
Jumping (J)5982
Standing (S)6043
Table 2. Performance of the ensembles for initial pool sizes of 50 (Accuracy/F1%).
Table 2. Performance of the ensembles for initial pool sizes of 50 (Accuracy/F1%).
Position50
DMGSOSENBestMeanWorst
waist95.9/95.885.5/85.974.6/74.660.5/60.3
chest94.4/94.482.4/82.672.1/72.352.7/52.9
right arm91.8/91.777.3/77.169.6/69.850.5/50.2
left ankle90.2/90.479.6/79.270.2/70.249.6/49.9
left wrist89.1/89.278.7/78.570.8/70.449.7/49.4
Table 3. Performance of the ensembles for initial pool sizes of 100 (Accuracy/F1%).
Table 3. Performance of the ensembles for initial pool sizes of 100 (Accuracy/F1%).
Position100
DMGSOSENBestMeanWorst
waist96.7/96.788.9/88.874.5/74.460.2/60.5
chest95.7/95.784.8/84.872.4/72.351.8/51.8
right arm92.4/92.481.5/81.270.1/70.449.7/47.9
left ankle90.5/90.483.7/83.470.6/70.349.3/49.5
left wrist89.3/89.384.1/84.571.2/71.549.1/49.3
Table 4. Performance of the ensembles for initial pool sizes of 150 (Accuracy/F1%).
Table 4. Performance of the ensembles for initial pool sizes of 150 (Accuracy/F1%).
Position150
DMGSOSENBestMeanWorst
waist95.5/95.489.1/89.474.1/74.460.3/60.4
chest95.2/95.184.9/84.672.4/72.650.7/50.7
right arm92.1/92.282.6/82.669.7/69.849.4/49.5
left ankle90.3/90.484.2/84.171.4/71.948.5/48.7
left wrist89.2/89.384.9/84.970.9/70.747.2/47.2
Table 5. Performance of the ensembles for initial pool sizes of 200 (Accuracy/F1%).
Table 5. Performance of the ensembles for initial pool sizes of 200 (Accuracy/F1%).
Position200
DMGSOSENBestMeanWorst
waist94.2/94.389.8/89.874.6/74.560.1/60.3
chest93.9/93.585.9/85.972.5/72.749.2/49.4
right arm92.7/92.583.8/83.969.6/69.447.7/47.6
left ankle90.1/90.285.2/85.271.6/71.546.6/46.7
left wrist88.9/88.984.3/84.470.3/70.345.7/45.7
Table 6. Confusion matrix for DMGSOSEN-based HAR on the data from the waist when the initial pool size is 100.
Table 6. Confusion matrix for DMGSOSEN-based HAR on the data from the waist when the initial pool size is 100.
WRGUGDJS
W60095422
R75768642
GU48550781
GD261156271
J02235750
S31112598
Table 7. Confusion matrix for DMGSOSEN-based HAR on the data from the chest when the initial pool size is 100.
Table 7. Confusion matrix for DMGSOSEN-based HAR on the data from the chest when the initial pool size is 100.
WRGUGD JS
W592119332
R1257212872
GU665411183
GD 481255654
J13245731
S12112592
Table 8. Confusion matrix for DMGSOSEN-based HAR on the data from the right arm when the initial pool size is 100.
Table 8. Confusion matrix for DMGSOSEN-based HAR on the data from the right arm when the initial pool size is 100.
WRGUGD JS
W577171614810
R145491018123
GU111653212132
GD61110523116
J446125473
S45347580
Table 9. Confusion matrix for DMGSOSEN-based HAR on the data from the left ankle when the initial pool size is 100.
Table 9. Confusion matrix for DMGSOSEN-based HAR on the data from the left ankle when the initial pool size is 100.
WRGUGDJS
W567211621166
R14535191792
GU151851723134
GD101815503193
J6714155389
S43243580
Table 10. Confusion matrix for DMGSOSEN-based HAR on the data from the left wrist when the initial pool size is 100.
Table 10. Confusion matrix for DMGSOSEN-based HAR on the data from the left wrist when the initial pool size is 100.
WRGUGDJS
W56236121777
R135282121172
GU109514242212
GD262022495167
J2562253311
S34243565
Table 11. Performance comparison with Adaboost and Bagging on 50 ELMs (Accuracy/F1%).
Table 11. Performance comparison with Adaboost and Bagging on 50 ELMs (Accuracy/F1%).
Position50
AdaboostnBaggingnDMGSOSENn
waist88.2/88.45086.4/86.55095.9/95.89
chest85.8/85.65084.1/84.25094.4/94.411
right arm79.5/79.45078.4/78.45091.8./91.714
left ankle83.8/83.75082.4/82.35090.2/90.417
left wrist84.8/84.75083.5/83.45089.1/89.215
Table 12. Performance comparison with Adaboost and Bagging on 100 ELMs (Accuracy/F1%).
Table 12. Performance comparison with Adaboost and Bagging on 100 ELMs (Accuracy/F1%).
Position100
AdaboostnBaggingnDMGSOSENn
waist88.6/88.510086.9/86.910096.7/96.716
chest86.4/86.210085.2/85.310095.7/95.719
right arm79.7/79.810079.6/79.610092.4/92.425
left ankle84.2/84.310083.3/83.410090.5/90.422
left wrist84.9/84.910084.2/84.310089.3/89.323
Table 13. Performance comparison with Adaboost and Bagging on 150 ELMs (Accuracy/F1%).
Table 13. Performance comparison with Adaboost and Bagging on 150 ELMs (Accuracy/F1%).
Position150
AdaboostnBaggingnDMGSOSENn
waist89.1/89.215087.3/87.215095.5/95.431
chest86.2/86.215085.4/85.315095.2/95.133
right arm79.8/79.915078.8/78.915092.1/92.242
left ankle84.8/84.615083.5/83.315090.3/90.438
left wrist84.9/84.815084.4/84.215089.2/89.336
Table 14. Performance comparison with Adaboost and Bagging on 200 ELMs (Accuracy/F1%).
Table 14. Performance comparison with Adaboost and Bagging on 200 ELMs (Accuracy/F1%).
Position200
AdaboostnBaggingnDMGSOSENn
waist89.4/89.520087.2/87.320094.2/94.338
chest86.8/86.720085.1/85.220093.9/93.543
right arm79.6/79.420079.4/79.420092.7/92.552
left ankle84.3/84.420083.8/83.720090.1/90.248
left wrist84.2/84.220083.5/83.620088.9/88.951
Table 15. Performance comparison (Accuracy/F1%) and number of ELMs after pruning achieved by comparative algorithms.
Table 15. Performance comparison (Accuracy/F1%) and number of ELMs after pruning achieved by comparative algorithms.
PositionDMGSOSENnAGOBnPOBEn
waist96.7/96.71688.6/88.62685.2/85.329
chest95.7/95.71986.5/86.43284.6/84.635
right arm92.4/92.42580.3/80.33879.4/79.438
left ankle90.5/90.42284.6/84.53484.2/84.235
left wrist89.3/89.32382.8/82.33581.7/81.732
Table 16. Performance comparison (Accuracy/F1%) and number of ELMs after pruning achieved by comparative algorithms.
Table 16. Performance comparison (Accuracy/F1%) and number of ELMs after pruning achieved by comparative algorithms.
PositionD-D-ELMnDF-D-ELMnGASENn
waist89.3/89.55089.7/89.64887.4/87.536
chest88.5/88.55085.3/85.35284.5/84.543
right arm81.3/81.45074.5/75.45776.6/76.638
left ankle84.3/84.35082.5/82.55185.2/85.341
left wrist82.9/82.95081.8/81.85184.7/84.744
Table 17. Performance comparison (Accuracy/F1%) and number of ELMs after pruning achieved by comparative algorithms.
Table 17. Performance comparison (Accuracy/F1%) and number of ELMs after pruning achieved by comparative algorithms.
PositionMOAGnRREnDivPn
waist81.2/81.32783.3/83.41989.4/89.311
chest80.6/80.62982.4/82.42387.3/87.317
right arm75.3/75.23574.5/74.53180.2/80.223
left ankle78.4/78.42579.8/79.82484.3/84.122
left wrist76.2/76.32879.1/79.22683.2/83.320
Table 18. Comparison with some previous studies in HAR.
Table 18. Comparison with some previous studies in HAR.
AuthorMethodPerformance (ACC/F1%)
Wang et al. [12]EEMD+FS+SVM81.2/-
Yuan et al. [17]ACELM95.02/-
Xu et al. [41]CELearning95.1/-
Ronao et al. [62]tFFT+Convnet95.75/-
Hassan et al. [63]KPCA+DBN95.85/-
Our proposed methodELM+DMGSOSEN96.7/96.7

Share and Cite

MDPI and ACS Style

Tian, Y.; Zhang, J.; Chen, L.; Geng, Y.; Wang, X. Selective Ensemble Based on Extreme Learning Machine for Sensor-Based Human Activity Recognition. Sensors 2019, 19, 3468. https://doi.org/10.3390/s19163468

AMA Style

Tian Y, Zhang J, Chen L, Geng Y, Wang X. Selective Ensemble Based on Extreme Learning Machine for Sensor-Based Human Activity Recognition. Sensors. 2019; 19(16):3468. https://doi.org/10.3390/s19163468

Chicago/Turabian Style

Tian, Yiming, Jie Zhang, Lingling Chen, Yanli Geng, and Xitai Wang. 2019. "Selective Ensemble Based on Extreme Learning Machine for Sensor-Based Human Activity Recognition" Sensors 19, no. 16: 3468. https://doi.org/10.3390/s19163468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop