Next Article in Journal
Skin Pigmentation Influence on Pulse Oximetry Accuracy: A Systematic Review and Bibliometric Analysis
Next Article in Special Issue
Past, Present and Future of Research on Wearable Technologies for Healthcare: A Bibliometric Analysis Using Scopus
Previous Article in Journal
A Systematic Literature Review on Machine and Deep Learning Approaches for Detecting Attacks in RPL-Based 6LoWPAN of Internet of Things
Previous Article in Special Issue
Hybrid-Pattern Recognition Modeling with Arrhythmia Signal Processing for Ubiquitous Health Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Human Activity Recognition Data Analysis: History, Evolutions, and New Trends

by
Paola Patricia Ariza-Colpas
1,2,*,
Enrico Vicario
3,
Ana Isabel Oviedo-Carrascal
2,
Shariq Butt Aziz
4,
Marlon Alberto Piñeres-Melo
5,
Alejandra Quintero-Linero
6 and
Fulvio Patara
3
1
Department of Computer Science and Electronics, Universidad de la Costa CUC, Barranquilla 080002, Colombia
2
Faculty of Engineering in Information and Communication Technologies, Universidad Pontificia Bolivariana, Medellín 050031, Colombia
3
Department of Information Engineering, University of Florence, 50139 Firenze, Italy
4
Department of Computer Science and IT, University of Lahore, Lahore 44000, Pakistan
5
Department of Systems Engineering, Universidad del Norte, Barranquilla 081001, Colombia
6
Microbiology Program, Universidad Popular del Cesar, Valledupar 200002, Colombia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(9), 3401; https://doi.org/10.3390/s22093401
Submission received: 31 October 2021 / Revised: 31 March 2022 / Accepted: 4 April 2022 / Published: 29 April 2022

Abstract

:
The Assisted Living Environments Research Area–AAL (Ambient Assisted Living), focuses on generating innovative technology, products, and services to assist, medical care and rehabilitation to older adults, to increase the time in which these people can live. independently, whether they suffer from neurodegenerative diseases or some disability. This important area is responsible for the development of activity recognition systems—ARS (Activity Recognition Systems), which is a valuable tool when it comes to identifying the type of activity carried out by older adults, to provide them with assistance. that allows you to carry out your daily activities with complete normality. This article aims to show the review of the literature and the evolution of the different techniques for processing this type of data from supervised, unsupervised, ensembled learning, deep learning, reinforcement learning, transfer learning, and metaheuristics approach applied to this sector of science. health, showing the metrics of recent experiments for researchers in this area of knowledge. As a result of this article, it can be identified that models based on reinforcement or transfer learning constitute a good line of work for the processing and analysis of human recognition activities.

1. Introduction

Currently, the number of older adults who require a caregiver due to various conditions related to neurodegenerative diseases has greatly increased. This situation constitutes a great problem both for society and for integrated health systems worldwide because there is not enough infrastructure to be able to massively attend to the increasing number of people with this type of condition. Due to the above, a line of research arises that relates the sensory and the processing of HAR (Human Activity Recognition) data, which allows supporting the management of these individuals.
In general, this type of experimentation uses a model as a representation of reality developed to study it. In most analyzes it is not necessary to consider all the details of reality, the model is not only a substitute for reality but also a simplification of it. According to the same author, the models are classified as iconic, analogous, symbolic, deterministic, stochastic, static, continuous, and discrete depending on the tools used [1]. On the other hand, Cramér [2] from the foundation of probability theory, specifies from the object of probability theory that a mathematical model makes a description of a certain class of observed phenomena.
Artificial intelligence (AI) is defined as a field of science and engineering concerned with the computational understanding of what is commonly called intelligent behavior with the creation of artifacts that exhibit such behavior [3]. These processes require the training of models from data sources, in some cases heterogeneous. Training with a learning view involves the acquisition and pre-processing of information and the application or construction of rules for the treatment and use of this, to generate reasoning. That is, it uses the rules to reach approximate or definitive conclusions and self-correction [4].
Artificial Intelligence has become more popular today thanks to Big Data, advanced algorithms, and computers with improved power and storage, systems based on artificial intelligence are becoming an integrated element of digital systems and more specifically they are generating a profound impact on human decision-making. As a result, there is a growing demand for information systems researchers to investigate and understand the implications of this, for decision making and contribute to the theoretical advancement and practical success of the applications of this area of knowledge [5].
In 1959, Arthur Samuel coined the term Machine Learning and defined it as “the field of study that gives computers the ability to learn without being explicitly programmed”. Machine learning is part of the field of Artificial Intelligence, and its objective is usually to recognize and fit statistics to models [6]. Along with Artificial Intelligence, Machine Learning has emerged as the method of choice for the development of practical software for image and speech recognition, natural language processing, robot control, and other applications like Human Activity Recognition (HAR). Many developers of Artificial Intelligence systems recognize that, for many applications, it may be easier to train a system by feeding it examples of the desired input and output behavior, than to manually program in advance the desired response for all possible inputs [7].
Machine Learning has been playing, in recent decades, an important role in the construction of models based on experience from processed data [8], enabling computers to build models from data For example, according to the automation of decision-making processes, based on the input data [6], for their part [8] affirms that the study and construction of algorithms are explored who can learn and make predictions from data. This systematic review of the literature locates the advances made in Human Activity Recognition in each of the automatic learning methods, their evolution, and results.
The recognition of human activities has become one of the most used areas of knowledge that has allowed many advances in the care of patients at home and the improvement of the quality of life of the elderly. That is why different ways have been used by which the data from the different datasets can be processed, among which are machine learning and metaheuristics. The HAR approach is based on the complexity associated with the different data inputs that can come from wearable sensors, object sensors, images, audio, among others. Many models have been developed to try to improve performance and quality metrics based on different experimentations. Motivated by this research eld, the main contribution of this paper is (a) Show researchers the datasets most used in the literature for experimentation processes, as well as detailing the most used algorithms in their analysis. (b) Identify for each one of the data processing approaches the results of the experimentations of the algorithms, as well as discriminate the quality metrics associated with said applications an (c) Suggest, based on the analysis of the literature, the different techniques to be used to obtain good results in the experiments and show future researchers what the results of current applications are for the improvement of their experiments.
For the development of this purpose, a compilation of 570 articles has been analyzed, extracted from specialized databases such as IEEE Xplorer, Scopus, Science Direct, WOS. These manuscripts were analyzed through a meta-analytic matrix that allowed extracting relevant information such as year of publication, the database used, techniques implemented, results of the quality metrics implemented.
This article is a review of the literature on the use of the technique of the different machine learning methods supervised, unsupervised, ensembled, deep learning, reinforcement learning. First, a concept map of HAR Approach Concept Maps is shown (Section 2). Second, conceptual information is displayed (Section 3). Third, the methodology for analyzing the information sources is detailed (Section 4). Fifth, is the Scientometric Analysis (Section 5). Sixth, Technical Analysis (Section 6). Seventh, Conclusions (Section 7). Finally, the Future Works are shown (Section 8).

2. HAR Approach Concept Maps

In the last decades, Machine Learning has evolved in different methods and techniques to address different challenges in different areas of knowledge. In Figure 1 you can see the discrimination of each of the data mining methods can be seen. From classic supervised or unsupervised-based machine learning. Among the most outstanding algorithms of Supervised Analysis (Section 3.1), the following can be highlighted: Decision Tree [9], Support Vectorial Machine [10], Naive Bayesian Classifier [11], Artificial Neural Networks [12], Decision Tables [13] and Logistic Models [14], among others. Regarding Unsupervised Learning (Section 3.2), several methods can be found, among which Clustering [15] and Association Rules [16], and Dimensionality Reduction [17] can be highlighted. As for Ensembled Learning (Section 3.3), techniques such as Stacking [18], Bagging [19], Boosting [20] can be highlighted. Later, it was emphasized in methods or techniques based on Deep Neural Networks (Section 3.4) [21] that have several levels of analysis for knowledge discovery. Nowadays, machine learning has evolved to analysis based on Reinforcement Learning (Section 3.5) [22], which allows the algorithm that is strengthened in a system of rewards and punishments to permeate the learning process. In (Section 3.6) Metaheuristic Techniques [23], are strategies for designing heuristic procedures. Therefore, the types of metaheuristics are established, in the first place, based on the type of procedures to which they refer. The following can be identified among other types of algorithms: Threshold Accepting [24], Memetic algorithms [25], MultiBoot Algorithms [26], CRO (Coral reef-based algorithms) [27], Swarm algorithms [28], Genetic algorithms [29], Scatter Search [30], Variable Neighborhood Search [31] and Ant Colony [32]. Finally, in (Section 3.7), we show the approach of Transfer Learning [33] to Human Activity Recognition using a different type of combination of Neural networks for the analysis.

3. Conceptual Information

3.1. Supervised Learning

Supervised learning is a technique for deducing a function from training data. Training data consists of pairs of objects (usually vectors): one component of the pair is the input data and the other is the desired results. The output of the function can be a numeric value (as in regression problems) or a class label (as in classification problems). The goal of supervised learning is to create a function capable of predicting the value corresponding to any valid input object after viewing a series of examples, the training data. To do this, you must generalize from the data presented to previously unseen situations. Among the techniques most used in machine learning, the following can be highlighted.

3.1.1. Decision Tree

According to Timaran [34], the quality of the decision tree depends on the size and precision of the classification. A subset of the dataset set (training) is chosen and a decision tree is created [9]. If it does not return the answer for the objects in the test set, a selection of exceptions is added to the dataset set, continuing the process until the correct decision set is found. The most used classification algorithms, from the decision trees category, are ID-3, C4.5, CART, Sprint, and j48 [34].

3.1.2. Support Vectorial Machine (SVM)

In SVMs [10] the input quantities are mapped non-linearly to a very high-dimensional feature space. In this feature space, a line decision surface is constructed [35]. According to Hang [36], SVM uses a non-linear mapping to transform the original training data into a higher dimension. Within this new dimension, it looks for the linear optimal separation hyperplane (that is, a “decision limit” that separates the tuples of one class from another). SVMs can be used for numerical prediction, as well as for classification. They have been applied to several areas, including handwritten digit recognition, object recognition, and speaker identification, as well as benchmark time series prediction tests.

3.1.3. Naïve Bayesian Classifier

Is a special typology of machine learning algorithms that address the task of classification [11]. The foundation of this is based on the “Bayes theorem”. In this algorithm, it is assumed that the variables that are used for the prediction are independent of each other. In other words, the presence of a series of characteristics in a data set is not related to the absence of another character in another data set.

3.1.4. Artificial Neural Networks-ANN

The fundamental processing elements of an ANN are artificial neurons (or nodes) that are interconnected by weighted links that form layers [12]. Normally in an ANN, there is an input layer and an output layer, and several hidden layers that vary depending on the complexity of the problem in question. Neurons transform weighted input into output, using an activation function that can take different linear and non-linear forms. The process by which the weights are adjusted is called learning. Several non-linear ANNs are known to perform function approximators. Several parameters define the architecture of a neural network: the type of connection, learning rule, and activation functions. Due to these conformation parameters, there are different types of ANN, for example, Multilayer Perceptron—MLP [37], Echo State Networks-ESN, radial basis function-RBFN, Boltzmann machine.

3.1.5. Decision Tables

The decision tables or also called decision rules achieve a synthetic representation of knowledge [13]. There are at least four sources of inconsistencies in decision tables, listed below: (1) hesitation in evaluating decision attribute values, (2) errors in recording, measurement, and observation, (3) condition attributes missing related to the evaluation of the decision attribute values, (4) the unstable nature of the system represented by the decision table and the like. These inconsistencies cannot be considered simple errors or noise. To acquire rules, from inconsistent decision tables, relative attribute reductions are needed. Skowron and Rauszer introduced the discernibility matrix method which became a popular approach for listing all reductions in the Rough set [26].

3.1.6. Tree-Based on the Logistic Model-LMT

This classification process mixes decision trees with logistic regression [14]. The classification process can be improved if characteristics selection techniques are used, these allow assigning the prioritization or relevance of the attributes using the class criterion, thus obtaining a structure of attributes that directly affect the model and that in turn. are increasingly relevant concerning classification.

3.2. Unsupervised Learning

Unsupervised learning is a Machine Learning method where a model is fit for observations. It is distinguished from supervised learning by the fact that there is no a priori knowledge. In unsupervised learning, a data set of input objects is processed. Thus, unsupervised learning typically treats input objects as a set of random variables, with a density model being constructed for the data set. There are different unsupervised learning methods, among which we can highlight: Clustering, Association Rules, and Dimensionality Reduction.

3.2.1. Clustering Methods

In the last decades, many are algorithms of grouping have been proposed and developed [38,39], from the hierarchical approach (Single Link, Complete Link, etc.) and partition (K-means, Gaussian Mixture, Density Estimation and Mode Seeking, etc.) among other methods. As data sets get larger and more varied, many of the dimensions are often irrelevant. These irrelevant dimensions can confuse traditional clustering algorithms.
Clustering is a technology used for many purposes because it simplifies massive data by extracting essential information, based on the relationship of subsequent analyzes or processes that make the process feasible or more efficient. For example, in information systems, grouping is applied to text documents or images to speed up indexing and retrieval [40,41]. Clustering can also be a stand-alone process and has been used as a technique for prototype-based supervised learning algorithms and different applications have also been made in non-vector data. The application of Clustering algorithms for the analysis of unsupervised data has become a useful tool to explore and solve the different application problems of data mining. Clustering methods [39,42] have been used to solve problems emanating from different contexts and disciplines, see Table 1.
Table 1. Clustering’s methods and applications.
Table 1. Clustering’s methods and applications.
MethodStrategyApplications
HierarchicalAgglomerativeNearest Neighbor [43]
Farthest Neighbor [44]
Average Linkage Pool [45]
Minimum Variance [46]
Median Method [47]
Divisive
Non-HierarchicalReassignmentCentroidsK-means [48], QuickCluster [49], Forgy Methods [50]
Medioidk-medioids [51], Clara [52]
DensityDynamic clouds [53]
Typological approximationModal Analysis [54], Taxmap Method [55], Fortin Method [56]
Probabilistic approximationWolf Methods [57]
DirectBlock Clustering [58]
ReductiveType Q Factor Analysis [59]

3.2.2. Association Rules Methods

The association rules base their analysis on the “if-then” algorithmic sentences, which allow supporting the different probabilities existing in the multiple elements of the data, found in large databases of different formats and types. The data mining techniques that are based on association rules throughout their evolution have had multiple applications, among which the sales and analysis of medical data sets can be highlighted.
Based on the algorithmic “if-then” sentences and based on established criteria such as support and trust, the association rules can identify the most important patterns. The support criterion gives the association rules the ability to know the frequency of the appearance of the elements in the data set is. As for the confidence criterion, it can determine the number of times the Boolean value of the “if-then” statement is true. There is also another common metric which is called Fit, which is fundamentally based on making a comparison between the expected confidence and the confidence that can be evidenced in the data. In the literature review, the progress of the association rules can be identified, as detailed below, see Table 2.
Table 2. Association Rules Evolutions.
Table 2. Association Rules Evolutions.
Based inAlgorithms
Frequent Itemsets MiningApriori [60]
Apriori-TID [61]
ECLAT TID-list [62,63]
FP-Growth [64]
Big Data AlgorithmsR-Apriori [65]
YAFIM [66]
ParEclat [67]
Par-FP (Parallel FP-Growth with Sampling) [68]
HPA (Hash Partitioned Apriori) [69]
Distributed algorithmsPEAR (Parallel Efficient Association Rules) [70]
Distributed algorithms for fuzzy association rule miningCount Distribution algorithm [71,72]

3.2.3. Dimensionality Reduction Methods

Dimensionality reduction methods are statistical techniques that map the data set to subspaces derived from the original space, of less dimension, which allow a description of the data at a lower cost. These techniques become important as many algorithms from various fields such as numerical analysis, machine learning or data mining tend to degrade their performance when used with high dimensional data. In external cases, the algorithm is no longer useful for the purpose for which it was designed. The curse of dimension refers to the various phenomena that arise when analyzing and organizing data from multi-dimensional spaces. Among the most important algorithms we can highlight.
Missing Values Ratio [73]: By examining the data, if we find that it contains many missing values, if there are few missing values, we can fill in the missing values or remove this variable directly; when the proportion of missing values in the dataset is too high, I usually choose directly Remove this variable because it contains too little information. Specific removal is not removed, how to remove depends on the situation, we can set a threshold, if the proportion of missing values is greater than the threshold, remove the column where it is. The higher the threshold. The more aggressive the dimensionality reduction method.
Low Variance Filter [74]: If the value of a column is the same in a dataset, that is, its variance is very low, we generally think that low-variance variables contain very little information, so you can eliminate it directly and put it into practice, which is to calculate all Variation Size variables and then eliminate the smallest of them.
High Correlation Filter [75]: If the two variables are highly correlated, this means that they have similar trends and can carry similar information. Similarly, the presence of such variables can reduce the performance of certain models (such as linear and logistic regression models). To solve such problems, we can calculate the correlation between independent variables. If the correlation coefficient exceeds a certain threshold, one of the variables is eliminated.
Random Forests/Ensemble Trees [76]: Random Forest is a widely used feature selection algorithm, it automatically calculates the importance of each feature, so no separate programming is required. This helps us choose a smaller subset of features. The advantages of the random forest: high precision, the introduction of randomness makes random forests not easy to overfit, the introduction of randomness makes the random forests have good anti-noise ability (can better handle outliers), can handle very high-dimensional data without feature selection, it can handle both discrete data and continuous data, and the data set does not need to be normalized, fast training speed, you can get the importance of variable classification and easy to parallelize. Disadvantages of the random forest: when there are many decision trees in the random forest, the space and time required for training will be large, and the interpretability of the random forest is poor.
Principal Component Analysis (PCA) [77]: PCA is a very common dimensionality reduction method. You can reduce the number of predictors by reducing the dimensionality of high-dimensional data while eliminating noise through dimensionality reduction. The most direct application is to compress data, mainly used in signal processing Noise reduction, and visualization after data dimensionality reduction.

3.3. Ensemble Learning

An ensemble is a set of machine learning models. Each model produces a different prediction. The predictions from the different models are combined to obtain a single prediction. The advantage we get from combining different models is that because each model works differently, its errors tend to be compensated for. This results in a better generalization error.

3.3.1. Voting by the Majority

Training multiple machine learning models with the same data [78]. When we have new data, we will get a prediction from each model. Each model will have a vote associated with it. In this way, we will propose as a final prediction what most of the models vote for. There is another way to combine voting. When machine learning models give a probability, we can use “soft-voting”. In soft voting, more importance is given to results in which some model is very confident. That is, when the prediction is very close to probability 0 or 1, more weight is given to the prediction of that model.

3.3.2. Bagging

Unlike majority voting, the way to get errors to compensate for each other is that each model is trained with subsets of the training set [79]. These subsets are formed by randomly choosing samples (with repetition) from the training set. The results are combined, for classification problems, as we have seen in majority voting, with the soft vote for the models that give probabilities. For regression problems, the arithmetic mean is normally used.

3.3.3. Boosting

In boosting, each model tries to fix the errors of the previous models [80]. For example, in the case of classification, the first model will try to learn the relationship between the input attributes and the result. You will surely make some mistakes. So, the second model will try to reduce these errors. This is achieved by giving more weight to poorly classified samples and less weight to well-classified samples. For regression problems, predictions with a higher mean square error will have more weight for the next model.

3.3.4. Stacking

When we talk about a stacking ensemble, we mean that we are stacking models [81]. When we stack models, what we are doing is using the output of multiple models as the input of multiple models.

3.4. Deep Learning

Deep Learning is a type of machine learning that is structured and inspired by the human brain and its neural networks [82]. Deep learning processes data to detect objects, recognize conversations, translate languages, and make decisions. Being a type of machine learning, this technology helps artificial intelligence learn continuously. Deep learning is based on the use of artificial neural networks. Within neural networks 3 types are the most used.

3.4.1. Convolutional Neural Networks (CNN)

Convolutional neural networks are artificial neural networks that have been designed to process structured matrices, such as images [83]. That is, they are responsible for classifying images based on the patterns and objects that appear in them, for example, lines, circles, or even eyes and faces.

3.4.2. Recurrent Neural Networks (RNN)

Recurrent neural networks are neural networks that use sequential data or time-series data [84]. These types of networks solve ordinal or temporal problems, such as language translation, speech recognition, Natural Language Processing (NLP, Natural Language Processing), and image capture. Therefore, these networks are in technologies such as Siri or Google translate. In this case, natural language processing recognizes a person’s speech. For example, it is distinguished if the person who is speaking is a man or a woman, an adult or a minor, if they have an Andalusian or Catalan accent, etc. In this way, the person’s way of speaking is analyzed, and their idiolect is reached.

3.4.3. Generative Adversarial Networks (GAN)

The antagonistic generative networks consist of using 2 artificial neural networks and opposing them to each other (that is why they are known as antagonistic) to generate new content or synthetic data that can be passed as real [85]. One of the networks generates and the other works as a “discriminator”. The discriminatory network (also known as an antagonistic network) has been trained to recognize real content and acts as a sensor for the network that generates content to make content that appears real.

3.5. Reinforcement Learning

The field of machine learning is the branch of Artificial Intelligence that encompasses techniques that allow machines to learn through their environment. This environment can be considered as the set of data that the algorithm has or obtains in the training stage. Reinforcement learning is the most common in nature. An individual has a connection with the environment with which he obtains information from the cause-effect relationships, the results of the actions carried out, and the strategy to follow to complete an objective [86].
The time difference method was introduced by Sutton [87] as a model-free method based on a bootstrapping update rule and consists of estimating the values of immediate and future rewards in a way like programming. Are dynamic and are denoted as TD (λ). Methods of time difference attempt to estimate the value function of a given state of a policy, and contrary to Monte Carlo methods, you do not need to wait at the end of an episode to make such an estimate. Some prominent algorithms are.

3.5.1. SARSA

One of the algorithms that derive from the method based on time difference is the SARSA algorithm [88] which is an on-policy method, that is, it has an initial policy and updates it at the end. of each episode.

3.5.2. Q-Learning

Q-learning is a value-based learning algorithm that focuses on optimizing the value function according to the environment or problem [89,90]. The Q in Q-learning represents the quality with which the model finds its next quality-improving action. The process can be automatic and simple. This technique is great to start your reinforcement learning journey. The model stores all the values in a table, which is Table Q. In simple words, the learning method is used for the best solution.

3.5.3. Deep Reinforcement Learning

Deep Reinforcement Learning [91], where reinforcement learning is integrated with neural networks. The DeepMind company began to use this type of learning to create agents that would learn to play Atari games from scratch without having any information about them, not even the rules of the video game.

3.6. Metaheuristic Learning

Metaheuristics are clever strategies to design or improve very general heuristic procedures with high performance. The term metaheuristic first appeared in Fred Glover’s seminal article on tabu search in 1986 [92]. Since then, many proposals for guidelines have emerged to design good procedures to solve certain problems that, by expanding their field of application, have adopted the denomination of metaheuristics.
Some of the main types are: Relaxation metaheuristics [93] refer to problem-solving procedures that use relaxations of the original model (that is, modifications of the model that make the problem easier to solve), the solution of which facilitates the solution of the original problem. The constructive metaheuristics [94] are oriented to the procedures that try to obtain a solution from the analysis and gradual selection of the components that form it. Search metaheuristics [95] guide procedures that use transformations or moves to traverse the space of alternative solutions and exploit the associated environment structures. Evolutionary metaheuristics [96] are focused on procedures based on solution sets that evolve over the solution space.

3.7. Transfer Learning

Deep Learning primarily emphasizes features, Reinforcement Learning primarily emphasizes feedback, and Transfer Learning primarily emphasizes adaptation. Traditional machine learning is about reaping the benefits of planting fruits and reaping the benefits of planting beans, while transfer learning can draw inferences from each other.
Artificial intelligence competition, from algorithm model development to data quality and data competition, these successful models and algorithms are mainly driven by supervised learning, and supervised learning consumes a lot of data and requires big data support (big data) to meet the precise requirements of the application. The development of artificial intelligence tends to satisfy the precise requirements of the applications without requiring massive data. Therefore, “small data learning” is becoming a new point of interest. Small data learning techniques represented by migration learning and reinforcement learning can better reflect artificial intelligence.
Since the transfer learning (TL) concept was proposed by Stevo Bozinovski and Ante Fulgosi in 1976 [97], it has received a great deal of attention from the academic community. The definition of transfer learning is too broad and a variety of specialized terms have appeared in related research, such as learning to learn, lifelong learning, multitasking learning, meta-learning, inductive transfer, knowledge transfer, context-sensitive learning, etc. Among them, transfer learning has the closest relationship with multitasking learning. Multitask learning learns multiple different tasks at the same time and discovers implicit common features to aid single-task learning.

3.8. Human Activity Recognition

Recognizing human activities consists of interpreting human gestures or movements through sensors to determine human action or activity [98]. For example, a HAR system can report activities performed by patients outside of hospital facilities, which makes it a useful tool for evaluating health interventions and therapy progress, and for clinical decision-making [99]. HAR can be supervised or unsupervised. The supervised HAR system requires prior training with a tagged data set, on the contrary, the unsupervised system does not require training but has a set of rules configured during development. In this particular work, we focused on a HAR system of the supervised type to recognize the following six human activities: walking (WK), climbing stairs (WU), descending stairs (WD), standing (ST), lying down (LD), and being sitting (SD). We name, in particular, the WK, WU, and WD activities as dynamic activities since they involve a voluntary movement that causes displacement and is reflected in the inertial sensors, and we call ST, LD, and SD activities. Given that they do not involve voluntary movements of the subject and there is no displacement of the person.
In HAR systems it is common to use signals and images that come from sensors that can be located in a specific physical space, such as in a room, or that can be placed or carried by people, like those found in smart cell phones or smartwatches. Smartphones are mobile phones that can perform tasks like those of a computer, such as the ability to store and process data and be able to navigate the Internet [100]. In addition, compared to personal computers, smartphones are widely accepted due to their small size, low weight, more personal device, and, especially, great connectivity that allows you to access at any time and place to information sites and social networks [101]. Other applications that are usually present are integrated cameras, contact management, multimedia software capable of playing music and being able to view photos and videos, and the use of navigation programs, and, in addition, having the ability to view business documents in different formats such as PDF and Microsoft Office [101].
Currently, different sensors are installed, such as positioning sensors, proximity sensors, temperature sensors, accelerometer, gyroscope, magnetometer, microphone, etc., as shown in Figure 2. This is currently a challenge carried out by different scientific communities, particularly in the fields of computer vision, signal processing, and machine learning. The sensors are usually operated by a microcontroller or microprocessor, which performs the function of a computer.
Inertial sensors are sensors based on the principle of inertia, the tendency of a body to conserve its speed (in the absence of an external influence, a body remains in a uniform rectilinear motion). There are different types of sensors to measure signals that can be used by HAR systems. Two of the most used are the accelerometer and the gyroscope. The accelerometer measures the acceleration (in meters per second squared, m/s2) based on the different variations that a capacitance makes inside the sensor. This capacitance is a microelectromechanical system (MEMS for its acronym in English microelectromechanical systems) that consists of the suspension of silicon particles that are located at a fixed point and are moved freely in the axis where they are measured. When acceleration occurs, the particles move and break with equilibrium in capacitance; this is measured to provide the information that occurs in a certain axis.
According to the type of sensors and the occupation of the indoor environments, a series of datasets have been built that have served to carry out different experiments based on machine learning techniques. The most prominent datasets are: UCI HAR [102], KU-HAR [103], Precis HAR [104], Fall-Up Dataset [105], VanKasteren [106], CASAS Multiresident [107], WISDM [108], DOMUS [109], Opportunity [110], CASAS Aruba [111], USC-HAD [112], MIT PlaceLab [113], Smart Environment—Ulster University [114], CASAS–Daily Life Kyoto [115], PAMAP2 [116], mHealth [117], DSADS [118], UJAmI SmatLab [119].

4. Methodology for Analyzing the Information

The methodology for the analysis of the publications is supported and defined by Kitchenham [120]. This methodology consists of identifying the main research problem, and then disaggregating each of its components by analyzing the different inclusions and exclusions, to determine a suitable search string to be used in scientific databases. Specifically for our case study, in addition to the Scientometric type variables, those related to the type of dataset used, techniques or algorithms implemented, as well as the quality of the results measured by the quality metrics, were identified. Kitchenham [120] defines different stages of the literature review process, among which the following can be highlighted: (a) Identification of the search parameters (search objectives, hypotheses identified) (b) Definition of search engines (selection of specialized databases where the study is to be developed) (c) Response to the hypotheses that were raised for the literature inquiry process.
By these previously defined phases, the first thing to do is to identify the central question of the inquiry process. For this literature review, it would be “What are the different techniques based on Machine Learning that support the analysis of dataset recognition of human activities?”. To carry out the literature review, the IEEE, Scopus, Science Direct, and WOS databases were used. To delimit the documentary findings, the following search string was used: (HAR OR ADL OR AAL) AND dataset AND (“indoor environment’’ OR “smart homes” OR “intelligent buildings” OR “ambient intelligence” OR “assisted living’’). In Figure 3 you can see the basic concept scheme for the review document filter. Then the references were analyzed by the machine learning technique that is implemented, which is described in Section 6.
It is important to specify that the order of the different terms that are observed in Figure 3 determine all those that are part of the domain of knowledge, which was previously tested in the different search engines of the scientific databases to eliminate the different noises from them. that can be generated at the time of the search and the exclusion of papers not related to the study area. Taking into account the previously explained methodology, different factors of the analytical order and high importance for those interested in this area of knowledge were described in the meta-analytic matrix, such as year of publication of the work (which is not greater than a window of 5 years), journal, conference or book where the publication was made, quartile in the case of publications in journals, country of origin of the first author as well as the university or research center. Other technical variables are taken into account in the same way for the development of this research, such as Name of the dataset, type of data collection, type of activities carried out, several individuals who define the occupation, data mining techniques used, hybridization of techniques, results of quality metrics.

5. Scientometric Analysis

In the results obtained from the 570 articles processed, different relevant variables were taken into account, among which are detailed: (1) the year of publication of the article see Figure 4, (2) the database where the publication can be found, (3) Type of publication if it is a journal, conference or book, (4) Quartile of the journal in the case of publications in this medium, (5) country of origin of the journal (6) country of origin of the first author of the article (7) University of the first author, (8) Dataset used for the experiments, (9) Techniques used for the discovery of information and (10) results of the metrics of each technique.
It can be identified that 2018 was where the most publications were generated in HAR’s line of work. In the same way, when discriminating the databases in which the publications are made, it is highlighted that most of the works have been published in the Science Direct database, then Scopus. Some publications are visible in different databases, as shown in Figure 5.
Of the total articles, analyzed 64% of them refer to conference publications, 4 & are books and 36% refer to journals, see Figure 6a,b.

6. Technical Analysis

6.1. Supervised Learning Applied to Human Activity Recognition Dataset

Regarding the application of Machine Learning techniques, to Human Activity Recognition Dataset, various experiments have been developed, but the most relevant ones found in the literature are highlighted below (see Table 3). Tasmin [121], carried out implementations in the UCI-HAR Dataset, through the implementation of supervised algorithms Nearest Neighbor, Decision Tree, Random Forest, and Naive Bayes, of the techniques used, the one with the best results in the detection of activities was the Bayesian with an accuracy of 76.9%. Igwe [122], concentrated his experimentations on the ARAS Data-set which was implemented in 2 different locations (House A and House B), CA-SAS Tulum created by WSU University, the author applied supervised techniques such as SVM, ANN, and MSA (Margin Setting Algorithm), demonstrating the effectiveness of the latter in identifying activities with an accuracy of 68.85%, 96.24% and 68% in the respective Datasets.
Subasi [123], performed analysis on the Meath Dataset, applying techniques such as K-NN, ANN, SVM, C4.5, CART. Random Forest and Rotation Forest obtained better results with SVM and Random Forest with 99.89%. Maswadi [124], firstly I prepare the Dataset using Sliding window segmentation techniques with a variable size in different Datasets such as WISDM with SCUT_NA-A, SCUT NA-.An only, PAMPA2 with Mhealth, SBHAR, WISDM, UTD-MHAD, Groupware, Free-living WISDM with Skoda, UniMiB SHAR, and Groupware, showing the superiority of this technique obtaining results greater than 80% accuracy. Other authors such as Damodaran [125], applied SVM, LSTM to the CSI-Data Dataset, where better results are shown in the use of SVM with 96%.
Other authors such as Saha [126] and Das [127], define the characteristics and process for the construction of their Dataset, to which a set of techniques are applied and it should be noted that both authors show that vector support machines show efficiency in the results of classification of human activities. Franco [128], uses techniques such as FFNN, SVM, and LSTM in the UK-Dale Dataset, showing the effectiveness of FFNN with 95.28% accuracy in quality metrics.
Bozkurt [129] and Wang [130], carry out supervised learning implementations in the UCI HAR Dataset, with various combined supervised techniques, and in the case of Bozkurt, they describe that using SVM + KNN obtains good results in the classification with an accuracy of 96.71% and Wang explains that using a combination of Decision Tree it is possible to count on the accuracy of 92.73%. Outreach [131], performs analysis on two Datasets of the set CASAS Tulum and Two, highlighting the use of BackPropagation with results 88.75% and 76.9%, respectively in accuracy.
Demrozi [132], performs multiple experiments of many supervised techniques in many widely known Datasets such as WISDM, DAPHNET, PAPAM, HHAR (Phone), HHAR (watch), Mhealth, RSSI, CSI. For this, different algorithms are implemented such as KNN, LDA, QDA, RF, DT, CNN. In the case of the algorithm WISDM, DAPHNET, PAPAM, HHAR (Phone), HHAR (watch) the RF algorithm obtains the best results with accuracy of 90%, 91%, 80%, 88%, 85% precision, and recall of 91%, 91%, 83%, 89%, 85%. For Mhealth, RSSI, the performance of the QDA algorithm is denoted with 91% and 92% in accuracy and 85% and 92% in precision and recall respectively.
Xu [133], applies compares techniques such as DT, SVM, KNN, AdaBoost, DCNN in Dataset CASAS Aruba, showing the superiority of ensembled techniques such as Adaboost with the accuracy of 98%, precision 96%, recall 95.9%, and f -measure 95.9%. Other authors such as Hussain [134] apply algorithms such as SVM, Random Forest, KNN to datasets such as SisFall, the SVM results being better with 97.77% accuracy. Finally, Liciotti [135], performs experimentation on a set of well-known CASAS project Datasets such as Milan, Cairo, Kyoto 2, Kyoto 3, Kyoto 4, of algorithms such as Naive Bayes, HMM + SVM, CRF, LSTM, showing the superiority of LSTM in the results.
Table 3. Supervised Techniques results.
Table 3. Supervised Techniques results.
DatasetTechniqueMetricsReferences
AccuracyPrecisionRecallF-Measure
UCI Machine LearningNearest Neighbor75.7---[121]
Decision Tree76.3---
Random Forest75.9---
Naive Bayes76.9---
Aras (House A)MSA (Margin Setting Algorithm)68.85---[122]
SVM66.90---
ANN67.32---
Aras (House B)MSA (Margin Setting Algorithm)96.24---
SVM94.81---
ANN95.42---
CASAS TulumMSA (Margin Setting Algorithm)68.00---
SVM66.6---
ANN67.37---
MhealthK-NN99.64--99.7[123]
ANN99.55--99.6
SVM99.89--100
C4.599.32--99.3
CART99.13--99.7
Random Forest99.89--99.89
Rotation Forest99.79--99.79
WISDM, SCUT_NA-ASliding window with variable size, S transform, and regularization based robust subspace (SRRS) for selection and SVM for Classification96.1---[124]
SCUT NA-ASliding window with fixed samples, SVM like a classifier, cross-validation91.21---
PAMPA2, MhealthSliding windows with fixed 2s, SVM, and Cross-validation84.10---
SBHARSliding windows with fixed 4s, SVM, and Cross-validation93.4---
WISDMMLP based on voting techniques with nb-Tree are used96.35---
UTD-MHADFeature level fusion approach& collaborative representation classifier79.1---
GroupwareMark Hall’s feature selection and Decision Tree 99.4---
Free-livingk-NN and Decision Tree95---
WISDM, SkodaHybrid Localizing learning (k-NN-LSS-VM)81---
UniMiB SHARLSTM and Deep Q-Learning95---
GroupwareSliding windows Gaussian Linear Filter and NB classifier89.5---
GroupwareSliding windows Gaussian Linear Filter and Decision Tree classifier99.99---
CSI-dataSVM96---[125]
LSTM89---
Built by the authorsIBK95---[126]
Classifier based ensemble98---
Bayesian network63---
Built by the authorsDecision Tree91.08--89.75[127]
Random Forest91.25--90.02
Gradient Boosting97.59--97.4
KNN93.76--93.21
Naive Bayes88.57--88.07
SVM92.7--91.53
XGBoost96.93--96.63
UK-DALEFFNN95.28---[128]
SVM93.84---
LSTM83.07---
UCI Machine LearningKNN90.7491.1590.2890.45[129]
SVM96.2796.4396.1496.23
HMM+SVM96.5796.7496.4996.56
SVM+KNN96.7196.7596.6996.71
Naive Bayes77.0379.2576.9176.72
Logistic Reg95.9396.1395.8495.92
Decision Tree87.3487.3986.9586.99
Random Forest92.392.492.0392.14
MLP95.2595.4995.1395.25
DNN96.8196.9596.7796.83
LSTM91.0891.3891.2491.13
CNN+LSTM93.0893.1793.1093.07
CNN+BiLSTM95.4296.5895.2695.36
Inception+ResNet95.7696.0695.6395.75
UCI Machine LearningNB-NB73.68--46.9[130]
NB-KNN85.58--61.08
NB-DT89.93--69.75
NB-SVM79.97--53.69
KNN-NB74.93--45
KNN-KNN79.3--49.82
KNN-DT87.01--60.98
KNN-SVM82.24--53.1
DT-NB84.72--60.05
DT-KNN91.55--73.11
DT-DT92.73--75.97
DT-SVM93.23--77.35
SVM-NB30.40---
SVM-KNN25.23---
SVM-DT92.43--75.31
SVM-SVM43.32---
CASAS TulumBack-Propagation88.75---[131]
SVM87.42---
DBM90.23---
CASAS TworBack-Propagation76.9---
SVM73.52---
DBM78.49---
WISDMKNN6978-78[132]
LDA4034-34
QDA6558-58
RF9091-91
DT7777-77
CNN6662-60
DAPHNETKNN9087-88
LDA9183-83
QDA9182-82
RF9191-91
DT9183-83
CNN9087-87
PAPAMKNN6566-66
LDA4545-45
QDA1519-19
RF8083-83
DT6060-60
CNN7376-73
HHAR(Phone)KNN8385-85
LDA4345-45
QDA4050-50
RF8889-89
DT6766-66
CNN8484-84
HHAR(watch)KNN7882-82
LDA5452-52
QDA2627-27
RF8585-85
DT6969-69
CNN8383-83
MhealthKNN7681-81
LDA3859-59
QDA9182-82
RF8585-85
DT7777-77
CNN8080-80
RSSIKNN9191-91
LDA9191-91
QDA9191-91
RF9191-91
DT9191-91
CNN9190-91
CSIKNN9393-93
LDA9393-93
QDA9292-92
RF9393-93
DT9393-93
CNN9292-92
Casas ArubaDT96.393.892.393[133]
SVM88.288.387.888.1
KNN89.287.885.986.8
AdaBoost989695.995.9
DCNN95.693.995.394.6
SisFallSVM97.7776.17 75.6[134]
Random Forest96.8279.99 79.95
KNN96.7193.99 68.36
CASAS MilanNaive Bayes76.65 [135]
HMM+SVM77.44
CRF61.01
LSTM93.42
CASAS CairoNaive Bayes82.79
HMM+SVM82.41
CRF68.07
LSTM83.75
CASAS Kyoto 2Naive Bayes63.98
HMM+SVM65.79
CRF66.20
LSTM69.76
CASAS Kyoto 3Naive Bayes77.5
HMM+SVM81.67
CRF87.33
LSTM88.71
CASAS Kyoto 4Naive Bayes63.27
HMM+SVM60.9
CRF58.41
LSTM85.57

6.2. Unsupervised Learning Applied to Human Activity Recognition Dataset

In the unsupervised learning applications in the literature, different applications of the algorithms can be observed that are measured with quality metrics associated with the groupings such as ARI, Jaccard Index, Silhouette Index, Euclidean, F1 Fisher’s discriminant (see Table 4). The following works developed by authors such as Wang [130] stand out, who uses various versions of the UCI-HAR Dataset, implementing algorithms such as K-means, HAC, FCM, both showing better results for the case of FCM. Mohmed [136] applies unsupervised algorithms like FCM to the Nottingham Trent University Dataset. Brena [137], applies his form developed by the author called PM Mo-del to perform unsupervised analysis to the Chest Sensor Dataset, Wrist Sensor Dataset, WISDM Dataset, and Smartphone Dataset, which he measures using the silhouette index. He [138], applies another method developed by the authors called wavelet tensor fuzzy clustering scheme (WTFCS) to the DSAD Dataset, obtaining an ARI index of 89.66%.
Wang [139], implements clustering-based algorithms such as Spectral Clustering, Single Linkage, Ward Linkage, Average Linkage, K-medioids to the UCI-HAR dataset, analyzing their Jaccard and Euclidean indices as shown in Table 4. In the same way, Bota [140] also makes experiments in the UCI-HAR and CADL Dataset with the K-means, K-Means 5, Spectral Clustering, Gaussian Mixture, DBSCAN algorithms analyzing its F1 Fisher’s discriminant rat.
Table 4. Unsupervised Techniques results.
Table 4. Unsupervised Techniques results.
DatasetTechniqueMetricsReferences
ARIJaccard IndexSilhouette IndexEuclideanF1 Fisher’s Discriminant
Ratio
UCI HAR SmartPhoneK-means0.77270.32460.4416 [130]
HAC0.42130.22240.5675
FCM0.83430.40520.4281
UCI HAR Single Chest-Mounted AccelerometerK-means0.88500.65440.6935
HAC0.59960.25630.6851
FCM0.91890.72300.7751
Nottingham Trent UniversityFCM---- [136]
Chest Sensor DatasetPM Model 25.8%- [137]
Wrist Sensor Dataset 64.3%-
WISDM Dataset 54%-
Smartphone Dataset 85%-
DSADwavelet tensor fuzzy clustering scheme (WTFCS)0.8966--- [138]
UCI HARSpectral Clustering 0.543 0.583 [139]
Single Linkage 0.807 0.851
Ward Linkage 0.770 0.810
Average Linkage 0.790 0.871
K-medioids 0.653 0.654
UCI HARK-means 52.1[140]
K-Means 5 50.7
Spectral Clustering 57.8
Gaussian Mixture 49.8
DBSCAN 16.4
CADLK-means 50.9
K-Means 5 50.5
Spectral Clustering 61.9
Gaussian Mixture 58.9
DBSCAN 13.9

6.3. Ensemble Learning Applied to Human Activity Recognition Dataset

In the lessons based on ensemble learning, the application of multiple techniques is usually carried out, which together offer better results (see Table 5). Below is a detailed description of the works found in the literature review that shows the application of these techniques in the recognition of human activities. Yacchirema [141], uses a combination of techniques such as Decision Tree, Ensemble, Logistic Regression, Deepnet to analyze the SisFall Dataset, explaining the results of the DeepNet algorithm with an accuracy of 99.06%. For his part, Manzi [142], uses a mixture of X-means and SV; to analyze the Cornell Activity Dataset and TST Dataset obtaining 98.4% and 92.7% respectively.
Ma [143], uses the model based on Multi-task deep clustering in the HHAR, MobiAct, MobiSense datasets, where the latter algorithm obtains an accuracy of 72.5%, a precision of 71.2%, and a recall of 70.7%. Budisteanu [144], describes the NTU-RGB + D Dataset, and implements the K-Means, GMM algorithms, obtaining 85.72% and 87.26% respectively. Xu [145], uses the well-known UCI-HAR Dataset, implementing the CELearning own technique, obtaining an accuracy of 96.88%.
Choudhury [146], also analyzes the UCI-HAR Dataset, with the algorithms RF, XGB, AdaB, GB, ANN, V. RNN, LSTM, DT, KNN, and NB, where the RF algorithm performs the best result in the ensemble models with 96.96%. Wang [147] for his part defines his Dataset to which he implements the algorithms GB, RFs, Bagging, XGB, AdaBoost, DT, MLP, LSVM, NLSVM, LR, KNNs, GNB, in which the RF algorithm obtains the best results with an accuracy of 83.9%. Jethanandani [148], works with the popular Dataset House A and House B, applying algorithms such as Bernoulli NB, Decision Tree, Logistic Regression, KNN. This experimentation shows the good results of the algorithms based on decision trees with 88% and 97.2% respectively.
Subasi [149], also uses the UCI-HAR Dataset, applying the algorithms SVM-AdaBoost, k-NN-AdaBoost, ANN-AdaBoost, NB-AdaBoost, RF-AdaBoost, CART-AdaBoost, C4.5-AdaBoost, REPTree-AdaBoost, LADTree-AdaBoost obtaining better results with the REPTree-AdaBoost combination with 99.95% accuracy. Padmaja [150], uses HAR Dataset, HAPT Dataset implementing the KN, CART, BAYES, RF, ET alalgorithms, a method proposed by the authors, demonstrating the superiority of the results of the method proposed by the authors.
Table 5. Ensembled Learning Techniques results.
Table 5. Ensembled Learning Techniques results.
DatasetTechniqueMetricsReferences
AccuracyPrecisionRecallF-Measure
SisFallDecision Tree97.48---[141]
Ensemble99.51---
Logistic Regression84.87---
Deepnet99.06---
Cornell Activity DatasetX-means-SVM98.495.095.8-[142]
TST Dataset92.795.691.1-
HHARMulti-task deep clustering 67.265.365.9[143]
MobiAct 68.369.166.8
MobiSense 72.571.270.7
NTU-RGB + DK-Means85.72---[144]
GMM87.26---
UCI HARCELearning96.88%---[145]
UCI HARRF96.9697.097.098[146]
XGB96.2969696
AdaB50.5615151
GB94.53959595
ANN92.51929392
V. RNN90.53909190
LSTM91.23909190
DT94.23959595
KNN96.59979797
NB80.67848181
Proposed DatasetGB84.184.184.284.1[147]
RFs83.983.984.183.9
Bagging838383.183
XGB80.480.580.480.4
AdaBoost77.277.377.377.3
DT76.9777777
MLP67.668.767.867.8
LSVM6565.765.164.9
NLSVM6363.363.262.8
LR59.660.259.859.4
KNNs58.960.159.258.9
GNB56.159.455.445.2
House ABernoulli NB78.764--[148]
Decision Tree8879.4--
Logistic Regression81.469.2--
KNN75.864.9-
House BBernoulli NB95.979.4-
Decision Tree97.286.4-
Logistic Regression96.582.7-
KNN93.179.8-
UCI HARSVM-AdaBoost99.9 99.9[149]
k-NN-AdaBoost99.43 99.4
ANN-AdaBoost99.33 99.33
NB-AdaBoost97.24 97.2
RF-AdaBoost99.98 100
CART-AdaBoost99.97 100
C4.5-AdaBoost99.95 100
REPTree-AdaBoost99.95 100
LADTree-AdaBoost98.84 98.8
HAR DatasetKNN90.3 [150]
CART84.9
BAYES77
RF92.7
HAPT DatasetKNN89.2
CART80.2
BAYES74.7
RF91
ET91.7
Proposed Method92.6

6.4. Deep Learning Applied to Human Activity Recognition Dataset

Implementations based on Deep Learning have become very useful for the identification of activities of daily life, especially those that include image processing [151,152] (see Table 6). Some relevant results of the literature review are detailed below. Wan [153], makes use of the UCI-HAR and PAMAP2 Dataset, implementing algorithms such as CNN, LSTM, BLSTM, MLP, SVM, in which the good results of CNN network implementation are shown with 92.71% and 91% respectively. Akula [154], configures its Dataset to which it applies the algorithms LBP-Naive Bayes, HOG-Naive Bayes, LBP-KNN, HOG-KNN, LBP-SVM, HOF-SVM obtaining better results with the implementation of HOF -SVM with 85.92% accuracy.
He [155], implements DeepConvLSTM, CNN in the UCI-HAR, and Wealky Datasets, showing good results of the implementation of Deep learning with 94.77% and 92.31% respectively. Long [156] in turn uses the Opportunity and Uni-MiB-SAHR Dataset with the algorithms HC, CBH, CBS, AE, MLP, CNN, LSTM, Hybrid, ResNet, and ARN were the results of RNA of 90.29% and 76.39%. Bozkurt [157], for his part, only analyzes the UCI-HAR Dataset, KNN, SVM, HMM + SVM, SVM + KNN, Naive Bayes, Logistic Regression, Decision Tree, Random Forest, MLP, DNN, LSTM, CNN + LSTM, CNN + BiLSTM, Inception + ResNet, the result of the DNN algorithm is shown with an accuracy of 96.81%.
Mekruksavanich [158], uses the Utwente Dataset and PAMAP2, applying the Naive Bayes, SVM, Deep Stacked Autoencoder, CNN-BiGRu techniques, showing better results with this last technique described. Papagiannaki [159] used the FrailSafe dataset with the implementation of CNN networks with an accuracy of 91.84%. Liciotti [139] uses techniques such as LSTM, Bi-LSTM, Casc-LSTM, ENs2-LSTM in the CASAS group dataset to show the dynamics of processes based on deep learning. Hassan [160], applied ANN, SVM and DBN in a proposal dataset for the development of a robust human activity recognition system based on the smartphone sensors’ data, obtaining the following accuracy results ANN 89.06%, SVM 94.12% and DBN 95.85%.
Table 6. Deep Learning Techniques results.
Table 6. Deep Learning Techniques results.
DatasetTechniqueMetricsReferences
AccuracyPrecisionRecallF-Measure
Uci HarCNN92.7193.2192.8292.93[154]
LSTM89.0189.1488.9988.99
BLSTM89.489.4189.3689.35
MLP86.8386.8386.5886.61
SVM89.8590.589.8689.85
PAMAP2CNN91.0091.6690.8691.16
LSTM85.8686.5184.6785.34
BLSTM89.5290.1989.0289.4
MLP82.0783.3582.1782.46
SVM84.0784.7184.2383.76
Propio Infrared ImagesLBP-Naive Bayes42.1---[155]
HOG-Naive Bayes77.01---
LBP-KNN53.261---
HOG-KNN83.541---
LBP-SVM62.34---
HOF-SVM85.92---
Uci HarDeepConvLSTM94.77---[156]
CNN92.76---
Weakly DatasetDeepConvLSTM92.31---
CNN85.17---
OpportunityHC85.69---[157]
CBH84.66---
CBS85.39---
AE83.39---
MLP86.65---
CNN87.62---
LSTM86.21---
Hybrid87.67---
ResNet87.67---
ARN90.29---
UniMiB-SAHRHC21.96---
CBH64.36---
CBS67.36---
AE68.39---
MLP74.82---
CNN73.36---
LSTM68.81---
Hybrid72.26---
ResNet75.26---
ARN76.39---
Uci HarKNN90.7491.1590.2890.48[158]
SVM96.2796.4396.1496.23
HMM+SVM96.5796.7406.4996.56
SVM+KNN96.7196.7596.6996.71
Naive Bayes77.0379.2576.9176.72
Logistic Regression95.9396.1395.8495.92
Decision Tree87.3487.3986.9586.99
Random Forest92.3092.492.0392.14
MLP95.2595.4995.1395.25
DNN96.8196.9596.7796.83
LSTM91.0891.3891.2491.13
CNN+LSTM93.0893.1793.1093.07
CNN+BiLSTM95.4295.5895.2695.36
Inception+ResNet95.7696.0695.6395.75
Utwente DatasetNaive Bayes---94.7[159]
SVM---91.6
Deep Stacked Autoencoder---97.6
CNN-BiGRu---97.8
PAMAP2DeepCOnvTCN---81.8
InceptionTime---81.1
CNN-BiGRu---85.5
FrailSafe datasetCNN91.84---[160]
CASAS MilanLSTM76.65---[135]
Bi-LSTM77.44---
Casc-LSTM61.01---
ENs2-LSTM93.42---
CASAS CairoLSTM82.79---
Bi-LSTM82.41---
Casc-LSTM68.07---
ENs2-LSTM83.75---
CASAS Kyoto 2LSTM63.98---
Bi-LSTM65.79---
Casc-LSTM66.20---
ENs2-LSTM69.76---
CASAS Kyoto 3LSTM77.5---
Bi-LSTM81.67---
Casc-LSTM87.33---
ENs2-LSTM88.71---
ProposalANN89.06---[160]
SVM94.12---
DBN95.85---

6.5. Reinforcement Learning Applied to Human Activity Recognition Dataset

Currently, there is a new trend in reinforcement-based learning processes where it is possible to have systems capable of learning by themselves from punishment and reward schemes, defined by behavioral psychology. It has been introduced in this new line of work for HAR. Which this review shows three highly relevant works (see Table 7). Ber-lin [161], made implementations in the Weizmann and KTH Datasets through the implementation of Spiking Neural Network showing promising results 94.44% and 92.50%. Lu [162] uses the DoMSEV Dataset using the Deep-shallow algorithm with an accuracy of 72.9% and Hossain [163], Pop used a new Dataset to which they implemented the Deep Q-Network algorithm with an accuracy of 83.26%.
Table 7. Reinforcement Learning Techniques results.
Table 7. Reinforcement Learning Techniques results.
DatasetTechniqueMetricsReferences
Accuracy
Weizmann datasetsSpiking Neural Network94.44[161]
KTH datasets92.50
DoMSEVDeep-Shallow72.9[162]
ProposalDeep Q-Network (DQN)83.26[163]
S.Yousefi-2017Reinforcement Learning Agent Recurrent Neural Network with Long Short-Term Memory80[156]
FallDeFi83
UCI HARReinforcement Learning + DeepConvLSTM98.36[164]
Proposal 79[165]
UCF-SportsQ-learning95[166]
UCF-10185
sub-JHMDB80
MHEALTHCluster-Q learning94.5[167]
PAMAP283.42
UCI HAR81.32
MARS85.92
DataEgoLRCN88[168]
ProposalMask Algorithm96.02[169]
ProposalLSTM-Reinforcement Learning90.50[169]
ProposalConvolutional Autoencoder87.7[170]

6.6. Metaheuristic Algorithms Applied to Human Activity Recognition Dataset

In the review of the state of the art, it was possible to identify different metaheuristic techniques that contribute to the identification of different algorithms. Among the most evident results are applications of Genetic Algorithms with the following results 96.43% [171], 87.5 [172], 95,71 [173], 99.75 [174], 98.00 [175] and 98.96 [175]. In many solutions, hybrid systems or new algorithms proposed by the authors are used, see Table 8.
Table 8. Metaheuristic Learning Techniques results.
Table 8. Metaheuristic Learning Techniques results.
DatasetTechniqueMetricsReferences
Accuracy
Cifar-100L4-Banched-ActionNet + EntACS + Cub-CVM98.00[175]
SbharptAnt-Colony, NB98.96[176]
UciharBee swarm optimization with a deep Q-network98.41[177]
MotionsenseBinary Grey Wolf Optimization93.95[178]
Mhealth96.83
Uci HarGenetic Algorithms-SVM96.43[171]
Ucf50Genetic Algorithms-CNN87.5[172]
SbharGA-PCA95,71[173]
MnistGA-CNN99.75[174]
Cifar-100Genetic Algorithms-SVM98.00[175]
SbharptGenetic Algorithms-CNN98.96[176]

6.7. Transfer Algorithms Applied to Human Activity Recognition Dataset

Transfer Learning TL transfers the parameters of the learned and trained model to a new model to help the training of the new model. Considering that most of the data or tasks are related, through transfer learning, the learned model parameters can be shared with the new model in a certain way to speed up and optimize the model learning efficiency. The basic motivation of TL is to try to apply the knowledge gained from one problem to a different but related problem, see Table 9.
Table 9. Transfer Learning Techniques results.
Table 9. Transfer Learning Techniques results.
DatasetTechniqueMetricsReferences
AccuracyPrecisionRecallF-Measure
CSIKNN98.3---[179]
SVM98.3---
CNN99.2---
OpportunityKNN+PCA60---[180]
GFK59---
STL65---
SA-GAN73---
USC-HADMMD80---[181]
DANN77---
WD72---
ProposalKNN-OS79.8485.8491.8888.61[182]
KNN-SS89.6494.4194.7694.52
SVM-OS77.1497.0479.2387.09
SVM-SS87.594.3992.6193.27
DT-OS87.594.6192.1693.14
DT-SS91.7995.1996.2695.71
JDA86.7992.7193.0792.89
BDA91.4395.995.1895.51
IPL-JPDA93.2197.0495.9796.48
KNN-OS79.8485.8491.8888.61
Wiezmann DatasetVGG-16 MODEL96.9597.0097.0097.00[183]
VGG-19 MODEL96.5497.0097.0096.00
Inception-v3 Model95.6396.0096.0096.00
PAMAP2DeepConvLSTM---93.2[184]
Skoda Mini Checkpoint---93
OpportunityPCA66.78---[185]
TCA68.43---
GFK70.87---
TKL70.21---
STL73.22---
TNNAR78.4---
PAMAP2PCA42.87---
TCA47.21---
GFK48.09---
TKL43.32---
STL51.22---
TNNAR55.48---
UCI DSADSPCA71.24---
TCA73.47---
GFK81.23---
TKL74.26---
STL83.76---
TNNAR87.41---
UCI HARCNN-LSTM90.8---[186]
DT76.73.--[187]
RF71.96---
TB75.65---
TransAct86.49---
MhealthDT48.02---
RF62.25---
TB66.48---
TransAct77.43---
Daily SportDT66.67...
RF70.38...
TB72.86.--
TransAct80.83---
ProposalWithout SVD (Singular Value Decomposition)63.13%---[188]
With SVD (Singular Value Decomposition)43.13%---
Transfer Accuracy97.5%---
PAMAP2CNN84.89---[189]
UCI HAR83.16---
UCI HARkNN77.28---[190]
DT72.16---
DA77.46---
NB69.93---
Transfer Accuracy83.7---
UCF Sports Action datasetVGGNet-1997.13---[191]
AMASSDeepConvLSTM87.46---[192]
DIP89.08---
DAR DatasetBase CNN85.38---[193]
AugToAc91.38---
HDCNN86.85---
DDC86.67---
UCI HARCNN_LSTM92.13---[194]
CNN_LSTM_SENSE91.55---
LSTM91.28---
LSTM_DENSE91.40---
ISPLCNN_LSTM99.06---
CNN_LSTM_SENSE98.43---
LSTM96.23---
LSTM_DENSE98.11---

7. Conclusions

The objective of this systematic literature review article is to provide HAR researchers with a set of recommendations, among which the different data sets that can be used depending on the type of research are highlighted. For the development of this analysis, different data sources were considered in an observation window between the years 2017 and 2021. Among the most representative databases, IEEE Xplorer can be highlighted with 256 articles, far surpassing other specialized databases such as Scopus, Science Direct, Web of Science, and ACM.
It is important to specify that 47% of the publications are due to proceedings of congresses or conferences and 36% to the specialized journal. Discriminating the quartiles where the articles are published, it is important to highlight that although the vast majority of publications are indeed focused on conference proceedings that do not have a specific category, 36% of the publications that were made in journals were are mostly in the first two quartiles Q1 and Q2.
In this article, technical analysis of different types of datasets that are used for experimentation processes with HAR was carried out. It should be noted that the creation of new data sets has increased. Some traditional approaches related to the use of indoor datasets based on the WSU Casas project remain. Also, public repositories such as UCI Machine learning have provided sets widely used in the literature such as Opportunity and UCI HAR. It should be noted that the processing of images and videos to the dataset has also been increased, allowing the application of different cutting-edge techniques, such as Weakly Dataset and UniMiB-SAHR.
In this review, different data processing approaches that have been used in this area of knowledge were used. For the specific case of supervised learning, the usability of algorithms based on decision trees such as RandomForest, Naive Bayes, and Support Vector Machine stands out. Regarding unsupervised learning, in most of the analyzed works, the use of techniques such as Spectral Clustering, Single Linkage, Ward Linkage, Average Linkage and K-medioids. Using ensembled learning, it was possible to demonstrate the use of different sets of techniques that allowed improving the results of the experiments, among which those based on classification and grouping can be highlighted. Another modern and widely used approach is the use of DeepLearning focused on datasets with massive image processing requirements, where the use of the following LSTM algorithms stands out, Bi-LSTM, Casc-LSTM, ENs2-LSTM. Other approaches based on Reinforcement learning use resources such as Q-learning and Cluster-Q with learning, in the experimentation processes. The metaheuristic-based approach shows the usability of different algorithms, among which the following stand out: L4-Banched-ActionNet+EntACS+Cub-CVM, Ant-Colony, N.B Bee swarm optimization with a deep Q-network and Genetic Algorithms.
It is important to point out that due to the high demand for data and information processing, it becomes increasingly necessary to implement techniques capable of improving performance and results, such as those based on Reinforcement Learning and Transfer Learning. Another challenge found in the literature is the processing of multi-occupancy datasets that make the use of computational resources and the identification of activities more expensive.

8. Future Works

Among the future works that can be implemented after this systematic review of the literature, the real-time analysis of the dataset not only with data from sensors but also images and sound, among which algorithms based on Reinforcement Learning and Transfer Learning can be highlighted. provide a wide range of competitive solutions, adding multi-occupancy in data sets.

Author Contributions

Definition of taxonomy, P.P.A.-C., F.P. and E.V.; Conceptualization, P.P.A.-C., A.I.O.-C. and M.A.P.-M.; Human Activity Recognition conceptual Information P.P.A.-C., F.P. and E.V.; Methodology P.P.A.-C. and M.A.P.-M.; Technical and Scientometric Analysis P.P.A.-C., M.A.P.-M. and F.P. and A.Q.-L.; Formal Conclusions P.P.A.-C. and A.I.O.-C.; Supervision F.P. and E.V.; Writing-Review & Editing, P.P.A.-C., S.B.A. and F.P. All authors have read and agreed to the published version of the manuscript.

Funding

European Union’s Horizon 2020 research and innovation program under the Marie Skłodowska-Curie grant agreement No. 734355.

Acknowledgments

This research has received funding under the REMIND project Marie Sklodowska-Curie EU Framework for Research and Innovation Horizon 2020, under Grant Agreement No. 734355. Furthermore, this research has been supported by the Spanish government by means of the projects RTI2018-098979-A-I00, PI-0387-2018 and CAS17/00292.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aracil, J.; Gordillo, F. Dinámica de Sistemas; Alianza Editorial: Madrid, Spain, 1997. [Google Scholar]
  2. Cramer, H.; Cansado, C. Métodos Matemáticos de Estadística; Aguilar: Madrid, Spain, 1968. [Google Scholar]
  3. Shapiro, S.C. Artificial intelligence. In Encyclopedia of Artificial Intelligence, 2nd ed.; Shapiro, S.C., Ed.; Wiley: New York, NY, USA, 1992; Volume 1. [Google Scholar]
  4. Rouse, M. Inteligencia Artificial, o AI. Available online: https://www.computerweekly.com/es/definicion/Inteligencia-artificial-o-IA (accessed on 30 October 2021).
  5. Duan, Y.; Edwards, J.S.; Dwivedi, Y.K. Artificial intelligence for decision making in the era of Big Data—Evolution, challenges and research agenda. Int. J. Inf. Manag. 2019, 48, 63–71. [Google Scholar] [CrossRef]
  6. Sekeroglu, B.; Hasan, S.S.; Abdullah, S.M. Comparison of Machine Learning Algorithms for Classification Problems. Adv. Intell. Syst. Comput. 2020, 944, 491–499. [Google Scholar]
  7. Jordan, M.I.; Jordan, M.T.M. Machine Learning: Trends, perspectives, and prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef]
  8. Ge, Z.; Song, Z.; Ding, S.X.; Huang, B. Data Mining and Analytics in the Process Industry: The Role of Machine Learning. IEEE Access 2017, 5, 20590–20616. [Google Scholar] [CrossRef]
  9. Amiribesheli, M.; Benmansour, A.; Bouchachia, A. A review of smart homes in healthcare. J. Ambient Intell. Humaniz. Comput. 2015, 6, 495–517. [Google Scholar] [CrossRef] [Green Version]
  10. Cook, D.J.; Youngblood, M.; Das, S.K. Amulti-agent approach to controlling a smart environment. In Designing Smart Homes; Springer: Berlin/Heidelberg, Germany, 2006; pp. 165–182. [Google Scholar]
  11. Andrew McCallum, K.N. A Comparison of Event Models for Naive Bayes Text Classification. In Proceedings of the AAAI-98 Workshop on Learning for Text Categorization, Menlo Park, CA, USA, 26–27 July 1998; Volume 752, p. 307. [Google Scholar] [CrossRef] [Green Version]
  12. Murata, N.; Yoshizawa, S.; Amari, S. Network information criterion-determining the number of hidden units for an artificial neural network model. IEEE Trans. Neural Netw. 1994, 5, 865–872. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Du, W.S.; Hu, B.Q. Approximate distribution reducts in inconsistent interval-valued ordered decision tables. Inf. Sci. 2014, 271, 93–114. [Google Scholar] [CrossRef]
  14. Chen, W.; Xie, X.; Wang, J.; Pradhan, B.; Hong, H.; Bui, D.T.; Duan, Z.; Ma, J. A comparative study of logistic model tree, random forest, and classification and regression tree models for spatial prediction of landslide susceptibility. CATENA 2017, 151, 147–160. [Google Scholar] [CrossRef] [Green Version]
  15. Saxena, A.; Prasad, M.; Gupta, A.; Bharill, N.; Patel, O.P.; Tiwari, A.; Er, M.J.; Ding, W.; Lin, C.-T. A review of clustering techniques and developments. Neurocomputing 2017, 267, 664–681. [Google Scholar] [CrossRef] [Green Version]
  16. Jones, F.W.; McLaren, I.P.L. Rules and associations. In Proceedings of the Twenty First Annual Conference of the Cognitive Science Society, Vancouver, BC, Canada, 23 December 2020; Psychology Press: East Sussex, UK, 2020; pp. 240–245. [Google Scholar]
  17. Van Der Maaten, L.; Postma, E.; Van den Herik, J. Dimensionality reduction: A comparative. J. Mach. Learn. Res. 2009, 10, 13. [Google Scholar]
  18. Divina, F.; Gilson, A.; Goméz-Vela, F.; García Torres, M.; Torres, J.F. Stacking Ensemble Learning for Short-Term Electricity Consumption Forecasting. Energies 2018, 11, 949. [Google Scholar] [CrossRef] [Green Version]
  19. Sewell, M. Ensemble learning. RN 2008, 11, 1–34. [Google Scholar]
  20. Svetnik, V.; Wang, T.; Tong, C.; Liaw, A.; Sheridan, R.P.; Song, Q. Boosting: An ensemble learning tool for compound classification and QSAR modeling. J. Chem. Inf. Modeling 2005, 45, 786–799. [Google Scholar] [CrossRef] [PubMed]
  21. Shen, Y.; Chen, T.; Xiao, Z.; Liu, B.; Chen, Y. High-Dimensional Data Clustering with Fuzzy C-Means: Problem, Reason, and Solution. In Proceedings of the International Work-Conference on Artificial Neural Networks, Virtual Event, 16–18 June 2021; Springer: Cham, Switzerland, 2021; pp. 89–100. [Google Scholar]
  22. Kaelbling, L.P.; Littman, M.L.; Moore, A.W. Reinforcement learning: A survey. J. Artif. Intell. Res. 1996, 4, 237–285. [Google Scholar] [CrossRef] [Green Version]
  23. Kaur, M.; Kaur, G.; Sharma, P.K.; Jolfaei, A.; Singh, D. Binary cuckoo search metaheuristic-based supercomputing framework for human behavior analysis in smart home. J. Supercomput. 2019, 76, 2479–2502. [Google Scholar] [CrossRef]
  24. Althöfer, I.; Koschnick, K.U. On the convergence of “Threshold Accepting”. Appl. Math. Optim. 1991, 24, 183–195. [Google Scholar] [CrossRef]
  25. Moscato, P.; Cotta, C.; Mendes, A. Memetic algorithms. In New Optimization Techniques in Engineering; Springer: Berlin/Heidelberg, Germany, 2004; pp. 53–85. [Google Scholar]
  26. Wesselkamper, J. Fail-Safe MultiBoot Reference Design; XAPP468; Xilinx: San Jose, CA, USA, 2009. [Google Scholar]
  27. Salcedo-Sanz, S.; Cuadra, L.; Vermeij, M. A review of Computational Intelligence techniques in coral reef-related applications. Ecol. Inform. 2016, 32, 107–123. [Google Scholar] [CrossRef]
  28. Krause, J.; Cordeiro, J.; Parpinelli, R.S.; Lopes, H.S. A survey of swarm algorithms applied to discrete optimization problems. In Swarm Intelligence and Bio-Inspired Computation; Elsevier: Amsterdam, The Netherlands, 2013; pp. 169–191. [Google Scholar]
  29. Kumar, M.; Husain, M.; Upreti, N.; Gupta, D. Genetic Algorithm: Review and Application. 2010. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3529843 (accessed on 30 October 2021).
  30. Glover, F.; Laguna, M.; Martí, R. Scatter search. In Advances in Evolutionary Computing; Springer: Berlin/Heidelberg, Germany, 2003; pp. 519–537. [Google Scholar]
  31. Hansen, P.; Mladenović, N. Variable neighborhood search. In Search Methodologies; Springer: Boston, MA, USA, 2005; pp. 211–238. [Google Scholar]
  32. Dorigo, M.; Birattari, M.; Stutzle, T. Ant colony optimization. IEEE Comput. Intell. Mag. 2006, 1, 28–39. [Google Scholar] [CrossRef]
  33. Torrey, L.; Shavlik, J. Transfer learning. In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques; IGI Global: Hershey, PA, USA, 2010; pp. 242–264. [Google Scholar]
  34. Timarán Pereira, S.R.; Hernández Arteaga, I.; Caicedo Zambrano, S.J.; Hidalgo Troya, A.; Alvarado Pérez, J.C. Descubrimiento de Patrones de Desempeño Académico con Árboles de Decisión en las Competencias Genéricas de la Formación Profesional; Ediciones Universidad Cooperativa de Colombia: Bogotá, Colombia, 2015. [Google Scholar] [CrossRef]
  35. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  36. Han, S.; Qubo, C.; Meng, H. Parameter selection in SVM with RBF kernel function. In Proceedings of the World Automation Congress 2012, Puerto Vallarta, Mexico, 24–28 June 2012; pp. 1–4. [Google Scholar]
  37. Gaikwad, N.B.; Tiwari, V.; Keskar, A.; Shivaprakash, N.C. Efficient FPGA implementation of multilayer perceptron for real-time human activity classification. IEEE Access 2019, 7, 26696–26706. [Google Scholar] [CrossRef]
  38. Yiyu, Y. Decision-theoretic rough set models. In Proceedings of the International Conference on Rough Sets and Knowledge Technology, Toronto, ON, Canada, 14–16 May 2007; Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Springer: Berlin/Heidelberg, Germany, 2007; pp. 1–12. [Google Scholar] [CrossRef]
  39. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification; John Wiley & Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  40. Jain, A.K.; Murty, M.N.; Flynn, P.J. Data clustering: A review. ACM Comput. Surv. (CSUR) 1999, 31, 264–323. [Google Scholar] [CrossRef]
  41. Li, J. Two-scale image retrieval with significant meta-information feedback. In Proceedings of the 13th Annual ACM International Conference on Multimedia, Singapore, 6–11 November 2005; pp. 499–502. [Google Scholar]
  42. Li, J.; Ray, S.; Lindsay, B.G. A Nonparametric Statistical Approach to Clustering via Mode Identification. J. Mach. Learn. Res. 2007, 8, 1687–1723. [Google Scholar]
  43. Xu, X. Survey of clustering algorithms. IEEE Trans. Neural Netw. 2005, 16, 645–678. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  45. Li, T.S.; Simon, J.D.; Drlica-Wagner, A.; Bechtol, K.; Wang, M.Y.; Garcia-Bellido, J.; Frieman, J.; Marshall, J.L.; James, D.J.; Strigari, L.; et al. Farthest Neighbor: The Distant Milky Way Satellite Eridanus II. Astrophys. J. Lett. 2017, 838, 8. [Google Scholar] [CrossRef]
  46. Huse, S.M.; Welch, D.M.; Morrison, H.G.; Sogin, M.L. Ironing out the wrinkles in the rare biosphere through improved OTU clustering. Environ. Microbiol. 2010, 12, 1889–1898. [Google Scholar] [CrossRef] [Green Version]
  47. McIntyre, R.M.; Blashfield, R.K. A Nearest-Centroid Technique for Evaluating The Minimum-Variance Clustering Procedure. Multivar. Behav. Res. 1980, 15, 225–238. [Google Scholar] [CrossRef]
  48. Ferrer, M.; Valveny, E.; Serratosa, F.; Bardají, I.; Bunke, H. Graph-based k-means clustering: A comparison of the set median versus the generalized median graph. In International Conference on Computer Analysis of Images and Patterns; Springer: Berlin/Heidelberg, Germany, 2009; pp. 342–350. [Google Scholar]
  49. Likas, A.; Vlassis, N.; Verbeek, J.J. The global k-means clustering algorithm. Pattern Recognit. 2003, 36, 451–461. [Google Scholar] [CrossRef] [Green Version]
  50. Kamen, J.M. Quick clustering. J. Mark. Res. 1970, 7, 199–204. [Google Scholar] [CrossRef]
  51. Redmond, S.; Heneghan, C. A method for initialising the K-means clustering algorithm using kd-trees. Pattern Recognit. Lett. 2007, 28, 965–973. [Google Scholar] [CrossRef]
  52. Park, H.-S.; Jun, C.-H. A simple and fast algorithm for K-medoids clustering. Expert Syst. Appl. 2009, 36, 3336–3341. [Google Scholar] [CrossRef]
  53. Gupta, T.; Panda, S.P. A comparison of k-means clustering algorithm and clara clustering algorithm on iris dataset. Int. J. Eng. Technol. 2018, 7, 4766–4768. [Google Scholar]
  54. Hadji, M.; Zeghlache, D. Minimum cost maximum flow algorithm for dynamic resource allocation in clouds. In Proceedings of the 2012 IEEE Fifth International Conference on Cloud Computing, Honolulu, HI, USA, 24–29 June 2012; pp. 876–882. [Google Scholar]
  55. Cara, F.J.; Carpio, J.; Juan, J.; Alarcón, E. An approach to operational modal analysis using the expectation maximization algorithm. Mech. Syst. Signal Processing 2012, 31, 109–129. [Google Scholar] [CrossRef] [Green Version]
  56. Gholami, E.; Jahromi, Y.M. Forecastingof the Value Added Tax from Tobacco Consumption Using Neural Network Method. J. Res. Econ. Model. 2015, 5, 55–72. [Google Scholar] [CrossRef] [Green Version]
  57. Fortin, M.; Fortin, A. A generalization of Uzawa’s algorithm for the solution of the Navier-Stokes equations. Commun. Appl. Numer. Methods 1985, 1, 205–208. [Google Scholar] [CrossRef]
  58. Niu, P.; Niu, S.; Liu, N.; Chang, L. The defect of the Grey Wolf optimization algorithm and its verification method. Knowl. -Based Syst. 2019, 171, 37–43. [Google Scholar] [CrossRef]
  59. Govaert, G.; Nadif, M. Clustering with block mixture models. Pattern Recognit. 2003, 36, 463–473. [Google Scholar] [CrossRef]
  60. De Roover, K. Finding Clusters of Groups with Measurement Invariance: Unraveling Intercept Non-Invariance with Mixture Multigroup Factor Analysis. Struct. Equ. Model. A Multidiscip. J. 2021, 28, 663–683. [Google Scholar] [CrossRef]
  61. Agrawal, R.; Srikant, R. Fast algorithms for mining association rules in large databases. In Proceedings of the 20th International Conference on Very Large Data Bases, Santiago, Chile, 12–15 September 1994; pp. 487–499. [Google Scholar]
  62. Hipp, J.; Güntzer, U.; Nakhaeizadeh, G. Algorithms for association rule mining—A general survey and comparison. ACM SIGKDD Explor. Newsl. 2000, 2, 58–64. [Google Scholar] [CrossRef]
  63. Zaki, M. Scalable algorithms for association mining. IEEE Trans. Knowl. Data Eng. 2000, 12, 372–390. [Google Scholar] [CrossRef] [Green Version]
  64. Han, J.; Pei, J.; Yin, Y. Mining frequent patterns without candidate generation. In Proceedings of the ACM International Conference on Management of Data, Dallas, TX, USA, 15–18 May 2000; pp. 1–12. [Google Scholar]
  65. Rathee, S.; Kaul, M.; Kashyap, A. R-Apriori: An efficient Apriori based algorithm on Spark. In Proceedings of the PIKM’15, ACM, Melbourne, Australia, 31 May–4 June 2015. [Google Scholar]
  66. Qiu, H.; Gu, R.; Yuan, C.; Huang, Y. YAFIM: A parallel frequent itemset mining algorithm with Spark. In Proceedings of the Parallel & Distributed Processing Symposium Workshops (IPDPSW), Phoenix, AZ, USA, 19–23 May 2014; pp. 1664–1671. [Google Scholar]
  67. Zaki, M.J.; Parthasarathy, S.; Ogihara, M.; Li, W. Parallel Algorithms for Discovery of Association Rules. Data Min. Knowl. Discov. 1997, 1, 343–373. [Google Scholar] [CrossRef]
  68. Cong, S.; Han, J.; Hoeflinger, J.; Padua, D. A sampling-based framework for parallel data mining. In Proceedings of the Tenth ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, Chicago, IL, USA, 15–17 June 2005; pp. 255–265. [Google Scholar]
  69. Shintani, T.; Kitsuregawa, M. Hash-based parallel algorithms for mining association rules. In Proceedings of the Fourth International Conference on Parallel and Distributed Information Systems, Miami Beach, FL, USA, 18–20 December 1996; pp. 19–30. [Google Scholar]
  70. Li, H.; Wang, Y.; Zhang, D.; Zhang, M.; Chang, E.Y. PFP: Parallel FP-growth for query recommendation. In Proceedings of the 2008 ACM Conference on Recommender Systems, Lausanne, Switzerland, 23–25 October 2008; pp. 107–114. [Google Scholar]
  71. Gabroveanu, M.; Cosulschi, M.; Slabu, F. Mining fuzzy association rules using MapReduce technique. In Proceedings of the International Symposium on INnovations in Intelligent SysTems and Applications, INISTA, Sinaia, Romania, 2–5 August 2016; pp. 1–8. [Google Scholar]
  72. Gabroveanu, M.; Iancu, I.; Coşulschi, M.; Constantinescu, N. Towards using grid services for mining fuzzy association rules. In Proceedings of the Ninth International Symposium on Symbolic and Numeric Algorithms for Scientific Computing, Timisoara, Romania, 26–29 September 2007; pp. 507–513. [Google Scholar]
  73. Palo, H.K.; Sahoo, S.; Subudhi, A.K. Dimensionality Reduction Techniques: Principles, Benefits, and Limitations. In Data Analytics in Bioinformatics: A Machine Learning Perspective; Willey: New York, NY, USA, 2021; pp. 77–107. [Google Scholar]
  74. Zhou, H.; Yu, K.-M.; Hsu, H.-P. Hybrid Modeling Method for Soft Sensing of Key Process Parameters in Chemical Industry. Sens. Mater. 2021, 33, 2789. [Google Scholar] [CrossRef]
  75. Priya, S.; Ward, C.; Locke, T.; Soni, N.; Maheshwarappa, R.P.; Monga, V.; Bathla, G. Glioblastoma and primary central nervous system lymphoma: Differentiation using MRI derived first-order texture analysis—A machine learning study. Neuroradiol. J. 2021, 34, 320–328. [Google Scholar] [CrossRef] [PubMed]
  76. Weerasuriya, A.U.; Zhang, X.; Lu, B.; Tse, K.T.; Liu, C.H. A Gaussian Process-Based emulator for modeling pedestrian-level wind field. Build. Environ. 2021, 188, 107500. [Google Scholar] [CrossRef]
  77. Uddin, M.P.; Mamun, M.A.; Afjal, M.I.; Hossain, M.A. Information-theoretic feature selection with segmentation-based folded principal component analysis (PCA) for hyperspectral image classification. Int. J. Remote Sens. 2021, 42, 286–321. [Google Scholar] [CrossRef]
  78. Bari, A.; Brower, W.; Davidson, C. Using Artificial Intelligence to Predict Legislative Votes in the United States Congress. In Proceedings of the 2021 IEEE 6th International Conference on Big Data Analytics (ICBDA), Xiamen, China, 5–8 March 2021; pp. 56–60. [Google Scholar]
  79. Nanehkaran, Y.A.; Chen, J.; Salimi, S.; Zhang, D. A pragmatic convolutional bagging ensemble learning for recognition of Farsi handwritten digits. J. Supercomput. 2021, 77, 13474–13493. [Google Scholar] [CrossRef]
  80. Kumari, P.; Toshniwal, D. Extreme gradient boosting and deep neural network based ensemble learning approach to forecast hourly solar irradiance. J. Clean. Prod. 2020, 279, 123285. [Google Scholar] [CrossRef]
  81. Cui, S.; Yin, Y.; Wang, D.; Li, Z.; Wang, Y. A stacking-based ensemble learning method for earthquake casualty prediction. Appl. Soft Comput. 2020, 101, 107038. [Google Scholar] [CrossRef]
  82. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  83. Kattenborn, T.; Leitloff, J.; Schiefer, F.; Hinz, S. Review on Convolutional Neural Networks (CNN) in vegetation remote sensing. ISPRS J. Photogramm. Remote Sens. 2021, 173, 24–49. [Google Scholar] [CrossRef]
  84. Kag, A.; Saligrama, V. Training Recurrent Neural Networks via Forward Propagation through Time. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 5189–5200. [Google Scholar]
  85. Yang, G.; Lv, J.; Chen, Y.; Huang, J.; Zhu, J. Generative Adversarial Networks (GAN) Powered Fast Magnetic Resonance Imaging—Mini Review, Comparison and Perspectives. arXiv 2021, arXiv:2105.01800. [Google Scholar]
  86. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  87. Rummery, G.A.; Niranjan, M. On-Line q-Learning Using Connectionist Systems; University of Cambridge, Department of Engineering: Cambridge, UK, 1994. [Google Scholar]
  88. Watkins, C. Learning from Delayed Rewards; King’s College: Cambridge, UK, 1989. [Google Scholar]
  89. Watkins, C.J.C.H.; Dayan, P. Technical Note: Q-Learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  90. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M.A. Playing atari with deep reinforcement learning. arXiv 2013, arXiv:1312.5602. [Google Scholar]
  91. Ann, O.C.; Theng, L.B. Human activity recognition: A review. In Proceedings of the 2014 IEEE International Conference on Control System, Computing and Engineering (ICCSCE 2014), Penang, Malaysia, 28–30 November 2014; pp. 389–393. [Google Scholar]
  92. Glover, F. Future paths for integer programming and links to artificial intelligence. Comput. Oper. Res. 1986, 13, 533–549. [Google Scholar] [CrossRef]
  93. Zhang, W.; Zhang, Y.; Gu, X.; Wu, C.; Han, L. Soft Computing. In Application of Soft Computing, Machine Learning, Deep Learning and Optimizations in Geoengineering and Geoscience; Springer: Singapore, 2022; pp. 7–19. [Google Scholar]
  94. Gavriluţ, V.; Pruski, A.; Berger, M.S. Constructive or Optimized: An Overview of Strategies to Design Networks for Time-Critical Applications. ACM Comput. Surv. 2022, 55, 1–35. [Google Scholar] [CrossRef]
  95. Sahoo, R.R.; Ray, M. Metaheuristic techniques for test case generation: A review. In Research Anthology on Agile Software, Software Development, and Testing; IGI Global: Hershey, PA, USA, 2022; pp. 1043–1058. [Google Scholar]
  96. Singh, R.M.; Awasthi, L.K.; Sikka, G. Towards Metaheuristic Scheduling Techniques in Cloud and Fog: An Extensive Taxonomic Review. ACM Comput. Surv. (CSUR) 2022, 55, 1–43. [Google Scholar] [CrossRef]
  97. Stevo, B.; Ante, F. The influence of pattern similarity and transfer learning upon the training of a base perceptron B2. In Proceedings of the Symposium Informatica, Gdańsk, Poland, 6–10 September 1976. [Google Scholar]
  98. Qi, W.; Su, H.; Yang, C.; Ferrigno, G.; De Momi, E.; Aliverti, A. A Fast and Robust Deep Convolutional Neural Networks for Complex Human Activity Recognition Using Smartphone. Sensors 2019, 19, 3731. [Google Scholar] [CrossRef] [Green Version]
  99. de Jesús, P.-V.R.; Palencia-Díaz, R. Teléfonos inteligentes y tabletas.¿ una herramienta o una barrera en la atención del paciente? Med. Interna De Mex. 2013, 29, 404–409. [Google Scholar]
  100. Organista-Sandoval, J.; McAnally-Salas, L.; Lavigne, G. El teléfono inteligente (smartphone) como herramienta pedagógica. Apertura 2013, 5, 6–19. [Google Scholar]
  101. Alonso, A.B.; Artime, I.F.; Rodríguez, M.Á.; Baniello, R.G. Dispositivos Móviles; EPSIG Ing. Telecomunicación Universidad de Oviedo: Oviedo, Spain, 2011. [Google Scholar]
  102. Anguita, D.; Ghio, A.; Oneto, L.; Parra Perez, X.; Reyes Ortiz, J.L. A Public Domain Dataset for Human Activity Recognition Using Smartphones. In Proceedings of the 21th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, ESANN 2013, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
  103. Sikder, N.; Nahid, A.-A. KU-HAR: An open dataset for heterogeneous human activity recognition. Pattern Recognit. Lett. 2021, 146, 46–54. [Google Scholar] [CrossRef]
  104. Popescu, A.-C.; Mocanu, I.; Cramariuc, B. PRECIS HAR. 2019. Available online: https://ieee-dataport.org/open-access/precis-har (accessed on 30 October 2021).
  105. Martínez-Villaseñor, L.; Ponce, H.; Brieva, J.; Moya-Albor, E.; Núñez-Martínez, J.; Peñafort-Asturiano, C. UP-Fall Detection Dataset: A Multimodal Approach. Sensors 2019, 19, 1988. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  106. Van Kasteren, T.; Noulas, A.; Englebienne, G.; Kröse, B. Accurate activity recognition in a home setting. In Proceedings of the 10th International Conference on Ubiquitous Computing, Seoul, Korea, 21–24 September 2008; pp. 1–9. [Google Scholar]
  107. Singla, G.; Cook, D.; Schmitter-Edgecombe, M. Recognizing independent and joint activities among multiple residents in smart environments. Ambient. Intell. Humaniz. Comput. J. 2010, 1, 57–63. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  108. Weiss, G.M.; Yoneda, K.; Hayajneh, T. Smartphone and Smartwatch-Based Biometrics Using Activities of Daily Living. IEEE Access 2019, 7, 133190–133202. [Google Scholar] [CrossRef]
  109. Gallissot, M.; Caelen, J.; Bonnefond, N.; Meillon, B.; Pons, S. Using the Multicom Domus Dataset; Research Report RR-LIG-020; LIG: Grenoble, France, 2011. [Google Scholar]
  110. Roggen, D.; Calatroni, A.; Rossi, M.; Holleczek, T.; Tröster, G.; Lukowicz, P.; Pirkl, G.; Bannach, D.; Ferscha, A.; Doppler, J.; et al. Collecting complex activity data sets in highly rich networked sensor environments. In Proceedings of the Seventh International Conference on Networked Sensing Systems (INSS’10), Kassel, Germany, 15–18 June 2010. [Google Scholar]
  111. Cook, D. Learning setting-generalized activity mdoels for smart spaces. IEEE Intell. Syst. 2010, 1. [Google Scholar] [CrossRef] [Green Version]
  112. Zhang, M.; Sawchuk, A.A. USC-HAD: A Daily Activity Dataset for Ubiquitous Activity Recognition Using Wearable Sensors. In Proceedings of the ACM International Conference on Ubiquitous Computing (UbiComp) Workshop on Situation, Activity and Goal Awareness (SAGAware), Pittsburgh, PA, USA, 5–8 September 2012. [Google Scholar]
  113. Logan, B.; Healey, B.J.; Philipose, J.M.; Tapia, E.M.; Intille, S. A long-term evaluation of sensing modalities for activity recognition. In Proceedings of the International Conference on Ubiquitous Computing, Taipei, Taiwan, 17–20 December 2007; Springer: Berlin/Heidelberg, Germany, 2007; pp. 483–500. [Google Scholar]
  114. Nugent, C.D.; Mulvenna, M.D.; Hong, X.; Devlin, S. Experiences in the development of a Smart Lab. Int. J. Biomed. Eng. Technol. 2009, 2, 319–331. [Google Scholar] [CrossRef]
  115. Schmitter-Edgecombe, M.; Cook, D.J. Assessing the Quality of Activities in a Smart Environment. Methods Inf. Med. 2009, 48, 480–485. [Google Scholar] [CrossRef] [Green Version]
  116. Reiss, A.; Stricker, D. Introducing a New Benchmarked Dataset for Activity Monitoring. In Proceedings of the 16th IEEE International Symposium on Wearable Computers (ISWC), Newcastle, UK, 18–22 June 2012. [Google Scholar]
  117. Banos, O.; Garcia, R.; Holgado, J.A.; Damas, M.; Pomares, H.; Rojas, I.; Saez, A.; Villalonga, C. mHealthDroid: A novel framework for agile development of mobile health applications. In Proceedings of the 6th International Work-conference on Ambient Assisted Living an Active Ageing (IWAAL 2014), Belfast, UK, 2–5 December 2014. [Google Scholar]
  118. Barshan, B.; Yüksek, M.C. Recognizing daily and sports activities in two open source machine learning environments using body-worn sensor units. Comput. J. 2014, 57, 1649–1667. [Google Scholar] [CrossRef]
  119. Espinilla, M.; Martínez, L.; Medina, J.; Nugent, C. The experience of developing theUJAmI Smart lab. IEEE Access. 2018, 6, 34631–34642. [Google Scholar] [CrossRef]
  120. Kitchenham, B.; Brereton, O.P.; Budgen, D.; Turner, M.; Bailey, J.; Linkman, S. Systematic literature reviews in software engineering—A systematic literature review. Inf. Softw. Technol. 2009, 51, 7–15. [Google Scholar] [CrossRef]
  121. Tasmin, M.; Ishtiak, T.; Ruman, S.U.; Suhan, A.U.R.C.; Islam, N.S.; Jahan, S.; Rahman, R.M. Comparative Study of Classifiers on Human Activity Recognition by Different Feature Engineering Techniques. In Proceedings of the 2020 IEEE 10th International Conference on Intelligent Systems (IS), Varna, Bulgaria, 28–30 August 2020; pp. 93–101. [Google Scholar]
  122. Igwe, O.M.; Wang, Y.; Giakos, G.C.; Fu, J. Human activity recognition in smart environments employing margin setting algorithm. J. Ambient Intell. Humaniz. Comput. 2020, 1–13. [Google Scholar] [CrossRef]
  123. Subasi, A.; Radhwan, M.; Kurdi, R.; Khateeb, K. IoT based mobile healthcare system for human activity recognition. In Proceedings of the 2018 15th Learning and Technology Conference (L&T), Jeddah, Saudi Arabia, 25–26 February 2018; pp. 29–34. [Google Scholar]
  124. Maswadi, K.; Ghani, N.A.; Hamid, S.; Rasheed, M.B. Human activity classification using Decision Tree and Naïve Bayes classifiers. Multimed. Tools Appl. 2021, 80, 21709–21726. [Google Scholar] [CrossRef]
  125. Damodaran, N.; Haruni, E.; Kokhkharova, M.; Schäfer, J. Device free human activity and fall recognition using WiFi channel state information (CSI). CCF Trans. Pervasive Comput. Interact. 2020, 2, 1–17. [Google Scholar] [CrossRef] [Green Version]
  126. Saha, J.; Chowdhury, C.; Biswas, S. Two phase ensemble classifier for smartphone based human activity recognition independent of hardware configuration and usage behaviour. Microsyst. Technol. 2018, 24, 2737–2752. [Google Scholar] [CrossRef]
  127. Das, A.; Kjærgaard, M.B. Activity Recognition using Multi-Class Classification inside an Educational Building. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Austin, TX, USA, 23–27 March 2020; pp. 1–6. [Google Scholar]
  128. Franco, P.; Martínez, J.M.; Kim, Y.C.; Ahmed, M.A. IoT based approach for load monitoring and activity recognition in smart homes. IEEE Access 2021, 9, 45325–45339. [Google Scholar] [CrossRef]
  129. Bozkurt, F. A Comparative Study on Classifying Human Activities Using Classical Machine and Deep Learning Methods. Arab. J. Sci. Eng. 2021, 47, 1507–1521. [Google Scholar] [CrossRef]
  130. Wang, A.; Zhao, S.; Zheng, C.; Chen, H.; Liu, L.; Chen, G. HierHAR: Sensor-Based Data-Driven Hierarchical Human Activity Recognition. IEEE Sens. J. 2020, 21, 3353–3365. [Google Scholar] [CrossRef]
  131. Oukrich, N. Daily Human Activity Recognition in Smart Home Based on Feature Selection, Neural Network and Load Signature of Appliances. Ph.D. Thesis, Université Mohamed V, Ecole Mohammadia d’Ingénieurs-Université Mohammed V de Rabat-Maroc, Rabat, Morocco, 2019. [Google Scholar]
  132. Demrozi, F.; Turetta, C.; Pravadelli, G. B-HAR: An open-source baseline framework for in depth study of human activity recognition datasets and workflows. arXiv 2021, arXiv:2101.10870. [Google Scholar]
  133. Xu, Z.; Wang, G.; Guo, X. Sensor-based activity recognition of solitary elderly via stigmergy and two-layer framework. Eng. Appl. Artif. Intell. 2020, 95, 10385. [Google Scholar] [CrossRef]
  134. Hussain, F.; Hussain, F.; Ehatisham-ul-Haq, M.; Azam, M.A. Activity-aware fall detection and recognition based on wearable sensors. IEEE Sens. J. 2019, 19, 4528–4536. [Google Scholar] [CrossRef]
  135. Liciotti, D.; Bernardini, M.; Romeo, L.; Frontoni, E. A sequential deep learning application for recognising human activities in smart homes. Neurocomputing 2019, 396, 501–513. [Google Scholar] [CrossRef]
  136. Mohmed, G.; Lotfi, A.; Langensiepen, C.; Pourabdollah, A. Clustering-based fuzzy finite state machine for human activity recognition. In UK Workshop on Computational Intelligence; Springer: Cham, Switzerland, 2018; pp. 264–275. [Google Scholar]
  137. Brena, R.F.; Garcia-Ceja, E. A crowdsourcing approach for personalization in human activities recognition. Intell. Data Anal. 2017, 21, 721–738. [Google Scholar] [CrossRef]
  138. He, H.; Tan, Y.; Zhang, W. A wavelet tensor fuzzy clustering scheme for multi-sensor human activity recognition. Eng. Appl. Artif. Intell. 2018, 70, 109–122. [Google Scholar] [CrossRef]
  139. Wang, X.; Lu, Y.; Wang, D.; Liu, L.; Zhou, H. Using jaccard distance measure for unsupervised activity recognition with smartphone accelerometers. In Proceedings of the Asia-Pacific Web (apweb) and Web-Age Information Management (WAIM) Joint Conference on Web and Big Data, Beijing, China, 7–9 July 2017; Springer: Cham, Switzerland, 2017; pp. 74–83. [Google Scholar]
  140. Bota, P.; Silva, J.; Folgado, D.; Gamboa, H. A Semi-Automatic Annotation Approach for Human Activity Recognition. Sensors 2019, 19, 501. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  141. Yacchirema, D.; de Puga, J.S.; Palau, C.; Esteve, M. Fall detection system for elderly people using IoT and ensemble machine learning algorithm. Pers. Ubiquitous Comput. 2019, 23, 801–817. [Google Scholar] [CrossRef]
  142. Manzi, A.; Dario, P.; Cavallo, F. A Human Activity Recognition System Based on Dynamic Clustering of Skeleton Data. Sensors 2017, 17, 1100. [Google Scholar] [CrossRef] [Green Version]
  143. Ma, H.; Zhang, Z.; Li, W.; Lu, S. Unsupervised Human Activity Representation Learning with Multi-task Deep Clustering. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021, 5, 1–25. [Google Scholar] [CrossRef]
  144. Budisteanu, E.A.; Mocanu, I.G. Combining Supervised and Unsupervised Learning Algorithms for Human Activity Recognition. Sensors 2021, 21, 6309. [Google Scholar] [CrossRef]
  145. Xu, S.; Tang, Q.; Jin, L.; Pan, Z. A Cascade Ensemble Learning Model for Human Activity Recognition with Smartphones. Sensors 2019, 19, 2307. [Google Scholar] [CrossRef] [Green Version]
  146. Choudhury, N.A.; Moulik, S.; Roy, D.S. Physique-Based Human Activity Recognition Using Ensemble Learning and Smartphone Sensors. IEEE Sens. J. 2021, 21, 16852–16860. [Google Scholar] [CrossRef]
  147. Wang, Z.; Wu, D.; Chen, J.; Ghoneim, A.; Hossain, M.A. A Triaxial Accelerometer-Based Human Activity Recognition via EEMD-Based Features and Game-Theory-Based Feature Selection. IEEE Sens. J. 2016, 16, 3198–3207. [Google Scholar] [CrossRef]
  148. Jethanandani, M.; Sharma, A.; Perumal, T.; Chang, J.R. Multi-label classification based ensemble learning for human activity recognition in smart home. Internet Things 2020, 12, 100324. [Google Scholar] [CrossRef]
  149. Subasi, A.; Dammas, D.H.; Alghamdi, R.D.; Makawi, R.A.; Albiety, E.A.; Brahimi, T.; Sarirete, A. Sensor Based Human Activity Recognition Using Adaboost Ensemble Classifier. Procedia Comput. Sci. 2018, 140, 104–111. [Google Scholar] [CrossRef]
  150. Padmaja, B.; Prasa, V.; Sunitha, K. A Novel Random Split Point Procedure Using Extremely Randomized (Extra) Trees Ensemble Method for Human Activity Recognition. EAI Endorsed Trans. Pervasive Health Technol. 2020, 6, e5. [Google Scholar] [CrossRef]
  151. Nweke, H.F.; Teh, Y.W.; Al-Garadi, M.A.; Alo, U.R. Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges. Expert Syst. Appl. 2018, 105, 233–261. [Google Scholar] [CrossRef]
  152. Ma, L.; Cheng, S.; Shi, Y. Enhancing learning efficiency of brain storm optimization via orthogonal learning design. IEEE Trans. Syst. Man Cybern. Syst. 2020, 51, 6723–6742. [Google Scholar] [CrossRef]
  153. Wan, S.; Qi, L.; Xu, X.; Tong, C.; Gu, Z. Deep Learning Models for Real-time Human Activity Recognition with Smartphones. Mob. Netw. Appl. 2019, 25, 743–755. [Google Scholar] [CrossRef]
  154. Akula, A.; Shah, A.K.; Ghosh, R. Deep learning approach for human action recognition in infrared images. Cogn. Syst. Res. 2018, 50, 146–154. [Google Scholar] [CrossRef]
  155. He, J.; Zhang, Q.; Wang, L.; Pei, L. Weakly supervised human activity recognition from wearable sensors by recurrent attention learning. IEEE Sens. J. 2018, 19, 2287–2297. [Google Scholar] [CrossRef]
  156. Long, J.; Sun, W.; Yang, Z.; Raymond, O.I. Asymmetric Residual Neural Network for Accurate Human Activity Recognition. Information 2019, 10, 203. [Google Scholar] [CrossRef] [Green Version]
  157. Ariza-Colpas, P.; Morales-Ortega, R.; Piñeres-Melo, M.A.; Melendez-Pertuz, F.; Serrano-Torné, G.; Hernandez-Sanchez, G.; Martínez-Osorio, H. Teleagro: Iot applications for the georeferencing and detection of zeal in cattle. In Proceedings of the IFIP International Conference on Computer Information Systems and Industrial Management, Belgrade, Serbia, 19–21 September 2019; pp. 232–239. [Google Scholar]
  158. Mekruksavanich, S.; Jitpattanakul, A. Deep convolutional neural network with rnns for complex activity recognition using wrist-worn wearable sensor data. Electronics 2021, 10, 1685. [Google Scholar] [CrossRef]
  159. Papagiannaki, A.; Zacharaki, E.I.; Kalouris, G.; Kalogiannis, S.; Deltouzos, K.; Ellul, J.; Megalooikonomou, V. Recognizing Physical Activity of Older People from Wearable Sensors and Inconsistent Data. Sensors 2019, 19, 880. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  160. Hassan, M.M.; Uddin, Z.; Mohamed, A.; Almogren, A. A robust human activity recognition system using smartphone sensors and deep learning. Futur. Gener. Comput. Syst. 2018, 81, 307–313. [Google Scholar] [CrossRef]
  161. Berlin, S.J.; John, M. R-STDP Based Spiking Neural Network for Human Action Recognition. Appl. Artif. Intell. 2020, 34, 656–673. [Google Scholar] [CrossRef]
  162. Lu, Y.; Li, Y.; Velipasalar, S. Efficient human activity classification from egocentric videos incorporating actor-critic reinforcement learning. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 564–568. [Google Scholar]
  163. Hossain, H.S.; Roy, N. Active deep learning for activity recognition with context-aware annotator selection. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 1862–1870. [Google Scholar]
  164. Zhou, X.; Liang, W.; Wang, K.I.-K.; Wang, H.; Yang, L.T.; Jin, Q. Deep-Learning-Enhanced Human Activity Recognition for Internet of Healthcare Things. IEEE Internet Things J. 2020, 7, 6429–6438. [Google Scholar] [CrossRef]
  165. Xu, W.; Miao, Z.; Yu, J.; Ji, Q. Deep Reinforcement Learning for Weak Human Activity Localization. IEEE Trans. Image Process. 2019, 29, 1522–1535. [Google Scholar] [CrossRef]
  166. Chen, K.; Yao, L.; Zhang, D.; Wang, X.; Chang, X.; Nie, F. A Semisupervised Recurrent Convolutional Attention Model for Human Activity Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1747–1756. [Google Scholar] [CrossRef]
  167. Possas, R.; Caceres, S.P.; Ramos, F. Egocentric activity recognition on a budget. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5967–5976. [Google Scholar]
  168. Liu, G.; Ma, R.; Hao, Q. A Reinforcement Learning Based Design of Compressive Sensing Systems for Human Activity Recognition. In Proceedings of the 2018 IEEE SENSORS, New Delhi, India, 28–31 October 2018; pp. 1–4. [Google Scholar]
  169. Shen, X.; Guo, L.; Lu, Z.; Wen, X.; Zhou, S. WiAgent: Link Selection for CSI-Based Activity Recognition in Densely Deployed Wi-Fi Environments. In Proceedings of the 2021 IEEE Wireless Communications and Networking Conference (WCNC), Nanjing, China, 29 March–1 April 2021; pp. 1–6. [Google Scholar]
  170. Campbell, C.; Ahmad, F. Attention-augmented convolutional autoencoder for radar-based human activity recognition. In Proceedings of the 2020 IEEE International Radar Conference (RADAR), Washington, DC, USA, 28–30 April 2020; pp. 990–995. [Google Scholar]
  171. Nguyen, T.D.; Huynh, T.T.; Pham, H.A. An improved human activity recognition by using genetic algorithm to optimize feature vector. In Proceedings of the 2018 10th International Conference on Knowledge and Systems Engineering (KSE), Ho Chi Minh City, Vietnam, 1–3 November 2018; pp. 123–128. [Google Scholar]
  172. Mocanu, I.; Axinte, D.; Cramariuc, O.; Cramariuc, B. Human activity recognition with convolution neural network using tiago robot. In Proceedings of the 2018 41st International Conference on Telecommunications and Signal Processing (TSP), Athens, Greece, 4–6 July 2018; pp. 1–4. [Google Scholar]
  173. El-Maaty, A.M.A.; Wassal, A.G. Hybrid GA-PCA feature selection approach for inertial human activity recognition. In Proceedings of the 2018 IEEE Symposium Series on Computational Intelligence (SSCI), Bangalore, India, 18–21 November 2018; pp. 1027–1032. [Google Scholar]
  174. Baldominos, A.; Saez, Y.; Isasi, P. Model selection in committees of evolved convolutional neural networks using genetic algorithms. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Madrid, Spain, 21–23 November 2018; Springer: Cham, Switzerland, 2018; pp. 364–373. [Google Scholar]
  175. Saba, T.; Rehman, A.; Latif, R.; Fati, S.M.; Raza, M.; Sharif, M. Suspicious Activity Recognition Using Proposed Deep L4-Branched-Actionnet With Entropy Coded Ant Colony System Optimization. IEEE Access 2021, 9, 89181–89197. [Google Scholar] [CrossRef]
  176. Li, J.; Tian, L.; Chen, L.; Wang, H.; Cao, T.; Yu, L. Optimal feature selection for activity recognition based on ant colony algorithm. In Proceedings of the 2019 14th IEEE Conference on Industrial Electronics and Applications (ICIEA), Xi’an, China, 19–21 June 2019; pp. 2356–2362. [Google Scholar]
  177. Fan, C.; Gao, F. Enhanced human activity recognition using wearable sensors via a hybrid feature selection method. Sensors 2021, 21, 6434. [Google Scholar] [CrossRef]
  178. Jalal, A.; Batool, M.; Kim, K. Stochastic Recognition of Physical Activity and Healthcare Using Tri-Axial Inertial Wearable Sensors. Appl. Sci. 2020, 10, 7122. [Google Scholar] [CrossRef]
  179. Arshad, S.; Feng, C.; Yu, R.; Liu, Y. Leveraging transfer learning in multiple human activity recognition using wifi signal. In Proceedings of the 2019 IEEE 20th International Symposium on “A World of Wireless, Mobile and Multimedia Networks” (WoWMoM), Washington, DC, USA, 10–12 June 2019; pp. 1–10. [Google Scholar]
  180. Soleimani, E.; Nazerfard, E. Cross-subject transfer learning in human activity recognition systems using generative adversarial networks. Neurocomputing 2021, 426, 26–34. [Google Scholar] [CrossRef]
  181. Ding, R.; Li, X.; Nie, L.; Li, J.; Si, X.; Chu, D.; Zhan, D. Empirical study and improvement on deep transfer learning for human activity recognition. Sensors 2018, 19, 57. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  182. Fu, Z.; He, X.; Wang, E.; Huo, J.; Huang, J.; Wu, D. Personalized Human Activity Recognition Based on Integrated Wearable Sensor and Transfer Learning. Sensors 2021, 21, 885. [Google Scholar] [CrossRef] [PubMed]
  183. Deep, S.; Zheng, X. Leveraging CNN and transfer learning for vision-based human activity recognition. In Proceedings of the 29th International Telecommunication Networks and Applications Conference (ITNAC), Auckland, New Zealand, 27–29 November 2019; pp. 1–4. [Google Scholar]
  184. Hoelzemann, A.; Van Laerhoven, K. Digging deeper: Towards a better understanding of transfer learning for human activity recognition. In Proceedings of the 2020 International Symposium on Wearable Computers, Virtual, 12–17 September 2020; pp. 50–54. [Google Scholar]
  185. Wang, J.; Zheng, V.W.; Chen, Y.; Huang, M. Deep transfer learning for cross-domain activity recognition. In Proceedings of the 3rd International Conference on Crowd Science and Engineering, Singapore, 28–31 July 2018; pp. 1–8. [Google Scholar]
  186. Mutegeki, R.; Han, D.S. Feature-representation transfer learning for human activity recognition. In Proceedings of the 2019 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Korea, 16–18 October 2019; pp. 18–20. [Google Scholar]
  187. Khan, M.A.A.H.; Roy, N. Transact: Transfer learning enabled activity recognition. In Proceedings of the 2017 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), Kona, HI, USA, 13–17 March 2017; pp. 545–550. [Google Scholar]
  188. Ding, X.; Jiang, T.; Li, Y.; Xue, W.; Zhong, Y. Device-free location-independent human activity recognition using transfer learning based on CNN. In Proceedings of the 2020 IEEE International Conference on Communications Workshops (ICC Workshops), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar]
  189. Kalouris, G.; Zacharaki, E.I.; Megalooikonomou, V. Improving CNN-based activity recognition by data augmentation and transfer learning. In Proceedings of the 2019 IEEE 17th International Conference on Industrial Informatics (INDIN), Helsinki, Finland, 22–25 July 2019; Volume 1, pp. 1387–1394. [Google Scholar]
  190. Rokni, S.A.; Ghasemzadeh, H. Autonomous Training of Activity Recognition Algorithms in Mobile Sensors: A Transfer Learning Approach in Context-Invariant Views. IEEE Trans. Mob. Comput. 2018, 17, 1764–1777. [Google Scholar] [CrossRef]
  191. Verma, K.K.; Singh, B.M. Vision based Human Activity Recognition using Deep Transfer Learning and Support Vector Machine. In Proceedings of the 2021 IEEE 8th Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON), Dehradun, India, 11–13 November 2021; pp. 1–9. [Google Scholar]
  192. Xiao, F.; Pei, L.; Chu, L.; Zou, D.; Yu, W.; Zhu, Y.; Li, T. A deep learning method for complex human activity recognition using virtual wearable sensors. In Proceedings of the International Conference on Spatial Data and Intelligence, Virtual, 8–9 May 2020; Springer: Cham, Switzerland, 2020; pp. 261–270. [Google Scholar]
  193. Faridee, A.Z.M.; Khan, M.A.A.H.; Pathak, N.; Roy, N. AugToAct: Scaling complex human activity recognition with few labels. In Proceedings of the 16th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services, Houston, TX, USA, 12–14 November 2019; pp. 162–171. [Google Scholar]
  194. Mutegeki, R.; Han, D.S. A CNN-LSTM approach to human activity recognition. In Proceedings of the 2020 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Fukuoka, Japan, 19–21 February 2020; pp. 362–366. [Google Scholar]
Figure 1. HAR Approach Concepts Maps.
Figure 1. HAR Approach Concepts Maps.
Sensors 22 03401 g001
Figure 2. Sensors of Human Activity Recognition.
Figure 2. Sensors of Human Activity Recognition.
Sensors 22 03401 g002
Figure 3. Relationship between concepts for the literature review.
Figure 3. Relationship between concepts for the literature review.
Sensors 22 03401 g003
Figure 4. Years of publication of the articles.
Figure 4. Years of publication of the articles.
Sensors 22 03401 g004
Figure 5. Years of publication of the articles.
Figure 5. Years of publication of the articles.
Sensors 22 03401 g005
Figure 6. (a) Publications division according to typology (b). Distribution by quartiles of publications.
Figure 6. (a) Publications division according to typology (b). Distribution by quartiles of publications.
Sensors 22 03401 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ariza-Colpas, P.P.; Vicario, E.; Oviedo-Carrascal, A.I.; Butt Aziz, S.; Piñeres-Melo, M.A.; Quintero-Linero, A.; Patara, F. Human Activity Recognition Data Analysis: History, Evolutions, and New Trends. Sensors 2022, 22, 3401. https://doi.org/10.3390/s22093401

AMA Style

Ariza-Colpas PP, Vicario E, Oviedo-Carrascal AI, Butt Aziz S, Piñeres-Melo MA, Quintero-Linero A, Patara F. Human Activity Recognition Data Analysis: History, Evolutions, and New Trends. Sensors. 2022; 22(9):3401. https://doi.org/10.3390/s22093401

Chicago/Turabian Style

Ariza-Colpas, Paola Patricia, Enrico Vicario, Ana Isabel Oviedo-Carrascal, Shariq Butt Aziz, Marlon Alberto Piñeres-Melo, Alejandra Quintero-Linero, and Fulvio Patara. 2022. "Human Activity Recognition Data Analysis: History, Evolutions, and New Trends" Sensors 22, no. 9: 3401. https://doi.org/10.3390/s22093401

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop