Next Article in Journal
Research on Vibration Amplitude of Ultra-Precision Aerostatic Motorized Spindle under the Combined Action of Rotor Unbalance and Hydrodynamic Effect
Next Article in Special Issue
An Enhanced Ensemble Deep Neural Network Approach for Elderly Fall Detection System Based on Wearable Sensors
Previous Article in Journal
A Hybrid Software and Hardware SDN Simulation Testbed
Previous Article in Special Issue
A Novel Privacy Preservation and Quantification Methodology for Implementing Home-Care-Oriented Movement Analysis Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Neural Network for the Detections of Fall and Physical Activities Using Foot Pressures and Inertial Sensing

1
Department of Electrical Engineering, Chang Gung University, Taoyuan 333, Taiwan
2
Department of Biomedical Engineering, Chang Gung University, Taoyuan 333, Taiwan
3
Neuroscience Research Center, Chang Gung Memorial Hospital, Linkou, Taoyuan 333, Taiwan
4
Department of Neurology, Chang Gung Memorial Hospital, Linkou, Taoyuan 333, Taiwan
5
School of Physical Therapy and Graduate Institute of Rehabilitation Science, College of Medicine, and Health Aging Research Center, Chang Gung University, Taoyuan 333, Taiwan
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(1), 495; https://doi.org/10.3390/s23010495
Submission received: 1 December 2022 / Revised: 16 December 2022 / Accepted: 28 December 2022 / Published: 2 January 2023
(This article belongs to the Special Issue Wearable Sensors for Gait and Falls Monitoring)

Abstract

:
Fall detection and physical activity (PA) classification are important health maintenance issues for the elderly and people with mobility dysfunctions. The literature review showed that most studies concerning fall detection and PA classification addressed these issues individually, and many were based on inertial sensing from the trunk and upper extremities. While shoes are common footwear in daily off-bed activities, most of the aforementioned studies did not focus much on shoe-based measurements. In this paper, we propose a novel footwear approach to detect falls and classify various types of PAs based on a convolutional neural network and recurrent neural network hybrid. The footwear-based detections using deep-learning technology were demonstrated to be efficient based on the data collected from 32 participants, each performing simulated falls and various types of PAs: fall detection with inertial measures had a higher F1-score than detection using foot pressures; the detections of dynamic PAs (jump, jog, walks) had higher F1-scores while using inertial measures, whereas the detections of static PAs (sit, stand) had higher F1-scores while using foot pressures; the combination of foot pressures and inertial measures was most efficient in detecting fall, static, and dynamic PAs.

1. Introduction

Increasingly, research and discussion have focused on the recognition of physical activity (PA) and fall detection in recent years [1,2]. PA interventions have been evidenced to reduce the risks of cardiometabolic syndrome, falls, depression, anxiety, and dementia [3]. PA interventions have also attracted increasing interest for their potential health benefits in various diagnostic populations [4]. Several studies have even developed disease-related PA markers to differentiate the distribution of activity levels in individuals with chronic cardiovascular disease from healthy individuals [5]. On the other hand, fall detection provides a practical solution for fall scenarios, calling for help automatically if they occur. In particular, falls are a debilitating problem among the elderly [6] and individuals with Parkinson’s disease [7], multiple sclerosis [8], and stroke [9]. Usually, a fear of falling restricts an individual’s participation in daily activities [10,11,12]. The incorporation of fall detectors into mobile assistive technology may enable older people to live independently at home [13,14].
Currently, vision-based [15,16], inertial sensing-based detections [1,2], and hybrid frameworks [17,18] are the major approaches for PA monitoring and fall detection. The vision-based approaches use single or multiple cameras to capture human postures or movements [19]. Rapid changes in human posture are linked with fall incidence [20,21], whereas specific postures or patterns of human movements are present in various types of PAs [15]. Recently, several advanced methods based on deep neural networks have been proposed to extract the discriminant patterns related to human postures or movements from images through machine learning [22,23,24]. Vision-based approaches are beneficial due to the characteristic of wear-free, unobstructed monitoring, but it is only available in spaces where cameras have been installed.
In contrast to the restrictions of the vision-based approaches, the inertial-based approaches attach wearable devices to the human body or limbs to record the static and dynamic activity of daily life (ADL) anywhere. Popular wearable devices, such as Fitbit® and JawboneUP®, employ an accelerometer to estimate activity levels, sleep quality, and energy expenditure. Recently, wearable PA monitors have undergone extensive development by attaching different varieties of sensors to the body to capture human motion data, from which several detection algorithms for various kinds of PAs have been developed. In these, inertial sensing based on accelerometers [25,26,27], paralleled with gyroscopic measurements [28,29], or further combined with magnetic field measurements [30,31], are used to detect various PAs.
Using similar sensor technology, accelerometers or gyroscopes are mostly used to detect falls during PAs. A great acceleration caused by the impact on the ground provides a direct marker for detecting falls [32,33,34,35,36,37]. Posture identification after impact can be used to reduce the false positive rate of fall detection [38,39,40,41]. In addition, a fall event can be characterized by the increasing angular velocity that peaks at the moment a knee or hip hits the ground [42]; therefore, gyroscopic measurements provide additional criteria for fall detection [43,44].
In recent years, deep learning neural networks have provided new ways to detect falls during PAs or classify various types of PAs. A convolutional neural network (CNN) is used to capture the discernible features from the inertial data. The succeeding fully connected network is then used as a classifier. The CNN and fully connected network weights are derived from machine learning based on the labeled data. The CNN-based model has been used to detect falls during PAs [45,46,47] and classify various types of PAs [48,49,50,51,52].
On the other hand, recurrent neuronal networks (RNNs) can aggregate the features of the inertial data at various time points to capture the sequential pattern of the fall or physical activity. The aggregated features are then used to detect the falls during PAs [48,53,54] and classify various types of PAs [55] through a fully connected network. In addition, a hybrid approach uses CNN in the first layer to extract the temporal features from the raw inertial data and applies RNN in the second layer to detect falls during PAs [56,57,58] and classify various types of PAs [59,60,61].
State-of-the-art PA classification and fall detection are illustrated in Table 1 and Table 2, respectively. Most of these studies placed the sensors on different parts of the body, including the waist, chest, back trunk, wrist, arm, leg, etc. Each placement has unique advantages in practical applications, e.g., sensors can be easily attached to certain areas, or it is easier to capture the main body motion or distinct features of specific movements using other areas. On the other hand, placing such sensors in shoes is also beneficial for detecting lower-extremity movements. A study attached an inertial measurement unit (IMU) containing an accelerometer and a gyroscope in a subject’s left shoe to detect different walking conditions (level walking, descending and ascending slope walking, walking downstairs, and walking upstairs) using a prediction model based on dorsi–plantar flexion angular velocities or a CNN based on raw inertial data [62]. Another study used a combination of inertial sensing on the trunk and thigh and foot pressure to classify postural transitions, locomotion, and walking type using a decision tree [63]. As for fall detection, one study used the embedded accelerometer in shoes to detect falls during PAs using empirical rules based on acceleration magnitude and the estimated posture [64]; another study employed an algorithm based on waist acceleration and the center of plantar pressure to detect falls during PAs [65].
As shoes are commonly worn during off-bed activities, they are worn during most outdoor activities. However, this methodology is rarely adopted for PAs and fall detection in the literature. In this work, we focus on the detection of falls and various types of PAs in parallel via footwear sensing and deep-learning technology. A novel footwear sensing system has been developed for instrumented shoes that are equipped with 11 force-sensitive resistors (FSRs) in the insole to measure foot pressures, and an IMU with an accelerometer, a gyroscope, and a magnetometer to measure foot inertial dynamics on each side. A novel deep-learning neural network based on a hybrid CNN and RNN model was used to learn the discernible features from the inertial signals and foot pressures and detect simulated falls and various types of PAs. To validate the feasibility of the proposed footwear approach, we define four kinds of simulated falls and seven types of activities of daily living, including jumping, jogging, level walking, walking downstairs, walking upstairs, sitting, and standing. The detections based on various combinations of signals and parameters were validated and compared based on a dataset collected in this study.

2. Materials and Methods

2.1. Subjects and Experimental Activities

A sample of 32 healthy participants (16 males and 16 females; age of 21.5 ± 2.0 yrs; height of 167.2 ± 7.1 cm; weight of 62.4 ± 11.3 kg) performed simulated falls and physical activities of daily life. Every participant performed four kinds of simulated falls and seven other types of PAs. Every simulated fall and PA were repeated three times by each of the participants. The protocol of this study was approved by the Research Ethics Committee of the Chang Gung Medical Foundation (IRB#201802118B0) in accordance with the Helsinki Declaration. All of the participants provided written informed consent.
The participants were protected by a thick mattress while undertaking the simulated falls. The four kinds of simulated falls were defined as follows:
(1)
A backward fall during stand-to-sit.
(2)
A forward fall, a lateral fall toward the left side, and a lateral fall toward the right side during walking.
(3)
A forward fall, a lateral fall toward the left side, and a lateral fall toward the right side from standing.
(4)
A forward fall, a lateral fall toward the left side, and a lateral fall toward the right side during sit-to-stand.
The seven types of PAs were defined as follows:
(1)
A single jump and a continuous jump.
(2)
A forward jog at an expected speed of 1.6 m/s and jogging in place with an expected time interval of 1.5 s.
(3)
Forward walking at two expected speeds (0.8 and 1.3 m/s), walking backward, and walking in place with an expected time interval of 2 s on level ground.
(4)
Walking downstairs.
(5)
Walking upstairs.
(6)
Sitting straight and sitting with legs stretched out.
(7)
Standing still, standing with hands raising, and standing with body swing in anteroposterior/left-right/up-down direction.

2.2. Sensing and Recording System

An ARM Cortex-M4 microcontroller (M451RG6AE, Nuvoton Tech. Corp., Hsinchu, Taiwan) received the digital data from an IMU (LSM9DS1, STMicroelectronics, Geneva, Switzerland) with a full-scale, ranged ±4 g tri-axial accelerometer, a full-scale, ranged ±500 degrees/s (dps) ±12 Gauss tri-axial magnetometer, via a serial peripheral interface bus at a sampling rate of 100 Hz. As shown in Figure 1, a customized insole with 11 FSRs (UNEO Incorporated, New Taipei City, Taiwan) was used to capture the foot pressures at the big toe, little toe, metatarsus (medial, middle, lateral), arches (medial, lateral), fore heels (medial, lateral), and heels (medial, lateral). The FSRs were constructed of a resistance-type piezo-resistive polymer composite made using processing and printing-based micromachining technology. Each FSR had a sensing range of 1 to 5 kg/cm2 and was individually calibrated using elastic-film pressurization to reduce the resistance variance between the sensors. The microcontroller digitized the transformed voltages from these FSRs through a built-in 12-bit analog-to-digital converter at a sampling rate of 100 Hz. All of the acquired samples were wirelessly transmitted to a notebook computer through a BLE 4.2 Bluetooth module (JDY-18, Shenzhen Innovation Technology, Shenzhen, China). The customized insole had a height of 260 mm, a metatarsus width of 850 mm, a heel width of 550 mm, and a 0.63 mm thickness. Therefore, we just included participants whose foot size could match the size of the customized insole as much as possible.
The sensing devices were fixed at the lateral aspect or insole of each shoe in parallel to measure the foot inertial data and foot pressures, respectively. A graphical user interface developed in the PyQt Designer (Riverbank Computing, Dorchester, UK) collected these data in parallel with a video recording from a Kinect V2 camera (Microsoft Corp., Redmond, WA, USA) at a rate of 30 frames/s.

2.3. Fall and PA Detection Network

A network to detect falls and various types of PAs was constructed based on the CNN and RNN, which are constructed using a deep residual network (DRN) and a bidirectional long short-term memory (LSTM) network. The inputs can be foot inertial data and/or three parameterized data from a single foot or both feet in a 3-s window. The inputs can also be foot pressures and/or the center of pressure (CoP) from a single foot or both feet. The combinations of these data and parameters are also used.
The foot inertial data and three parameterized data are described as follows:
(1)
Tri-axial accelerations: ax(i), ay(i), az(i)
(2)
Tri-axial angular velocities: ωx(i), ωy(i), ωz(i)
(3)
Acceleration amplitude (AM) defined as the square root of the sum of tri-axial accelerations [32]:
AM ( i ) = a x 2 ( i ) + a y 2 ( i ) + a z 2 ( i )
(4)
Acceleration cubic-product-root magnitude (ACM) defined as the cube root of the product of tri-axial absolute accelerations [44]:
ACM ( i ) = | a x ( i ) × a y ( i ) × a z ( i ) | 3
(5)
Angular velocity cubic-product-root magnitude (AVCM) defined as the cube root of the product of tri-axial absolute angular velocities [44]:
AVCM ( i ) = | ω x ( i ) × ω y ( i ) × ω z ( i ) | 3
Figure 2 shows the tri-axial accelerations, tri-axial angular velocities, and the corresponding inertial parameters (AM, ACM, and AVCM) while a participant was undertaking a forward fall, a jump, a forward walk, and a forward jog. Both falling and jumping produced abrupt changes in accelerations and angular velocities. Two distinct acceleration peaks were generated particularly while jumping off ground and back to ground. Both walking and jogging produced repeated changes in accelerations and angular velocities.
The foot pressures and their CoP are described as follows:
(1)
Eleven individual foot pressures: fp1(i), fp2(i), …, fp11(i)
(2)
CoP in horizontal direction (CoPx)
(3)
CoP in vertical direction (CoPy)
CoPx and CoPy of the left or right foot are, respectively, defined as the sum of the products of the eleven individual foot pressures and their x and y positions divided by the sum of the eleven foot pressures:
CoPx ( i ) = ( f p n ( i ) x n ) f p n ( i )
CoPy ( i ) = ( f p n ( i ) y n ) f p n ( i )
where xn and yn indicate the location of the centroid of the n-th FSR relative to the local reference frame.
The CoP of both feet is defined as the weighted sum of the left-side CoP and right-side CoP. When non-ground contact is detected, CoP is set to the center of a single foot or the center of both feet.
Figure 3 shows the foot pressures at the metatarsus and heel areas, and the CoP trajectories during the same activity events. The foot pressures were close to zero after the fall incidence. There were also two distinct peak pressures at the metatarsus area while jumping off ground and back to ground. The foot pressures at the metatarsus and heel areas were interlaced while walking or jogging, which were linked to toe-off and heel-strike. Moreover, various CoP trajectory patterns were presented in falling, jumping, walking, and jogging. In particular, both walking and jogging produced butterfly patterns.
The inputs, therefore, have a size of M (channels) × 300 (points). As shown in Figure 4, a DRN is constructed by a stack of three residual units (RUs). Each RU is composed of six convolutional layers and a skip layer. In each convolutional layer, an equivalent number of 1-D filters are used to capture the temporal patterns of the inputs individually, preserving the temporal dimensions (stride 1, same padding); the temporal patterns are then summed along the channel dimension and outputted through a ReLU activation function. Each convolutional layer has multiple sets of 1-D filters and outputs one feature channel per set. A max-pooling layer subsamples the outputs of the last convolutional layer to halve the dimensionality of the features. The skip layer adds the inputs of the RU directly to the halved output features through a convolution layer with a stride of 2 and the right number of output channels. The detail of the deep residual network is listed in Table 3. The kernel size is the same within each RU but differs across RUs.
The benefits of the residual network have been shown through its easier optimization and the accuracy obtained from the increased depth [66]. The purpose of using three RUs is to extract the low-, middle-, and high-level features sequentially from the signals. The kernel size of these three RUs was 11, 9, and 5, respectively. With the increased level, the number of feature maps is increased, and the dimensionality of the features is halved.
The output of the deep residual network can be viewed as a sequence of feature vectors (38 vectors with vector size 108) such that it is inputted to a bidirectional LSTM network. The current input vector and previous short-term state vector are fed into a fully connected layer on which the LSTM cell generates a short-term vector and a long-term vector (each has a vector size of 64) to the next state. After the bidirectional LSTM network has been traversed in a forward direction, it is then traversed backward. The short-term vector of the last state is used as an input to a terminal fully connected network, which contains one hidden layer of 32 neurons and one output layer of 8 neurons corresponding to a fall and 7 types of PAs. The activation function in the hidden layer is chosen as the ReLU function, where the output layer uses a softmax function to represent a categorical probability distribution.
A batch containing 96 randomly chosen instances was inputted into the deep neural network. The corresponding output vectors were calculated via forward propagation. Therefore, the cross-entropy loss was computed based on the forward output vectors and the one-hot target vectors. A backpropagation of the errors was subsequently applied to update the network weights using the Adam optimization algorithm, which estimates the updates using a running average of the first and second moment of the gradient [67].
The settings of the hyper-parameters with a brief description of the forward processing and consideration are summarized in the following. These settings were determined through various tests on our dataset and achieved the optimal prediction performance.
(1)
A batch size of 96 to achieve a balance between the robustness of stochastic gradient descent and the efficiency of batch gradient descent.
(2)
Three RUs to extract low-, middle-, and high-level features from signals: 1D convolutions with a kernel size of 11 in the first RU to capture the temporal patterns of signals (~110 ms), 9 in the second RU to capture the low-level temporal features, and 5 in the third RU to aggregate the middle-level features.
(3)
36, 72, and 108 feature maps to generate sufficient low-, middle- and high-level features.
(4)
Two sequential 64-unit LSTM cells to cluster the high-level features at various time points.
(5)
A fully connected network containing a hidden layer of 32 neurons to aggregate the last-state short-term vectors and an output layer of 8 neurons to classify falls and 7 types of PAs.
The proposed detection network was implemented in the Visual Studio Code Ver. 1.57.0 (Microsoft Corp., Redmond, WA, USA) using Python Ver. 3.8.10 (Python Software Foundation, Wilmington, DE, USA) and the Keras API of Tensorflow 2.0 (Google Brain, Mountain View, CA, USA). The networks were trained, validated, and tested in a server computer with an 8-core CPU (Intel® CoreTM i9-9900K, Santa Clara, CA, USA) and a 24-GB GPU (Titan RTX, Nvidia Corp., Santa Clara, CA, US) using CUDA technology.

2.4. Network Training and Testing

In order to generate the data for network training and testing, foot pressures and foot inertial data with a duration of 4.5 s that covered a fall or a PA were selected as a basis for data augmentation.
(1)
Several 3-s fall segments were obtained by choosing various onset times at 0, 0.5, 1, and 1.5 s and four additional multiples by randomly shifting the onset time within the three intervals (0–0.5 s, 0.5–1 s, 1–1.5 s). Therefore, a total of 3 × 10 × 16 segments (10 falls, each fall had 3 trials, 16-time data augmentation) were obtained from each participant and labeled as falls (class 0).
(2)
Several 3-s PA segments were obtained by choosing various onset times at 0, 0.5, 1, and 1.5 s and four additional multiples by randomly shifting the onset time within three intervals (0–0.5 s, 0.5–1 s, 1–1.5 s). The number of the augmented segments depended on the type of PA such that a total of 3 × 160 segments (each movement had 3 trials) were obtained from each PA for each participant and labeled as a specific PA (class 1 to class 7). Therefore, an evenly balanced dataset between falls and various types of PA was generated.
The dataset, therefore, contained 32 × 3840 instances (32 participants, 480 instances per class in 8 classes related to fall and seven types of PAs). Each instance was constituted by 34 channels × 300 samples (3-s segment with 3-axial accelerations, 3-axial angular velocities, and 11 foot pressures on each shoe).
The foot pressures were normalized by the sum of the eleven standing foot pressures collected from each participant. The inertial data, the normalized foot pressures, and their parameterized data were separately standardized between 0 and 1 across participants.
Leave-one-out cross-validation (LOOCV) was used to validate the neural network. The data of one participant were chosen to validate the network determined by the data from the remaining participants (training set). The LOOCV was repeated to allow each participant’s combination of falls and PAs to be the validation data once.
A total of 50 iterations were applied to train the networks using an Adam optimization algorithm based on the cross-entropy loss function. A fraction of the hidden units in the terminal fully connected network was randomly dropped at every iteration with a probability of 0.5 to force the network to learn general and robust patterns from the data to prevent overfitting.
All of the validation data were fed forward to the trained model. The output values were then compared with the labeled values. The recall, precision, and F1-score for the fall or each PA were calculated based on all LOOCVs. The recall was defined as the percentage of true positives (fall or PA was correctly classified) among all validation data, while the precision was defined as the number of true positives divided by the number of true positives and false positives (the misclassified fall or the misclassified PA). The F1-score was computed by 2 × recall × precision/(recall + precision).

3. Results

Table 4 lists the results of the fall detection and various types of PA based on foot pressures, CoP, and their combination obtained from both feet. We used F1-score as an index to evaluate the detection performance. Falls were well detected, particularly when using the combination of foot pressures and CoP. Static PAs (sitting, standing) were also best detected using this combination. However, the incorporation of foot pressures to detect walking upstairs and downstairs was limited. The poor detection was improved when using CoP alone. Similarly, other dynamic PAs (jumping, jogging, level walking) were best detected using CoP only.
Table 5 lists the results of the detections of the falls and various types of PA based on the inertial data, inertial parameters, and their combination obtained from both feet. The use of inertial data provided better detections (higher F1-scores) for falls and most PAs except jogging. The combination of inertial data and inertial parameters improved the detection of sitting, jumping, jogging, and walking downstairs. Compared to the foot pressure-based measurements (Table 2), the inertial-based measurements detected fall and dynamic PAs better but had worse detection for static PAs (sitting and standing).
As listed in Table 6, falls and various types of PA were well detected when the foot pressure-based measurements and inertial-based measurements were used in combination.
Table 7 summarizes the detection of particular performances based on foot pressures, CoP, and the inertial-based measures obtained from the left foot, right foot, or both feet. The application of right foot data performed best when detecting falls and various types of PA.
Table 8 lists the results of the classifications of falls and various types of PAs using data from both feet. The number of each true class is 480 for each participant. The listed values are the average misclassifications over 32 participants when the inertial data and inertial parameters (1st row), CoP (2nd row) and inertial data, inertial parameters, foot pressures, and CoP (3rd row) are used, respectively. Overall, 14.7% (70.63/480) and 19.8% (94.84/480) of the results were misclassifications between sitting and standing, which were found by using the inertial data and inertial parameters. The misclassifications were improved when the inertial data, inertial parameters, CoP, and foot pressures were used in combination. The use of inertial data and inertial parameters resulted in more than 2% misclassifications from jogging, upstairs walking to jumping. The use of CoP alone led to more than 2% misclassifications from falling to jumping, from jumping to falling or standing, from jogging to downstairs walking, from upstairs walking to falling, and among level walking, upstairs walking and downstairs walking. These misclassifications were improved when the inertial data, inertial parameters, CoP, and foot pressures were used in combination.
Table 9 lists the computational complexity of the proposed deep neural model with various inputs. The number of weights and the number of computations slightly increased with the number of input features. The time for model training in each leave-one-out cross-validation is 30 min. A total of 16 h was needed to complete 32 cross-validations, which allowed every participant’s data to test the model trained by the other 31 participants’ data.

4. Discussion

Falls and PAs are detected or classified based on different rationales. Falls are usually characterized by impact-on-ground and post-impact lying such that several parameter-based methods have been developed to capture these characteristics, whereas PAs have a variety of kinematic patterns depending on the included PA types. Deep-learning neural networks provide a way to extract multiple varieties of PA-related features through network training on labeled data. In this study, we detected falls and various types of PAs based on a deep-learning architecture that is constituted using three networks. First, several multi-layer 1-D convolutional networks extract the low- to high-level features. Second, a skip connection feeds the raw data or features to the output of multiple convolutional layers to reduce feature degradation. Third, a bidirectional LSTM network clusters the output features at various time points to capture the sequential pattern of a fall or PA.
Inertial sensing on the chest, waist, or wrist is the commonly used measurement for fall detection in the literature. The measurement based on foot pressures and foot inertial data provides an alternative approach to distinguish falls from PA. One study built empirical rules based on the foot’s acceleration magnitude (AM) and inactivity duration on the ground [64], but the rules were hand-crafted and only validated on six participants. Another study employed a decision tree to determine the threshold of the waist’s AM or the CoP of both feet for fall detection, which led to a better performance (area under receiver operating characteristic curve) using the waist’s AM and an improved performance when both features were used [65]. However, the performance using foot inertial data was not reported. In this study, we investigated the availability of deep neural networks based on foot-pressure-based or/and foot-inertial-based measures on fall detection. The F1-score was used because it was an overall index between the sensitivity of fall detection and the ratio of correctness in the detected falls. Our result showed a recall of 0.971, a precision of 0.943, and an F1-score of 0.980 in fall detection based on foot pressures and CoP, which supported the existence of fall-related information from foot pressures in accordance with different temporal foot pressure distributions during fall relative to PAs [65]. In addition, using foot inertial data provided a recall of 0.999, a precision of 0.998, and an F1-score of 0.998 for fall detection. The excellent performance of foot inertial data could be attributed to the significant changes in both acceleration and angular velocities during falls, particularly during ground impact.
In this study, we extended fall detection to various types of PAs and applied a deep-learning model to handle these multiple classifications. Our results showed that the inertial-based measures outperformed the foot-pressure-based measures on the detection of dynamic PAs, i.e., the inertial-based measures presented more distinct patterns among jumping, jogging, level walking, upstairs walking, and downstairs walking than the foot-pressure-based measures. On the contrary, using inertial-based measures to detect static PAs was limited, whereas using the foot-pressure-based measures worked well. Therefore, the combination of foot-pressure-based and inertial-based measures was suggested because it achieved a good performance in detecting both static PAs and dynamic PAs.
It is not easy to compare the efficiency of the detection model among various studies because there are different PA types in different datasets. Nevertheless, our model also produced a similar trend in the F1-score between PA classification and fall detection, as in the previous works listed in Table 1 and Table 2; that is, the accuracy of fall detection was higher than those of PA detection. We speculate that falls create quite different temporal patterns in contrast to PAs, whereas the movements of some PAs are somewhat similar, e.g., among level, upstairs and downstairs walks. Therefore, misclassifications among these walks were higher than their false negatives as other PAs or false positives from other PAs, as shown in the confusion matrix (Table 8).
A limitation of this study is that all of the data were collected from young, healthy participants. However, it is not rational or ethical to invite elderly subjects to perform simulated falls for safety reasons. In practical applications, the detection model can be pre-trained based on a large healthy dataset; once a smaller elderly dataset is obtained, transfer learning can be applied to fine-tune the model.
In this study, the performance of fall and PA detections based on a single foot was similar or even higher on the right foot than that based on both feet. In fact, a single-foot measurement is more readily applicable in daily life since a stand-alone detection device can be mounted on one shoe without the need for data transmission between the left and right sides. In cases such as post-stroke hemiplegia or Parkinson’s disease, the application of the device on the unaffected side would be especially easier. We speculate that the unaffected side would produce a pattern closer to younger subjects if compared to the affected side. Similarly, transfer learning can be employed to fine-tune the trained model.
CoP is commonly given by the sum of the products of individual foot pressures and their x and y positions divided by the sum of the foot pressures. This calculation yields a high-precision estimation while using insoles with high-density foot-pressure sensing. It is also applied to the CoP estimation using instrumental insoles with eight foot-pressure sensors on each side [68], 10 sensors on both feet [65], the optimal placement of eight sensors on the left foot [69] or the optimal 13 sensors on both feet [70]. In order to improve the precision of the CoP estimation by the reduced number of foot-pressure sensing, a linear regression calibration [70] or a feed-forward neural network [68] was proposed. Our work focused on fall detection and PA classification based on the customized insole with 11 FSR sensors and demonstrated that the proposed deep neural network could learn discernible features from the crudely estimated CoP for the detection of falls and various types of PAs. Further study is needed to clarify whether calibrated CoP can improve detection.
Healthcare monitoring based on wearable devices has received considerable interest in the management of physiology and psychology. It is worth noting that the measured physiological and behavioral information can be gathered via Internet of Things (IoT) technology, and the increased amount of the gathered information requires further processing to broaden its application through deep learning techniques [71,72]. Wearable IoT sensors provide a solution for the objective remote monitoring of real-life ADL and real fall events for activity-level evaluation, fall prevention, and risk assessment in the elderly and subjects with dementia, Parkinson’s disease, cardiovascular disease, and frailty [73].
Detection methods that rely on the inertial sensing of the human body and limbs are light, easy to wear, and low-cost and can be practically implemented in the real world, but the applications may be limited when the subjects are not willing to or forget to wear these devices. The proposed footwear approach is beneficial in that the sensors can be embedded in shoes, which are commonly worn during off-bed activities, particularly outdoor activities. The cost and complexity of the proposed footwear device can be further reduced by considering the use of one-foot measurements with a smaller number of foot pressures. Further study is needed to investigate the effect of dimensionality reduction on detection performance in future work. In addition to the proposed fall and PA detection, foot-pressure monitoring also provides plantar-pressure information, which is used as a biomechanical assessment for body balance and ergonomics posture during static or dynamic gait [74].
Two potential problems may affect the detection performance while applying the trained deep-learning model in a real scenario. The first problem is the existence of overfitting at the training stage. The trained model cannot efficiently predict the results on the new data when there exists variance with the trained data. Several works have addressed this problem and utilized some approaches to avoid overfitting. For instance, a penalty term is added to the loss function to optimize the boundary of features [75,76]; data augmentation is used to allow the model to more accurately catch different data structures [77]; dropout is applied to prevent neurons from co-adapting too much [78]; early stopping is adopted to reduce unnecessary computing [79]. The above strategies allow the model reduce the focus on some rare specific features during the training phase, thereby keeping the balance and flexibility of the model. In our work, we expanded the collected data by sixteen multiples to enable the model to gather more detail across various situations. Considering the computational complexity of the proposed CNN + LSTM model (Table 9), we used the 50% dropout of neurons randomly at each training epoch to avoid overfitting as much as possible and reduced the computational complexity of model training as well. Additionally, early stopping was applied when the validation loss increased.
The second problem is the effect of noise on the detection performance. At present, the data collected by the inertial sensors and FSRs are not proven to be perfect. Every type of measured signal (accelerations, angular velocities, and foot pressures) in our device possibly has interferences from movement disturbances. A study applied the Dempster–Shafer theory [80] to conduct decision-level image fusion to improve the crack detection accuracy with high robustness to the noise effects [81]. Therefore, it is possible to incorporate the Dempster–Shafer algorithm to fuse data from different sensors [82] for the detection of falls and PAs in our deep neural model in future work.

5. Conclusions

We focused on the employment of foot-pressure-based and foot inertial-based measures to detect falls and various types of PAs using a deep neural network that used CNN to extract discernible features and RNN to cluster the features at various time points. Foot-pressure-based measures, as well as foot inertial-based measures, performed well in fall detection. Foot inertial-based measures led to better performance in detecting dynamic PAs (jumping, jogging, walking), while the foot-pressure-based measures yielded better performance in detecting static PAs (sitting, standing). The combination of foot-pressure-based and foot-inertial-based measures allowed for the detection of both static PAs and dynamic PAs. Although the capability of a deep neural network based on foot pressures and foot inertial sensing to detect falls, static PAs, and dynamic PAs was demonstrated based on the collected data from young participants, the model can be fine-tuned by transfer learning for practical application to the elderly. Moreover, the investigation of one-foot measurement with a fewer number of foot pressures can be conducted to reduce the complexity of the proposed footwear approach in the future.

Author Contributions

Conceptualization, H.-L.C. and Y.-J.C.; methodology, Y.O. and Y.-H.L.; software, Y.-H.L.; validation, C.-C.K. and G.-S.L.; formal analysis, W.-Y.H.; investigation, R.-S.C.; resources, Y.-J.C.; data curation, Y.O. and R.-S.C.; writing—original draft preparation, H.-L.C.; writing—review and editing, H.-L.C., Y.-J.C., and W.-Y.H.; project administration, Y.-J.C. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to acknowledge the support and grants from the Chang Gung Memorial Hospital, Taoyuan, Taiwan (CMRPD2I0033, CMRPD2M0051) and the National Science and Technology Council, Taiwan (109-2224-E-182-001, 111-2224-E-182-013).

Institutional Review Board Statement

The protocol of this study was approved by the Research Ethics Committee of the Chang Gung Medical Foundation (IRB#201802118B0) in accordance with the Helsinki Declaration.

Informed Consent Statement

All of the participants provided written informed consent.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Conflicts of Interest

The authors declared no potential conflict of interest with respect to the research, authorship, and/or publication of this article.

References

  1. Yang, C.-C.; Hsu, Y.-L. A review of accelerometry-based wearable motion detectors for physical activity monitoring. Sensors 2010, 10, 7772–7788. [Google Scholar] [CrossRef] [PubMed]
  2. Bet, P.; Castro, P.C.; Ponti, M.A. Fall detection and fall risk assessment in older person using wearable sensors: A systematic review. Int. J. Med. Inform. 2019, 130, 103946. [Google Scholar] [CrossRef]
  3. Bauman, A.; Merom, D.; Bull, F.C.; Buchner, D.M.; Fiatarone Singh, M.A. Updating the evidence for physical activity: Summative reviews of the epidemiological evidence, prevalence, and interventions to promote “active aging”. Gerontol. 2016, 56, S268–S280. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Schrack, J.A.; Gresham, G.; Wanigatunga, A.A. Understanding physical activity in cancer patients and survivors: New methodology, new challenges, and new opportunities. Cold Spring Harb. Mol. Case Stud. 2017, 3, a001933. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Šimaitytė, M.; Petrėnas, A.; Kravčenko, J.; Kaldoudi, E.; Marozas, V. Objective evaluation of physical activity pattern using smart devices. Sci. Rep. 2019, 9, 2006. [Google Scholar] [CrossRef] [Green Version]
  6. Kosk, K.; Luukinen, H.; Laippala, P.; KivelÄ, S.-L. Physiological factors and medications as predictors of injurious falls by elderly people: A prospective population-based study. Age Ageing 1996, 25, 29–38. [Google Scholar] [CrossRef] [Green Version]
  7. Allen, N.E.; Schwarzel, A.K.; Canning, C.G. Recurrent falls in Parkinson’s disease: A systematic review. Park. Dis. 2013, 2013, 906274. [Google Scholar] [CrossRef] [Green Version]
  8. Giannì, C.; Prosperini, L.; Jonsdottir, J.; Cattaneo, D. A systematic review of factors associated with accidental falls in people with multiple sclerosis: A meta-analytic approach. Clin. Rehabil. 2014, 28, 704–716. [Google Scholar] [CrossRef]
  9. Batchelor, F.A.; Mackintosh, S.F.; Said, C.M.; Hill, K.D. Falls after stroke. Int. J. Stroke 2012, 7, 482–490. [Google Scholar] [CrossRef]
  10. Vellas, B.J.; Wayne, S.J.; Romero, L.J.; Baumgartner, R.N.; Garry, P.J. Fear of falling and restriction of mobility in elderly fallers. Age Ageing 1997, 26, 189–193. [Google Scholar] [CrossRef]
  11. Auais, M.; French, S.; Alvarado, B.; Pirkle, C.; Belanger, E.; Guralnik, J. Fear of falling predicts incidence of functional disability 2 years later: A perspective from an international cohort study. J. Gerontol. Ser. A 2018, 73, 1212–1215. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Jonasson, S.B.; Nilsson, M.H.; Lexell, J.; Carlsson, G. Experiences of fear of falling in persons with Parkinson’s disease—A qualitative study. BMC Geriatr. 2018, 18, 44. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Blaschke, C.M.; Freddolino, P.P.; Mullen, E.E. Ageing and technology: A review of the research literature. Br. J. Soc. Work 2009, 39, 641–656. [Google Scholar] [CrossRef]
  14. Williams, V.; Victor, C.R.; McCrindle, R. It is always on your mind: Experiences and perceptions of falling of older people and their carers and the potential of a mobile falls detection device. Curr. Gerontol. Geriatr. Res. 2013, 2013, 295073. [Google Scholar] [CrossRef] [Green Version]
  15. Gutiérrez, J.; Rodríguez, V.; Martin, S. Comprehensive review of vision-based fall detection systems. Sensors 2021, 21, 947. [Google Scholar] [CrossRef]
  16. Clark, R.A.; Mentiplay, B.F.; Hough, E.; Pua, Y.H. Three-dimensional cameras and skeleton pose tracking for physical function assessment: A review of uses, validity, current developments and Kinect alternatives. Gait Posture 2019, 68, 193–200. [Google Scholar] [CrossRef]
  17. Kwolek, B.; Kepski, M. Human fall detection on embedded platform using depth maps and wireless accelerometer. Comput. Methods Programs Biomed. 2014, 117, 489–501. [Google Scholar] [CrossRef]
  18. Doukas, C.N.; Maglogiannis, I. Emergency fall incidents detection in assisted living environments utilizing motion, sound, and visual perceptual components. IEEE Trans. Inf. Technol. Biomed. 2011, 15, 277–289. [Google Scholar] [CrossRef]
  19. Shieh, W.-Y.; Huang, J.-C. Falling-incident detection and throughput enhancement in a multi-camera video-surveillance system. Med. Eng. Phys. 2012, 34, 954–963. [Google Scholar] [CrossRef]
  20. De Miguel, K.; Brunete, A.; Hernando, M.; Gambao, E. Home camera-based fall detection system for the elderly. Sensors 2017, 17, 2864. [Google Scholar] [CrossRef]
  21. Stone, E.E.; Skubic, M. Fall detection in homes of older adults using the Microsoft Kinect. IEEE J. Biomed. Health Inform. 2015, 19, 290–301. [Google Scholar] [CrossRef] [PubMed]
  22. Lu, N.; Wu, Y.; Feng, L.; Song, J. Deep Learning for fall detection: Three-dimensional CNN combined with LSTM on video kinematic data. IEEE J. Biomed. Health Inform. 2019, 23, 314–323. [Google Scholar] [CrossRef] [PubMed]
  23. Sultana, A.; Deb, K.; Dhar, P.K.; Koshiba, T. Classification of indoor human fall events using deep learning. Entropy 2021, 23, 328. [Google Scholar] [CrossRef]
  24. Lee, P.; Chen, T.-B.; Liu, C.-H.; Wang, C.-Y.; Huang, G.-H.; Lu, N.-H. Identifying the posture of young adults in walking videos by using a fusion artificial intelligent method. Biosensors 2022, 12, 295. [Google Scholar] [CrossRef]
  25. Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Activity recognition using cell phone accelerometers. ACM SigKDD Explor. Newsl. 2011, 12, 74–82. [Google Scholar] [CrossRef]
  26. Micucci, D.; Mobilio, M.; Napoletano, P. UniMiB SHAR: A dataset for human activity recognition using acceleration data from smartphones. Appl. Sci. 2017, 7, 1101. [Google Scholar] [CrossRef] [Green Version]
  27. Sztyler, T.; Stuckenschmidt, H. On-body localization of wearable devices: An investigation of position-aware activity recognition. In Proceedings of the 2016 IEEE International Conference on Pervasive Computing and Communications (PerCom), Sydney, Australia, 14–19 March 2016; pp. 1–9. [Google Scholar]
  28. Zhang, M.; Sawchuk, A.A. USC-HAD: A daily activity dataset for ubiquitous activity recognition using wearable sensors. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing, Pittsburgh, PA, USA, 5–8 September 2012; pp. 1036–1043. [Google Scholar]
  29. Chavarriaga, R.; Sagha, H.; Calatroni, A.; Digumarti, S.T.; Tröster, G.; Millán, J.d.R.; Roggen, D. The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition. Pattern Recognit. Lett. 2013, 34, 2033–2042. [Google Scholar] [CrossRef] [Green Version]
  30. Garcia-Gonzalez, D.; Rivero, D.; Fernandez-Blanco, E.; Luaces, M.R. A public domain dataset for real-life human activity recognition using smartphone sensors. Sensors 2020, 20, 2200. [Google Scholar] [CrossRef] [Green Version]
  31. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A public domain dataset for human activity recognition using smartphones. In Proceedings of the European Symposium on Artificial Neural Networks, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
  32. Karantonis, D.M.; Narayanan, M.R.; Mathie, M.; Lovell, N.H.; Celler, B.G. Implementation of a real-time human movement classifier using a triaxial accelerometer for ambulatory monitoring. IEEE Trans. Inf. Technol. Biomed. 2006, 10, 156–167. [Google Scholar] [CrossRef]
  33. Bourke, A.K.; O’Brien, J.V.; Lyons, G.M. Evaluation of a threshold-based tri-axial accelerometer fall detection algorithm. Gait Posture 2007, 26, 194–199. [Google Scholar] [CrossRef]
  34. Ojetola, O.; Gaura, E.I.; Brusey, J. Data set for fall events and daily activities from inertial sensors. In Proceedings of the 6th ACM Multimedia Systems Conference, MMSys 2015, Portland, OR, USA, 18–20 March 2015; pp. 243–248. [Google Scholar]
  35. Medrano, C.; Igual, R.; Plaza, I.; Castro, M. Detecting falls as novelties in acceleration patterns acquired with smartphones. PLoS ONE 2014, 9, e94811. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Sucerquia, A.; López, J.D.; Vargas-Bonilla, J.F. SisFall: A fall and movement dataset. Sensors 2017, 17, 198. [Google Scholar] [CrossRef] [Green Version]
  37. Vavoulas, G.; Pediaditis, M.; Spanakis, E.G.; Tsiknakis, M. The MobiFall dataset: An initial evaluation of fall detection algorithms using smartphones. In Proceedings of the 13th IEEE International Conference on BioInformatics and BioEngineering, Chania, Greece, 10–13 November 2013; pp. 1–4. [Google Scholar]
  38. Kangas, M.; Konttila, A.; Lindgren, P.; Winblad, I.; Jämsä, T. Comparison of low-complexity fall detection algorithms for body attached accelerometers. Gait Posture 2008, 28, 285–291. [Google Scholar] [CrossRef]
  39. Chao, P.-K.; Chan, H.-L.; Tang, F.-T.; Chen, Y.-C.; Wong, M.-K. A comparison of automatic fall detection by the cross-product and magnitude of tri-axial acceleration. Physiol. Meas. 2009, 30, 1027–1037. [Google Scholar] [CrossRef]
  40. Bourke, A.K.; van de Ven, P.; Gamble, M.; O’Connor, R.; Murphy, K.; Bogan, E.; McQuade, E.; Finucane, P.; ÓLaighin, G.; Nelson, J. Evaluation of waist-mounted tri-axial accelerometer based fall-detection algorithms during scripted and continuous unscripted activities. J. Biomech. 2010, 43, 3051–3057. [Google Scholar] [CrossRef] [PubMed]
  41. Casilari, E.; Santoyo-Ramón, J.A.; Cano-García, J.M. Analysis of a smartphone-based architecture with multiple mobility sensors for fall detection. PLoS ONE 2016, 11, e0168069. [Google Scholar] [CrossRef] [Green Version]
  42. Nyan, M.N.; Tay, F.E.H.; Tan, A.W.Y.; Seah, K.H.W. Distinguishing fall activities from normal activities by angular rate characteristics and high-speed camera characterization. Med. Eng. Phys. 2006, 28, 842–849. [Google Scholar] [CrossRef]
  43. Bourke, A.K.; Lyons, G.M. A threshold-based fall-detection algorithm using a bi-axial gyroscope sensor. Med. Eng. Phys. 2008, 30, 84–90. [Google Scholar] [CrossRef] [PubMed]
  44. Wang, F.-T.; Chan, H.-L.; Hsu, M.-H.; Lin, C.-K.; Chao, P.-K.; Chang, Y.-J. Threshold-based fall detection using a hybrid of tri-axial accelerometer and gyroscope. Physiol. Meas. 2018, 39, 105002. [Google Scholar] [CrossRef] [PubMed]
  45. Santos, G.L.; Endo, P.T.; Monteiro, K.H.; Rocha, E.D.; Silva, I.; Lynn, T. Accelerometer-based human fall detection using convolutional neural networks. Sensors 2019, 19, 1644. [Google Scholar] [CrossRef]
  46. Ribeiro, N.F.; André, J.; Costa, L.; Santos, C.P. Development of a strategy to predict and detect falls using wearable sensors. J. Med. Syst. 2019, 43, 134. [Google Scholar] [CrossRef] [PubMed]
  47. Casilari, E.; Lora-Rivera, R.; García-Lagos, F. A Study on the application of convolutional neural networks to fall detection evaluated with multiple public datasets. Sensors 2020, 20, 1466. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Qi, W.; Su, H.; Yang, C.; Ferrigno, G.; De Momi, E.; Aliverti, A. A fast and robust deep convolutional neural networks for complex human activity recognition using smartphone. Sensors 2019, 19, 3731. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Almaslukh, B.; Artoli, A.M.; Al-Muhtadi, J. A robust deep learning approach for position-independent smartphone-based human activity recognition. Sensors 2018, 18, 3726. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Avilés-Cruz, C.; Ferreyra-Ramírez, A.; Zúñiga-López, A.; Villegas-Cortéz, J. Coarse-fine convolutional deep-learning strategy for human activity recognition. Sensors 2019, 19, 1556. [Google Scholar] [CrossRef] [Green Version]
  51. Russell, B.; McDaid, A.; Toscano, W.; Hume, P. Moving the Lab into the mountains: A pilot study of human activity recognition in unstructured environments. Sensors 2021, 21, 654. [Google Scholar] [CrossRef]
  52. Huang, J.; Lin, S.; Wang, N.; Dai, G.; Xie, Y.; Zhou, J. TSE-CNN: A two-stage end-to-end CNN for human activity recognition. IEEE J. Biomed. Health Inform. 2020, 24, 292–299. [Google Scholar] [CrossRef]
  53. Mauldin, T.R.; Canby, M.E.; Metsis, V.; Ngu, A.H.H.; Rivera, C.C. SmartFall: A smartwatch-based fall detection system using deep learning. Sensors 2018, 18, 3363. [Google Scholar] [CrossRef] [Green Version]
  54. Luna-Perejón, F.; Domínguez-Morales, M.J.; Civit-Balcells, A. Wearable fall detector using recurrent neural networks. Sensors 2019, 19, 4885. [Google Scholar] [CrossRef] [Green Version]
  55. Murad, A.; Pyun, J.-Y. Deep recurrent neural networks for human activity recognition. Sensors 2017, 17, 2556. [Google Scholar] [CrossRef]
  56. Delgado-Escaño, R.; Castro, F.M.; Cózar, J.R.; Marín-Jiménez, M.J.; Guil, N.; Casilari, E. A cross-dataset deep learning-based classifier for people fall detection and identification. Comput. Methods Programs Biomed. 2020, 184, 105265. [Google Scholar] [CrossRef] [PubMed]
  57. Liu, L.; Hou, Y.; He, J.; Lungu, J.; Dong, R. An energy-efficient fall detection method based on FD-DNN for elderly people. Sensors 2020, 20, 4192. [Google Scholar] [CrossRef] [PubMed]
  58. Theodoridis, T.; Solachidis, V.; Vretos, N.; Daras, P. Human fall detection from acceleration measurements using a recurrent neural network. In Proceedings of the Precision Medicine Powered by pHealth and Connected Health, Thessaloniki, Greece, 10–13 November 2017; pp. 145–149. [Google Scholar]
  59. Fridriksdottir, E.; Bonomi, A.G. Accelerometer-based human activity recognition for patient monitoring using a deep neural network. Sensors 2020, 20, 6424. [Google Scholar] [CrossRef]
  60. Ankita; Rani, S.; Babbar, H.; Coleman, S.; Singh, A.; Aljahdali, H.M. An efficient and lightweight deep learning model for human activity recognition using smartphones. Sensors 2021, 21, 3845. [Google Scholar] [CrossRef] [PubMed]
  61. Nafea, O.; Abdul, W.; Muhammad, G.; Alsulaiman, M. Sensor-based human activity recognition with spatio-temporal deep learning. Sensors 2021, 21, 2141. [Google Scholar] [CrossRef]
  62. Chen, W.-H.; Lee, Y.-S.; Yang, C.-J.; Chang, S.-Y.; Shih, Y.; Sui, J.-D.; Chang, T.-S.; Shiang, T.-Y. Determining motions with an IMU during level walking and slope and stair walking. J. Sport. Sci. 2020, 38, 62–69. [Google Scholar] [CrossRef]
  63. Moufawad el Achkar, C.; Lenoble-Hoskovec, C.; Paraschiv-Ionescu, A.; Major, K.; Büla, C.; Aminian, K. Instrumented shoes for activity classification in the elderly. Gait Posture 2016, 44, 12–17. [Google Scholar] [CrossRef]
  64. Zitouni, M.; Pan, Q.; Brulin, D.; Campo, E. Design of a smart sole with advanced fall detection algorithm. J. Sens. Technol. 2019, 9, 71–90. [Google Scholar] [CrossRef] [Green Version]
  65. Lee, C.M.; Park, J.; Park, S.; Kim, C.H. Fall-detection algorithm using plantar pressure and acceleration data. Int. J. Precis. Eng. Manuf. 2020, 21, 725–737. [Google Scholar] [CrossRef]
  66. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  67. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference for Learning Representations, San Diego, CA, USA, 7–9 May 2015. arXiv:1412.6980. [Google Scholar]
  68. Zhang, H.; Zanotto, D.; Agrawal, S.K. Estimating CoP trajectories and kinematic gait parameters in walking and running using instrumented insoles. IEEE Robot. Autom. Lett. 2017, 2, 2159–2165. [Google Scholar] [CrossRef]
  69. Lin, C.-W.; Ruan, S.-J.; Hsu, W.-C.; Tu, Y.-W.; Han, S.-L. Optimizing the sensor placement for foot plantar center of pressure without prior knowledge using deep reinforcement learning. Sensors 2020, 20, 5588. [Google Scholar] [CrossRef] [PubMed]
  70. Choi, A.; Jung, H.; Lee, K.Y.; Lee, S.; Mun, J.H. Machine learning approach to predict center of pressure trajectories in a complete gait cycle: A feedforward neural network vs. LSTM network. Med. Biol. Eng. Comput. 2019, 57, 2693–2703. [Google Scholar] [CrossRef] [PubMed]
  71. Lakshmanna, K.; Kaluri, R.; Gundluru, N.; Alzamil, Z.S.; Rajput, D.S.; Khan, A.A.; Haq, M.A.; Alhussen, A. A review on deep learning techniques for IoT data. Electronics 2022, 11, 1604. [Google Scholar] [CrossRef]
  72. Manogaran, G.; Shakeel, P.M.; Fouad, H.; Nam, Y.; Baskar, S.; Chilamkurti, N.; Sundarasekar, R. Wearable IoT smart-log patch: An edge computing-based Bayesian deep learning network system for multi access physical monitoring system. Sensors 2019, 19, 3030. [Google Scholar] [CrossRef] [Green Version]
  73. Stavropoulos, T.G.; Papastergiou, A.; Mpaltadoros, L.; Nikolopoulos, S.; Kompatsiaris, I. IoT wearable sensors and devices in elderly care: A literature review. Sensors 2020, 20, 2826. [Google Scholar] [CrossRef]
  74. Zulkifli, S.S.; Loh, W.P. A state-of-the-art review of foot pressure. Foot Ankle Surg. 2020, 26, 25–32. [Google Scholar] [CrossRef]
  75. Zheng, Q.; Yang, M.; Yang, J.; Zhang, Q.; Zhang, X. Improvement of generalization ability of deep CNN via implicit regularization in two-stage training process. IEEE Access 2018, 6, 15844–15869. [Google Scholar] [CrossRef]
  76. Jin, B.; Cruz, L.; Gonçalves, N. Pseudo RGB-D face recognition. IEEE Sens. J. 2022, 22, 21780–21794. [Google Scholar] [CrossRef]
  77. Yao, T.; Qu, C.; Liu, Q.; Deng, R.; Tian, Y.; Xu, J.; Jha, A.; Bao, S.; Zhao, M.; Fogo, A.B.; et al. Compound figure separation of biomedical images with side loss. In Proceedings of the Deep Generative Models, and Data Augmentation, Labelling, and Imperfections, Strasbourg, France, 1 October 2021; pp. 173–183. [Google Scholar]
  78. Srivastava, N.; Hinton, G.E.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R.J.J.M.L.R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  79. Zhao, M.; Jha, A.; Liu, Q.; Millis, B.A.; Mahadevan-Jansen, A.; Lu, L.; Landman, B.A.; Tyska, M.J.; Huo, Y. Faster Mean-shift: GPU-accelerated clustering for cosine embedding-based cell segmentation and tracking. Med. Image Anal. 2021, 71, 102048. [Google Scholar] [CrossRef]
  80. Dempster, A.P. Upper and lower probabilities induced by a multivalued mapping. In Classic Works of the Dempster-Shafer Theory of Belief Functions; Yager, R.R., Liu, L., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 57–72. [Google Scholar]
  81. Yu, Y.; Samali, B.; Rashidi, M.; Mohammadi, M.; Nguyen, T.N.; Zhang, G. Vision-based concrete crack detection using a hybrid framework considering noise effect. J. Build. Eng. 2022, 61, 105246. [Google Scholar] [CrossRef]
  82. Högger, A. Dempster Shafer Sensor Fusion for Autonomously Driving Vehicles: Association Free Tracking of Dynamic Objects. Master’s Thesis, KTH Royal Institute of Technology, Stockholm, Sweden, 2016. [Google Scholar]
Figure 1. Instrumental shoes with inertial sensing and foot pressure measurements at the big toe, little toe, metatarsus (medial, middle, lateral), arches (medial, lateral), fore heels (medial, lateral), and heels (medial, lateral).
Figure 1. Instrumental shoes with inertial sensing and foot pressure measurements at the big toe, little toe, metatarsus (medial, middle, lateral), arches (medial, lateral), fore heels (medial, lateral), and heels (medial, lateral).
Sensors 23 00495 g001
Figure 2. Tri-axial accelerations, tri-axial angular velocities, and the corresponding inertial parameters including acceleration amplitude (AM), acceleration cubic-product-root magnitude (ACM), and angular velocity cubic-product-root magnitude (AVCM) while a subject was undertaking a forward fall, a jump, a forward walk, and a forward jog.
Figure 2. Tri-axial accelerations, tri-axial angular velocities, and the corresponding inertial parameters including acceleration amplitude (AM), acceleration cubic-product-root magnitude (ACM), and angular velocity cubic-product-root magnitude (AVCM) while a subject was undertaking a forward fall, a jump, a forward walk, and a forward jog.
Sensors 23 00495 g002
Figure 3. Relative foot pressures at the metatarsus, heel areas, and center of plantar pressure while a subject was undertaking a forward fall, a jump, a forward walk, and a forward jog.
Figure 3. Relative foot pressures at the metatarsus, heel areas, and center of plantar pressure while a subject was undertaking a forward fall, a jump, a forward walk, and a forward jog.
Sensors 23 00495 g003
Figure 4. Deep neural network for detecting fall and 7 types of physical activities based on foot pressures and foot inertial data (M channels).
Figure 4. Deep neural network for detecting fall and 7 types of physical activities based on foot pressures and foot inertial data (M channels).
Sensors 23 00495 g004
Table 1. Methods for physical activity (PA) classifications.
Table 1. Methods for physical activity (PA) classifications.
AuthorsDataset/
Subject No.
SensorPlacementFeatureModelPA TypeAccuracy
Kwapisz et al. [25]WISDM/29Alegstatistical featuresMLPjog, stand, sit, walk, stairs0.917
Micucci et al. [26]UniMiB-SHAR/30AthighsignalskNNjump, jog, lay down, stand up, sit down, walk, stairs0.830
Anguita et al. [31]UCI HAR/30A, Gwaisttime and freq featuresSVMlay, stand, sit, walk, stairsF1 = 0.960
Sztyler & Stuckenschmidt [27]Realworld HAR/15Achest, head, waist, arm, thigh, shintime and freq featuresRFjump, jog, lay, stand, sit, walk, stairsF1 = 0.890
Chavarriaga et al. [29]12A, Gtrunk, arm, knee, footprincipal componentskNNlay, stand, sit, walkF1 = 0.850
Garcia-Gonzalez et al. [30]15A, Glong-term smartphonestatistical featuresSVMinactive, active, walk, drive0.744
Moufawad el Achkar et al. [63]10A, G, B, FPtrunk, thigh, foot DTstand, sit, walk, stairs, uphill, downhill 0.970
Qi et al. [48]20A, G, Mwaistprincipal componentsCNN;
LSTM
jump, jog, lay, stand, walk, stairs0.953;
0.964
Almaslukh et al. [49]Realworld HAR/15 [27]Achest, head, waist, arm, thigh, shinsignalsCNNjump, jog, lay, stand, sit, walk, stairsF1 = 0.860
Avilés-Cruz et al. [50]UCI HAR/30 [31]A, GwaistsignalsCNNlay, stand, sit, walk, stairs1
Russell et al. [51]-AchestsignalsCNNlay, sit, walk, climb0.982
Huang et al. [52]WISDM/36 [25]AlegsignalsCNNjog, stand, sit, walk, stairsF1 = 0.940
Chen et al. [62]30A, GfootsignalsCNNUphill, downhill, slope, walk, stairs0.877
Murad & Pyun [55]USC HAD/14 [28]A, GhipsignalsRNNjump, jog, lay, stand, sit, walk, stairsF1 = 0.970
Fridriksdottir & Bonomi [59]20AtrunksignalsCNN + LSTMstand, sit, walk, stairsF1 = 0.946
Ankita [60]UCI HAR/30 [31]A, Gwaisttime and freq featuresCNN + LSTMlay, stand, sit, walk, stairs0.979
Methods in the first block use the hand-crafted features that are obtained from foot pressures (FP), barometer (B), inertial signals by accelerometer (A), gyroscope (G), and magnetometer (M). Multilayer perceptron (MLP), support vector machine (SVM), k-nearest neighbors (kNN), decision tree (DT), and random forest (RF) are therefore used to classify PA based on the hand-crafted features or signals. Methods in the second block use convolutional neural networks (CNNs), recurrent neural networks (RNNs), long short-term memory (LSTM), and their combination (CNN + LSTM) to learn discriminant features form inertial signals and classify PA.
Table 2. Methods for fall detections.
Table 2. Methods for fall detections.
AuthorsDataset/
Subject No.
SensorPlacementFeatureModelActivity of Daily LifeAccuracy
Karantonis et al. [32]6AwaistAM, postureempiricalwalk, PT0.908
Bourke et al. [33]20Achest; thighAMEmpiricalwalk, PT1 ; 0.913
Kangas et al. [38]3Awaist, wrist, headAM, postureempiricalwalk, stairs, PT0.990
Chao et al. [39]7Achest;
thigh
AC, postureempiricalwalk, squat, PT0.991 ;
0.996
Bourke et al. [40]20AwaistAC, VV, postureempiricalwalk, PT1
Sucerquia et al. [36]SisFall/38AwaistAMempiricaljump, jog, walk, PT0.978
Medrano et al. [35]DFNAPAS/10ApocketAMSVMreal-life ADL0.939
Ojetola et al. [34]Cognent/32A, GchestAMempiricalstand, lay, sit, walk, crouch, PTF1 = 0.940
Nyan et al. [42]10Gchest, waistangular velocityempiricalwalk, PT0.987
Bourke et al. [43]20Gchestangular velocity and accelerationempiricalwalk, PT1
Casilari et al. [41]UMAFall/17A, Gchest, waist, thigh, wristAM, DA, postureempiricaljump, walk, stairs, PT0.886
Wang et al. [44]15;
Cognent/32 [34];
UMAFall/17 [41]
A, GchestAM, ACM, AVCMempiricaljump, walk, PT;
crouch, walk, PT;
jump, walk, stairs, PT
0.991 ; 0.973 ;
0.896
Zitouni et al. [64]6AfootAM, DA, postureempiricalstand, lay, sit, walk, sport, PT0.966
Lee et al. [65]9A, FPfootAM, DA, CoPDTstand, lay, sit, walk, PT0.951
Santos et al. [45]URFD/5 [17];
SmartFall/7 [53]
A, Ghip;
wrist
signalsCNNlay, sit, walk, crouch;
jog, sit, hand move
0.999;
0.999
Ribeiro et al. [46]12A, G, Mback, thigh, footprincipal componentsCNNwalk0.927
Casilari et al. [47]SisFall/38 [36];
UMAFall/17 [41]
AwaistAM, accelerationsCNNjump, jog, walk, PT;jump, walk, stairs, PT0.988;
0.821
Mauldin et al. [53]SmartFall/7AwristsignalsRNNjog, sit, hand move0.850
Luna-Perejón et al. [54]SisFall/38 [36]AwaistsignalsRNNjump, jog, walk, PT0.967
Theodoridis et al. [58]URFD/5 [17]AhipsignalsCNN + LSTMlay, sit, walk, crouch0.986
Delgado-Escaño et al. [56]DFNAPAS/10 [35];
SisFall/38 [36];
UniMiB-SHAR/30 [26]
Apocket;
waist;
thigh
signalsCNN + LSTMreal-life ADL;
jump, jog, walk, PT;
jump, jog, walk, stairs, PT
0.997;
0.972;
0.873
Liu [57]SisFall/38 [36]
MobiFall/24 [37]
A, GwaistsignalsCNN + LSTMjump, jog, walk, stairs0.992
Methods in the first block use hand-crafted features that are obtained from foot pressures (FP) and inertial signals by accelerometer (A), gyroscope (G), and magnetometer (M). The features including acceleration amplitude (AM), acceleration cross-product (AC), differential acceleration (DA), vertical velocity (VV), acceleration cubic-product-root magnitude (ACM), angular velocity cubic-product-root magnitude (AVCM), angular velocity, angular acceleration, and posture are therefore used to distinguish falls from activities of daily life (ADL) based on empirical rules or a support vector machine (SVM) or decision tree (DT). Methods in the second block use convolutional neural networks (CNNs), recurrent neural networks (RNNs), long short-term memory (LSTM), and their combination (CNN + LSTM) to learn discriminant features form inertial signals and distinguish falls from ADL. indicates the accuracy (acc) is calculated by sensitivity × specificity .
Table 3. Deep residual network architecture.
Table 3. Deep residual network architecture.
Residual Unit (RU)/LayerThe 1st RU
Kernel Size 11, Stride 1
The 2nd RU
Kernel Size 9, Stride 1
The 3rd RU
Kernel Size 5, Stride 1
Input ChannelOutput ChannelSizeInput ChannelOutput ChannelSizeInput ChannelOutput ChannelSize
Conv1dM63003642150727875
Conv1d6123004248150788475
Conv1d12183004854150849075
Conv1d18243005460150909675
Conv1d243030060661509610275
Conv1d3036300667215010210875
Max pooling
kernel size 2
stride 2
363615072727510810838
Skip layer
kernel size 1
stride 2
M361503672757210838
Table 4. Results on the detections of fall and physical activities based on foot pressures and center of pressure obtained from both feet.
Table 4. Results on the detections of fall and physical activities based on foot pressures and center of pressure obtained from both feet.
FallSitStandJumpJogLWUSDSMacro
Foot pressures
Recall
Precision
F1-score
0.935
0.973
0.950
0.992
0.963
0.973
0.850
0.922
0.875
0.837
0.825
0.812
0.898
0.846
0.853
0.755
0.819
0.764
0.644
0.569
0.555
0.622
0.465
0.487
0.817
0.798
0.784
Center of pressure
Recall
Precision
F1-score
0.953
0.888
0.913
0.856
0.937
0.877
0.923
0.869
0.885
0.876
0.951
0.902
0.945
0.977
0.958
0.922
0.882
0.887
0.762
0.877
0.783
0.848
0.893
0.844
0.886
0.909
0.881
Foot pressures + center of pressure
Recall
Precision
F1-score
0.971
0.943
0.980
0.976
0.989
0.980
0.946
0.906
0.921
0.891
0.924
0.897
0.900
0.901
0.881
0.852
0.869
0.841
0.602
0.638
0.580
0.574
0.671
0.592
0.839
0.855
0.831
Macro average used to calculate recall, precision, and F1-score of multiple classes; LW, level walking; US, upstairs walking; DS, downstairs walking.
Table 5. Results on the detections of fall and physical activities based on inertial data and parameters obtained from both feet.
Table 5. Results on the detections of fall and physical activities based on inertial data and parameters obtained from both feet.
FallSitStandJumpJogLWUSDSMacro
Inertial data
Recall
Precision
F1-score
0.999
0.998
0.998
0.724
0.887
0.770
0.862
0.781
0.800
0.976
0.957
0.963
0.915
0.962
0.930
0.974
0.965
0.968
0.972
0.974
0.968
0.987
0.994
0.990
0.926
0.940
0.923
Inertial parameters
Recall
Precision
F1-score
0.995
0.969
0.980
0.769
0.790
0.759
0.774
0.788
0.767
0.941
0.956
0.940
0.932
0.958
0.938
0.963
0.940
0.946
0.937
0.990
0.954
0.961
0.973
0.958
0.909
0.921
0.905
Inertial data + inertial parameters
Recall
Precision
F1-score
0.994
0.988
0.990
0.854
0.819
0.804
0.742
0.887
0.762
0.968
0.968
0.966
0.943
0.958
0.946
0.962
0.964
0.961
0.964
0.973
0.964
0.990
0.995
0.993
0.927
0.964
0.923
Macro average used to calculate recall, precision, and F1-score of multiple classes; LW, level walking; US, upstairs walking; DS, downstairs walking.
Table 6. Results on the detections of fall and physical activities based on Inertial data and parameters obtained from both feet.
Table 6. Results on the detections of fall and physical activities based on Inertial data and parameters obtained from both feet.
FallSitStandJumpJogLWUSDSMacro
Foot pressures + center of pressure + inertial data + inertial parameters
Recall
Precision
F1-score
0.996
0.997
0.996
0.987
0.995
0.990
0.990
0.979
0.983
0.981
0.965
0.971
0.924
0.953
0.931
0.917
0.957
0.960
0.919
0.948
0.924
0.979
0.987
0.983
0.968
0.973
0.967
Center of pressure + inertial data + inertial parameters
Recall
Precision
F1-score
0.998
0.991
0.994
0.904
0.961
0.924
0.958
0.920
0.934
0.991
0.980
0.985
0.956
0.957
0.953
0.966
0.960
0.961
0.926
0.964
0.938
0.981
0.991
0.985
0.960
0.966
0.959
Macro average used to calculate recall, precision, and F1-score of multiple classes; LW, level walking; US, upstairs walking; DS, downstairs walking.
Table 7. Results on the detections of fall and physical activities based on foot pressures, center of pressure, inertial data and inertial parameters obtained from a single foot or both feet.
Table 7. Results on the detections of fall and physical activities based on foot pressures, center of pressure, inertial data and inertial parameters obtained from a single foot or both feet.
FallSitStandJumpJogLWUSDSMacro
Left foot
Recall
Precision
F1-score
0.998
0.995
0.996
0.969
0.988
0.972
0.978
0.970
0.970
0.922
0.929
0.919
0.883
0.926
0.894
0.961
0.935
0.945
0.946
0.963
0.950
0.981
0.979
0.979
0.955
0.961
0.953
Right foot
Recall
Precision
F1-score
0.998
0.996
0.997
0.977
0.998
0.985
0.990
0.971
0.979
0.977
0.951
0.960
0.920
0.967
0.935
0.972
0.962
0.966
0.952
0.963
0.951
0.982
0.994
0.988
0.971
0.975
0.970
Double feet
Recall
Precision
F1-score
0.996
0.997
0.996
0.987
0.995
0.990
0.990
0.979
0.983
0.981
0.965
0.971
0.924
0.953
0.931
0.917
0.957
0.960
0.919
0.948
0.924
0.979
0.987
0.983
0.968
0.973
0.967
Macro average used to calculate recall, precision, and F1-score of multiple classes; LW, level walking; US, upstairs walking; DS, downstairs walking.
Table 8. Confusion matrix of the detections of fall and physical activities based on data from both feet.
Table 8. Confusion matrix of the detections of fall and physical activities based on data from both feet.
True ClassPredicted Class
FallSitStandJumpJogLWUSDS
Fall479.03
457.38
479.19
0.03
2.88
0.28
0.44
4.78
0.09
0.06
13.28
0.13
0
0.06
0
0.31
1.19
0.13
0.09
0.09
0.03
0.03
0.34
0.16
Sit0
5.16
0
385.16
410.84
434.00
94.84
64.00
46.00
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
Stand0.19
6.16
0
70.63
25.75
17.63
404.00
443.03
460.00
0.38
3.09
2.38
0
0
0
4.81
1.97
0
0
0
0
0
0
0
Jump0.50
42.50
0.25
0
0
0
0.03
16.78
0.44
462.50
420.66
475.84
9.59
0.09
3.13
5.72
0.06
0.25
1.06
0.06
0.03
0.59
0.09
0.06
Jog0.41
1.25
0
0
0
0
0.16
0
0
17.84
3.78
6.09
446.47
453.78
458.66
2.69
2.69
1.69
7.84
2.38
12.31
4.28
16.13
1.25
Walk0.22
6.09
0.06
0
0
0
5.69
1.28
2.19
3.78
1.19
0.72
4.59
3.75
7.09
461.34
442.75
463.66
3.19
19.97
4.56
0.88
4.97
1.72
US0.50
16.19
0.25
0
0
0
0
3.84
1.56
10.50
0
2.34
3.78
0.13
10.59
8.47
67.00
19.78
455.06
350.88
444.25
1.69
41.97
1.22
DS0
8.59
4.81
0
0
0
0
0
0
0.03
1.44
0.03
7.66
9.22
2.34
4.22
17.81
0.47
2.84
35.72
1.38
465.25
407.22
470.97
The number of each true class is 480 for each participant. The listed values are the average over 32 participants when inertial data and inertial parameters (1st row), center of pressure (2nd row) and foot pressures, center of pressure, inertial data, and parameters (3rd row) were, respectively, used. LW, level walking; US, upstairs walking; DS, downstairs walking.
Table 9. Computational complexity of the deep neural networks.
Table 9. Computational complexity of the deep neural networks.
Input FeaturesInput ShapeNumber of WeightsComputations
(MFLOPs)
Foot pressures22,300462,23882.9
Center of pressure (CoP)4300460,51082.0
Foot pressures + CoP26,300462,62283.1
Inertial data12,300461,27882.4
Inertial parameters6300460,70282.1
Inertial (data + parameters) = 18,300461,85482.7
Foot pressures + CoP +
inertial (data + parameters)
44,300464,35083.9
CoP + inertial (data + parameters)22,300462,23882.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chan, H.-L.; Ouyang, Y.; Chen, R.-S.; Lai, Y.-H.; Kuo, C.-C.; Liao, G.-S.; Hsu, W.-Y.; Chang, Y.-J. Deep Neural Network for the Detections of Fall and Physical Activities Using Foot Pressures and Inertial Sensing. Sensors 2023, 23, 495. https://doi.org/10.3390/s23010495

AMA Style

Chan H-L, Ouyang Y, Chen R-S, Lai Y-H, Kuo C-C, Liao G-S, Hsu W-Y, Chang Y-J. Deep Neural Network for the Detections of Fall and Physical Activities Using Foot Pressures and Inertial Sensing. Sensors. 2023; 23(1):495. https://doi.org/10.3390/s23010495

Chicago/Turabian Style

Chan, Hsiao-Lung, Yuan Ouyang, Rou-Shayn Chen, Yen-Hung Lai, Cheng-Chung Kuo, Guo-Sheng Liao, Wen-Yen Hsu, and Ya-Ju Chang. 2023. "Deep Neural Network for the Detections of Fall and Physical Activities Using Foot Pressures and Inertial Sensing" Sensors 23, no. 1: 495. https://doi.org/10.3390/s23010495

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop