You are currently on the new version of our website. Access the old version .
SensorsSensors
  • Article
  • Open Access

29 March 2017

Posture Detection Based on Smart Cushion for Wheelchair Users

,
,
and
1
School of Logistics Engineering, Wuhan University of Technology, Wuhan, 430070, China
2
Department of Informatics, Modeling, Electronics and Systems, University of Calabria, Rende, 87036, Italy
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Body Sensor Networks: Sensors, Systems, and Applications

Abstract

The postures of wheelchair users can reveal their sitting habit, mood, and even predict health risks such as pressure ulcers or lower back pain. Mining the hidden information of the postures can reveal their wellness and general health conditions. In this paper, a cushion-based posture recognition system is used to process pressure sensor signals for the detection of user’s posture in the wheelchair. The proposed posture detection method is composed of three main steps: data level classification for posture detection, backward selection of sensor configuration, and recognition results compared with previous literature. Five supervised classification techniques—Decision Tree (J48), Support Vector Machines (SVM), Multilayer Perceptron (MLP), Naive Bayes, and k-Nearest Neighbor (k-NN)—are compared in terms of classification accuracy, precision, recall, and F-measure. Results indicate that the J48 classifier provides the highest accuracy compared to other techniques. The backward selection method was used to determine the best sensor deployment configuration of the wheelchair. Several kinds of pressure sensor deployments are compared and our new method of deployment is shown to better detect postures of the wheelchair users. Performance analysis also took into account the Body Mass Index (BMI), useful for evaluating the robustness of the method across individual physical differences. Results show that our proposed sensor deployment is effective, achieving 99.47% posture recognition accuracy. Our proposed method is very competitive for posture recognition and robust in comparison with other former research. Accurate posture detection represents a fundamental basic block to develop several applications, including fatigue estimation and activity level assessment.

1. Introduction

The continuous progress of ICT (Information and Communication Technology) is rapidly changing our daily life. Thanks to the high processing capabilities, reduced dimensions and low cost of modern microelectronics, a plethora of wearable devices are being realized to non-invasively support and monitor several human daily activities. Such devices play a central role in the so-called ubiquitous sensing [1], which aims at gathering data from the sensors and mining knowledge from them. Body Area Networks (BANs) can measure physical activity and biophysical signals [2] and pose the basis for a wide range of new user-centric applications in the domains of health care [3], remote monitoring [4], activity measurement [5], activity level assessment [6], user-centric wheelchairs [7], and health care service security [8].
In order to cope with the problems that elderly and mobility disabled individuals are facing, more effective care services, especially for wheelchair users, need to be provided. Several supportive devices have already been proposed to help such people keep an independent lifestyle. For this purpose, mobility-assistive devices, including smart wheelchairs, have recently begun to emerge. A smart wheelchair augments the functionalities of a traditional wheelchair by means of various technologies such as BANs [9], cloud computing [10,11], and multi sensor fusion [12].
Current research on smart wheelchairs is mainly focused on vital signals and physical activity monitoring [13], human–machine interface to control the wheelchair [14], obstacle detection and navigation [15], and wheelchair movement condition estimation [16].
This work is focused on a wheelchair assist system for mobility-impaired individuals that can recognize postures using a cushion based system. Pressure sensors are used to obtain data of the wheelchair user’s postures, several classification methods have been compared to determine an efficient classifier, and former sensor deployment methods have been compared against our proposed deployment to determine an optimal deployment method. In addition, some practical applications are discussed in the paper. If dangerous situations such as long-term abnormal posture are detected, the system can issue an alarm immediately; in addition, posture detection is a basic block towards activity level assessment.
The main contributions of this work are the following:
  • An optimization method of pressure sensor deployment is proposed to more accurately detect sitting postures.
  • An in-depth comparison of several classification techniques has been carried out to identify the best posture classifier based on pressure sensors’ smart cushion.
  • In contrast with previous studies, the Body Mass Index (BMI) is among the considered parameters to evaluate the generality and robustness of the proposed deployment method across different body shapes.
The reminder of the paper is organized as follows. Section 2 discusses the related work on posture recognition using smart cushion; we describe and analyze the two main production technologies of smart cushions for pressure detection, the first is based on pressure sensor array and the second relies on the use of a few individual sensors deployed on the seat and backrest. Section 3 describes the architecture of the proposed system. Section 4 reports the system evaluation; data collection and experimental settings as well as optimal sensor selection methods are discussed here. Section 5 is about the results and performance analysis; we compared several classifiers and the results of our optimal sensor deployment are analyzed in comparison with the state-of-the-art. Finally, Section 6 concludes the paper and some future directions are anticipated.

3. System Design

In the following, we describe our method for monitoring postures of smart wheelchair users. In particular, we designed a smart cushion to monitor sitting postures. The method serves as a basis for the development of a smart wheelchair system that can e.g., warn the user of long-term wrong/dangerous postures and, in case of emergency situations, alert relatives and caregivers to help the assisted user on time. In the following sections, we will describe the system in detail.

3.1. System Architecture

A layered architecture of the proposed posture detection system is depicted in Figure 1. The posture detection layer represents the core of the proposed system and it has been already implemented; several application services—outside the scope of this work—can be easily developed on top.
Figure 1. Architecture of the proposed system.
It is composed of two main layers:
  • Posture Detection Layer: it includes two subsystems, (i) the Sensing subsystem, which uses pressure sensors deployed on the wheelchair to collect data generated by the weight of the body and (ii) the Recognition subsystem that, based on the Arduino platform (see Section 3.2), runs the posture recognition algorithm on the collected pressure sensor data. Posture detection results can therefore be fed to the application service layer.
  • Application Service Layer: several (mobile and cloud-based) applications can be developed on top of the posture recognition subsystem and exploit the geo-location from dedicated mobile device services (such as GPS, WiFi, or cellular tower signal strength) to locate the user. Applications can display the sensing results, perform activity level assessment, alert the user of dangerous postures and, if necessary, send his/her geo-location and make emergency automatic voice calls to caregivers.

3.2. Hardware Design of the Smart Cushion

The smart cushion is composed of three main modules:
  • data sensing (cushion equipped with pressure sensors);
  • data processing (Arduino-based unit);
  • data transmission (Bluetooth shield for Arduino).
The hardware design of the smart cushion is shown in Figure 2.
Figure 2. Hardware design of the smart cushion.
The data sensing module is based on Force Sensitive Resistor pressure sensors (Interlink FSR-406) [47]. The resistance value is inversely proportional to the pressure applied on the sensing patch. Such sensors can be easily embedded into the chair textile or foam filling. We chose Arduino DUE [48] to construct the data processing module for its low energy consumption, high sampling rate and processing capabilities. After sensor data collection and processing, results are sent to the smartphone in wireless, using an Arduino Bluetooth HC-06 shield [49] that is attached to the processing module.
As previously mentioned, we deployed 12 pressure sensors on the wheelchair (as shown in Figure 3) to evaluate our system. The list of required components along with its market price has been reported in Table 3. The total price of our prototype is $228. In comparison, the commercially available Body Pressure Measurement System (BPMS) using the sensors array developed by TekScan [27] has a sensor matrix of 42 × 48 , and it is sold for over ten thousand US dollars. PoSeat [35] was reported at a price of $460; another marketed smart wheelchair cushion called SENSIMAT [46] is currently sold at $599.
Figure 3. Sensor deployment: (a) schematic diagram of the sensor deployment; (b) sensors deployed on the real wheelchair.
Table 3. List of the components required by the prototype system.

4. System Evaluation

Instead of using sensor arrays, we chose to design our cushion with fewer pressure sensors individually connected to the processing unit; this has advantages in terms of economic cost and power consumption.
In Section 2.2, we provide an overview of different kinds of pressure sensor deployment methods. Most of the works listed in Table 2 did not systematically analyzed (or at least described) the deployment methods. In this study, we adopt two systematic principles to determine the sensor deployment; one uses empirical analysis to find the most significant sensors and the other tries to take into account all of the analyzed literature sensor deployment methods.
Here, we adopt a sensor deployment configuration as shown in Figure 3. The deployment is suitable for the body mass distribution. Actually, using the pressure sensors on the seat is enough to detect most of the sitting postures; however, we decided to apply pressure sensors on the backrest too, as a supplementary tool to more comprehensively monitor users’ postures.
Assuming a conventional size for the seat and backrest of 40 cm × 40 cm, we split the seat and backrest into 5 × 5 square zones. We performed a preliminary empirical analysis to exclude the least significant pressure points and found that the most effective initial deployment configuration to detect pressure changes due to different postures is the one depicted in Figure 3. It is worth noting that F S R 2 and F S R 4 have been deployed to the left and right sides of the seat to better deal with individual physiognomy differences.
In our previous studies [44], we used two pressure sensors on the seat and one pressure sensor on the backrest. However, we noticed that if the user does not sit perfectly at the center of seat, some posture transitions are not correctly detected. Therefore, with the aim of finding a better deployment to cope with different sitting positions, the smart cushion we designed in this work can record up to 12 pressure points simultaneously (the limitation is due to the number of analog inputs of the Arduino DUE platform). All of the FSR sensors (marked as F S R 0 , , F S R 11 ) were sampled at a frequency of 2 Hz (i.e., sampling rate of two samples per second) because it is enough to capture posture transitions and has the property of low energy consumption [44].
The electrical circuitry highlighting the connections of the FSR sensors to the Arduino platform is shown in Figure 4.
Figure 4. Electrical circuitry of the pressure sensor system.

4.1. Posture Definition and Data Collection

We carried out experimental tests in our laboratory involving 12 participants (seven males and five females) with ages in the range of [22, 36] and BMI in the range of [16, 34]. The BMI, whose distribution in our experiment sample is shown in Table 4, is an important parameter to consider in order to realize a user-independent posture recognition system.
Table 4. Body Mass Index (BMI) distribution of the subjects participating in the experiments.
To simulate a real-life setting, we asked each participant to perform five different postures that are typical for wheelchair users, as listed in Table 5. When the user sits on the wheelchair, the embedded sensors can measure multiple pressure points due to body weight. The posture data were manually labeled during the experiments; we asked to keep each posture for three minutes; in addition, at the beginning of the experiment and between each posture, we allow 30 s for the participant to adjust the posture in the wheelchair.
Table 5. Wheelchair user postures of interest.
The participants were free to choose their favorite seating position and each subject performed the experiment only once. The final dataset, consisting of data collected from all 12 of the test subjects, counts a total of 36,000 different posture recordings (see Table 5).
Pressure sensors signals are marked as { P 0 , P 1 , , P 11 } . At the time t, we obtained the instance vector P t = { P t 0 , P t 1 , , P t 11 } . Using this instance vector, we will recognize the sitting postures defined before.

4.2. Classification

Studies have shown that a lot of different algorithms can be used for the classification of sitting postures with satisfactory accuracy ranging from 90.9% to 99.5% [36,37,40,44]. Since the performance of the classification results is highly dependent on the used data, we compared the five algorithms as referred to before. In order to choose the best classifier for the classification, we analyzed five supervised machine learning algorithms applied for the recognition of sitting postures. Specifically, we compared J48 decision tree, Support Vector Machine, Multilayer Perceptron, Naive Bayes and k-Nearest Neighbors, respectively, briefly described in the following.
J48 [50] is a specific decision tree implementation of the well known C4.5 algorithm using the concept of information entropy. The classifier model is generated by a training procedure that uses a set of pre-classified samples. Each sample is a p-dimensional vector also known as a feature vector. Each node of the tree represents a decision (typically a comparison against a threshold value); at each node, J48 chooses the feature (i.e., an attribute) of the data that most effectively splits its set of samples into distinct subsets according to the normalized information gain (NIG). In particular, the algorithm chooses the feature with the highest NIG to generate the decision node. The main parameters used to tune the classifier generation are the pruning confidence (C) and the minimum number of instances per leaf (M), as summarized in Table 6.
Table 6. Tested machine learning algorithms.
Support Vector Machine (SVM) [51] is a supervised learning model used for both classification and regression analysis with associated learning algorithms that analyze data to recognize patterns. Given a set of training instances, new samples will be classified with one of the defined categories. The standard SVM builds a model that assigns new instances into one of two possible categories, so the SVM is a type of non-probabilistic binary linear classifier. The key parameters of the SVM classifier are reported in Table 6. In our analysis, we used a specific type of SVM model called regularized support vector classification (C-SVC) and we set the C parameter to 1. We used Radial Basis Function (RBF) as the kernel type with its degree set to 3.
Multilayer perceptron (MLP) [52] is a feed-forward artificial neural network model that maps sets of input data onto a set of appropriate outputs. An MLP typically involves three or more layers. MLP adopts a supervised learning technique called back propagation for network training. Except for the input nodes, each node is a neuron (or processing element) with a nonlinear activation function. Nodes are connected with weight values. The network output will be decided by node connection, weight values and activation function. The main parameters of the MLP classifier are reported in Table 6. For our purpose, we need 12 attributes and five classes (corresponding to the five sitting positions), so the model is composed of nine hidden layers. Default Momentum Rate and Learning Rate have been selected for the back-propagation algorithm.
Naive Bayes [53] is a simple technique for constructing classifiers: models that assign class labels to problem instances, represented as vectors of feature values, where the class labels are drawn from some finite set. It consists of a family of algorithms based on the assumption that the value of a particular feature is independent from the value of any other feature, given the class variable. An advantage of Naive Bayes is that it only requires a small amount of training data to estimate the parameters necessary for classification. Using different parameters, we trained two classifiers, one using default parameters and the other is BayesNet.
k-Nearest Neighbors algorithm (k-NN) [54] is a simple supervised learning method method used for classification and regression. In both cases, the input consists of the k closest training examples in the feature space. The parameter k plays the role of a capacity control. If k = 1, it means that the object is simply assigned to the class of that single nearest neighbor.
The five classifiers were set up with various configurations resulting in seven classifier configurations; each classifier was evaluated using the same dataset.

4.3. Sensor Deployment Optimization Using the Backward Selection Method

Our goal is to investigate the contribution of each individual sensor to the recognition accuracy and determine the best sensor deployment. This study was performed using the backward selection procedure [55]. It starts with fitting a model with all the variables of interest (i.e., all the 12 pressure sensors active). Then, sensors are excluded one at a time from the base configuration; we choose the best configuration, and the least significant variable is dropped. We iterate this procedure by successively re-fitting reduced models and applying the same rule until only one sensor is left.

5. Performance Evaluation

In this section, we present and discuss the results obtained from each classifier and choose the best sensor configuration using the backward selection method. To accurately study the performance evaluation, five relevant metrics (Accuracy, Precision, Recall, F-measure, and Model Build Time) have been compared.

5.1. Performance Evaluation of Each Classifier

We analyzed the seven classifiers as in Table 6 using WEKA (Waikato Environment for Knowledge Analysis) data mining toolbox [56] with the dataset described in Section 4.1. Data from all of the 12 pressure sensors are therefore taken into account in this phase. In this work, 10-fold cross validation was used to separate the training from the test set. The classifier models were obviously trained using the training set, while, in the test step, the estimated classes were compared to the true classes in order to compute the classification accuracy.
The Accuracy measure is used to evaluate the classifiers performance; specifically, it measures the proportion of correctly classified instances. In the case of binary classification, the formula of accuracy can be expressed as follows:
A c c u r a c y = T p + T n T p + T n + F p + F n
where T n (true negatives) is the correct classifications of negative examples, T p (true positives) is the correct classifications of positive examples. F n (false negatives) and F p (false positives) are, respectively, the positive examples incorrectly classified into the negative classes and the negative examples incorrectly classified into the positive classes.
For evaluating the classifiers, F-measure accuracy (overall accuracy) of the test data has been computed to evaluate recognition performance. F-measure represents the combination of precision and recall, defined, respectively, as follows:
p r e c i s i o n = T p T p + F p
r e c a l l = T p T p + F n
F m e a s u r e = 2 · r e c a l l · p r e c i s i o n r e c a l l + p r e c i s i o n
Table 7 summaries the results of our comparison. We recorded the same amount of samples for each different posture. The class imbalance problem was not taken into account for these classifiers. However, the main idea of this paper is to determine which classifier is more suitable for our system.
Table 7. Summary of classifiers’ performance.
J48 shows better performance than other classifiers. It is a classifier that predicts the posture classes of a new sample based on the threshold feature values. Furthermore, it is lightweight and easy to implement in embedded devices. We will eventually choose a J48 pruned tree for the implementation on the Arduino platform.
The SVM shows the worst performance among the analyzed classifiers; while it is very effective for binary classification, setting its key parameters for multi classification is much harder. In addition, it is computatively demanding and therefore not easily implemented in embedded and mobile devices. The MLP takes a longer time to build the train model than the J48 and its performance results are lower than J48. MLP also requires high processing resources, making it not ideal on embedded and mobile devices. We did not achieve good results with the Naive Bayes either, especially with configuration No. 4 (see Table 6), and it shows the worst performance among the classifiers we took into consideration. kNN achieved slightly worse accuracy than J48. We observed that different k values did not affect the recognition accuracy: our hypothesis is that the clusters identified by our feature set are internally consistent and well separated among each other. kNN requires less time to build the model (essentially the time required to split the dataset to obtain the training set), but it requires high computation load for classification, so it does not fit lightweight requirements imposed by embedded platforms.

5.2. Sensor Selection Using the Backward Selection Method

Given the results described in Section 5.1, we choose J48 as our classifier in the following. It is worth mentioning that the accuracy reported in the table refers to a configuration with all 12 pressure sensors. As described in Section 4.3, sensors were excluded one at a time from the base configuration and accuracy of the recognition was evaluated by a 10-fold cross-validation; then, the configuration with highest accuracy in this step became the base configuration. The procedure then iterates until only one sensor is left. The testing has been performed with a decimation factor d = 1 .
Results of the backward selection of various sensor configurations are shown in Table 8. The column represents the number of active sensors in a given configuration. The row represents the ID excluded sensor in a given configuration (let us mention here that once excluded from a configuration, with the backward selection that sensor cannot be re-added to any following iteration). The initial configuration has 12 active sensors achieving an accuracy of 99.48%, as shown in Table 7. At the first iteration of the backward selection procedure, the least significant sensor is F S R 10 because the best accuracy that can be achieved with 11 sensors is 99.50% and corresponds to the exclusion of F S R 10 . Then, the sensors’ configuration during the second iteration is { F S R 0 , F S R 1 , F S R 2 , F S R 3 , F S R 4 , F S R 5 , F S R 6 , F S R 7 , F S R 8 , F S R 9 , F S R 11 } ; this time, the best configuration can be achieved excluding F S R 6 , with a maximum possible accuracy of 99.51% (see the bold value in column 10). The procedure continues to iterate excluding, sequentially, F S R 8 , F S R 9 , F S R 2 , F S R 5 , and F S R 0 . At this point, the current best configuration has five sensors ( { F S R 1 , F S R 3 , F S R 4 , F S R 7 , F S R 11 } ) with an accuracy of 99.47% as shown in Figure 5. We can conclude that this is the most effective sensor deployment configuration in terms of tradeoff between posture recognition accuracy and number of required sensors (which influences the system complexity and its economic cost). Excluding a further sensor, in fact, would decrease the accuracy below 99%.
Table 8. Backward selection of the best sensor configuration. Bold font highlights the best accuracy for each number of active sensors.
Figure 5. Selected sensors of the best configuration.

5.3. Comparison of Recognition Results with Different BMI Values

Using the best sensor configuration determined in Section 5.2, we analyzed the recognition results grouped by the three weight categories classified by BMI (see Section 4.1). In addition, we analyzed the possible advantage of taking the BMI value as one of the features in the dataset.
The results grouped by BMI category obtained without considering the BMI as a feature of the dataset are shown in Table 9 while Table 10 depicts the result of classification accuracy using the dataset with the added BMI feature. Given that we obtained similar results, we can conclude that our sensor deployment can successfully fit different body physiognomies without the need to include user-dependent information (such as his/her BMI) into the classifier model.
Table 9. Results with different BMI values without considering the BMI feature.
Table 10. Results with different BMI values using the BMI feature in the J48 tree.

5.4. Comparison of the Proposed Method with Previous Studies

To compare our current work with the state-of-the-art, we applied the J48 classifier to datasets generated using sensor deployment configurations proposed in other literature works, and we summarized the results in Table 11.
Table 11. Recognition results of different sensor deployment.
Using the same J48 classifier, we compared the quality of different sensor deployments. It can be clearly seen that our novel proposed deployment method allows to achieve the highest accuracy.

6. Conclusions

In this paper, we have proposed an Arduino-based pressure cushion to detect sitting postures of wheelchair users and a sensor selection method to obtain optimal deployment on the wheelchair. We extracted pressure data and performed initial offline analysis using WEKA. Laboratory experiments were run over a diversified sample of subjects (in terms of BMI category) with different sitting habits. Using the obtained dataset, we compared five different machine learning approaches (Decision Tree (J48), Support Vector Machines (SVM), Multilayer Perceptron (MLP), Naive Bayes, and k-Nearest Neighbor (k-NN)) and selected the most effective sensor deployment among the possible configurations given by our smart cushion. We evaluated seven different classifiers to recognize five sitting postures. Results showed that the lightweight J48 algorithm can efficiently and accurately ( 99 . 47 % ) be used for posture recognition of seated users. An advantage of the decision tree-based approach is that trained models are very lightweight and easily implemented on embedded devices. Ongoing efforts are focused on improving current end-user applications based on posture detection. For example, activity level assessment can be achieved based on the analysis of posture transitions. Future research will be devoted to acquiring more data from heterogeneous sensors deployed on the chair, applying multi sensor fusion techniques [57] to better detect the postures, and assessing user activity on a smart wheelchair more accurately.
Finally, we plan to improve our BodyCloud-based Wheelchair System [10,24,58] with the work presented in this paper.

Acknowledgments

The research is financially supported by the China–Italy Science & Technology Cooperation project “Smart Personal Mobility Systems for Human Disabilities in Future Smart Cities” (China-side Project ID: 2015DFG12210, Italy-side Project ID: CN13MO7), and the National Natural Science Foundation of China (Grant Nos: 61571336 and 61502360). This work has been also carried out under the framework of “INTER-IoT” Project financed by the European Union’s Horizon 2020 Research & Innovation Programme under Grant 687283. The authors would like to thank the volunteers who participated in the experiments for their efforts and time.

Author Contributions

Congcong Ma designed the hardware platform and the experiment protocol, implemented the methodologies and drafted the manuscript. Raffaele Gravina contributed to the technical and scientific aspects, to the analysis of the experimental data, and to a substantial revision of the manuscript. Wenfeng Li and Giancarlo Fortino supervised the study and contributed to the overall research planning and assessment. All the authors contributed to the proofreading and approved the final version of the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Perez, A.J.; Labrador, M.A.; Barbeau, S.J. G-sense: A scalable architecture for global sensing and monitoring. IEEE Netw. 2010, 24, 57–64. [Google Scholar] [CrossRef]
  2. Bellifemine, F.; Fortino, G.; Giannantonio, R.; Gravina, R.; Guerrieri, A.; Sgroi, M. SPINE: A domain-specific framework for rapid prototyping of WBSN applications. Softw. Pract. Exp. 2011, 41, 237–265. [Google Scholar] [CrossRef]
  3. Banos, O.; Villalonga, C.; Damas, M.; Gloesekoetter, P.; Pomares, H.; Rojas, I. PhysioDroid: Combining Wearable Health Sensors and Mobile Devices for a Ubiquitous, Continuous, and Personal Monitoring. Sci. World J. 2014, 2014, 490824. [Google Scholar] [CrossRef] [PubMed]
  4. Bourouis, A.; Feham, M.; Bouchachia, A. A new architecture of a ubiquitous health monitoring system: A prototype of cloud mobile health monitoring system. Int. J. Comput. Sci. 2012, 9, 434–438. [Google Scholar]
  5. Alshurafa, N.; Xu, W.; Liu, J.J. Designing a robust activity recognition framework for health and exergaming using wearable sensors. IEEE J. Biomed. Health Inform. 2014, 18, 1636–1646. [Google Scholar] [CrossRef] [PubMed]
  6. Ma, C.; Gravina, R.; Li, W.; Zhang, Y.; Li, Q.; Fortino, G. Activity Level Assessment of Wheelchair Users Using Smart Cushion. In Proceedings of the 11th International Conference on Body Area Networks (BodyNets 2016), Turin, Italy, 15–16 December 2016. [Google Scholar]
  7. Yang, L.; Li, W.; Ge, Y.; Fu, X.; Gravina, R.; Fortino, G. People-Centric Service for mHealth of Wheelchair Users in Smart Cities. In Internet of Things based on Smart Objects: Technology, Middleware and Applications; Springer International Publishing: Cham, Switzerland, 2014; pp. 163–179. [Google Scholar]
  8. Tong, Y.; Sun, J.; Chow, S.M.; Li, P. Cloud-Assisted Mobile-Access of Health Data with Privacy and Auditability. IEEE J. Biomed. Health Inform. 2014, 18, 419–429. [Google Scholar] [CrossRef] [PubMed]
  9. Fortino, G.; Giannantonio, R.; Gravina, R.; Kuryloski, P.; Jafari, R. Enabling Effective Programming and Flexible Management of Efficient Body Sensor Network Applications. IEEE Trans. Hum.-Mach. Syst. 2013, 43, 115–133. [Google Scholar] [CrossRef]
  10. Fortino, G.; Fatta, G.; Pathan, M.; Vasilakos, A. Cloud-assisted Body Area Networks: State-of-the-art and Future Challenges. Wirel. Netw. 2014, 20, 1925–1938. [Google Scholar] [CrossRef]
  11. Gravina, R.; Ma, C.; Pace, P.; Aloi, G.; Russo, W.; Li, W.; Fortino, G. Cloud-based Activity-aaService cyberphysical framework for human activity monitoring in mobility. Futur. Gener. Comput. Syst. 2016. [Google Scholar] [CrossRef]
  12. Fortino, G.; Guerrieri, A.; Bellifemine, F.; Giannantonio, R. Platform-independent development of collaborative wireless body sensor network applications: SPINE2. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, 2009 (SMC 2009), San Antonio, TX, USA, 11–14 October 2009; pp. 3144–3150. [Google Scholar]
  13. Chou, H.; Wang, Y.; Chang, H. Design intelligent wheelchair with ECG measurement and wireless transmission function. Technol. Health Care 2015, 24, 345–355. [Google Scholar] [CrossRef] [PubMed]
  14. Srivastava, P.; Chatterjee, S.; Thakur, R. Design and Development of Dual Control System Applied to Smart Wheelchair using Voice and Gesture Control. Int. J. Res. Electr. Electron. Eng. 2014, 2, 1–9. [Google Scholar]
  15. Dryvendra, D.; Ramalingam, M.; Chinnavan, E.; Puviarasi, P. A Better Engineering Design: Low Cost Assistance Kit for Manual Wheelchair Users with Enhanced Obstacle Detection. J. Eng. Technol. Sci. 2015, 47, 389–405. [Google Scholar] [CrossRef]
  16. Sonenblum, S.E.; Sprigle, S.; Caspall, J.; Lopez, R. Validation of an accelerometer-based method to measure the use of manual wheelchairs. Med. Eng. Phys. 2012, 34, 781–786. [Google Scholar] [CrossRef] [PubMed]
  17. Postolache, O.; Viegas, V.; Pereira, J.M.D. Toward developing a smart wheelchair for user physiological stress and physical activity monitoring. In Proceedings of the International Conference on Medical Measurements and Applications (MeMeA), Lisbon, Portugal, 11–12 June 2014; pp. 1–6. [Google Scholar]
  18. Hiremath, S.V.; Intille, S.; Kelleher, A. Detection of physical activities using a physical activity monitor system for wheelchair users. Med. Eng. Phys. 2015, 37, 68–76. [Google Scholar] [CrossRef] [PubMed]
  19. Wu, Y.H.; Wang, C.C.; Chen, T.S. An Intelligent System for Wheelchair Users Using Data Mining and Sensor Networking Technologies. In Proceedings of the 2011 IEEE Asia-Pacific Services Computing Conference (APSCC), Jeju Island, Korea, 12–15 December 2011; pp. 337–344. [Google Scholar]
  20. Nakane, H.; Toyama, J.; Kudo, M. Fatigue detection using a pressure sensor chair. In Proceedings of the International Conference on Granular Computing (GrC2011), Kaohsiung, Taiwan, 8–10 November 2011; pp. 490–495. [Google Scholar]
  21. Ding, D.; Cooper, R.A.; Cooper, R.; Kelleher, A. Monitoring Seat Feature Usage among Wheelchair Users. In Proceedings of the 29th International Conference on Engineering in Medicine and Biology Society (EMBS 2007), Lyon, France, 22–26 August 2007; pp. 4364–4367. [Google Scholar]
  22. Fortino, G.; Gravina, R.; Li, W.; Ma, C. Using Cloud-assisted Body Area Networks to Track People Physical Activity in Mobility. In Proceedings of the 10th International Conference on Body Area Networks (BodyNets 2015), Sydney, Australia, 28–30 September 2015; pp. 85–91. [Google Scholar]
  23. Fortino, G.; Parisi, D.; Pirrone, V.; Fatta, G.D. BodyCloud: A SaaS approach for community Body Sensor Networks. Future Gener. Comput. Syst. 2014, 35, 62–79. [Google Scholar] [CrossRef]
  24. Ma, C.; Li, W.; Cao, J.; Gravina, R.; Fortino, G. Cloud-based Wheelchair Assist System for Mobility Impaired Individuals. In Proceedings of the 9th International Conference on Internet and Distributed Computing Systems (IDCS 2016), Wuhan, China, 28–30 September 2016; pp. 107–118. [Google Scholar]
  25. Ma, C.; Li, W.; Cao, J.; Wang, S.; Wu, L. A Fatigue Detect System Based on Activity Recognition. In Proceedings of the International Conference on Internet and Distributed Computing Systems, Calabria, Italy, 22–24 September 2014; pp. 303–311. [Google Scholar]
  26. Xu, W.; Huang, M.; Amini, N.; He, L.; Sarrafzadeh, M. eCushion: A Textile Pressure Sensor Array Design and Calibration for Sitting Posture Analysis. IEEE Sens. J. 2013, 13, 3926–3934. [Google Scholar] [CrossRef]
  27. Tekscan Website. Available online: http://www.tekscan.com/ (accessed on 20 March 2017).
  28. Tan, H.Z.; Slivovsky, L.A.; Pentland, A. A sensing chair using pressure distribution sensors. IEEE/ASME Trans. Mechatron. 2001, 6, 261–268. [Google Scholar] [CrossRef]
  29. Mota, S.; Picard, R.W. Automated Posture Analysis for Detecting Learner’s Interest Level. In Proceedings of the Conference on Computer Vision and Pattern Recognition Workshop, 2003 (CVPRW ’03), Madison, WI, USA, 16–22 June 2003; Volume 5, p. 49. [Google Scholar]
  30. Meyer, J.; Arnrich, B.; Schumm, J.; Troster, G. Design and modeling of a textile pressure sensor for sitting posture classification. IEEE Sens. J. 2010, 10, 1391–1398. [Google Scholar] [CrossRef]
  31. Mutlu, B.; Krause, A.; Forlizzi, J.; Guestrin, C.; Hodgins, J. Robust, low-cost, non-intrusive sensing and recognition of seated postures. In Proceedings of the 20th Annual ACM Symposium on User Interface Software and Technology, Newport, RI, USA, 7–10 October 2007; pp. 149–158. [Google Scholar]
  32. Kamiya, K.; Kudo, M.; Nonaka, H.; Toyama, J. Sitting posture analysis by pressure sensors. In Proceedings of the 19th International Conference on Pattern Recognition (ICPR2008), Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar]
  33. Xu, L.; Chen, G.; Wang, J.; Shen, R.; Zhao, S. A sensing cushion using simple pressure distribution sensors. In Proceedings of the 2012 IEEE Conference on. Multisensor Fusion and Integration for Intelligent Systems (MFI), Hamburg, Germany, 13–15 September 2012; pp. 451–456. [Google Scholar]
  34. Fard, F.; Moghimi, S.; Lotfi, R. Evaluating Pressure Ulcer Development in Wheelchair-Bound Population Using Sitting Posture Identification. Engineering 2013, 5, 132–136. [Google Scholar] [CrossRef]
  35. Yu, H.; Stoelting, A.; Wang, Y.; Yi, Z.; Sarrafzadeh, M. Providing a cushion for wireless healthcare application development. IEEE Potentials 2010, 29, 19–23. [Google Scholar] [CrossRef]
  36. Benocci, M.; Farella, E.; Benini, L. A context-aware smart seat. In Proceedings of the 2011 4th IEEE International Workshop on Advances in Sensors and Interfaces (IWASI), Brindisi, Italy, 28–29 June 2011; pp. 104–109. [Google Scholar]
  37. Bao, J.; Li, W.; Li, J.; Ge, Y.; Bao, C. Sitting Posture Recognition based on data fusion on pressure cushion. TELKOMNIKA Indones. J. Electr. Eng. 2013, 11, 1769–1775. [Google Scholar] [CrossRef]
  38. Arias, D.E.; Pino, E.J.; Aqueveque, P.; Curtis, D.W. Unobtrusive support system for prevention of dangerous health conditions in wheelchair users. Mobile Inform. Syst. 2016, 2016, 1–14. [Google Scholar] [CrossRef]
  39. Min, S.D. System for Monitoring Sitting Posture in Real-Time Using Pressure Sensors. U.S. Patent 20,160,113,583, 28 April 2016. [Google Scholar]
  40. Zemp, R.; Tanadini, M.; Plüss, S.; Schnüriger, K.; Singh, N.B.; Taylor, W.R.; Lorenzetti, S. Application of Machine Learning Approaches for Classifying Sitting Posture Based on Force and Acceleration Sensors. Biomed Res. Int. 2016, 2016, 1–9. [Google Scholar] [CrossRef] [PubMed]
  41. Barba, R.; Madrid, Á.P.; Boticario, J.G. Development of an inexpensive sensor network for recognition of sitting posture. Int. J. Distrib. Sens. Netw. 2015, 2015, 969237. [Google Scholar] [CrossRef]
  42. Fu, T.; Macleod, A. IntelliChair: An Approach for Activity Detection and Prediction via Posture Analysis. In Proceedings of the International Conference on Intelligent Environments, Washington, DC, USA, 30 June–4 July 2014; pp. 211–213. [Google Scholar]
  43. Kumar, R.; Bayliff, A.; De, D.; Evans, A.; Das, S.K.; Makos, M. Care-Chair: Sedentary Activities and Behavior Assessment with Smart Sensing on Chair Backrest. In Proceedings of the 2016 IEEE International Conference on Smart Computing, St Louis, MO, USA, 18–20 May 2016; pp. 1–8. [Google Scholar]
  44. Ma, C.; Li, W.; Gravina, R.; Fortino, G. Activity Recognition and Monitoring for Smart Wheelchair Users. In Proceedings of the 2016 IEEE Computer Supported Cooperative Work in Design (CSCWD), Nanchang, China, 4–6 May 2016; pp. 664–669. [Google Scholar]
  45. Darma Website. Available online: http://darma.co/ (accessed on 20 March 2017).
  46. Sensimat Website. Available online: http://www.sensimatsystems.com/ (accessed on 20 March 2017).
  47. Fsr406 Website. Available online: http://www.interlinkelectronics.com (accessed on 20 March 2017).
  48. Arduino DUE Website. Available online: https://www.arduino.cc (accessed on 20 March 2017).
  49. HC06 Bluetooth Website. Available online: https://www.olimex.com (accessed on 20 March 2017).
  50. Patil, T.; Sherekar, S. Performance analysis of Naive Bayes and J48 classification algorithm for data classification. Int. J. Comput. Sci. Appl. 2013, 6, 256–261. [Google Scholar]
  51. Chang, C.; Lin, C. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  52. Pirttikangas, S.; Fujinami, K.; Nakajima, T. Feature Selection and Activity Recognition from Wearable Sensors. Ubiquitous Comput. Syst. 2006, 4239, 516–527. [Google Scholar]
  53. Hall, M.A.; Frank, E. Combining Naive Bayes and Decision Tables. FLAIRS Conf. 2008, 2118, 318–319. [Google Scholar]
  54. Zhang, M.; Zhou, Z. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef]
  55. Sazonov, E.S.; Fulk, G.; Hill, J.; Schütz, Y.; Browning, R. Monitoring of posture allocations and activities by a shoe-based wearable sensor. IEEE Trans. Bio-Med. Eng. 2011, 58, 983–990. [Google Scholar] [CrossRef] [PubMed]
  56. Mark, H.; Eibe, F.; Geoffrey, H.; Bernhard, P.; Peter, R.; H, W.I. The WEKA data mining software: An update. ACM SIGKDD Explor. Newsl. 2009, 11, 10–18. [Google Scholar]
  57. Gravina, R.; Alinia, P.; Ghasemzadeh, H.; Fortino, G. Multi-Sensor Fusion in Body Sensor Networks: State-of-the-art and research challenges. Inf. Fusion 2017, 35, 68–80. [Google Scholar] [CrossRef]
  58. Fortino, G.; Gravina, R.; Guerrieri, A.; Fatta, G.D. Engineering Large-Scale Body Area Networks Applications. In Proceedings of the 8th International Conference on Body Area Networks (BodyNets 2013), Boston, MA, USA, 30 September–2 October 2013; pp. 363–369. [Google Scholar]

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.