Next Article in Journal
Label-Free Fluorescence Assay of S1 Nuclease and Hydroxyl Radicals Based on Water-Soluble Conjugated Polymers and WS2 Nanosheets
Next Article in Special Issue
Normalized Metadata Generation for Human Retrieval Using Multiple Video Surveillance Cameras
Previous Article in Journal
Recognition of Banknote Fitness Based on a Fuzzy System Using Visible Light Reflection and Near-infrared Light Transmission Images
Previous Article in Special Issue
Networked Fusion Filtering from Outputs with Stochastic Uncertainties and Correlated Random Transmission Delays

Sensors 2016, 16(6), 864; https://doi.org/10.3390/s16060864

Article
Gyro Drift Correction for An Indirect Kalman Filter Based Sensor Fusion Driver
1
School of Computer Science and Engineering, Chung-Ang University, Seoul 156-756, Korea
2
Department of Vehicle Components, LG Electronics, Seoul 073-36, Korea
*
Author to whom correspondence should be addressed.
Academic Editors: Xue-Bo Jin, Feng-Bao Yang, Shuli Sun and Hong Wei
Received: 24 April 2016 / Accepted: 8 June 2016 / Published: 11 June 2016

Abstract

:
Sensor fusion techniques have made a significant contribution to the success of the recently emerging mobile applications era because a variety of mobile applications operate based on multi-sensing information from the surrounding environment, such as navigation systems, fitness trackers, interactive virtual reality games, etc. For these applications, the accuracy of sensing information plays an important role to improve the user experience (UX) quality, especially with gyroscopes and accelerometers. Therefore, in this paper, we proposed a novel mechanism to resolve the gyro drift problem, which negatively affects the accuracy of orientation computations in the indirect Kalman filter based sensor fusion. Our mechanism focuses on addressing the issues of external feedback loops and non-gyro error elements contained in the state vectors of an indirect Kalman filter. Moreover, the mechanism is implemented in the device-driver layer, providing lower process latency and transparency capabilities for the upper applications. These advances are relevant to millions of legacy applications since utilizing our mechanism does not require the existing applications to be re-programmed. The experimental results show that the root mean square errors (RMSE) before and after applying our mechanism are significantly reduced from 6 . 3 × 10 - 1 to 5 . 3 × 10 - 7 , respectively.
Keywords:
sensor fusion; indirect Kalman filter; accuracy improvement; gyro drift correction

1. Introduction

Currently, the mobile applications era is looking forwards to the next generation where a virtual personal assistant (VPA) performs the central unified framework, which integrates separate applications to provide humans with context-aware personalized information and facilities [1]. Cooperating with various advanced technologies, sensor fusion plays an important role in improving the accuracy of multi-sensing information from the surrounding environment, which is a vital input data for a VPA to correctly respond to the user’s requests or for the user experience features for interactive relaxation applications, etc. [2]. By utilizing the Internet of Things (IoT) infrastructure, we are able to interconnect with many kinds of sensors. However, within the human activities of daily living (HADL) focus, accelerometer and gyroscope sensors are the most popular objects, which have already been installed in billions of smartphones nowadays. Moreover, this progress is forecast to continue for at least the next decade [3].
However, due to the rapid upgrading of smartphones, mobile applications are facing difficulties in that the code requires modification to adapt to additional hardware components in new devices. This problem negatively affects billions of existing smartphones in the world and millions of applications in the market stores [4]. For instance, the pedometer sensor was introduced in the Google Nexus 5 (Google, San Francisco, CA, USA), Samsung Galaxy S5 (Samsung, Seoul, Korea), and Apple iPhone 5S (Apple, Cupertino, CA, USA) generations. Therefore, step-count information provided by the pedometer cannot be taken into account without modifications of their code. Moreover, even if the developers try to modify and re-compile the code to utilize the sensor fusion technique for upgrading their applications, another problem arises because the applications separately apply sensor fusion using their own approach. This not only burdens the smartphone’s performance, but also creates more latency in the processing time [5].
Although there are existing solutions which implemented sensor fusion in kernel space as well as user space, typical implementations of Kalman-based sensor fusion require external feedback loops between the components that are requesting and providing the services. Some solutions are installed at firmware level according to different underlying hardware; as a consequence, they are inflexible and inconvenient to be updated. On the other hand, the solutions implemented in the application level often exhibit lower performance compared to kernel space solutions in many operating systems; even worse, all deployed applications have to be re-programmed when the underlying hardware is changed or updated. This is critical because there are billions of deployed applications nowadays.
Attempts to address these aforementioned problems have achieved significant success by using some interesting approaches [6]. The existing solutions have applied a variety of theoretical frameworks to build effective fusion algorithms for imperfect data (see Section 2 for more detail). However, almost all approaches concentrate only on improving the accuracy and fusion calculating performance without any consideration of implementation position or facilitation for developers, which is very important in the emerging mobile applications era [7].
From this point of view, we have proposed a new solution that implements the well-known Kalman filter for sensor fusion in the device-driver layer. The experimental results show that our solution provides greater convenience, benefiting the developer by making the applications independent of the underlying sensor hardware upgrade. A part of our work has been presented at the third International workshop on smartphone applications and services, 2011 [8]. Based on these achievements, we enhance the accuracy and processing performance by using a quaternion based indirect Kalman filter and develop additional feedback components to correct the gyro drift problem. The gyroscope and accelerometer values are pre-calculated before reaching the corresponding applications.
Our main contributions in this paper are summarized as follows:
  • We have proposed a new software architecture for sensor fusion driver utilizing the quaternion based indirect Kalman filter in conjunction with additional feedback components for gyro drift correction. These components do not only handle the external feedback loop issue between the device driver and applications, but also cancel the non-gyro signal in the measured state vector.
  • The developed sensor fusion driver abstracts underlying sensor hardware to provide an unified framework for mobile applications. The multi-sensing information is facilitated without any requirement of re-programming or modification in the existing applications. It supports backward compatibility for the legacy applications as well.
  • The implementation in the device driver layer provides greater performance up to 10 times from 538 to 5,347 samples per second, and lower latency in calculation time from 1.8579 ns to 0.18702 ns. The duplication of sensor fusion process among applications is completely addressed.
The remainders of this paper are organized as follows. Section 2 classifies the related work of different approaches into corresponding categories. Section 3 defines the problem statement and our approach to resolve it. The proposed solution is described in Section 4. Section 5 shows the experimental results and discussion. Conclusions are drawn in Section 6.

2. Related Work

As mentioned above, in the scope of this paper, we are concerned about sensor fusion problems that happen in smartphone environments where VPA applications are closely installed. Therefore, the survey of related work has concentrated on existing solutions for imperfect data from multi-sensing information. We classify the related works into two groups by systematical designs, including centralized calculation and local calculation. In both groups, the existing solutions are also divided into theoretical approaches: probabilistic, statistic, and artificial intelligence.

2.1. Systematical Designs

The partial-centralized and fully centralized calculation approaches are typical models that were introduced in past decades. Based on the network architecture, within the clustering and tree-topology models, the multi-sensing information might be roughly pre-filtered at the clustering head nodes or the i-th level root nodes before reaching the centralized processing server [5,9,10]. The centralized calculation approach provides the outstanding advantages of high performance, complex mathematical solutions, social information sharing, unified IoT framework, etc. [11,12,13]. However, in some special cases, e.g., VPA applications, it reveals limitations of response latency, assisted-information localization, and personal data privacy.
In contrast, the local calculation approach is able to compensate for the above limitations. However, due to its natural disadvantages, nowadays, local calculation models only play a supplemental role in the centralized solution as a local agents [10,11,14]. The combination of two approaches almost satisfies the HADL applications’ requirements.

2.2. Theoretical Approaches

Based on theoretical approaches, our taxonomy classifies the existing related work into three categories, including a probabilistic method, statistical method, and artificial intelligence application method.
Within the probabilistic methods, the probability distribution functions of multi-sensing data are fused together using Bayesian analysis. The rationale of this method is that since the sensing information from related sources has properties of concrete probability distributions, combinations of this information will be able to better correct the imperfect data [4,6]. This approach is only effective when the related data has well-known distributions and the environment’s conditions are stable. One exceptional case of Bayesian analysis is the well-known Kalman filter and its variations. The orientation correction and error reduction of the Kalman filter based sensor fusion technique were significantly improved by a variety of research such as the effective solution proposed by Van Der Merwe in [15]. Nowadays, the Kalman filter is one of the most popular fusion methods, which is implemented in various sensing domains due to its simplicity and acceptable accuracy [16,17].
Applying the statistical method, the evidence of sensing data from multi sensors are collected and analyzed during the processing time [4]. Using the Dempster-Shafer evidence accumulation rule, the probability mass function is fused to recognize the properties of the data and predict the trend of the next sensing value. The Depster-Shafer framework allows different information levels from multi resources to be combined, depending on its weight factor [5,18]. It makes more flexible to reproduce the desired sensing data. Besides that, the context-aware, semantic information matching [19], and fuzzy reasoning methods [20,21] are also utilized to adjust the output data as required.
Recently, the artificial intelligence method is becoming popular in the information fusion field. Almost proposed solutions are centralized models where the multi-sensing data is gathered into a logical central processing server. The complex algorithms are performed using neural networks [22,23] and machine learning techniques [24] for analyzing and combining sensing data to obtain the desired information. Moreover, the emerging cloud-fog computing infrastructure also provides better performance for artificial intelligence algorithms.

3. Preliminaries

3.1. Problem Statement

Following the above survey, since the outstanding properties of the Kalman filter and its variations are simplicity in implementation and acceptable accuracy for almost all HADL applications, nowadays, these filters are the most popular sensor fusion methods and are widely integrated into personal devices such as smartphones, smartwatchs, and wearable devices. The position at which the Kalman filters are implemented is also an interesting topic inspiring researchers to pay attention. The dominant strategies focus on the application layer (i.e., internal process of stand alone applications) and remote processing (i.e., cloud and fog computing). The general architecture of legacy sensor fusion methods is described in Figure 1. Whenever the application requests sensing information, it must simultaneously contact multi sensors through separate corresponding drivers. The overhead in request and response operations rapidly increases, proportionally to the number of related sensors. Besides that, the fusion process is calculated in the application layer, which might cause significant latency and cannot be acceptable by real-time applications.
The strategy to implement sensor fusion in the lower layer is even more challenging due to the difference in protocols and interface standards among sensor components. For instance, within our proposal, we intend to apply the indirect Kalman filter at the device driver level to obtain better performance and lower latency and overhead, we are facing some challenges as follows:
  • External feedback loop: There is a feedback loop between the modules of the filter. If the indirect Kalman filter were directly applied, it would result in a feedback loop signal between the applications and the device driver, which is not a desirable configuration. When the Kalman filter is implemented in the device driver level, it would use Euler angle kinematics in calculation, which is not linear. The orientation value is transformed to gyro value before coming out of the Kalman filter, then it has to be converted to orientation value again inside the applications. The repeated transformations may cause error accumulations because Euler angle kinematics is not linear. Note that the feedback loop is outside of the Kalman filter.
  • Non-gyro signal in the measured state vector: Within the original indirect Kalman filter, a non-gyro signal always appears in the state vector after measurement. It is not hard to resolve this problem in the desired applications, but it wastes more time.

3.2. Our Approach

In the emerging mobile application era, especially focusing on the VPA framework, quick and exact response capability is very important. Moreover, since the applications are implemented in mobile and wearable devices, which have limited hardware resources and battery, the processing overhead must also be considered. To satisfy these requirements, we utilized a driver-based approach. Our proposed sensor fusion architecture was developed at the driver layer to provide a unified framework where the measured sensing data is collected and processed before reaching the upper applications.
Due to its popularity and simplicity, the indirect Kalman filter is integrated into the framework to perform information fusing functions. In order to address the external feedback loop and non-gyro signal in the measured state vector issues, we apply the quaternion based method on the indirect Kalman filter incorporating additional modules to return the required gyro and accelerometer values at the corresponding output interfaces.

4. Proposed Solution

4.1. Sensor Fusion Driver Architecture

Figure 2 shows our proposed sensor fusion driver architecture, which consists of a quaternion based indirect Kalman filter and three additional modules named T q n , T q a , and C q . The gyro interface and accelerometer interface separately serve applications using their expected gyro and accelerometer information, respectively.
In the figure, Ω, δ Ω , and Δ Ω represent the angular velocity, gyro bias error, and gyro drift rate noise, respectively, and q m is the measured value received from the accelerometer sensor. The quaternion value q ^ is defined as
q ^ = q 4 + q 1 i + q 2 j + q 3 k
where i, j, and k are hyper-imaginary numbers, q 4 represents the real part of the quaternion, which is equal to
q 4 = cos ( θ / 2 )
where θ is the angle of rotation [25].
Denote the measured value received from the gyro sensor by w m . The noise model of the gyro sensor is derived as follows
w m = Ω + δ Ω + Δ Ω
The purposes of gyro error compensation modules I and II is to remove the δ Ω and Δ Ω signals, respectively. The error state vector x ^ of the indirect Kalman filter can be expressed as
x ^ = δ q 1 δ q 2 δ q 3 δ q x δ q y δ q z
where δ q = δ q 1 δ q 2 δ q 3 .

4.2. Gyro Drift Correction

We resolved the external feedback loop by deploying two additional modules, called C q and T q a , into the system. The module C q computes the orientation from a given gyro value and the module T q a transforms the orientation to the accelerometer value that is expected from the accelerometer-assisted application.
For non-gyro signals in the measured state vector, we integrated an extra module called T q n . The role of T q n is to convert the δ q value, which includes elements other than the gyro error, to Δ Ω , i.e., the gyro rate noise. The quaternion based indirect Kalman filter is utilized due to the linear properties of quaternion kinematics [26]. In this way, the error accumulation during the repeated transformation is significantly reduced.
The computation of the C q module is based on an assumption that during a time interval Δ t = t k + 1 - t k , the angular velocity Ω = ( Ω x , Ω y , Ω z ) is stable. Therefore, we can consider the zero-th order quaternion integration as a quaternion product [27] given by
q ^ = Ω Ω sin ( Ω 2 Δ t ) cos ( Ω 2 Δ t ) q ^ k - 1
The purpose of module T q a is to transform the orientation into the accelerometer value. T q a takes the quaternion q ^ as an input and translates it into an Euler orientation. After that, the Euler orientation is translated into an accelerometer value. The details for this transformation can be found in [25].
The transformation from δ Ω to Δ Ω is performed in the module T q a . From Equation (4), we can derive δ q k + 1 as follows
δ q 1 δ q 2 δ q 3 δ q 4 k + 1 = δ q 4 - δ q 3 δ q 2 δ q 1 δ q 3 δ q 4 - δ q 1 δ q 2 - δ q 2 δ q 1 δ q 4 δ q 3 - δ q 1 - δ q 2 - δ q 3 δ q 4 k Δ Ω x Δ Ω sin ( Δ Ω 2 Δ t ) Δ Ω y Δ Ω sin ( Δ Ω 2 Δ t ) Δ Ω z Δ Ω sin ( Δ Ω 2 Δ t ) cos ( Δ Ω 2 Δ t ) k
where δ q 4 is approximated to unity since the incremental quaternion is assumed to be a very small rotation. Calculating Equation (6), we derive
Δ Ω x = 2 cos - 1 ( γ 4 ) Δ t 1 - γ 4 2 γ 1
Δ Ω y = 2 cos - 1 ( γ 4 ) Δ t 1 - γ 4 2 γ 2
Δ Ω z = 2 cos - 1 ( γ 4 ) Δ t 1 - γ 4 2 γ 3
where Δ Ω x , Δ Ω y , and Δ Ω z are the gyro rate noises on the X, Y, and Z axis, respectively. In the above equation, the γ value can be computed as
γ 1 γ 2 γ 3 γ 4 = δ q 4 δ q 3 - δ q 2 δ q 1 - δ q 3 δ q 4 δ q 1 δ q 2 δ q 2 - δ q 3 δ q 4 δ q 3 - δ q 1 - δ q 2 - δ q 3 δ q 4 k - 1 δ q 1 δ q 2 δ q 3 δ q 4 k + 1

5. Experimental Results and Discussions

5.1. Preparation

In order to evaluate the performance of our proposed solution, we have conducted experiments with 41,500 samples measured by a pair consisting of a gyroscope sensor and an accelerometer sensor. The evaluations are performed based on an Android mobile device whose specifications are described in Table 1. The results are compared among the proposed approach, the legacy approach, and the reference standard provided by an encoder [28]. The true orientation values were recorded by the encoder, which can precisely capture orientations of the experimented object. A conceptual diagram of the encoder is shown in Figure 3.

5.2. Accuracy Evaluation

We assess the accuracy among the experimental approaches by calculating the root mean square error (RMSE) as following
R M S E = i = 1 n | V m i - V t i | 2 n
where V m i and V t i are the value measured by the sensor and the actual value of the Euler angle at the i-th time, respectively, and n is the total number of sampled sensor data.
Figure 4a–c show the results of Euler pitch angle calculation by using a separate accelerometer sensor, our proposed solution, and the encoder, respectively. The values and line graph shapes indicate the equivalence between our proposed solution and the encoder. However, the result of the separate accelerometer sensor is different, except the trend of the line graph. The reason is the accelerometer sensor does not respond quickly to changes of the status during high speed motion. In this Figure 4b, with our proposed solution, even though the angles are calculated by using the acceleration sensor, the result can follow the progress measured by the encoder very well without any negative effects resulting from the translational motion.
Under the same analysis as applied to the accelerometer sensor, Figure 5a–c show the results of the Euler pitch angle calculation by using a separate gyroscope sensor, our proposed solution, and the encoder, respectively. Unlike the accelerometer sensor, the separate gyroscope sensor can follow the progress of angle changes. However, due to the shortcoming physical characteristics, the gyroscope sensor has a gyro drift problem that distorts the shape of the graph over time (see Figure 5a). On the other hand, with our proposed solution, the sensing information from the gyroscope sensor and accelerometer sensor is fused to compensate for each other. As a result, the output value is approximately the same as the value provided by the encoder (see Figure 5b,c).
Figure 6 shows the calculated orientations comparison among the true value, our proposed mechanism, and the sensor fusion device driver (SFDD) mechanism [8]. The X-axis represents the time-lines at which the measurements were performed and the Y-axis represents the orientation values in degrees. It is observed that the proposed method provides a more accurate estimation of the orientation values than the results of the SFDD mechanism. As the recorded values are closer to each end of the range of ( - 30 , 30 ) , the difference between our proposed mechanism and SFDD mechanism becomes much clear. It is noteworthy that the gyro drift is almost corrected by the proposed method. The RMSE of the SFDD mechanism and the proposed mechanism are decreased from 6 . 3 × 10 - 1 to 5 . 3 × 10 - 7 , respectively.
The dependence of the calculated values on the angle range are summarized in Table 2. The calculated values are compared with the reference standard values received by the encoder.

5.3. Performance Evaluation

To evaluate the performance of the proposed solution, we calculated the average processing speed per sample over 4500 experimental data samples. The result shows that the proposed solution significantly increases the processing speed up to 10 times in comparison with the existing application-based methods, from 1.8579 ns to 0.18702 ns, respectively. This improvement is achieved by the movement of the fusion process from the application layer to the driver layer on the kernel level. Generally, the sample rates provided by the Android API are 4 samples/s in normal mode and 50 samples/s in fastest mode. The experiments show that the application-based method and our proposed method are able to handle 538 and 5347 samples per second, respectively.

6. Conclusions

In this paper, we proposed a solution to implement the sensor fusion technique at the driver layer of a mobile device. The quaternion based indirect Kalman filter and three additional modules are integrated to provide better performance and accuracy of sensing information. The external feedback loop and non-gyro error elements contained in the state vector issues are addressed before the required sensing data is delivered to the desired corresponding applications. The experimental results show that our proposed solution achieves better performance and accuracy. Moreover, any change in the device driver does not negatively affect the upper applications. The applications do not need to modify their code or be re-programmed and re-compiled to utilize the sensor fusion features. In the future research, the proposed architecture will be expanded to interact with multiple sensors and be implemented based on different devices’ hardware and operating systems.

Acknowledgments

This research was supported by the Chung-Ang University Research Scholarship Grants in 2015, Korea Electric Power Corporation through Korea Electrical Engineering & Science Research Institute (grant number: R15XA03-69), and a National Research Foundation of Korea Grant funded by the Korean Government 231 (No. 2011-0013924, 2010-0015636, and 2014R1A2A2A01005519).

Author Contributions

Chan-Gun Lee and D. Kim conceived and developed the algorithm; Nhu-Ngoc Dao and Sungrae Cho surveyed the related work and wrote the manuscript; Deokhwan Kim and Yonghun Kim performed the experiments; Nhu-Ngoc Dao analyzed the data; Seonmin Jang performed additional studies after revisions; Chan-Gun Lee and Sungrae Cho verified the results and finalized the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kalns, E.T.; Freitag, D.B.; Mark, W.S.; Ayan, N.F.; Wolverton, M.J.; Lee, T.J. Rapid Development of Virtual Personal Assistant Applications. U.S. Patent 9,081,411, 7 July 2015. [Google Scholar]
  2. Ranjan, R.; Wang, M.; Perera, C.; Jayaraman, P.P.; Zhang, M.; Strazdins, P.; Shyamsundar, R.K. City Data Fusion: Sensor Data Fusion in the Internet of Things. Int. J. Distrib. Syst. Technol. 2016, 7, 15–36. [Google Scholar]
  3. Pires, I.M.; Garcia, N.M.; Pombo, N.; Flórez-Revuelta, F. From Data Acquisition to Data Fusion: A Comprehensive Review and a Roadmap for the Identification of Activities of Daily Living Using Mobile Devices. Sensors 2016, 16. [Google Scholar] [CrossRef] [PubMed]
  4. Qin, X.; Gu, Y. Data fusion in the Internet of Things. Proced. Eng. 2011, 15, 3023–3026. [Google Scholar] [CrossRef]
  5. Shivaprasad Yadav, S.G.; Chitra, A.; Lakshmi Deepika, C. Reviewing the process of data fusion in wireless sensor network: a brief survey. Int. J. Wirel. Mob. Comput. 2015, 8, 130–140. [Google Scholar] [CrossRef]
  6. Khaleghi, B.; Khamis, A.; Karray, F.O.; Razavi, S.N. Multisensor data fusion: A review of the state-of-the-art. Inf. Fusion 2013, 14, 8–44. [Google Scholar] [CrossRef]
  7. Ganti, R.K.; Srinivasan, S.; Gacic, A. Multisensor Fusion in Smartphones for Lifestyle Monitoring. In Proceedings of the 2010 International Conference on Body Sensor Networks, Singapore, 7–9 June 2010.
  8. Gim, D.H.; Min, S.H.; Lee, C.G. A Novel Technique for Composing Device Drivers for Sensors on Smart Devices. In IT Convergence and Services; Springer Lecture Notes in Electrical Engineering; Springer: Gwangju, Korea, 2011; Volume 107, pp. 671–677. [Google Scholar]
  9. Alamri, A.; Ansari, W.S.; Hassan, M.M.; Hossain, M.S.; Alelaiwi, A.; Hossain, M.A. A survey on sensor-cloud: Architecture, applications, and approaches. Int. J. Distrib. Sens. Netw. 2013, 2013, 1–18. [Google Scholar] [CrossRef]
  10. La, H.M.; Sheng, W. Distributed sensor fusion for scalar field mapping using mobile sensor networks. IEEE Trans. Cybern. 2013, 43, 766–778. [Google Scholar] [PubMed]
  11. Rawat, P.; Singh, K.D.; Chaouchi, H.; Bonnin, J.M. Wireless sensor networks: a survey on recent developments and potential synergies. J. Supercomput. 2014, 68, 1–48. [Google Scholar] [CrossRef]
  12. Mishra, S.; Thakkar, H. Features of WSN and data aggregation techiques in WSN: A survey. Int. J. Eng. Innov. Technol. 2012, 1, 264–273. [Google Scholar]
  13. Tripathi, A.; Gupta, S.; Chourasiya, B. Survey on data aggregation techniques for wireless sensor networks. Int. J. Adv. Res. Comput. Commun. Eng. 2014, 3, 7366–7371. [Google Scholar]
  14. Novak, D.; Riener, R. A survey of sensor fusion methods in wearable robotics. Robot. Auton. Syst. 2015, 73, 155–170. [Google Scholar] [CrossRef]
  15. Van der Merwe, R.; Wan, E.A.; Julier, S. Sigma-point Kalman filters for nonlinear estimation and sensor-fusion: Applications to integrated navigation. In Proceedings of the AIAA Guidance, Navigation & Control Conference, Providence, RI, USA, 16–19 August 2004; pp. 16–19.
  16. Sun, S.L.; Deng, Z.L. Multi-sensor optimal information fusion Kalman filter. Automatica 2004, 40, 1017–1023. [Google Scholar] [CrossRef]
  17. Chen, S.Y. Kalman filter for robot vision: A survey. IEEE Trans. Ind. Electron. 2012, 59, 4409–4420. [Google Scholar] [CrossRef]
  18. Enders, R.H.; Brodzik, A.K.; Pellegrini, M.R. Algebra of Dempster-Shafer evidence accumulation. In Proceedings of the SPIE 6968, Signal Processing, Sensor Fusion, and Target Recognition XVII, 696810, Orlando, FL, USA, 16 March 2008.
  19. Jenkins, M.P.; Gross, G.A.; Bisantz, A.M.; Nagi, R. Towards context aware data fusion: Modeling and integration of situationally qualified human observations to manage uncertainty in a hard + soft fusion process. Inf. Fusion 2015, 21, 130–144. [Google Scholar] [CrossRef]
  20. Chen, S.; Deng, Y.; Wu, J. Fuzzy sensor fusion based on evidence theory and its application. Appl. Artif. Intell. Int. J. 2013, 27, 235–248. [Google Scholar] [CrossRef]
  21. Izadi, D.; Abawajy, J.H.; Ghanavati, S.; Herawan, T. A data fusion method in wireless sensor networks. Sensors 2015, 15, 2964–2979. [Google Scholar] [CrossRef] [PubMed]
  22. Maren, A.J.; Craig, H.T.; Robert, P.M. Handbook of Neural Computing Applications; Academic Press: San Diego, CA, USA, 2014. [Google Scholar]
  23. Paul, P.S.; Varadarajan, A.S. A multi-sensor fusion model based on artificial neural network to predict tool wear during hard turning. J. Eng. Manuf. 2012, 226, 853–860. [Google Scholar] [CrossRef]
  24. Pinto, A.R.; Montez, C.; Araujo, G.; Vasques, F.; Portugal, P. An approach to implement data fusion techniques in wireless sensor networks using genetic machine learning algorithms. Inf. Fusion 2014, 15, 90–101. [Google Scholar] [CrossRef]
  25. Diebel, J. Representing Attitude: Euler Angles, Unit Quaternions, and Rotation Vectors; Technical report; Stanford University: Stanford, CA, USA, 2006. [Google Scholar]
  26. Chonkroun, D.; Bar-Itzhack, I.Y.; Oshman, Y. Novel quaternion Kalman filter. IEEE Trans. Aerosp. Electron. Syst. 2006, 42, 174–190. [Google Scholar] [CrossRef]
  27. Trawny, N.; Roumeliotis, S. Indirect Kalman Filter for 3D Attitude Estimation; University of Minnesota: Minneapolis, MN, USA, TR 2005-002, Rev. 57; March 2005. [Google Scholar]
  28. Lee, H.J.; Jung, S. Gyro sensor drift compensation by Kalman filter to control a mobile inverted pendulum robot system. In Proceedings of the IEEE International Conference on Industrial Technology, Gippsland, Australia, 10–13 February 2009; pp. 1–6.
Figure 1. Sensor fusion processes in the legacy methods.
Figure 1. Sensor fusion processes in the legacy methods.
Sensors 16 00864 g001
Figure 2. The proposed sensor fusion driver architecture.
Figure 2. The proposed sensor fusion driver architecture.
Sensors 16 00864 g002
Figure 3. The encoder for recording the true orientation value.
Figure 3. The encoder for recording the true orientation value.
Sensors 16 00864 g003
Figure 4. The Euler pitch angles calculated using a separate accelerometer sensor, our proposed solution, and the encoder. (a) Separate accelerometer sensor; (b) Our proposed solution; (c) The encoder.
Figure 4. The Euler pitch angles calculated using a separate accelerometer sensor, our proposed solution, and the encoder. (a) Separate accelerometer sensor; (b) Our proposed solution; (c) The encoder.
Sensors 16 00864 g004
Figure 5. The Euler pitch angles calculated using a separate gyroscope sensor, our proposed solution, and the encoder. (a) Separate gyroscope sensor; (b) Our proposed solution; (c) The encoder.
Figure 5. The Euler pitch angles calculated using a separate gyroscope sensor, our proposed solution, and the encoder. (a) Separate gyroscope sensor; (b) Our proposed solution; (c) The encoder.
Sensors 16 00864 g005
Figure 6. The comparison of calculated orientations among the true value, our proposed mechanism, and the sensor fusion device driver (SFDD) mechanism.
Figure 6. The comparison of calculated orientations among the true value, our proposed mechanism, and the sensor fusion device driver (SFDD) mechanism.
Sensors 16 00864 g006
Table 1. Specifications of the experimental mobile device.
Table 1. Specifications of the experimental mobile device.
SpecificationValue
CPUS5PV210 ARM-CORTEX A8 [1 GHz]
Memory512M DDR SDRAM
KernelLinux kernel 2.6.32 (Android OS)
Accelerometer sensor3-axis accelerometer sensor
Gyroscope sensor2-axis gyroscope sensor
Sampling frequency100 Hz
Angle range ( - 90 , 90 )
Table 2. Calculation bias of the sensing information by using difference methods.
Table 2. Calculation bias of the sensing information by using difference methods.
MethodSensor ( - 30 , 30 ) ( - 60 , 60 ) ( - 90 , 90 )
Separate sensorAccelerometer22.354621.958935.5267
Gyroscope70.958072.6149307.1894
Our proposed methodsAccelerometer0.048750.05370.2537
Gyroscope1.87021.93098.6471
Back to TopTop