Next Article in Journal
Comparison of Active Drag Using the MRT-Method and the MAD-System in Front Crawl Swimming
Previous Article in Journal
The Influence of the Grip Acceleration on Club Head Rotation during a Golf Swing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

An Automated Method to Extract Three-Dimensional Position Data Using an Infrared Time-of-Flight Camera †

Centre for Sports Engineering, Sheffield Hallam University, Sheffield S10 2LX, UK
*
Author to whom correspondence should be addressed.
Presented at the 12th Conference of the International Sports Engineering Association, Brisbane, Queensland, Australia, 26–29 March 2018.
Proceedings 2018, 2(6), 502; https://doi.org/10.3390/proceedings2060502
Published: 11 February 2018

Abstract

:
Traditional motion capture systems can be prohibitive in healthcare settings from time, cost, space and user-expertise perspectives. Ideally, movement analysis technologies for healthcare should be low-cost, quick, simple and usable in small spaces. This study demonstrates a simple, low-cost and close-range time-of-flight depth-camera system, for automatic gait analysis. A method to automatically track three-dimensional position and orientation of retro-reflective marker-triads in real-time was developed. A marker-triad was applied to a participant (self-selected walking pace): thigh angle (wrt. global-vertical) was calculated. Trials were concurrently recorded using a motion capture system. Root-mean-square error was 2.5°, 1.3° and 2.2° for depth-camera distances of 0.8 m, 1.1 m and 1.4 m respectively. Results indicate that walking distances of 1.1 m are optimal for the current system. Further development and investigation into potential healthcare applications (e.g., low-cost, close-range gait analysis) is warranted.

1. Introduction

In post-stroke patients, impaired cerebral cortex function limits coordinated movement and typically manifests by pronounced asymmetric deficits [1]. The restoration of symmetric gait is a major goal of post-stroke rehabilitation, to allow independent walking and living [2]. Current stroke patient rehabilitation assessments, whether in a clinic or at home, involve subjective, task orientated assessments. Whilst these are useful, visual observation alone cannot sufficiently characterise subtle changes to gait [3]. This is an important consideration as the progress of patient rehabilitation will inform the direction and continuation of physical therapy interventions, particularly if those interventions are home-based. Traditional, marker-based motion capture systems can be prohibitive to healthcare applications from cost, space, time and user-expertise perspectives. In particular, the portability of motion capture systems is a fundamental consideration for practitioners conducting off-site assessments, where a patient’s functional mobility is assessed in their own home. Movement analysis technologies for healthcare applications should therefore aim to be low-cost, portable, simple-to-use and suitable for use in small spaces (e.g., a patient’s home). An objective tool, capable of measuring sensitive gait parameters (e.g., knee joint angle, gait symmetry, etc.), would support home-based assessments of post-stroke gait.
Depth-cameras represent a low-cost alternative to traditional motion capture as calibrated, three-dimensional measurements are captured automatically by a single-camera device (e.g., Figure 1b). The Microsoft Xbox 360 Kinect and Xbox One Kinect are low-cost depth-camera devices, originally developed for use in the home. Both the Xbox 360 and Xbox One Kinect provide a software development kit (SDK) for automated 3D human-pose estimation [4]. Automatic human-pose estimation using Microsoft’s SDK has been assessed in many sport and health applications, including functional movement assessment [5] and gait analysis [6]. Whilst automatic human-pose estimation is an extremely useful tool for movement-based analyses, measurement agreement has limited its use as a surrogate motion capture system [5]. Moreover, whilst joint position estimates contain error [4], the model-based nature of estimation can exacerbate errors for persons deviating from this model (i.e., stature, shape, etc.).
Marker-tracking methods, based on infrared disparity-mapping depth cameras (e.g., Xbox 360 Kinect), have demonstrated promising results [7]. Macpherson and colleagues [7] used an Xbox 360 Kinect to process three-dimensional point cloud data based on four retro-reflective markers, mounted to participant’s posterior pelvic region. When compared to three-dimensional motion capture, Macpherson and colleagues [7] reported 95% limits of agreement as ≤9.9 mm and ≤4.6° for linear and angular motions of the pelvis and trunk during treadmill-based walking and running. Whilst reasonable agreement was demonstrated, limitations associated with processing three-dimensional point cloud data were highlighted as a necessary area of development to improve measurement accuracy.
The advent of infrared time-of-flight depth-cameras (e.g., Xbox One Kinect), which provide higher resolution sensors with increased depth measurement fidelity [8], might yield improved accuracy, addressing previous limitations [7]. Therefore, the use of infrared time-of-flight depth cameras might expand potential applications of depth-cameras for healthcare specialists. Recently, bespoke analysis software, capable of automatic and real-time, three-dimensional position measurement using an infrared time-of-flight depth-camera, has been developed. The aim of this study was to demonstrate a low-cost, portable and simple-to-use system for the measurement of basic gait parameters.

2. Materials and Methods

All procedures were approved by the Research Ethics Committee of the Faculty of Health and Wellbeing, Sheffield Hallam University. A single female participant (age = 24 years; stature = 1.68 m; mass = 53 kg) was recruited and written informed consent was obtained. A retro-reflective marker-triad was secured to the lateral aspect of the participant’s right thigh (e.g., Figure 2b).
Bespoke analysis software, developed using the .NET (C#) framework, automatically measured the three-dimensional position and orientation of the marker triad in real-time. Infrared images captured (~30 Hz) by an Xbox One Kinect (Microsoft, Redmond, USA) are first converted to logical images (based on a threshold), allowing retro-reflective markers to be tracked (e.g., Figure 1a). Depth data—based on the phase-shift distance of modulated infrared light [8]—cannot be derived from marker locations directly as retro-reflective material scatter infrared light. Therefore, image locations of retro-reflective markers (e.g., Figure 1a) are used to track three coplanar image targets (e.g., red, green and blue tracking crosshairs highlighted by ‘L’ shape in Figure 1a) within infrared images. The three-dimensional, camera space coordinates of coplanar targets are then extracted and a marker-triad transformation matrix (comprising a rotation matrix and translation vector) calculated for each received frame. For the purpose of assessment, thigh angle, defined as the vertical axis of the marker triad (e.g., blue vector superimposed on Figure 1a) was calculated.
The participant was asked to walk at a self-selected pace (perpendicular to the depth-camera) from a stationary start, at three close-range (0.8, 1.1 and 1.4 m) and three long-range (2.3, 3.2 and 3.6 m) camera-distances (e.g., Figure 2a). Three trials at each camera-distance were recorded. A six-camera motion capture system (Motion Analysis Corporation, Santa Rosa, CA, USA) was used to concurrently record the three-dimensional position of retro-reflective markers at 150 Hz. Immediately prior to walking trials, the participant was asked to perform a small vertical jump, to allow the temporal alignment of kinematic data derived by both measurement systems. All position data were reported in the laboratory coordinate system (e.g., Figure 2a). Depth-camera data were resampled to 150 Hz using linear interpolation and smoothed using a discretised smoothing spline, based on generalised cross-validation [9]. Thigh angle was reported with respect to the global vertical axis. Agreement was assessed using Bland-Altman 95% limits of agreement (LOA). In the case of heteroscedastic data distribution (i.e., |r2| > 0.1), ratio LOA (dimensionless) was also reported. Further, root-mean square error (RMSE) was assessed using the following:
R M S E = i 1 N ( X i R X i r ) 2
where XiR is the criterion, Xir is the estimate and N is the number of data points.

3. Results

For close-range camera-distances, 95% of depth-camera thigh angle estimates were between ±4.9°. Further, RMSE for thigh angle estimates was <2.5° for close-range camera distances. Thigh angle differences for 0.8 and 1.1 m camera-distances were heteroscedastic; 95% of ratios were between 34.6 and 23.8% of the mean ratio respectively (Table 1).
For long-range camera distances, 95% of depth-camera thigh angle estimates were between ± 9.5°. Further, RMSE for thigh angle estimates was <6.7°. Thigh angle differences for 2.3 and 3.2 m camera distances were heteroscedastic; 95% of ratios were between 54.2 and 60.4% of the mean ratio respectively (Table 1). Figure 3 presents thigh angle (wrt. global vertical axis) time histories for a stand-to-walk trial, measured using Xbox One Kinect and bespoke software (solid red line) and motion capture (dashed blue line).

4. Discussion

The purpose of this study was to demonstrate a low-cost, portable and simple-to-use system for the measurement of basic gait parameters. Using an Xbox One Kinect and bespoke software, thigh angle (wrt. global vertical axis) for a participant walking at a self-selected pace, was automatically captured and analysed in real-time. For long-range camera distances (i.e., 2 m < camera distance < 4 m), systematic thigh angle differences increased with camera distance. This reflects known depth-accuracy limits of depth-cameras [8] as well as decreasing resolution of tracking targets (i.e., markers). However, close-range measurement errors were not related to camera distance; rather, an optimal camera-distance emerged (Table 1). Using a camera-distance of 1.1 m, systematic thigh angle differences were <0.1° and 95% of differences between ± 2.65°. Furthermore, thigh angle RMSE was 1.35°. However, thigh angle differences at a camera distance of 1.1 m did exhibit heteroscedasticity. This potentially highlights a limitation of a relatively low but also variable sample rate (e.g., ~30 Hz). For example, Figure 3 indicates that thigh angle was overestimated at ~2.1 s. It is reasonable to assume that thigh angle magnitude at this instant (indicated by the criterion) was not beyond the system’s measurement range. This therefore might indicate sample rate as a potential contributor to thigh angle overestimation in this case.
Potential clinical applications of a portable, close-range, three-dimensional measurement system might include home-based assessments of gait. Figure 4 illustrates equipment used in the current study. Clinicians have highlighted that such a tool would help supplement their assessment of stroke survivors’ walking gait and provision of rehabilitative care. However, whilst three-dimensional measurements are derived by the current system, thigh angle—a predominantly sagittal plane motion—was assessed as a demonstra. The calculation of more complex kinematic measures, such as functionally defined joint angles, would aid clinician’s assessment of patients. Future development should therefore consider functional joint angle definitions as well as potential effects of multi-planar triad rotations.

5. Conclusions

A portable, simple-to-use and low-cost depth-camera system automatically measured thigh angle during walking gait. Future development toward the measurement of functionally defined joint angles is warranted. The use of a simple, portable and close-range gait analysis system might support clinicians in home-based assessments of post-stroke rehabilitation patients.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Belda-Lois, J.-M.; Horno, S.; Bermejo-Bosch, I.; Moreno, J.; Pons, J.; Farina, D.; Iosa, M.; Tamburella, F.; Ramos, A.; Caria, A.; et al. Rehabilitation of gait after stroke: A review towards a top-down approach. J. Neuroeng. Rehabil. 2011, 8, 16–85. [Google Scholar] [CrossRef] [PubMed]
  2. Lindquist, A.; Prado, C.; Barros, R.; Mattioli, R.; Lobo da Costa, P.; Salvini, T. Gait training combining partial body-weight support, a treadmill, and functional electrical stimulation: Effects on poststroke gait. Phys. Ther. 2007, 87, 1144–1154. [Google Scholar] [CrossRef] [PubMed]
  3. Beauchet, O.; Annweiler, C.; Allali, G.; Berrut, G.; Herrmann, F.; Dubost, V. Recurrent falls and dual task-related decrease in walking speed: Is there a relationship? J. Am. Geriatr. Soc. 2008, 56, 1265–1269. [Google Scholar] [CrossRef] [PubMed]
  4. Shotton, J.; Fitzgibbon, A.; Cook, M.; Sharp, T.; Finocchio, M.; Moore, R.; Kipman, A.; Blake, A. Real-time human pose recognition in parts from single depth images. In Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, Colorado Springs, CO, USA, 20–25 June 2011; pp. 1297–1304. [Google Scholar]
  5. Eltoukhy, M.; Kelly, A.; Kim, C.Y.; Jun, H.P.; Campbell, R.; Kuenze, C. Validation of the Microsoft Kinect® camera system for measurement of lower extremity jump landing and squatting kinematics. Sports Biomech. 2016, 15, 89–102. [Google Scholar] [CrossRef] [PubMed]
  6. Geerse, D.J.; Coolen, B.H.; Roerdink, M. Kinematic validation of a multi-Kinect v2 instrumented 10-meter walkway for quantitative gait assessments. PLoS ONE 2015, 10, 1–15. [Google Scholar] [CrossRef] [PubMed]
  7. Macpherson, T.W.; Taylor, J.; McBain, T.; Waston, M.; Spears, I. Real-time measurement of pelvis and trunk kinematics during treadmill locomotion using a low-cost depth-sensing camera: A concurrent validity study. J. Biomech. 2015, 49, 464–478. [Google Scholar] [CrossRef] [PubMed]
  8. Wang, Q.; Kurillo, G.; Ofli, F.; Bajcsy, R. Evaluation of pose tracking accuracy in the first and second generations of Microsoft Kinect. In Proceedings of the IEEE International Conference on Healthcare Informatics, Dallas, TX, USA, 4–7 June 2015; pp. 380–389. [Google Scholar]
  9. Garcia, D. Robust smoothing of gridded data in one and higher dimensions with missing values. Comput. Stat. Data Anal. 2010, 54, 1167–1178. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) Infrared image allows retro-reflective marker tracking due to pixel saturation (tracked markers highlighted by red, green and blue circles); (b) corresponding three-dimensional data.
Figure 1. (a) Infrared image allows retro-reflective marker tracking due to pixel saturation (tracked markers highlighted by red, green and blue circles); (b) corresponding three-dimensional data.
Proceedings 02 00502 g001
Figure 2. (a) Schematic of laboratory setup; (b) marker-triad configuration when applied to participant.
Figure 2. (a) Schematic of laboratory setup; (b) marker-triad configuration when applied to participant.
Proceedings 02 00502 g002
Figure 3. Thigh angle during a stand-to-walk trial measured by depth-camera (solid red line) and motion capture (dashed blue line) for self-paced walking at 1.1 m camera distance.
Figure 3. Thigh angle during a stand-to-walk trial measured by depth-camera (solid red line) and motion capture (dashed blue line) for self-paced walking at 1.1 m camera distance.
Proceedings 02 00502 g003
Figure 4. Low-cost, automatic and simple-to-use system for home-based gait analysis. System comprises a Microsoft Xbox One Kinect, laptop (analysis software) and retro-reflective marker triad.
Figure 4. Low-cost, automatic and simple-to-use system for home-based gait analysis. System comprises a Microsoft Xbox One Kinect, laptop (analysis software) and retro-reflective marker triad.
Proceedings 02 00502 g004
Table 1. LOA (absolute and ratio) and RMSE for thigh angle at six depth-camera distances.
Table 1. LOA (absolute and ratio) and RMSE for thigh angle at six depth-camera distances.
Camera Distance (m)Absolute LOA (°)|r2|Ratio LOARMSE (°)
Close-range0.80.07 ± 4.890.140.99 (×/÷1.52)2.49
1.1−0.02 ± 2.650.131.00 (×/÷1.32)1.35
1.40.48 ± 4.440.04-2.31
Long-range2.3−1.6 ± 9.460.260.94 (×/÷2.05)5.09
3.2−2.41 ± 9.240.140.92 (×/÷1.32)5.29
3.6−5.16 ± 8.320.05-6.68
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dunn, M.; Pavan, D.; Ramirez, P.; Rava, S.; Sharin, A. An Automated Method to Extract Three-Dimensional Position Data Using an Infrared Time-of-Flight Camera. Proceedings 2018, 2, 502. https://doi.org/10.3390/proceedings2060502

AMA Style

Dunn M, Pavan D, Ramirez P, Rava S, Sharin A. An Automated Method to Extract Three-Dimensional Position Data Using an Infrared Time-of-Flight Camera. Proceedings. 2018; 2(6):502. https://doi.org/10.3390/proceedings2060502

Chicago/Turabian Style

Dunn, Marcus, Davide Pavan, Paul Ramirez, Silvia Rava, and Atiqah Sharin. 2018. "An Automated Method to Extract Three-Dimensional Position Data Using an Infrared Time-of-Flight Camera" Proceedings 2, no. 6: 502. https://doi.org/10.3390/proceedings2060502

APA Style

Dunn, M., Pavan, D., Ramirez, P., Rava, S., & Sharin, A. (2018). An Automated Method to Extract Three-Dimensional Position Data Using an Infrared Time-of-Flight Camera. Proceedings, 2(6), 502. https://doi.org/10.3390/proceedings2060502

Article Metrics

Back to TopTop