Next Article in Journal
Derivation of All Attitude Error Governing Equations for Attitude Filtering and Control
Next Article in Special Issue
Sport-Related Human Activity Detection and Recognition Using a Smartwatch
Previous Article in Journal
Detection of Participation and Training Task Difficulty Applied to the Multi-Sensor Systems of Rehabilitation Robots
Previous Article in Special Issue
Practical and Durable Flexible Strain Sensors Based on Conductive Carbon Black and Silicone Blends for Large Scale Motion Monitoring Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Model-Based System for Real-Time Articulated Hand Tracking Using a Simple Data Glove and a Depth Camera

1
Beijing Key Laboratory of Network System Architecture and Convergence, School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
Beijing Laboratory of Advanced Information Networks, Beijing 100876, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2019, 19(21), 4680; https://doi.org/10.3390/s19214680
Submission received: 29 September 2019 / Revised: 22 October 2019 / Accepted: 25 October 2019 / Published: 28 October 2019
(This article belongs to the Special Issue Multi-Modal Sensors for Human Behavior Monitoring)

Abstract

:
Tracking detailed hand motion is a fundamental research topic in the area of human-computer interaction (HCI) and has been widely studied for decades. Existing solutions with single-model inputs either require tedious calibration, are expensive or lack sufficient robustness and accuracy due to occlusions. In this study, we present a real-time system to reconstruct the exact hand motion by iteratively fitting a triangular mesh model to the absolute measurement of hand from a depth camera under the robust restriction of a simple data glove. We redefine and simplify the function of the data glove to lighten its limitations, i.e., tedious calibration, cumbersome equipment, and hampering movement and keep our system lightweight. For accurate hand tracking, we introduce a new set of degrees of freedom (DoFs), a shape adjustment term for personalizing the triangular mesh model, and an adaptive collision term to prevent self-intersection. For efficiency, we extract a strong pose-space prior to the data glove to narrow the pose searching space. We also present a simplified approach for computing tracking correspondences without the loss of accuracy to reduce computation cost. Quantitative experiments show the comparable or increased accuracy of our system over the state-of-the-art with about 40% improvement in robustness. Besides, our system runs independent of Graphic Processing Unit (GPU) and reaches 40 frames per second (FPS) at about 25% Central Processing Unit (CPU) usage.

1. Introduction

The articulated hand tracking has been studied for decades, due to its wide range of applications, like computer graphics, animation, human-computer interaction, rehabilitation, and robotics. Nowadays, with the boom of virtual reality (VR) and augmented reality (AR), more natural interaction with the digital world is desired to increase the sense of presence and immersion. Fully articulated hand tracking holds the potential to become a first-class input mechanism [1]. Recent works have put more attention on the task of fully articulated hand tracking, aiming to recover the detailed motion of a user’s hands in real-time. However, tracking detailed hand motion is still a challenge, facing many factors like large variations in hand shapes, small hand size, viewpoint changes, many degrees of freedom (DoFs), fast movement, self-similarity, and occlusions [2].
Judging by the input devices, we can broadly categorize existing works into the wearable-based and the camera-based. For the wearable-based works, the data gloves that can record hand pose stably and directly, are the most representative. When it comes to those camera-based researches, an approach to recover hand pose is using discriminative (appearance-based) methods [3]. These methods search the correspondence between input images and pose parameters from a large amount of data using deep-learning. Another approach is the use of generative (model-based) methods [4,5,6] that apply iterative model-fitting optimization to camera images. Due to their heavy dependence upon the proper initialization, recent works [7,8] utilize the appearance-based methods to provide initialization and re-initialization. Apart from the wearable-based and camera-based methods, some researchers [9,10] try to make improvements through multi-model input devices.
Although the methods above have made many significant contributions, there is still a long way for fully articulated hand tracking to become the first choice of the user interface. The following are several factors. For data-glove based systems, a complex, tedious, non-automatic calibration is inevitable to reduce inaccuracies, especially for applications that require high levels of accuracy [11]. Those expensive commercial gloves, i.e., 5DT Data Glove [12], Cyber Glove [13], are always unaffordable. They also add on additional hardware which may cause discomfort to the user for prolonged use. For camera-based works, discriminative methods still generalize poorly to unseen hand shapes [14,15] because existing hand data-sets have many shortcomings, e.g., low variation in hand shape, annotation accuracy, and limited in scale. Most generative methods are computationally expensive and a top-end consumer GPU is essential for heavy parallelization to achieve real-time performance. The inherent issue of visual occlusions also makes those camera-based works still lack sufficient accuracy and robustness for fine manipulation, like surgical applications. As for those multi-model researches, their systems are always burdensome because of their poor integration of various inputs. All those factors mean there is still room to make further improvement in flexibility, accuracy, robustness, efficiency, and cost.
In this paper, we present a system with the synergy of a simple data glove and a depth camera to recover articulated hand motion by iterative model-based registration progress. We aim to achieve more accurate and robust tracking results and overcome some downsides of existing works, e.g., tedious calibration or high dependency on GPU. The contributions of this paper are as follows:
  • Design and implementation of a multi-model articulated hand tracking system that runs in real-time without GPU and improves by about 40% of robustness with comparable or increased accuracy over the state-of-the-art.
  • The re-definition and simplification on the function of the data glove as an approximate initialization in Section 3.2.2 and a strong pose-space regularization in Section 3.3.3 that increases the robustness of our system and frees our data glove from tedious calibration and heavily hampering hand movement.
  • A new proposal for DoFs setting that can avoid potential artificial error in the kinematic chain; see Section 3.1.
  • The new fitting terms with a simplified approach for computing tracking correspondences that can reduce computational cost without the loss of accuracy; see Section 3.3.1.
  • A new strategy to consider the shape adjustment of the triangular mesh hand model that includes a tailored shape integration term in Section 3.3.2 for better fitting the input images and an adoptive collision prior consistent with the shape adjustment in Section 3.3.3 to prevent self-intersection and produce plausible hand poses.
The following structure of the paper is: We survey related works in Section 2. In Section 3, we introduce the components of our system, including the hand model, the setting of the kinematic chain and DoFs, data acquirement and processing, and the objective function. In Section 4, we quantitatively and qualitatively analyze the performance of our hand tracking system and provide comparisons with the state-of-the-art. We conclude in Section 5 with a discussion of future works.

2. Related Works

In this section, we put our focus on the relevant introduction of the two mainstream camera-based works, i.e., appearance-based methods and model-based methods. We also briefly introduce some multi-model methods. Dipietro et al. [11] have done elaborate research for all kinds of data gloves and relevant applications. We refer the readers to [11] for a detailed review of glove-based works.

2.1. Appearance-Based Methods

Appearance-based methods train a classifier or a regressor to map image features to hand poses. Nearest neighbor search and decision trees [16,17] are widely used in early works. In recent years, convolutional neural network (CNN)-based discriminative methods [3,18,19,20,21,22] are state-of-the-art which estimate 3D joint positions directly from depth images. Besides, kinematics and geometric constraints are considered to avoid joint estimations violating kinematic constraints. Malik et al. [2] embedded a novel hand pose and shape layer inside CNN to produce not only 3D joint positions but also hand mesh information. For a more comprehensive analysis and investigation of the state-of-the-art along-with future challenges, we refer the readers to [15]. The biggest limitation of appearance-based methods is the training data. Existing benchmarks [17,23,24,25,26] are not perfect enough to ensure well generalize to unseen hand shapes. We refer to [14] for a detailed analysis of the drawbacks of existing data-sets. Considering this limitation, our system follows the model-based approaches that do not rely on massive data-sets.

2.2. Model-Based Methods

Despite the considerable advance in learning-based hand tracking, systems that employ generative models of explicit hand kinematics and surface geometry and fit these models to depth data using local optimization have produced the most compelling results [1]. The most common problems for model-based methods are a good enough initialization point, an expressive enough hand model and a discriminative object function that minimizes the error between the 3D hand model and the observed data.

2.2.1. Initialization

A good enough Initialization has been proven critical to the robustness [23], which enables faster converge and better resistant to local optima. There exist many initialization methods. Some works [5,23,27] were initialized by the fingertip detection. Besides, Tagliasacchi et al. [27] and Tkach et al. [5] also detected a color wristband as a first alignment. The use of simple geometric heuristics for initialization can sometimes be impractical for those gestures which contain occlusions or difficult hand orientations. For this reason, most of the previous studies concentrated on exploiting the given image data with the train-based methods. Taylor et al. [28] generated candidate’s hand poses quickly by a retrieval forest [29]. Taylor et al. [1,30] trained a decision forest classifier on a synthetic training set to generate an initial pose estimate. Sanchez-Riera et al. [7] trained a convolutional neural network for initialization with 243,000 tuples of images. Sharp et al. [6] inferred a hierarchical distribution over hand pose with a layered discriminative model. However, initialization errors often occur due to imperfect training data-sets, mentioned in Section 2.1, which may cause tracking failure. In our system, it is more reliable and robust to provide an approximate initialization by a simple data glove.

2.2.2. Hand Model

The human hand model serves as the medium of computation and the presentation of algorithm results. A detailed and accurate generative model tends to deepen the good local minima and widen their basins of convergence [1]. Many hand models have been proposed, see Figure 1.
Early works [27,31,32,33] used the capsule mode made by two basic geometric primitives: a sphere and a cylinder. Qian et al. [23] built the hand model using a number of spheres. Melax et al. [34] used a union of convex bodies for hand tracking. Sridhar et al. [35] modeled the volumetric extent of the hand as a 3D sum of an-isotropic Gaussian model. These approaches can model a broad spectrum of hand shape variations and enable fast evaluation of distances and a high degree of computational parallelism. However, they only roughly approximate hand shape even if Tkach et al. [5] proposed the use of sphere-meshes as a novel geometric representation. An alternative is a triangulated mesh model [4,6,7,8,28,30,36,38] with linear blend skinning (LBS) that is more realistic and fits image data better. But these triangulated meshes cost more computational effort and are hard to deal with the collision. There also exist some implicit templates except these explicit models. Schmidt et al. [37] voxelized each shape-primitive and computed a signed distance function for the local coordinate frame. Taylor et al. [1] constructed the hand as an articulated signed distance function that allows fast calculation of the distance to the hand surface. To explain the input data better and explicitly visualize the tracking result, our system uses an expressive triangular mesh hand model.
Apart from modeling the hand model detailed and realistic, hand model personalizing is a core ingredient in model-based methods. Joseph Tan et al. [38] quantitatively demonstrated for the first time that detailed personalized models improve the accuracy of hand tracking. In some works [7,27], only the simple uniform scaling of the model was considered. Makris et al. [33] investigated the calibration of a cylinder model through particle swarm optimization. Ballan et al. [4,8] reconstructed a personalized template mesh offline with a multi-view stereo method for more detail model calibration. Taylor et al. [30] presented a fast, practical method to acquire detailed hand models from as few as 15 frames of depth data. Moreover, their work was extended in [36], which simplifies hand shape variation by linear shape-spaces. Joseph Tan et al. [38] went further and presented a fast, practical method for personalizing a hand shape basis to an individual user’s detailed hand shape using only a small set of depth images. Similarly, Remelli et al. [39] presented a robust algorithm for personalizing a sphere-mesh tracking model to a user from a collection of depth measurements. However, those methods suffer a major drawback: the template must be created during a controlled calibration stage, where the hand is scanned in several static poses (i.e., offline). Tkach et al. [40] yielded a fully automatic, real-time hand tracking system that jointly estimates pose and shape for their sphere-mesh hand model. It remains unsolved for the triangular mesh model to adjust its shape during tracking. Our system tries to adapt the conclusion in [40] for the shape adjustment of our triangular mesh model to fit the shape of input images.

2.2.3. Objective Function

The objective function measures the discrepancy between the hand model and input depth, as well as the validity of the hand model [23]. In general, the objective function is made by fitting terms and prior terms.
Fitting terms measure how well the hand parameters explain the input frames. Oikonomidis et al. [31] and Sharp et al. [6] formed the fitting terms as the discrepancy between the observed images and the rendered images from a given hand pose hypothesis. Then solutions were searched by a slow-converge stochastic optimizer like PSO. Most works [5,7,23,27,28,30,32,36] modeled the fitting terms as the least-squares error between the effectors and their target positions, and solved it by gradient-based approaches, such as Gauss–Newton and Levenberg–Marquardt. However, finding corresponding points is difficult and time-consuming, especially for triangular hand models. Taylor et al. [28,30,36,38] subdivided the mesh to produce a smooth surface function for evaluating both pose and corresponding points. Down-sampling the point cloud randomly is also a good way to reduce computation effort [23,28]. Except for the 3D fitting term, a 2D registration in most works [4,5,23,27,28,39,40] is also important, which pushes the hand model to lie within the visual sensor hull. Our system also adopts the 3D and 2D fitting terms but deals with the time-consuming problem of computing corresponding points to reduce the computational cost and enable our system independent of GPU.
Prior terms regularize the solution to produce realistic hand poses. Every model-based works adopt a joint limitation term extracted from a database to constrain the posture parameters within plausible value ranges. Self-intersection is a big problem. Oikonomidis et al. [31] penalized abduction–adduction angles of adjacent fingers. Some works [5,27,39,40] restricted the distance between cylinders to solve this problem. Qian et al. [23] limited the distance of spheres in neighboring fingers. For triangular hand models, it becomes harder. In [4], a repulsion term was computed in the form of a 3D-3D correspondence that pushes the vertex back. Taylor et al. [28] defined a set of spheres that approximate the volume of the fingers to simplify this question. However, these collision terms did not consider the shape adjustment during tracking. Our system gets this point and introduces an adaptive collision term consistent with shape adjustment. Besides, the pose-space prior in [5,27,28,38], obtained by performing dimension reduction on the training data, provides implicitly constrains to the recovered hand postures. In general, the pose-space prior covers a large pose space and can not constrain the hand pose tightly. Our system integrates a data glove in it to produce a stronger implicit restriction. There are also lots of prior terms, including temporal priors to prevent the tracked hand to jitter [5,27,28], ARAP regularization to penalized large shape deformations [30,36], fingertips prior term to guarantee each detected fingertip should have a model fingertip nearby [28].

2.3. Multi Model Systems

For multi-model systems, the core idea is that different input models each have their limitations, but may complement each other. For example, wearable-based systems can fill in the data gap that occurs with vision-based systems during camera occlusions, and the vision-based device provides an absolute measurement of hand state [9]. Arkenbout et al. [9] integrated the hand pose of 5DT date glove and the Nimble VR system through a Kalman filter and shows substantial improvement in accuracy. Ponraj et al. [10] increased the accuracy of fingertips tracking in occluded cases by combining the leap motion control with a Sensorized Glove. Tannous et al. [41] proposed a fusion scheme between inertial and visual motion capture sensors to improve the estimation accuracy of knee joint angles. Sun et al. [42] reached a higher gesture recognition rate using the Kinect and Electromyogram signals. Pacchierotti et al. [43] placed a novel wearable cutaneous device on the proximal phalanx to improve the tracking of the fingertips on commercially-available tracking systems, such as the Leap Motion controller or the Kinect sensor. These methods show improvement through multi-model inputs. These multi-model systems show improvement in accuracy and robustness, but they retain the cumbersome setting of wearable devices, e.g., calibration and uncomfortable hardware, which makes their system more burdensome. Our system can reduce the effect of additional hardware by re-defining and simplifying the function of the data glove to make our system lightweight.

3. Method

The overview of our system can be found in Figure 2. In this section, we will introduce our system in detail. We first present the hand model and the setting of the kinematic chain before tracking in Section 3.1. Then, we introduce step 1 during tracking, i.e., how we acquire and process the input data from the depth camera and the simple data glove, in Section 3.2. Finally, we describe in detail the step 2 during tracking, i.e., the objective function constructing, in Section 3.3. The iterative optimization step by Levenberg–Marquardt approach is not included in this Section because it is a very common gradient-based solution.

3.1. Hand Model

We use the publicly available MANO hand model [44] for pose and shape tracking, see Figure 3a,b. There are several reasons: (1) it is learned from around 1000 high-resolution 3D scans of hands of 31 subjects in a wide variety of hand poses. (2) As an articulated triangular model, it is more expressive than those hand models [1,2,4,7,8,14,15] made by basic geometric primitives. It also reduces the artifacts of LBS, i.e., mesh collapse around joints. (3) Romero et al. [44] provide not only a set of shape offset vectors to generate different hand shapes but also the sparse linear joints locations regressor to generate corresponding hand skeleton, which is important for shape tracking.
The general formulation M β , θ of MANO, taken from the original paper [44] for completeness, is as follows:
M β , θ = W T β , θ , J β , θ , W
where β and θ control the shape and pose respectively, W is a linear blend skinning function applied to a template hand triangulated mesh T, the hand template T is obtained by deforming a mean mesh through β and θ , J is a sparse linear joints locations regressor learned from mesh vertices, and W is the blend weights. For more details about MANO, please refer to [44].
Given the MANO hand joints, we build our kinematic chain, Figure 3c.
The kinematic chain is not fixed because the shape parameters θ of MANO will affect the locations of the joints. Each joint J k of the kinematic chain, except the wrist joint, is defined with its previous joints J k p a r e n t . Moreover, each joint J k is associated with an orthogonal frame T ¯ k according to which local transformations are specified. We automatically set the local coordinate systems T ¯ k according to the relationship and structure of joints in the mean shape of the hand ( β is set to 0). For the sake of simplicity, these preset local coordinate systems T ¯ k do not change with the location change of the joints, considering the relatively small effect of β on hand structure. Incorrectly specified kinematic frames can be highly detrimental to tracking quality [5]. Relaxing the restrictions on DoFs helps reduce the impact of incorrectly setting frames, see Figure 4b.
So each joint J k in our kinematic chain has three DoFs, which results in 51 DoFs in total with the six DoFs global transformation. However, more DoFs mean larger search space, which is dealt with in Section 3.3.

3.2. Data Acquisition and Processing

3.2.1. Camera Data

We get the input of depth images from Intel RealSense SR300 (Intel, Santa Clara, CA, USA), a consumer short-range RGBD camera. From the raw depth images, we obtain the segment of hand by performing classification with standard random decision forests [25]. A 2D silhouette image S s is extracted directly from the segment. Besides, we extract also a 3D point cloud X s that contains about 200 pixels by performing a stochastic sampling, which has been proved by [28] to enable CPU optimization without significant loss of precision.

3.2.2. Glove Data

We measure the approximate pose of the hand, using a simple and cheap prototype data glove provided by our cooperation company [45], Figure 5a,b. The glove is affixed with 11 inertial measurement units (IMUs) according to the anatomical structure of the hand. Figure 5c shows the location, initial coordinate systems, and orientation of those IMUs.
The output of our glove is a 100 Hz stream of the orientation of IMUs in the form of unit quaternions:
Q i = q i x , q i y , q i z , q i w , i 1 11 ,
where i is the index of IMUs in Figure 5c. Q 11 records the global rotation information of the hand. Q 1 and Q 2 represent the movements of the distal and proximal phalange of the thumb, respectively. The rest quaternions represent the movements of the medial and proximal phalanges of other fingers. Theoretically, all IMUs’ orientation is set to be the same in the initial pose, see Figure 5c and Figure 6a, which means initially:
Q i i n i t i a l = Q j i n i t i a l i , j 1 , 2 , , 11 .
Based on this premise, we can easily measure the angular values of joints, using the relative rotation between IMUs:
Q i , j = Q i Q j 1 ,
where i represents the index of IMUs, j represents the index of the parent IMU of the ith IMU, Q i , j is also a unit quaternion that contains the rotation information of the joint between the two phalanges where the ith and jth IMU locate in. See Figure 6 for example.
Then we convert the quaternion Q i , j to the Euler angle in XYZ rotation order and map to the 51 DoFs pose parameters mentioned in Section 3.1 to provide an initial hand pose:
θ i j x = t a n 1 2 p i j w p i j x + p i j y p i j z 1 2 p i j x 2 + p i j y 2 θ i j y = s i n 1 2 p i j w p i j y p i j x p i j z θ i j z = t a n 1 2 p i j w p i j z + p i j x p i j y 1 2 p i j y 2 + p i j z 2 .
The movements of the metacarpal phalanx of the thumb and the distal phalanx of other fingers are not considered independently but calculated according to the proximal phalanx of the thumb and the medial phalanx of other fingers respectively.
Our simple data glove only provides approximate poses, see Figure 7, because the fixed positions of IMUs may not fit everyone, and shifting of IMUs may appear when incorrect wearing the glove or making various gestures.
We redefine and simplify the function of the data glove to be a robust initialization, re-initialization, and a strong prior. So we think the approximate pose from our simple data glove is acceptable and do not design a complex calibration progress for it to keep our system light-weighted and easy to use.

3.3. Objective Function

Given the 2D silhouette image S s , 3D point cloud X s , and initialization form data glove, we aim to find the pose θ and shape β parameters that make our hand model a good explanation of the absolute measurement from the camera accurately and efficiently. We formulate this goal as a minimum on the following objective function with references to recent works:
min θ , β E 3 D + E 2 D F i t t i n g t e r m s + E s h a p e + E p o s e s p a c e + E t e m p o r a l + E c o l l i s i o n + E b o u n d s P r i o r t e r m s ,
where fitting terms determine the pose θ and shape β parameters in each frame, E s h a p e elegantly integrates shape information β from frames in different poses, those prior terms regularize the solution to ensure the recovered pose is plausible. We will put our focus on the novelties to meet our premise on efficiency, accuracy, and robustness, and give only a brief introduction on the unchanged terms.

3.3.1. Fitting terms

3D Registration. The E 3 D registers the hand model M to the point cloud X s . In recent works, it is always formulated in the spirit of ICP as:
E 3 D = ω 3 D x X s x Π M ( x ) 2 2
where x represents a 3D point of the point cloud X s , Π M ( x ) is the corresponding point of x on the hand model M .
Finding the corresponding points is the most critical, time-consuming, and challenging step, especially for triangular mesh models. In order to improve efficiency and reduce the computational cost, we simplify this process and re-formulate the 3D registration without loss of accuracy as:
E 3 D = ω 3 D x X s x Π V ^ ( x , θ , β ) 2 2
where the difference is that we directly search the corresponding point Π V ^ ( x , θ , β ) among the visible vertices V ^ (about half of the original 778 vertices) for each 3D point x X s . The reasons behind are two-fold:
  • The MANO hand model is more detailed than the hand models in [4,6,7,8,28,30,36,38] even though it is made by only 778 vertices and 1554 triangular faces. We also try to produce a more detailed model by the Loop subdivision [36], see Figure 8. The result shows no significant improvement, which enables the vertex v V account for these corresponding points around it.
  • We conducted another way to find more detailed corresponding points on the hand model M using loop subdivision and compare the performance with our simplified approach, see Figure 9. The result shows that our approach will not decrease accuracy.
With the down-sampled point cloud X s , this approach dramatically reduces the amount of calculation and allows us to search all the corresponding points for the point cloud X s in less than 1 ms on CPU with a single thread.
2D Registration. The E 2 D provides the supplementary registration that the point cloud alignment does not take into account. Ganapathi et al. [46] show that the depth map provides evidence not only for the existence of a model surface in the form of point cloud X s but also for the non-existence of surface between the point cloud X s and the camera. Thus, the E 2 D , also called “free space” constraint, is non-trivial in registration. We adapt the formula mentioned in [27] for this “free space” constraint as:
E 2 D = ω 2 D p S r ( θ , β ) p Π S s ( p ) 2 2 ,
where p is a 2D point of the projection S r ( θ , β ) of the visible vertices V ^ with pose θ and shape β , which means we need not rasterizing the model and significantly decrease the run-time complexity. Π S s and Π S s ( p ) remain unchanged and denote the image-space distance transform [46] and the closest 2D point in the input 2D silhouette S s , respectively.

3.3.2. Shape Integration

To explain the full input image well, we should take shape β into account as well, which has been shown to have a considerable impact on the accuracy [5]. Shape information β is so weakly constrained in any given frame that sufficient information must be gathered from different frames capturing different hand poses [40]. We refer to [40] to integrate shape information β from frames in different poses elegantly. Tkach et al. [40] do similar work on their 112 manually designed explicit shape parameters of the sphere-mesh model, i.e., the length of fingers and the radius of circles. However, our shape parameters β provided by [44] are the top 10 shape principal components analysis (PCA) components that control the shape of the hand model implicitly. In order to adapt the method in [40] to our triangular hand model, we make the following feasibility analysis:
We imitate [40] to abstract the hand shape/pose estimation problem from a single frame into one of a simpler 2D stick-figure, Figure 10.
The co-variance of shape parameters is derived from the Hessian Matrix of the registration energies. The co-variance represents the confidence or uncertainty of shape parameters. Figure 10 shows a similar conclusion in [40] that the co-variance of the implicit shape parameters is also conditional on the pose of the current frame, and the co-variance decreases with a bent finger. Besides, the implicit shape parameters can produce better convergence results even when the blend angle is small. This analysis means that we can easily adapt the joint cumulative regression scheme in [40] without changing the form:
E shape = ω s h a p e Σ ^ n 1 1 / 2 ( β n β n 1 ^ ) 2 2 ,
and update the shape parameters β n and the cumulative co-variance Σ ^ n in the nth frame using the kalman-like style mentioned in [40]:
β ^ n = Σ n ( Σ ^ n 1 + Σ n ) 1 β ^ n 1 + Σ ^ n 1 ( Σ ^ n 1 + Σ n ) 1 β n Σ ^ n = ( Σ ^ n 1 1 + Σ n 1 ) 1 .
For more details about the kalman-like shape integration, we refer the readers to [40].

3.3.3. Prior Terms

Pose Space prior. The E p o s e s p a c e is a data-driven prior that limits the pose space (except the six global components of θ ) within a reasonable range. We extend the pose space prior in [27] with the restriction of data glove to narrow the pose searching space.
We first construct the standard pose space prior in [27] on the publicly available database of [44] by PCA, see Figure 11a, which results in a 45 × 45 matrix V of eigen-vectors and a set of 45 eigenvalues λ = λ 1 , , λ 45 . Taking the top N pose components, we have:
θ ˜ = C T ( θ μ ) θ ˜ N ( 0 , Σ ) ,
where μ is the 45-dimensional mean pose, Σ is a diagonal matrix that contains the variance of the PCA basis, C is a 45 × N matrix made by the top N eigen-vectors in V corresponding to the N largest eigenvalues. Tagliasacchi et al. [27] hold that the estimated pose should lie in this data-driven space to take on reasonable poses.
Then we take the input of data glove θ g l o v e into account. We think that we should also search the exact pose around θ g l o v e . Thus, we project the input glove data θ g l o v e recorded when we recover the hand pose on the database of [44] into θ ˜ g l o v e with the same 45 × N matrix C, see Figure 11b. Moreover, we build a multivariate Gaussian model for the difference between the glove data θ ˜ g l o v e and corresponding ground truth θ ˜ in the low-dimensional subspace, see Figure 11c:
θ ˜ d i f f = θ ˜ g l o v e θ ˜ θ ˜ d i f f N ( μ d i f f , Σ d i f f ) Σ d i f f = Σ d i f f + Σ n o i s e ,
where μ d i f f is the N-dimensional mean difference, the Σ d i f f is co-variance matrix and Σ n o i s e represents the noise of data glove.
So given a specific input of data glove θ g l o v e i , we search the ground-truth pose in the distribution of both N ( 0 , Σ ) and N ( θ ˜ g l o v e i μ d i f f , Σ d i f f ) . We thus merge the two distribution, see Figure 12:
N ( μ m e r g e , Σ m e r g e ) = N ( 0 , Σ ) N ( θ ˜ g l o v e i μ d i f f , Σ d i f f ) .
We rewrite the PCA prior as:
E p o s e s p a c e = ω p o s e s p a c e ( C T θ μ m e r g e ) T Σ m e r g e 1 ( C T θ μ m e r g e )
After Cholesky decomposition on Σ m e r g e 1 = L L T , we convert the PCA prior to a squared form:
E p o s e s p a c e = ω p o s e s p a c e L ( C T θ μ m e r g e ) 2 2
Temporal prior. In order to mitigate jitter of hand on time series, we adopt a very efficient and effective prior from [27]. We build a set K that contains 50 randomly selected vertices from the hand model, and penalize their velocity and acceleration the same as [27]:
E t e m p o r a l = ω t e m p o r a l ( v i K v i v i 1 2 2 + v i K v i 2 v i 1 + v i 2 2 2 ) .
Collision prior. The E c o l l i s i o n is used to avoid self-intersections of fingers. Unlike the hand model made by simple geometry, the MANO hand model is a triangular mesh model. It is a difficult problem to judge whether self-crossing occurs with little cost. To solve this problem, we follow the ideas in [28] and approximate the volume of the fingers in the MANO hand model with a set of spheres, see the right picture in Figure 13.
In order to be consistent with the pose and shape deformation of the hand model, we let those spheres have radius r s θ , β and locations c s θ , β specified by the vertices of hand model with pose θ and shape β . For each phalanx, we automatically set four spheres, one is root and the other three from interpolation:
c r o o t = J θ , β r r o o t = m i n d i , J θ , β i V J c i i n t e r p o l a t i o n = J θ , β + i J J c h i l d N + 1 r i i n t e r p o l a t i o n = r c h i l d r o o t + N + 1 i N + 1 ( r r o o t r c h i l d r o o t ) ,
where for the root spheres, J is the regressor of joints, J θ , β is the location of one joint, V J is the top 20 vertices that have the greatest impact on the regression of this joint in the regressor J, d i , J θ , β represents the Euclidean distance between the ith vertex in V J and the joint J θ , β ; for the interpolated spheres, N represents the number of spheres you want to interpolate, i 1 , , N is the index of the interpolated sphere, J c h i l d is position of the child joint of J according the kinematic tree. See Figure 13 for details.
Given the set of spheres, we penalize self-intersection with:
E c o l l i s i o n = ω c o l l i s i o n i , j X i , j p i p j 2 2
where p i and p j are two point on the ith sphere and jth sphere and play the role as original point and target point respectively, X i , j indicates whether collision between ith sphere and jth sphere happened:
X i , j = 0 o t h e r w i s e 1 c i , c j 2 < r i + r j 2
Joints bounds prior. To prevent the hand from reaching an impossible posture by over-bending the joints, we limit the angles of the hand model and adopt the same function in [27]:
E b o u n d s = ω b o u n d s θ i θ X ̲ i θ i θ i ̲ 2 + X ¯ i ( θ i θ i ¯ 2 ,
where each hand joint is associated with limitations θ i ̲ , θ i ¯ . Because of our different settings of DoFs, we extract the limitations for each DOF from the detailed hand database in [44]. The extracted limitations can be see as Table 1. X ̲ i and X ¯ i are indicator functions:
X ̲ i = 0 o t h e r w i s e 1 θ i < θ i ̲ X ¯ i = 0 o t h e r w i s e 1 θ i > θ i ¯

4. Experiments and Discussion

In this section, we evaluate our system in many aspects, e.g., the robustness to noise and occlusion, the accuracy on various poses and shapes, and the improvement over the-state-of-art. The quantitatively experiments are conducted on the synthetic data-set, Handy (Teaser [5] and GuessWho [40]) data-set, and the NYU data-set [25] for self-evaluation, the comparison with model-based techniques [5,39,40], and the comparison with appearance-based works [2,20,21,22,47,48,49] , respectively. We also qualitatively show the real-time performance and comparison with [2,5,39,40].

4.1. Data-Sets

Synthetic data-set The Synthetic data-set was generated based on a sequence of real hand motion. Firstly, we tracked a sequence of real hand motion and recorded the recovered shape and pose parameters with the inputs of data glove synchronously. Secondly, we chose five different hand shapes from the data-set of [44]. Then we applied those pose parameters to our hand model along with the five different shape parameters and produced five sets of synthetic image sequences. Each sequence contains 1129 synthetic depth images. Moreover, we applied Gaussian noise to shape 1 with a standard deviation ranging from 0 to 12 mm. The influence of noise can be seen in Figure 14.
Handy/Teaser and Handy/GuessWho data-set. the Handy data-set was created by [5] for the evaluation of high-precision generative tracking algorithms. It contains the full range of hand motion that has been studied in previous researches. The recording device was an Intel RealSense SR300, which was the same camera used in our system. The Teaser sequence [5] among Handy contains 2800 images of one subject with various range of hand motion. The GuessWho sequence [40] among Handy contains hand movement sequences of 12 different users. For each subject, there are about 1000 to 3000 images in Handy/GuessWho data-set.
NYU data-set. The NYU data-set, introduced in [25], records a great amount of high noise real data with quite an accurate annotation. It covers a good range of complex hand poses and a wide range of viewpoints. Because the NYU data-set is full of noise and miss pixels, it is a very challenging data-set. We only use the test set of the NYU data-set that contains 8225 images from two different actors.

4.2. Quantitative Comparison Metrics

The metrics are chosen to explain the difference between the recovered hand motion with the original hand motion. Different data-sets offer the original hand motion in different ways. Thus the following metrics vary with each data-set.
Metrics for the synthetic data-set. The synthetic data-set is generated from our hand model. We record the original full 3D hand motion in terms of the location of vertices and joints of the hand model. So, we compute the mean errors of the vertices and joints between the recovered hand model and the ground truth to show the difference:
V e r r o r = i = 1 N v V r e c o v e r e d i V g r o u n d t r u t h i 2 N v J e r r o r = i = 1 N j J r e c o v e r e d i J g r o u n d t r u t h i 2 N j ,
where N v = 778 is the number of vertices of our hand model, N j = 16 is the number of joints of our hand model.
Metrics for the Handy data-set. The Handy data-set records depth and color images for the original hand motion. Those model-based techniques [5,39,40] also provide their tracking results in the form of rendered depth images. To compare with those methods, we choose the algorithm agnostic metrics E 3 D and E 2 D proposed by [5] for evaluating the discrepancy between the input depth image and the rendered depth image. The E 3 D can be formulated as follows:
E 3 D = i = 1 N 3 D p i p c l o s e s t i 2 N 3 D ,
where the p i is the ith point in the 3D point cloud from the input depth image, the p c l o s e s t i is the closest correspondence point of p i in the 3D point cloud from the rendered depth image, the N 3 D is the total number of points in the 3D point cloud from the input depth image.
The E 2 D is evaluated using follow equation:
E 2 D = i = 1 N 2 D p r e n d e r i p c l o s e s t i 2 N o u t s i d e
where the p r e n d e r i is the ith point in the 2D hand image rendered from the hand model, p c l o s e s t i is the 2D closest point of p r e n d e r i in the silhouette of the input image. p c l o s e s t i = p r e n d e r i if p r e n d e r i lies inside the silhouette of the input image, the N 2 D is the number of 2D rendered points, the N o u t s i d e counts only the number of 2D rendered points outside the silhouette of the input image.
Metrics for the NYU data-set. The NYU data-set provides a quite accurate annotation of joints as the ground truth of the original hand motion. Those appearance-based works [2,20,21,22,47,48,49] also predict the hand joints as the tracking results. For the comparison with [2,20,21,22,47,48,49] on the NYU data-set, we choose the widely used metric among appearance-based methods, i.e., the mean error of the 3D joint locations:
E m e a n = i = 1 N J i J c o r r e s p o n d i 2 N
where the J i is the ith ground truth joint, the J c o r r e s p o n d i is the corresponding joint of J i , and N is the total number of ground truth joints.

4.3. Quantitative Experiments

Synthetic data-set. We make an overall evaluation of our system on various poses, hand shapes, and noise. Figure 15 shows the performance of our system on the five hand shapes and various hand poses. It is acceptable that the V e r r o r and J e r r o r fluctuate within 3 mm for different hand shapes. Over 95 percent of the V e r r o r and J e r r o r are within 5 mm, and almost 100 percent of the the V e r r o r and J e r r o r are within 10 mm. Few outliers appear but stay in 16 mm. Besides, the robustness of our system to different hand shapes and poses is illustrated by the steep curves of the V e r r o r and J e r r o r , and the narrow interquartile range(IQR) in the box plots of the V e r r o r and J e r r o r . Figure 16 illustrates the robustness of our system on different noise. Our system is robust as the standard deviation of noise gradually increases from 0 mm to 8 mm. The V e r r o r and J e r r o r slightly increase but are still within 10 mm. With more severe noise, the details of the point cloud from depth image disappear, Figure 14, and there is no wonder that the performance declines sharply. However, we can still give an approximate hand pose from the data glove even though the noise makes the point cloud a mess, see Figure 14.
Handy data-set. We compare our system with state-of-the-art generative methods [5,39,40] in the aspects of accuracy and robustness on various hand poses and shapes. We acquire the tracking result of [5] on the Handy/Teaser sequence though their released code on the Internet. The authors in [39,40] provide their tracking results on the Handy/GuessWho sequence.
Figure 17 shows the comparison between our system with [5] on the Handy/Teaser sequence. It can be seen in Figure 17 that our system achieves a visually noticeable improvement on the E 3 D and a comparable result on the E 2 D . The numerical improvements can be found in Table 2. The improvement of accuracy on the E 3 D is attributed to our hand model with the relaxed DoFs. Our triangular hand model deformed by LBS can express the surface geometry of the human hand better than the hand model made by sphere mesh. The relaxed DoFs introduced in Section 3.1 provide more possibilities to register the point cloud well. The comparable result on the E 2 D indicates that our system successfully tracks the hand shape in the Handy/Teaser sequence. There is an improvement of robustness on the E 3 D and E 2 D to our data glove. We owe it to our data glove. Our data glove can provide stable initialization and restriction, which reduces tracking failures, i.e., the outliers in the box plot of Figure 17. The p-values of significance tests for the results in the Handy/Teaser in Table 3 validate that our improvements are statistically significant.
Figure 18 gives the comparison with [39,40] on all subjects in the Handy/GuessWho sequence. The numerical improvements can be found in Table 4. We can draw a similar conclusion in terms of E 3 D . When comes to the E 2 D , our system shows no significant improvement, which indicates that our system does not track the hand shape well. We ascribe this to our implicit shape parameters from PCA which may overly constrain the hand shape and are hard to perform as well as the explicit 114 DoFs shape parameters on the detail of hand shape. For the comparison of each subject in the Handy/GuessWho sequence can be found in Figure 19 and Figure 20. The p-values of significance tests for the results in the Handy/GuessWho in Table 3 validate that our improvements of the E 3 D are statistically significant, and the performance of our system and [40] on the E 2 D is comparable with no statistical difference.
NYU data-set. We compare our system with state-of-the-art appearance-based methods [2,20,21,22,47,48,49] on the NYU data-set, see Figure 21. The results of [2,20,21,22,47,48,49] are publicly available on the Internet. Since the annotation scheme of NYU is different from ours, we manually choose a subset of 15 joints from the NYU annotation for the comparisons. Figure 22 shows the difference between the chosen joints and the joints of our hand model. Figure 21 shows the comparisons between our system with those appearance-based methods [2,20,21,22,47,48,49]. The numerical improvements can be found in Table 5. Our system achieves a comparable accuracy over those appearance-based methods. We expect better results for consistent annotation schemes. Besides, our system significantly improves about 40% on robustness. The p-values of significance tests for the results of the NYU dataset in Table 6 validate that the our improvements are statistically significant. Malik et al. [2] evaluated not only the hand pose but also the hand shape by deep-learning methods. We also conducted a comparison with [2] on the E 3 D and E 2 D metrics, see Figure 23. We outperform than [2] both on E 3 D and E 2 D , which indicates the superiority of our model-based system on recovering the hand shape.

4.4. Qualitative Experiments

We first qualitatively show our real-time performance in Figure 24. We can see that our system acts well on various hand poses and the rendered depth image from the recovered hand model looks almost the same as the input depth image of the hand. We also present the robustness of our system with occlusion in Figure 25. We do not conduct extra segments for those small items, which means those small items will affect the completeness of the point cloud and confuse the tracking progress. Figure 25 shows that our system stays robust to those influences thanks to the strong prior to our data glove. Then we exhibit the qualitative comparisons with Tkach et al. [5] on the Handy/Teaser, Remelli et al. [39] and Tkach et al. [40] on the Handy/GuessWho, and Malik et al. [2] on the NYU data-set, in Figure 26, Figure 27 and Figure 28. Those qualitative comparisons also show comparable performance with state-of-the-art.

4.5. System Efficiency

The machine we use is a laptop with a 6-core Intel Core i7 2.2 GHz CPU, 8 G RAM, and one GPU of NVIDIA GTX1050Ti. Our system does not use GPU and occupies about 25% of CPU and about 400 MB RAM, which leaves enough CPU, RAM, and GPU for other applications. Our system can reach real-time performance at around 40 FPS. Data Acquisition and Preprocessing take less than 500 us. We solve the model-fitting of our system iteratively with a Levenberg-Marquardt approach. In general, five iterations are enough. Each iteration costs within 5 ms. In each iteration, tracking correspondences searching costs less than 1 ms. Comparing with those generative methods [5,39,40] which rely on heavy parallelization and high-end GPU hardware, it is reasonable to believe that our system is more efficient.

5. Conclusions and Future Works

In this paper, we propose a model-based system for real-time articulated hand tracking with the synergy of a simple data glove and a depth camera. We redefine and simplify the data glove as a strong priority to ensure robustness and keep our system light-weight and easy-to-use. To improve accuracy and efficiency, we present several novelties to deal with DoFs setting, hand shape adjustment, self-interaction, tracking corresponding computation, and pose searching space constriction. These contributions make our system take the essence of wearable-based methods and model-based approaches, i.e., robustness and accuracy, but overcomes their downsides, i.e., tedious calibration, occlusions, and high dependency on GPU. Experimental results demonstrate that our system performs better than the state-of-the-art approach in the aspect of accuracy, robustness, and efficiency.
There are some factors in our system that should be considered in the future. When we compare our system with Remelli et al. [39] and Tkach et al. [40], we find that our system does not perform well on detailed hand shape tracking. We ascribe this weakness to the over constraints on the hand shape with the top 10 principal components of PCA as our shape parameters. More detailed and efficient shape parameters need to be designed for our triangular mesh model in the future. Besides, we do not claim to have “solved” the occlusion problem completely. When the occlusion occurs, we use the hand pose from the data glove to prevent tracking failure. The hand pose from our simple data glove is only approximate, and we expect an adoptive auto-calibration for the simple data glove by online learning in the future.

Author Contributions

Conceptualization, L.J. and H.X.; funding acquisition, H.X. and C.G.; investigation, L.J.; methodology, L.J.; project administration, H.X. and C.G.; supervision, H.X. and C.G.; validation, L.J.; writing—original draft, L.J.; writing—review and editing, H.X. and C.G.

Funding

This work was funded by National Natural Science Foundation of China (Grant No. 61976022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HCIHuman–computer interaction
DoFDegree of freedom
VRVirtual reality
ARAugmented reality
IMUInertial measurement unit
CNNConvolutional neural network
LBSLinear blend skinning
ICPIterative closest point
PSOParticle swarm optimization
PCAPrincipal components analysis
FPSFrames per second
GPUGraphics Processing Unit
CPUCentral Processing Unit
SDFSigned Distance Function

References

  1. Taylor, J.; Tankovich, V.; Tang, D.; Keskin, C.; Kim, D.; Davidson, P.; Kowdle, A.; Izadi, S. Articulated distance fields for ultra-fast tracking of hands interacting. ACM Trans. Graphics (TOG) 2017, 36, 244. [Google Scholar] [CrossRef]
  2. Malik, J.; Elhayek, A.; Nunnari, F.; Varanasi, K.; Tamaddon, K.; Heloir, A.; Stricker, D. Deephps: End-to-end estimation of 3d hand pose and shape by learning from synthetic depth. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 110–119. [Google Scholar]
  3. Oberweger, M.; Lepetit, V. Deepprior++: Improving fast and accurate 3d hand pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 585–594. [Google Scholar]
  4. Ballan, L.; Taneja, A.; Gall, J.; Van Gool, L.; Pollefeys, M. Motion capture of hands in action using discriminative salient points. In Proceedings of the European Conference on Computer Vision, Firenze, Italy, 7–13 October 2012; pp. 640–653. [Google Scholar]
  5. Tkach, A.; Pauly, M.; Tagliasacchi, A. Sphere-meshes for real-time hand modeling and tracking. ACM Trans. Graphics (TOG) 2016, 35, 222. [Google Scholar] [CrossRef]
  6. Sharp, T.; Keskin, C.; Robertson, D.; Taylor, J.; Shotton, J.; Kim, D.; Rhemann, C.; Leichter, I.; Vinnikov, A.; Wei, Y.; et al. Accurate, robust, and flexible real-time hand tracking. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, Seoul, Korea, 18–23 April 2015; pp. 3633–3642. [Google Scholar]
  7. Sanchez-Riera, J.; Srinivasan, K.; Hua, K.L.; Cheng, W.H.; Hossain, M.A.; Alhamid, M.F. Robust RGB-D hand tracking using deep learning priors. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 2289–2301. [Google Scholar] [CrossRef]
  8. Tzionas, D.; Ballan, L.; Srikantha, A.; Aponte, P.; Pollefeys, M.; Gall, J. Capturing Hands in Action using Discriminative Salient Points and Physics Simulation. Int. J. Comput. Vision 2016, 118, 172–193. [Google Scholar] [CrossRef]
  9. Arkenbout, E.; de Winter, J.; Breedveld, P. Robust hand motion tracking through data fusion of 5DT data glove and nimble VR Kinect camera measurements. Sensors 2015, 15, 31644–31671. [Google Scholar] [CrossRef] [PubMed]
  10. Ponraj, G.; Ren, H. Sensor Fusion of Leap Motion Controller and Flex Sensors Using Kalman Filter for Human Finger Tracking. IEEE Sens. J. 2018, 18, 2042–2049. [Google Scholar] [CrossRef]
  11. Dipietro, L.; Sabatini, A.M.; Dario, P. A survey of glove-based systems and their applications. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2008, 38, 461–482. [Google Scholar] [CrossRef]
  12. 5DT. Available online: http://5dt.com/5dt-data-glove-ultra/ (accessed on 26 October 2019).
  13. CyberGlove. Available online: http://www.cyberglovesystems.com/ (accessed on 26 October 2019).
  14. Yuan, S.; Ye, Q.; Stenger, B.; Jain, S.; Kim, T.K. Bighand2. 2m benchmark: Hand pose dataset and state of the art analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 4866–4874. [Google Scholar]
  15. Yuan, S.; Garcia-Hernando, G.; Stenger, B.; Moon, G.; Yong Chang, J.; Mu Lee, K.; Molchanov, P.; Kautz, J.; Honari, S.; Ge, L.; et al. Depth-based 3d hand pose estimation: From current achievements to future goals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2636–2645. [Google Scholar]
  16. Tang, D.; Yu, T.H.; Kim, T.K. Real-time articulated hand pose estimation using semi-supervised transductive regression forests. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 3224–3231. [Google Scholar]
  17. Tang, D.; Chang, H.J.; Tejani, A.; Kim, T. Latent Regression Forest: Structured Estimation of 3D Hand Poses. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1374–1387. [Google Scholar] [CrossRef]
  18. Deng, X.; Yang, S.; Zhang, Y.; Tan, P.; Chang, L.; Wang, H. Hand3d: Hand pose estimation using 3d neural network. arXiv 2017, arXiv:1704.02224. [Google Scholar]
  19. Guo, H.; Wang, G.; Chen, X. Two-stream convolutional neural network for accurate rgb-d fingertip detection using depth and edge information. arXiv 2016, arXiv:1612.07978. [Google Scholar]
  20. Rad, M.; Oberweger, M.; Lepetit, V. Feature mapping for learning fast and accurate 3d pose inference from synthetic images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4663–4672. [Google Scholar]
  21. Du, K.; Lin, X.; Sun, Y.; Ma, X. CrossInfoNet: Multi-Task Information Sharing Based Hand Pose Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9896–9905. [Google Scholar]
  22. Oberweger, M.; Wohlhart, P.; Lepetit, V. Generalized Feedback Loop for Joint Hand-Object Pose Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2019. [Google Scholar] [CrossRef] [PubMed]
  23. Qian, C.; Sun, X.; Wei, Y.; Tang, X.; Sun, J. Realtime and robust hand tracking from depth. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 1106–1113. [Google Scholar]
  24. Sun, X.; Wei, Y.; Liang, S.; Tang, X.; Sun, J. Cascaded hand pose regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 824–832. [Google Scholar]
  25. Tompson, J.; Stein, M.; Lecun, Y.; Perlin, K. Real-time continuous pose recovery of human hands using convolutional networks. ACM Trans. Graphics (TOG) 2014, 33, 169. [Google Scholar] [CrossRef]
  26. Wetzler, A.; Slossberg, R.; Kimmel, R. Rule of thumb: Deep derotation for improved fingertip detection. arXiv 2015, arXiv:1507.05726. [Google Scholar] [Green Version]
  27. Tagliasacchi, A.; Schroeder, M.; Tkach, A.; Bouaziz, S.; Botsch, M.; Pauly, M. Robust Articulated-ICP for Real-Time Hand Tracking. Comput. Graphics Forum (Symp. Geom. Process.) 2015, 34, 101–114. [Google Scholar] [CrossRef] [Green Version]
  28. Taylor, J.; Bordeaux, L.; Cashman, T.; Corish, B.; Keskin, C.; Sharp, T.; Soto, E.; Sweeney, D.; Valentin, J.; Luff, B.; et al. Efficient and precise interactive hand tracking through joint, continuous optimization of pose and correspondences. ACM Trans. Graphics (TOG) 2016, 35, 143. [Google Scholar] [CrossRef]
  29. Valentin, J.; Dai, A.; Nießner, M.; Kohli, P.; Torr, P.; Izadi, S.; Keskin, C. Learning to navigate the energy landscape. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 323–332. [Google Scholar]
  30. Taylor, J.; Stebbing, R.; Ramakrishna, V.; Keskin, C.; Shotton, J.; Izadi, S.; Hertzmann, A.; Fitzgibbon, A. User-specific hand modeling from monocular depth sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 644–651. [Google Scholar]
  31. Oikonomidis, I.; Kyriazis, N.; Argyros, A.A. Efficient model-based 3D tracking of hand articulations using Kinect. BMVC 2011, 1, 3. [Google Scholar] [CrossRef]
  32. Fleishman, S.; Kliger, M.; Lerner, A.; Kutliroff, G. Icpik: Inverse kinematics based articulated-icp. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 28–35. [Google Scholar]
  33. Makris, A.; Argyros, A. Model-based 3D hand tracking with on-line hand shape adaptation. In Proceedings of the BMVC, Swansea, UK, 7–10 September 2015; pp. 77.1–77.12. [Google Scholar]
  34. Melax, S.; Keselman, L.; Orsten, S. Dynamics based 3D skeletal hand tracking. In Proceedings of the Graphics Interface 2013. In Proceedings of the Canadian Information Processing Society, Regina, SK, Canada, 29–31 May 2013; pp. 63–70.
  35. Sridhar, S.; Rhodin, H.; Seidel, H.P.; Oulasvirta, A.; Theobalt, C. Real-time hand tracking using a sum of anisotropic gaussians model. In Proceedings of the 2014 2nd International Conference on 3D Vision, Tokyo, Japan, 8–11 December 2014; Volume 1, pp. 319–326. [Google Scholar]
  36. Khamis, S.; Taylor, J.; Shotton, J.; Keskin, C.; Izadi, S.; Fitzgibbon, A. Learning an efficient model of hand shape variation from depth images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2540–2548. [Google Scholar]
  37. Schmidt, T.; Newcombe, R.; Fox, D. DART: dense articulated real-time tracking with consumer depth cameras. Auton. Robots 2015, 39, 239–258. [Google Scholar] [CrossRef]
  38. Joseph Tan, D.; Cashman, T.; Taylor, J.; Fitzgibbon, A.; Tarlow, D.; Khamis, S.; Izadi, S.; Shotton, J. Fits like a glove: Rapid and reliable hand shape personalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5610–5619. [Google Scholar]
  39. Remelli, E.; Tkach, A.; Tagliasacchi, A.; Pauly, M. Low-dimensionality calibration through local anisotropic scaling for robust hand model personalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2535–2543. [Google Scholar]
  40. Tkach, A.; Tagliasacchi, A.; Remelli, E.; Pauly, M.; Fitzgibbon, A. Online generative model personalization for hand tracking. ACM Trans. Graphics (TOG) 2017, 36, 243. [Google Scholar] [CrossRef]
  41. Tannous, H.; Istrate, D.; Benlarbi-Delai, A.; Sarrazin, J.; Gamet, D.; Ho Ba Tho, M.; Dao, T. A new multi-sensor fusion scheme to improve the accuracy of knee flexion kinematics for functional rehabilitation movements. Sensors 2016, 16, 1914. [Google Scholar] [CrossRef]
  42. Sun, Y.; Li, C.; Li, G.; Jiang, G.; Jiang, D.; Liu, H.; Zheng, Z.; Shu, W. Gesture Recognition Based on Kinect and sEMG Signal Fusion. Mob. Netw. Appl. 2018, 23, 797–805. [Google Scholar] [CrossRef]
  43. Pacchierotti, C.; Salvietti, G.; Hussain, I.; Meli, L.; Prattichizzo, D. The hRing: A wearable haptic device to avoid occlusions in hand tracking. In Proceedings of the 2016 IEEE Haptics Symposium (HAPTICS), Philadelphia, PV, USA, 8–11 April 2016; pp. 134–139. [Google Scholar]
  44. Romero, J.; Tzionas, D.; Black, M.J. Embodied hands: Modeling and capturing hands and bodies together. ACM Trans. Graphics (TOG) 2017, 36, 245. [Google Scholar] [CrossRef]
  45. Sensfusion. Available online: http://www.sensfusion.com/ (accessed on 26 October 2019).
  46. Ganapathi, V.; Plagemann, C.; Koller, D.; Thrun, S. Real-time human pose tracking from range data. In Proceedings of the European Conference on Computer Vision, Firenze, Italy, 7–13 October 2012; pp. 738–751. [Google Scholar]
  47. Poier, G.; Opitz, M.; Schinagl, D.; Bischof, H. MURAUER: Mapping unlabeled real data for label austerity. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Hilton Waikoloa Village, HI, USA, 7–11 January 2019; pp. 1393–1402. [Google Scholar]
  48. Ge, L.; Ren, Z.; Yuan, J. Point-to-point regression pointnet for 3d hand pose estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 475–491. [Google Scholar]
  49. Chen, X.; Wang, G.; Zhang, C.; Kim, T.K.; Ji, X. Shpr-net: Deep semantic hand pose regression from point clouds. IEEE Access 2018, 6, 43425–43439. [Google Scholar] [CrossRef]
Figure 1. Hand models. (a) Capsule model [31,32,33]. (b) Cylinder model [27]. (c) Sphere model [23]. (d) Convex bodies for tracking [34]. (e) Sum of an-isotropic Gaussian model [35]. (f) Sphere-mesh model [5]. (g) Triangular hand model [4]. (h) Triangular mesh [8]. (i) Loop subdivision Surface of a triangular control mesh [36]. (j) Articulated Signed Distance Function (SDF) for a voxelized shape-primitive hand model [37]. (k) Articulated signed distance function for a hand model [1]. Images reproduced from the cited papers or their supplementary videos.
Figure 1. Hand models. (a) Capsule model [31,32,33]. (b) Cylinder model [27]. (c) Sphere model [23]. (d) Convex bodies for tracking [34]. (e) Sum of an-isotropic Gaussian model [35]. (f) Sphere-mesh model [5]. (g) Triangular hand model [4]. (h) Triangular mesh [8]. (i) Loop subdivision Surface of a triangular control mesh [36]. (j) Articulated Signed Distance Function (SDF) for a voxelized shape-primitive hand model [37]. (k) Articulated signed distance function for a hand model [1]. Images reproduced from the cited papers or their supplementary videos.
Sensors 19 04680 g001
Figure 2. Overview of our system. Before tracking, the hand model has been prepared. During tracking, the workflow of the system is as follows: firstly, we acquire and process the inputs from the camera and the glove. For each acquired image, we extract a 3D point cloud and a 2D silhouette from the depth to provide the absolute measure. For the glove input, we get a rough hand pose for initialization. Secondly, we construct the objective function to measure the discrepancy between the hand model and the extracted 3D point cloud and 2D silhouette. Finally, we iteratively optimize the objective function by Levenberg–Marquardt approach to get the recovered hand motion. If the optimization failed, the input of the data glove provides a guarantee to ensure robustness.
Figure 2. Overview of our system. Before tracking, the hand model has been prepared. During tracking, the workflow of the system is as follows: firstly, we acquire and process the inputs from the camera and the glove. For each acquired image, we extract a 3D point cloud and a 2D silhouette from the depth to provide the absolute measure. For the glove input, we get a rough hand pose for initialization. Secondly, we construct the objective function to measure the discrepancy between the hand model and the extracted 3D point cloud and 2D silhouette. Finally, we iteratively optimize the objective function by Levenberg–Marquardt approach to get the recovered hand motion. If the optimization failed, the input of the data glove provides a guarantee to ensure robustness.
Sensors 19 04680 g002
Figure 3. The MANO hand model and the kinematic chain. (a) The MANO hand model. (b) The triangular mesh of MANO. (c) The kinematic chain and human hand anatomy. The explanations of capital letters in (c) : A represents the Wrist joint. B represents the Carpometacarpal (CMC) joint. C represents the Metacarpophalangeal (MCP) joints. D represents the Proximal Interphalangeal (PIP) joints. E represents the Distal Interphalangeal (DIP) joints. F represents the Interphalangeal (IP) joint.
Figure 3. The MANO hand model and the kinematic chain. (a) The MANO hand model. (b) The triangular mesh of MANO. (c) The kinematic chain and human hand anatomy. The explanations of capital letters in (c) : A represents the Wrist joint. B represents the Carpometacarpal (CMC) joint. C represents the Metacarpophalangeal (MCP) joints. D represents the Proximal Interphalangeal (PIP) joints. E represents the Distal Interphalangeal (DIP) joints. F represents the Interphalangeal (IP) joint.
Sensors 19 04680 g003
Figure 4. Fitting MANO hand model to the data set in [44]. Green one represents the data set, while red one is the fitted MANO model. (a) The IP joint with only one degree of freedom (DoF). (b) The IP joint with three DoFs.
Figure 4. Fitting MANO hand model to the data set in [44]. Green one represents the data set, while red one is the fitted MANO model. (a) The IP joint with only one degree of freedom (DoF). (b) The IP joint with three DoFs.
Sensors 19 04680 g004
Figure 5. Data glove. (a) The back side of the data glove. (b) The front side of the data glove. (c) The location of inertial measurement units (IMUs) on the hand. We attach 11 IMUs on the glove made by breathable textile lining. Two IMUs on the distal, medial phalange of thumb; two IMUs on the medial, proximal phalanges of each of the other fingers; one IMU on the back of the hand.
Figure 5. Data glove. (a) The back side of the data glove. (b) The front side of the data glove. (c) The location of inertial measurement units (IMUs) on the hand. We attach 11 IMUs on the glove made by breathable textile lining. Two IMUs on the distal, medial phalange of thumb; two IMUs on the medial, proximal phalanges of each of the other fingers; one IMU on the back of the hand.
Sensors 19 04680 g005
Figure 6. The example of how the IMUs attached on the index finger measure the rotation. (a) The initial poses of the 3rd and 4th IMU are equal Q 3 = Q 4 , when Index finger is stretching. (b) The relative pose Q 3 , 4 = Q 3 Q 4 1 of the 3rd and 4th IMU, when the index finger is blend. Q 3 , 4 can represent the 3rd IMU rotate 90 degrade counterclockwise around axis x related to the 4th IMU.
Figure 6. The example of how the IMUs attached on the index finger measure the rotation. (a) The initial poses of the 3rd and 4th IMU are equal Q 3 = Q 4 , when Index finger is stretching. (b) The relative pose Q 3 , 4 = Q 3 Q 4 1 of the 3rd and 4th IMU, when the index finger is blend. Q 3 , 4 can represent the 3rd IMU rotate 90 degrade counterclockwise around axis x related to the 4th IMU.
Sensors 19 04680 g006
Figure 7. Some hand poses given by the simple data glove.
Figure 7. Some hand poses given by the simple data glove.
Sensors 19 04680 g007
Figure 8. The original MANO hand model, the hand model applied the Loop subdivision, and four different views of the subdivided hand model with original MANO in wire-frame.
Figure 8. The original MANO hand model, the hand model applied the Loop subdivision, and four different views of the subdivided hand model with original MANO in wire-frame.
Sensors 19 04680 g008
Figure 9. The comparison of two different corresponding points finding ways when we fit an input image from the Handy/Teaser data-set [5]. Each part from left to right is the 3D image, mixed 2D silhouette, and mixed depth image applied pseudo-color enhancement. The 3D image contains point cloud X s in green, hand model in red, and corresponding points on the hand model in blue. The red part in the mixed 2D silhouette is the rendered silhouette of the hand model, while the green part is the silhouette of input. The mixed depth shows how the rendered depth matches the original depth map, using pseudo-color enhancement for better viewing. The E r r o r 3 D and E r r o r 2 D are the metrics mentioned in Section 4.
Figure 9. The comparison of two different corresponding points finding ways when we fit an input image from the Handy/Teaser data-set [5]. Each part from left to right is the 3D image, mixed 2D silhouette, and mixed depth image applied pseudo-color enhancement. The 3D image contains point cloud X s in green, hand model in red, and corresponding points on the hand model in blue. The red part in the mixed 2D silhouette is the rendered silhouette of the hand model, while the green part is the silhouette of input. The mixed depth shows how the rendered depth matches the original depth map, using pseudo-color enhancement for better viewing. The E r r o r 3 D and E r r o r 2 D are the metrics mentioned in Section 4.
Sensors 19 04680 g009
Figure 10. A visualization of the co-variance estimate for simple stick model as we vary the blend angles θ 1 , θ 2 . the co-variance ellipsoids are centered at the corresponding θ 1 , θ 2 location. The explicit shape parameters L 1 , L 2 , L 3 are the length of each stick, and the implicit shape parameters α , β are the top two principal components analysis (PCA) components extract on 1000 random generated data.
Figure 10. A visualization of the co-variance estimate for simple stick model as we vary the blend angles θ 1 , θ 2 . the co-variance ellipsoids are centered at the corresponding θ 1 , θ 2 location. The explicit shape parameters L 1 , L 2 , L 3 are the length of each stick, and the implicit shape parameters α , β are the top two principal components analysis (PCA) components extract on 1000 random generated data.
Sensors 19 04680 g010
Figure 11. An illustration of the Gaussian distribution in the two-dimensional subspace of PCA. (a) The result θ ˜ of ground-truth hand pose(except the six global components) θ on the database of [44]. (b) The projection θ ˜ g l o v e of the glove data θ g l o v e into the same subspace in red dots. (c) The multivariate Gaussian model in blue color N ( μ d i f f , Σ d i f f ) for the difference between the glove data θ ˜ g l o v e and corresponding ground truth θ ˜ in the two-dimensional subspace.
Figure 11. An illustration of the Gaussian distribution in the two-dimensional subspace of PCA. (a) The result θ ˜ of ground-truth hand pose(except the six global components) θ on the database of [44]. (b) The projection θ ˜ g l o v e of the glove data θ g l o v e into the same subspace in red dots. (c) The multivariate Gaussian model in blue color N ( μ d i f f , Σ d i f f ) for the difference between the glove data θ ˜ g l o v e and corresponding ground truth θ ˜ in the two-dimensional subspace.
Sensors 19 04680 g011
Figure 12. An illustration of the influence of N ( 0 , Σ ) and N ( μ d i f f , Σ d i f f ) and their fusion on a given glove input θ g l o v e i . The red hand model is the initialization from our data glove, while the green one is the ground-truth from the data-set [44]. The distribution of N ( μ d i f f , Σ d i f f ) is convert to N ( θ ˜ g l o v e i μ d i f f , Σ d i f f ) with the specific input of glove θ g l o v e i . The red ellipse represents the distribution of N ( μ m e r g e , Σ m e r g e ) , which shows a smaller search area and a closer μ m e r g e toward the ground-truth. We also show the comparison between the initialization and convergence with the ground-truth on the right side.
Figure 12. An illustration of the influence of N ( 0 , Σ ) and N ( μ d i f f , Σ d i f f ) and their fusion on a given glove input θ g l o v e i . The red hand model is the initialization from our data glove, while the green one is the ground-truth from the data-set [44]. The distribution of N ( μ d i f f , Σ d i f f ) is convert to N ( θ ˜ g l o v e i μ d i f f , Σ d i f f ) with the specific input of glove θ g l o v e i . The red ellipse represents the distribution of N ( μ m e r g e , Σ m e r g e ) , which shows a smaller search area and a closer μ m e r g e toward the ground-truth. We also show the comparison between the initialization and convergence with the ground-truth on the right side.
Sensors 19 04680 g012
Figure 13. The process of generating a set of collision spheres to approximate the volume of the fingers in MANO hand model.
Figure 13. The process of generating a set of collision spheres to approximate the volume of the fingers in MANO hand model.
Sensors 19 04680 g013
Figure 14. The influence of the standard deviation of noise on the point cloud from synthetic depth images. The visible performance of our system with those noise.
Figure 14. The influence of the standard deviation of noise on the point cloud from synthetic depth images. The visible performance of our system with those noise.
Sensors 19 04680 g014
Figure 15. The V e r r o r and J e r r o r of our system on the five hand shapes and various hand poses in the synthetic data-set.
Figure 15. The V e r r o r and J e r r o r of our system on the five hand shapes and various hand poses in the synthetic data-set.
Sensors 19 04680 g015
Figure 16. The V e r r o r and J e r r o r of our system with different noise in the synthetic data-set.
Figure 16. The V e r r o r and J e r r o r of our system with different noise in the synthetic data-set.
Sensors 19 04680 g016
Figure 17. The E 3 D and E 2 D results of comparison between our system with [5] on Handy/ Teaser sequence.
Figure 17. The E 3 D and E 2 D results of comparison between our system with [5] on Handy/ Teaser sequence.
Sensors 19 04680 g017
Figure 18. The E 3 D and E 2 D total results of comparison between our system with Remelli et al. [39] and Tkach et al. [40] on Handy/GuessWho sequence.
Figure 18. The E 3 D and E 2 D total results of comparison between our system with Remelli et al. [39] and Tkach et al. [40] on Handy/GuessWho sequence.
Sensors 19 04680 g018
Figure 19. The E 3 D result of comparison between our system with Remelli et al. [39] and Tkach et al. [40] on the 12 subjects in Handy/GuessWho sequence.
Figure 19. The E 3 D result of comparison between our system with Remelli et al. [39] and Tkach et al. [40] on the 12 subjects in Handy/GuessWho sequence.
Sensors 19 04680 g019
Figure 20. The E 2 D result of comparison between our system with Remelli et al. [39] and Tkach et al. [40] on the 12 subjects in Handy/GuessWho sequence.
Figure 20. The E 2 D result of comparison between our system with Remelli et al. [39] and Tkach et al. [40] on the 12 subjects in Handy/GuessWho sequence.
Sensors 19 04680 g020
Figure 21. Experiment results on the NYU data-set. (a) The joints errors on the NYU data-set. (b) The box plot of the joints errors on the NYU data-set.
Figure 21. Experiment results on the NYU data-set. (a) The joints errors on the NYU data-set. (b) The box plot of the joints errors on the NYU data-set.
Sensors 19 04680 g021
Figure 22. The subset of 15 joints chosen from the NYU annotation and the joints of our hand model in a simple hand pose. In this pose, the mean error of joints is 8.30 mm.
Figure 22. The subset of 15 joints chosen from the NYU annotation and the joints of our hand model in a simple hand pose. In this pose, the mean error of joints is 8.30 mm.
Sensors 19 04680 g022
Figure 23. The E 3 D and E 2 D comparisons between our system with Malik et al. [2] on the NYU data-set.
Figure 23. The E 3 D and E 2 D comparisons between our system with Malik et al. [2] on the NYU data-set.
Sensors 19 04680 g023
Figure 24. The real-time performance of our system. There are two rows. For each row, the upper one shows the hand segment of the input depth after pseudo-color enhancement, while the lower one shows the rendered depth image from the recovered hand model after pseudo-color enhancement.
Figure 24. The real-time performance of our system. There are two rows. For each row, the upper one shows the hand segment of the input depth after pseudo-color enhancement, while the lower one shows the rendered depth image from the recovered hand model after pseudo-color enhancement.
Sensors 19 04680 g024
Figure 25. The real-time performance of our system with occlusion from some small items. There are two rows. For each row, the upper one shows the hand segment of the input depth after pseudo-color enhancement, while the lower one shows the rendered depth image from the recovered hand model after pseudo-color enhancement.
Figure 25. The real-time performance of our system with occlusion from some small items. There are two rows. For each row, the upper one shows the hand segment of the input depth after pseudo-color enhancement, while the lower one shows the rendered depth image from the recovered hand model after pseudo-color enhancement.
Sensors 19 04680 g025
Figure 26. The qualitative comparison between our system with Tkach et al. [5] on Handy/Teaser. The upper row shows the hand segment of the input depth after pseudo-color enhancement; the middle row shows the rendered depth image from the recovered hand model of Tkach et al. [5]; the bottom row shows the rendered depth image from the recovered hand model of our system after pseudo-color enhancement.
Figure 26. The qualitative comparison between our system with Tkach et al. [5] on Handy/Teaser. The upper row shows the hand segment of the input depth after pseudo-color enhancement; the middle row shows the rendered depth image from the recovered hand model of Tkach et al. [5]; the bottom row shows the rendered depth image from the recovered hand model of our system after pseudo-color enhancement.
Sensors 19 04680 g026
Figure 27. The qualitative comparison between our system with Remelli et al. [39] and Tkach et al. [40] on the Handy/GuessWho sequence of user4. The first row shows the hand segment of the input depth after pseudo-color enhancement; the second row shows the rendered depth image from the recovered hand model of Remelli et al. [39]; the third row shows the rendered depth image from the recovered hand model of Tkach et al. [40]; the bottom row shows the rendered depth image from the recovered hand model of our system.
Figure 27. The qualitative comparison between our system with Remelli et al. [39] and Tkach et al. [40] on the Handy/GuessWho sequence of user4. The first row shows the hand segment of the input depth after pseudo-color enhancement; the second row shows the rendered depth image from the recovered hand model of Remelli et al. [39]; the third row shows the rendered depth image from the recovered hand model of Tkach et al. [40]; the bottom row shows the rendered depth image from the recovered hand model of our system.
Sensors 19 04680 g027
Figure 28. The qualitative comparison between our system with Malik et al. [2] on the NYU data-set. The first row shows the hand segment of the input depth after pseudo-color enhancement; the second row shows the rendered depth image from the recovered hand model of Malik et al. [2]; the bottom row shows the rendered depth image from the recovered hand model of our system.
Figure 28. The qualitative comparison between our system with Malik et al. [2] on the NYU data-set. The first row shows the hand segment of the input depth after pseudo-color enhancement; the second row shows the rendered depth image from the recovered hand model of Malik et al. [2]; the bottom row shows the rendered depth image from the recovered hand model of our system.
Sensors 19 04680 g028
Table 1. The extracted limitations of pose parameters. We extract those limitations by fitting MANO model to the database provided by [44] using our kinematic structure.
Table 1. The extracted limitations of pose parameters. We extract those limitations by fitting MANO model to the database provided by [44] using our kinematic structure.
ThumbCMC_xCMC_yCMC_zDCP_xDCP_yDCP_zIP_xIP_yIP_z
Max18.37°29.34°78.16°16.36°33.72°21.28°17.59°78.29°19.26°
Min3.08°−33.06°−24.01°−18.85°−75.38°−19.70°−13.11°−93.88°−15.56°
IndexDCP_xDCP_yDCP_zPIP_xPIP_yPIP_zDIP_xDIP_yDIP_z
Max18.53°22.72°89.40°16.74°7.66°101.89°11.71°5.96°75.80°
Min−15.41°−31.75°89.40°−11.08°−4.19°−34.83°−13.58°−4.43°−42.08°
MiddleDCP_xDCP_yDCP_zPIP_xPIP_yPIP_zDIP_xDIP_yDIP_z
Max16.65°18.74°105.90°14.43°4.17°101.65°6.98°6.27°77.22°
Min−13.19°−23.33°−39.66°−13.86°−6.03°−30.38°−7.28°−1.24°−28.96°
RingDCP_xDCP_yDCP_zPIP_xPIP_yPIP_zDIP_xDIP_yDIP_z
Max17.53°46.82°106.52°16.60°6.69°101.61°7.27°6.28°82.49°
Min−20.31°−30.92°−60.50°−19.58°−5.70°−24.83°−15.63°−5.45°−30.86°
PinkyDCP_xDCP_yDCP_zPIP_xPIP_yPIP_zDIP_xDIP_yDIP_z
Max18.47°31.52°102.32°2.48°7.11°103.65°10.22°6.63°82.53°
Min−16.75°−19.83°−39.75°−16.44°−2.81°−29.03°−11.20°−0.12°−32.14°
Table 2. The numerical improvements between our system and [5] on the Handy/Teaser sequence. We consider the the mean value and the standard deviation (SD) of the E 3 D and E 2 D .
Table 2. The numerical improvements between our system and [5] on the Handy/Teaser sequence. We consider the the mean value and the standard deviation (SD) of the E 3 D and E 2 D .
[5]Proposed MethodImprovement on [5]
E 3 D Mean (mm)3.181.9937%
SD (mm)1.610.6162%
E 2 D Mean (pixel)2.272.126%
SD (pixel)1.060.6440%
Table 3. The results of significance tests on the E 3 D and E 2 D for the Handy data-set.
Table 3. The results of significance tests on the E 3 D and E 2 D for the Handy data-set.
Handy/Teaser on [5]Handy/GuessWho on [39]Handy/GuessWho on [40]
E 3 D E 2 D E 3 D E 2 D E 3 D E 2 D
p-value 1.53 × e 262 6.57 × e 12 0 8.43 × e 156 00.1907
Table 4. The numerical improvements between our system and [39,40] on the Handy/GuessWho sequence. We consider the the mean value and the SD of the E 3 D and E 2 D .
Table 4. The numerical improvements between our system and [39,40] on the Handy/GuessWho sequence. We consider the the mean value and the SD of the E 3 D and E 2 D .
[39][40]Proposed MethodImprovement on [39]Inprovement on [40]
E 3 D Mean (mm)3.243.002.5422%15%
SD (mm)1.020.970.5050%48%
E 2 D Mean (pixel)1.731.891.90−9%−0.5%
SD (pixel)0.710.710.684%4%
Table 5. The numerical improvements between our system with [2,20,21,22,47,48,49] on the NYU data-set. We consider the the mean value and the standard deviation(SD) of the joints error.
Table 5. The numerical improvements between our system with [2,20,21,22,47,48,49] on the NYU data-set. We consider the the mean value and the standard deviation(SD) of the joints error.
Ours[2][20][21][22][47][48][49]
Mean (mm)10.4514.427.4410.0810.899.479.0510.78
SD (mm)3.648.304.446.676.225.277.028.00
Mean
Improvement
27%−40%−3%4%−10%−15%3%
SD
Improvement
56%18%45%41%30%48%54%
Table 6. The results of significance tests between our system with [2,20,21,22,47,48,49] on the joints error of the NYU data-set.
Table 6. The results of significance tests between our system with [2,20,21,22,47,48,49] on the joints error of the NYU data-set.
With [2]With [20]With [21]With [22]With [47]With [48]With [49]
p-value00 1.07 × e 5 1.74 × e 8 1.13 × e 43 7.39 × e 58 0.0007

Share and Cite

MDPI and ACS Style

Jiang, L.; Xia, H.; Guo, C. A Model-Based System for Real-Time Articulated Hand Tracking Using a Simple Data Glove and a Depth Camera. Sensors 2019, 19, 4680. https://doi.org/10.3390/s19214680

AMA Style

Jiang L, Xia H, Guo C. A Model-Based System for Real-Time Articulated Hand Tracking Using a Simple Data Glove and a Depth Camera. Sensors. 2019; 19(21):4680. https://doi.org/10.3390/s19214680

Chicago/Turabian Style

Jiang, Linjun, Hailun Xia, and Caili Guo. 2019. "A Model-Based System for Real-Time Articulated Hand Tracking Using a Simple Data Glove and a Depth Camera" Sensors 19, no. 21: 4680. https://doi.org/10.3390/s19214680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop