Next Article in Journal
A Real-Time Monitoring Method for Droplet Transfer Frequency in Wire-Filled GTAW Based on Arc Sensing
Previous Article in Journal
Assistance Device Based on SSVEP-BCI Online to Control a 6-DOF Robotic Arm
Previous Article in Special Issue
Human Pose Estimation Based on Efficient and Lightweight High-Resolution Network (EL-HRNet)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Model for Markerless Pose Estimation Based on Keypoint Augmentation: What Factors Influence Errors in Biomechanical Applications?

by
Ana V. Ruescas-Nicolau
1,*,
Enrique Medina-Ripoll
1,
Helios de Rosario
1,
Joaquín Sanchiz Navarro
1,
Eduardo Parrilla
1 and
María Carmen Juan Lizandra
2
1
Instituto de Biomecánica-IBV, Universitat Politècnica de València, Edifici 9C, Camí de Vera s/n, 46022 Valencia, Spain
2
Instituto Universitario de Automática e Informática Industrial, Universitat Politècnica de València, Edifici 1F, Camí de Vera, s/n, 46022 Valencia, Spain
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1923; https://doi.org/10.3390/s24061923
Submission received: 18 February 2024 / Revised: 7 March 2024 / Accepted: 13 March 2024 / Published: 17 March 2024

Abstract

:
In biomechanics, movement is typically recorded by tracking the trajectories of anatomical landmarks previously marked using passive instrumentation, which entails several inconveniences. To overcome these disadvantages, researchers are exploring different markerless methods, such as pose estimation networks, to capture movement with equivalent accuracy to marker-based photogrammetry. However, pose estimation models usually only provide joint centers, which are incomplete data for calculating joint angles in all anatomical axes. Recently, marker augmentation models based on deep learning have emerged. These models transform pose estimation data into complete anatomical data. Building on this concept, this study presents three marker augmentation models of varying complexity that were compared to a photogrammetry system. The errors in anatomical landmark positions and the derived joint angles were calculated, and a statistical analysis of the errors was performed to identify the factors that most influence their magnitude. The proposed Transformer model improved upon the errors reported in the literature, yielding position errors of less than 1.5 cm for anatomical landmarks and 4.4 degrees for all seven movements evaluated. Anthropometric data did not influence the errors, while anatomical landmarks and movement influenced position errors, and model, rotation axis, and movement influenced joint angle errors.

1. Introduction

In the clinical setting, the analysis of human movement is important for understanding and managing diseases of the musculoskeletal system. For example, the analysis of the mobility of the cervical spine can facilitate the treatment of conditions such as low back pain or cervical pain [1,2,3]. Similarly, gait analysis allows the diagnosis and subsequent planning and evaluation of treatment of neuromusculoskeletal pathology [4,5].
Marker-based photogrammetry motion capture (MoCap) is considered the gold standard for the analysis of human movement, demonstrating the ability to provide comprehensive, accurate, robust, and reproducible data. However, despite being the gold standard, its daily clinical practice application is constrained by certain drawbacks associated with this technique.
The main reason discouraging movement assessment using marker-based photogrammetry systems is the high cost, both financially and in terms of time, with instrumentation of the participants being one of the main problems [6]. Normally, it is necessary to identify a series of anatomical landmarks and to fix reflective markers in order to calculate the anatomical axes and kinematics of each joint [7,8]. This operation requires highly qualified personnel and is very time consuming. In addition, differences in marker placement directly affect the reproducibility and accuracy of the measurements [9,10,11].
In order to overcome these limitations, commercial markerless motion capture systems have been developed in recent years, such as Theia3D (www.theiamarkerless.ca, accessed on 18 January 2024), Captury (www.captury.com, accessed on 18 January 2024), Fast Move AI 3D (www.fastmove.cn, accessed on 18 January 2024), Kinatrax Motion Capture (www.kinatrax.com, accessed on 18 January 2024), and OpenCap (www.opencap.ai, accessed on 18 January 2024). Many of these systems accurately estimate kinematics in the principal plane [12] but tend to have larger errors in the measurement of secondary angles. Additionally, many of these approaches require multiple cameras, specific software, and specialized computing resources [13,14].
An alternative to vision systems is inertial systems. Inertial measurement units are portable and can accurately estimate kinematics [15,16]. However, commercial sensors still present similar problems: long instrumentation time, the use of proprietary algorithms, and the low-cost sensors are often prone to errors of misalignment, orthogonality, and offset, which require correction methods to achieve accurate position and orientation measurements [17].
In recent years, new methods of human pose detection based on images have been developed [18,19,20,21,22,23]. Thus, by triangulating the keypoints identified by these pose estimation algorithms in various videos, their 3D positions can be obtained [24,25].
These pose detection methods are usually trained using image datasets where the labeling (localization of keypoints) was performed manually, which may result in inconsistencies and low accuracy [26].
There are datasets in which the images were labeled with reliable and accurate marker-based MoCap [27,28,29,30] but the collected images feature subjects that are altered by the presence of external elements, such as reflective spheres. This alteration prevents their use for training general computational models. In such cases, the detector would be at risk of learning to localize the presence of external elements rather than the anatomical location of keypoints, especially when they are arranged in a fixed pattern [31].
According to [26], open-source pose estimation algorithms were never designed for biomechanical applications. In fact, they usually provide the centers of the main joints but not the key anatomical landmarks needed to fully characterize the translations and rotations of all body segments. There have been different ways to improve the accuracy of pose estimation methods so that they can be used in the analysis of human motion from a biomechanical point of view.
The most widespread solutions involved generating proprietary image datasets for training the pose estimator. For example, the markerless Theia3D system was trained with 500k images in the wild labeled by a group of expert taggers [32,33,34]. Other solutions were based on generating image datasets of specific movements, such as [35,36], who created ENSAM and its extended version in order to fine-tune the pose estimator for gait applications. Another example is presented in work [30], where the GPJATK dataset is introduced, specifically created for markerless 3D motion tracking and 3D gait recognition.
Another approach involves feeding a biomechanical model with the 3D landmarks detected in synchronized videos. For example, in [37], the 3D reconstructed keypoints were used to drive the motion of a constrained rigid-body kinematic model in OpenSim, from which the position and orientation of each segment could be obtained [38].
A recent approach is based on augmenting the keypoints of a pose estimator in images with anatomical markers [25]. In this approach, the authors proposed using a marker augmenter, a long short-term memory (LSTM) network, that estimates a set of anatomical landmarks from the reconstructed keypoints detected in videos. Two LSTM networks were trained for an arm and full body, with 3D video keypoints, weight and height, and 3D anatomical landmarks from 108 h of motion captures processed in OpenSim 4.3.
We were inspired by the latter strategy, as it leverages pose estimation technology from images to solve a different aspect of the problem, and it is also suitable for application to a wide range of movements. However, we wondered if this strategy could be applied with any marker augmentation model and what other factors would influence its performance. Therefore, we present a comparative study in which we assess three marker augmenters for lower limbs across eight different poses and movements with 14 subjects, using the MediaPipe Pose Landmarker model as the pose detector [23]. We utilized the “Human tracking dataset of 3D anatomical landmarks and pose keypoints” [39], which is a collection of 3D reconstructed video keypoints associated with 3D anatomical landmarks extracted from 567 sequences of 71 participants in A-pose and performing seven movements. This dataset was designed for biomechanical applications and is freely available for the research community.
This study had a dual purpose. The first was to assess the performance of the different proposed marker augmentation models. The second was aimed at determining the factors that most influence anatomical landmark estimation and joint angles errors. The factors we considered included the type of marker augmentation model, the movement, the anatomical landmark, the joint, the axis, and the anthropometric characteristics of the subject (height, weight, and sex).
The following sections present the characteristics of the data used in this work, the architectures of the marker augmentation models, and the results of the error evaluation and statistical analysis. Subsequently, the obtained results are discussed, and the conclusions of the work are presented.

2. Materials and Methods

In this section, we provide a description of the dataset used in this work, elaborate on the procedure and details of the proposed marker augmentation models, discuss the metrics employed for evaluation, and present the statistical analysis of the errors.

2.1. Dataset

A portion of the dataset “Human tracking dataset of 3D anatomical landmarks and pose keypoints” [39] has been used. Section 2.1.1 briefly introduces the workflow followed to obtain the whole dataset. Section 2.1.2 specifies the portion employed and its use in training the models.

2.1.1. Dataset Generation Overview

The subjects who participated in the dataset collection were scanned with Move4D (www.move4d.net, accessed on 31 January 2024) [40], which is composed of a photogrammetry-based 3D temporal scanner (or 4D scanner) and a processing software. It generates a series of rigged and textured meshes, referred to as homologous meshes, which maintain a consistent vertex-to-vertex correspondence between subjects and through the movement sequence as shown in Figure 1.
Move4D was validated as a good alternative to marker-based MoCap using photogrammetry [41] and its homologous mesh topology allows for considering a set of fixed vertices as representative of anatomical landmark locations [42].
The subjects were scanned at high resolution (49,530 vertices per mesh and a texture image size of 4 megapixels) performing a calibration A-pose and seven movements: gait (60 fps), running (60 fps), squats (30 fps), jumping jacks (j-jacks, 30 fps), counter movement jump (jump, 60 fps), forward jump (f-jump, 30 fps), and turn jump (t-jump, 30 fps). During the sessions, the participants wore tight garments and were asked to remove any clothing or wearable that impeded or distorted their body shape.
The result of processing each scan in each timeframe was a mesh with a realistic texture. These meshes served, on the one hand, to render virtual images from 48 different points of view, and on the other hand, to export the position of the vertices representing the anatomical landmarks, our ground truth. The 3D keypoints were obtained by applying linear triangulation to the 2D keypoints estimated in the virtual images using MediaPipe.
Further details and the complete followed procedure for generating the whole dataset are described in [43], and the data are available in [39].

2.1.2. Description of the Data Used in Training and Validation

As mentioned in the previous section, the whole dataset contains 3D keypoints obtained from two different procedures. The first corresponds to the 3D positions triangulated from 2D keypoints estimated by the MediaPipe pose network from various points of view. The second corresponds to the 3D anatomical landmarks extracted from the homologous mesh for the same subject and timeframe.
From the whole dataset, we specifically utilized data related to the lower limbs. In particular, we utilized 13 out of the 33 MediaPipe keypoints and 18 out of the 53 available anatomical landmarks. Figure 2 illustrates the selection of keypoints used in this work.
Anthropometric data, including weight, height, and sex, were used as inputs in all the models.
The dataset was randomly split into two subsets: the training set, composed of 57 subjects, and the test set, composed of 14 subjects. Table 1 shows the anthropometric characteristics of each subset. The code for the subjects in the test subset used in this work is detailed in Appendix A.

2.2. Data Pre-Processing

We first substituted R-PSIS and L-PSIS with a new anatomical landmark, SACR, defined as SACR = (R-PSIS + L-PSIS)/2. We carried this out following the recommendation of the International Society of Biomechanics (ISB) to calculate the reference framework for the pelvis, which allows for the use of the midpoint between R-PSIS and L-PSIS [7].
The position of the 3D keypoints and anatomical landmarks was expressed with respect to the midpoint of the hip keypoints and was also scaled by the subject’s height. Finally, min–max normalization was applied to height and weight inputs, considering their value ranges in the whole dataset.
Data augmentation was performed on the training subset sequences by applying multiple rotations around the vertical axis.

2.3. Marker Augmentation Models

To investigate the effect of the type of marker augmentation model on the estimation errors of position and joint angles, we utilized three marker augmentation models. These models estimate the anatomical landmarks required for calculating the joint kinematics of the lower limb. We selected these models to represent neural networks of varying complexity, enabling us to examine their suitability for marker augmentation purposes. The models presented are listed below:
  • MLP: a multilayer perceptron with a rectified linear unit activation function (ReLU). This model is less accurate but very lightweight, allowing its implementation on devices with low resources such as mobile phones or other low-cost devices.
  • LSTM: an adaptation of the long-short term memory neural network used in OpenCap for the full body [25]. It uses temporal but not spatial information.
  • Transformer [44]: designed for a comprehensive understanding of the problem, capturing long-range dependences in the global context. This model improves upon the previous one by incorporating spatial information. Transformers have recently been used successfully in different problems. It is more resource intensive but more accurate.
The same numbers of inputs and outputs were set for all the models. They had 42 inputs: sex, weight, height, the 3 × 13 coordinates of the 3D keypoints, and 51 outputs, which were the 3 × 17 coordinates of the selected anatomical landmarks (Figure 2). The architectures and training parameters of the models are described in the following subsections. The development and the training of the models were carried out in Python 3.8.10 using Keras 2.7.0 and Tensorflow backend 2.7.0 [45,46].

2.3.1. MLP Model

The multilayer perceptron model we developed consisted of several blocks (see Figure 3). Each block comprised a dense layer, followed by a batch normalization layer and an activation layer. Keras Tuner was used to determine the network’s hyperparameters [47]. Specifically, an optimization was performed to determine the number of blocks, the number of units in the dense layer, the activation function (hyperbolic tangent or ReLU), and the learning rate. This process took about one week using an Nvidia GTX 1050 Ti.
The final architecture comprised three blocks. The first one had 256 units in the dense layer with a hyperbolic tangent activation function. The activation functions for the second and third blocks were ReLU, with 128 and 224 units in the dense layer, respectively. The input layer consisted of 42 units (weight, height, sex, and the 3D coordinates of 13 keypoints from a specific frame). The output layer consisted of 51 units (3D coordinates of 17 anatomical points). The learning rate was set to 2 × 10−5, the optimizer selected was RMSprop, and the loss function was the mean squared error (MSE). The training process took 100 epochs with batch size of 64 (approximately one hour of training using an Nvidia GTX 1050 Ti).

2.3.2. LSTM Model

We adapted the input and output shape of the LSTM body model presented in [25] with a sequence length of 16 frames. This model comprised 2 LSTM layers with 128 units in each layer, followed by a dense layer with linear activation. The learning rate was set to 7 × 10−5, the optimizer selected was Adam, and the loss function was mean squared error (MSE), while the other parameters remained at their default values. The training process took 200 epochs with batch size of 32 (approximately twelve hours of training using an Nvidia GTX 1050 Ti). Figure 4 shows a diagram of the proposed LSTM.

2.3.3. Transformer Model

The model’s input consists of a series of 3D keypoints and anthropometric data. This 3D information tensor was initially transposed so that the number of keypoints matched the channel dimension, allowing for linear projection to the Transformer’s hidden dimension using a Conv2D with a kernel size of (1, 1).
Next, the sine positional embedding layer added spatial information to the matrix [48], which was then flattened to match the Transformer’s input shape. These projected features were passed through 6 self-attention layers with 14 heads, and then reshaped to match the input of the MLP head. The MLP head outputs the sequence of 3D anatomical landmark time series with a linear activation (see Figure 5).
The learning rate, optimizer hyperparameters, and loss function were the same as those used in the LSTM model. The training process took 300 epochs with batch size of 64 (approximately one day of training using two Nvidia RTX 3090).

2.4. Metrics

The Euclidean distance between observed and reconstructed anatomical landmarks, averaged over every movement sequence across all subjects in the test subset, was used to characterize the errors in the estimation of the anatomical landmarks positions, whereas errors in the joint angles were parametrized as the root mean squared error (RMSE), as in [12,25,49,50].
The anatomical axes and angles of the hip, knee, and ankle were calculated according to [7,51] using the positions of the anatomical landmarks. The hip joint center was obtained following the procedure described in [52].
The trochanterion landmarks were used as technical markers on the thighs in the calculations of the joint angles.

2.5. Error Analysis

Means and standard deviations of those errors were calculated, and their distributions for different movements, anatomical landmarks, and axes were compared across models. In order to quantify the influence of those factors on the size of the errors, linear mixed models (LMMs) were fitted using the subject as a random factor and the following fixed factors:
  • Prediction model, movement, and anatomical landmarks for the prediction of anatomical landmark locations;
  • Prediction model, side of the body joint, and rotation axis for the calculation of joint angles.
The LMMs also accounted for possible interactions between the effects of: (a) model and movements for both types of errors; (b) model and anatomical landmarks for errors in anatomical landmark locations; (c) model, joint, and rotation axis for errors in joint angles.
An analysis of variance (ANOVA) was used to compare those LMMs to others that also included the characteristics of the subjects (sex, height, and weight) as fixed factors in order to test whether the errors could be assumed to be independent of the subject’s anthropometry.
The analysis was carried out with the R 4.1.3. software for statistical computing [53], using the packages lmerTest, performance, and phia [54,55,56].

3. Results

In this section, we present the errors obtained with each model and the results of the statistical analysis performed.

3.1. Anatomical Landmark Position Errors

The average errors per anatomical landmark and movement fell within the following ranges: [0.77, 3.75] cm for the MLP model, [0.64, 2.74] cm for LSTM model, and [0.55, 2.11] cm for Transformer model.
All three evaluated models estimated the anatomical landmarks located around the pelvis (L/R-ASIS, L/R-TRO, SACR) with the greatest errors (ranging from 1.88 to 2.23 cm), while those around the ankles (L/R-LM, L/R-MM, L/R-CAL) were estimated with the lowest errors (ranging from 0.88 to 1.36 cm).
Regarding the errors per movement, the greatest errors were observed in the running movement, with mean errors of 2.73 cm for the MLP model, 1.84 cm for the LSTM model, and 1.81 cm for the Transformer model. Errors for the A-pose were the smallest, around 1.1 cm for all models, followed by jumps and squats, for which errors were approximately 1.57 for MLP, 1.31 cm for LSTM, and 1.36 cm for Transformer. The complete anatomical landmark position errors are shown in Table 2 (MLP), Table 3 (LSTM), and Table 4 (Transformer).

3.2. Joint Angle Errors

The RMSDs per movement and axis were limited within the following ranges: [2.52, 15.35] degrees for the MLP model, [1.78, 9.32] degrees for the LSTM model, and [1.91, 7.13] degrees for the Transformer model.
While the worst results in the MLP model were found in the ankle angles (over 5.8 degrees), the LSTM and Transformer models obtained the worst results in the hip angles. The knee angles were generally well estimated (under 4.4 degrees), except for the rotation in the MLP model (9.06 degrees).
All the models obtained their worst results in the running movement (ranging from 4.32 to 9.29 degrees) and the best results in jump and squat movements (ranging from 3.19 to 4.93 degrees).
Table 5, Table 6 and Table 7 show detailed joint angle errors for the MLP, LSTM, and Transformer models, respectively.

3.3. Factors Influencing the Errors

The subjects’ characteristics had no significant influence on the errors (p = 0.573 for anatomical landmark errors, p = 0.758 for joint angle errors). Table 8 and Table 9 show the results of the ANOVA for the LMM fitted without those characteristics; values of the statistical tests are omitted, since due to the large amount of data points, the null hypothesis (no effect of the factors) would always be rejected even for negligible effect sizes.
The R2 values in those tables show that around half of the variance was random error not explained by the considered factors and that the random influence of the subjects (difference between conditional and marginal R2) was also small [57]. The sums of squares show that, for anatomical landmark position errors, the effect of the model was smaller than the effects of the movement and the anatomical landmark and that the effects of the interactions were one order of magnitude smaller. For joint angle errors, on the other hand, the model was the greatest source of variation, and there were also important interactions between effects, especially between those of the joint and the rotation axis; the side of the body, however, barely affected the results.
Figure 6, Figure 7, Figure 8 and Figure 9 represent the distributions of the observed errors and their expected values according to the LMM, accounting for different factors and their interactions with the model. An advantage of the LSTM and the Transformer over the MLP model can be observed for both anatomical landmark position and joint angle errors, although the improvement is less than 1 cm and 3 degrees, respectively (Table 10). The performances of the LSTM and the Transformer models are similar, with a small advantage (<1 degree on average) for the Transformer in gait and running joint angles, mostly due to differences in hip flexion–extension.

4. Discussion

This study aimed to assess the suitability of using a marker augmentation model to convert the keypoints detected in images by standard pose estimation networks into anatomical landmarks, enabling the calculation of joint kinematics. Additionally, the study aimed to identify the factors that mainly affect their performance.

4.1. Size of the Joint Angle and Landmark Position Errors

To gain a comprehensive understanding of the results obtained, we reviewed studies that focus on comparing errors between markerless and marker-based photogrammetry MoCap systems and referred to research on the effect of marker placement to assess the size of errors in our results [58].
Markerless vs. marker-based studies typically reported errors in the positions of landmarks, joint centers, and joint angles along various axes (such as hip, knee, and ankle flexo-extension and hip abduction–adduction and rotation). In all cases, our comparisons were always performed with the lowest reported errors, such as those from studies using the highest-resolution pose estimators and the maximum number of cameras, or those reported under conditions similar to our work (e.g., studies involving sports clothing).
The anatomical landmark position errors of all the models in the reference A-pose were of the same order of magnitude as the intra-examiner position errors reported in work [58], indicating that our models and marker-based MoCap systems yielded similar levels of anatomical landmark position uncertainty.
We took the joint angle errors observed in gait movement as the reference values with which to compare the effects of marker position reported in work [58]. The magnitude of joint angle errors was similar to or lower than the uncertainty in joint rotations typically introduced by inter-examiner marker positioning errors in all three models, except for hip flexion–extension in the LSTM model (9.32 degrees vs. 5 degrees) and ankle flexion–extension in the MLP model (5.33 degrees vs. 3.3 degrees).
We found several references reporting anatomical landmarks or joint position errors for gait movement [12,25,36,59], with the smallest average error being 1.22 cm [36]. The Transformer model was the only one that achieved comparable errors (1.42 cm). MLP and LSTM errors were of the same order of magnitude as or smaller than the errors in the rest of the studies (ranging between an average of 1.81 and 2.97 cm).
We could also compare the anatomical landmark position errors in running to those reported in [60]. LSTM and Transformer models exhibited smaller errors compared to the literature (1.84 vs. 2.32 cm), whereas the MLP model did not.
With respect to squats and jumps, we verified that the anatomical landmark position errors of all three models (each under 1.6 cm) improved upon those reported in the literature [25] (over 2.2 cm).
When considering joint angle errors, we observed that for gait movement, the errors for the Transformer model (average of 3.37 degrees) were lower than those found in the literature. Next, the performance of LSTM model was comparable to that reported in [25] (4.77 degrees vs. 4.76 degrees). The MLP model achieved joint angle errors of 5.68 degrees, slightly lower compared to those reported in other literature (ranging from 6.9 degrees and above) [12,49,59,61].
The joint angle errors in running for the Transformer model were the lowest found, with an average of 4.32 degrees, followed by the errors in LSTM (5.29 degrees) and those reported in [61] (6.26 degrees). The results of the MLP model (9.29 degrees) and [49] were far from the best achieved.
The comparison conducted for the squats and jumps led to similar conclusions. All three models improved the errors in ankle (equal to or under 4.17 degrees) and knee (equal to or under 3.4 degrees) flexo-extension, as well as hip abduction–adduction (equal to or under 3.1 degrees). However, only the Transformer model showed improvement in hip flexion–extension error.

4.2. Factors Influencing the Errors

The statistical analysis revealed that the anthropometric characteristics of the subjects did not significantly affect the errors in the positions of anatomical landmarks or joint angles. Therefore, all the variation that could be attributed to the physical characteristics of the subjects was fully captured by the model and did not have a significant influence on the errors.
Only half of the error variance could be explained by the considered factors, which included model, movement, and anatomical landmark for anatomical landmark position errors, and model, movement, joint, axis, and side for joint angle errors. The factors that most influenced the errors were the anatomical landmark and movement for anatomical landmark position errors, and model, rotation axis, and movement for joint angle errors. The interaction between joint and rotation axis was particularly relevant.
In general, anatomical landmark position errors were greater for the MLP model and smaller for the LSTM and Transformer models, depending on the movement.
The anatomical landmark position errors reported for all the models followed a common pattern. Errors of the anatomical landmarks located on the pelvis (approximately 1.95 cm for ASIS, TRO, and SACR) were slightly greater than those of landmarks on the knees (LFE and MFE) and TOE3 (ranging from 1.43 to 1.80 cm). The smallest errors (ranging from 1.05 to 1.60 cm) were found in the anatomical landmarks in the ankle area (MM, LM, and CAL). Regarding anatomical landmark position errors per movement, the greatest errors were observed in running in all models (ranging from 1.81 to 2.73 cm), while the smallest errors were found in jumps and squats (ranging from 1.31 to 1.58 cm). These error patterns were consistent with those found in the literature.
The joint angle errors in the Transformer model were generally the smallest for all the movements, while MLP provided the highest errors.
The axial rotation angle had the greatest error in all models, ranging from 3.64 to 9.06 degrees. Flexion–extension errors averaged 3.56 degrees for the Transformer and 4.62 degrees for LSTM, which were larger than the abduction–adduction errors, except for MLP (5.18 degrees). The abduction–adduction axis yielded the best results for the LSTM (3.22 degrees) and Transformer models (2.86 degrees).
Similar to anatomical landmark position errors, errors in the running movement were the largest for all models (9.29 degrees for MLP, 5.29 degrees for LSTM, and 4.32 degrees for Transformer). Conversely, joint angle errors were smallest in jump and squat movements (5.03 degrees for MLP, 3.55 degrees for LSTM, and 3.29 degrees for Transformer).
Finally, we observed that flexion–extension angle errors at the ankles tended to be the lowest, averaging between 2.92 and 5.84 degrees, whereas the abduction–adduction angle error at the hips was the smallest, with errors ranging from 2.13 to 3.32 degrees. In the knee, rotation angles were commonly the worst estimated, with errors ranging between 3.64 and 9.06 degrees.

4.3. Other Considerations

The data used for this study were specifically designed to minimize the 3D reconstruction errors of keypoints in images. It is worth noting that 48 different points of view were generated per timeframe. Further research will focus on assessing the proposed models using a real setup with a specific number of cameras that provide an equivalent reconstruction error.
An interesting finding was that the proposed Transformer model exhibited similar errors for anatomical landmark position compared to the LSTM model, while demonstrating smaller errors for joint angles.
The Transformer model, by performing self-attention for each patch of the input sequence in an N-to-N manner, outputs a set of context-enriched features, thus obtaining a global understanding of the problem [44]. This comprehensive understanding enables the model to effectively capture long-range dependencies and intricate relationships between keypoints in 3D space. By assigning varying degrees of importance to different patches based on their relevance to one another, the Transformer excels in discerning the underlying structure of the keypoints, ultimately leading to superior performance in understanding augmented sets of 3D anatomical landmarks and achieving lower angle errors.
As discussed, the Transformer model improved the anatomical landmark position errors of the LSTM model in f-jump, gait, and running movements, the ones that precisely involve the translation of the body in the anteroposterior axis.
The analysis of the anatomical landmark position and the joint angle errors indicated that the movement factor was particularly relevant. Our results revealed that errors in running were consistently 10% to 45% bigger than errors in other movements. Therefore, it may be reasonable to develop a model specifically tailored for running assessment.

5. Conclusions

An assessment of three deep learning models for marker augmentation was conducted, revealing distinct performance levels across each model.
Through the testing and comparison across seven movements and three models (MLP, LSTM, Transformer), we identified the MLP model as the one which provides lower accuracy in terms of both anatomical landmark position errors and joint angle errors. The LSTM and Transformer models provided similar results, surpassing the Transformer model the LSTM model in the joint angle errors. While the joints angle errors in the LSTM model ranged from 3.5 to 5.29 degrees, in the Transformer model, they ranged from 3.19 to 4.32 degrees. The Transformer model might achieve a global understanding of the keypoints’ 3D relationships utilizing self-attention mechanisms. However, the selection of the model depends on the final application and the increase in accuracy is usually accompanied by an increase in model parameters.
The anthropometric characteristics of the subjects had no significant impact on the errors associated with anatomical landmarks or joint angles, suggesting that the models’ performance is robust across different body types and sizes.
The analysis revealed that errors in running were consistently higher than in other movements, manifesting the influence of movement on the behavior of the models.
Anatomical landmarks were another factor that influenced the magnitude of errors. We observed that those placed on the pelvis are prone to having the biggest errors, whereas those on the ankle had the lowest.
Hip abduction–adduction and ankle flexion–extension angles were the best estimated at each joint for all models. Conversely, knee rotation angles were poorly estimated by all three models.
This work introduces a new framework to the research community and is expected to contribute to the enhancement of markerless MoCap models.

Author Contributions

Conceptualization, A.V.R.-N., E.M.-R., H.d.R., E.P. and M.C.J.L.; methodology, A.V.R.-N., E.M.-R. and E.P.; software, E.M.-R., H.d.R. and J.S.N.; validation, E.M.-R. and H.d.R.; formal analysis, A.V.R.-N., E.M.-R. and H.d.R.; investigation, A.V.R.-N., E.M.-R. and J.S.N.; resources, A.V.R.-N. and E.P.; data curation, A.V.R.-N. and E.M.-R.; writing—original draft preparation, A.V.R.-N., E.M.-R., H.d.R. and J.S.N.; writing—review and editing, E.M.-R., H.d.R., E.P. and M.C.J.L.; visualization, A.V.R.-N. and J.S.N.; supervision, E.M.-R., H.d.R., E.P. and M.C.J.L.; project administration, E.M.-R., H.d.R. and E.P.; funding acquisition, A.V.R.-N. All authors have read and agreed to the published version of the manuscript.

Funding

Research activity supported by Instituto Valenciano de Competitividad Empresarial (IVACE) and Valencian Regional Government (GVA), IMAMCA/2024; and project IMDEEA/2024, funding requested to Instituto Valenciano de Competitividad Empresarial (IVACE), call for proposals 2024, for Technology Centers of the Valencian Region, funded by European Union.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The “Human tracking dataset of 3D anatomical landmarks and pose keypoints” is available at https://data.mendeley.com/datasets/493s6f753v/2, accessed on 31 January 2024. Neural network architecture, data pre-processing, and training hyperparameters are described within the article. Appendix A details the content in the test subset and details the anatomical landmarks taken from the dataset.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

The following table relates the denomination used in this work to the names of the anatomical landmarks denoted in the dataset used.
Anatomical LandmarksName in Dataset
L-ASISLt ASIS
L-CALLt Calcaneous Post
L-LFELt Femoral Lateral Epicn
L-LMLt Lateral Malleolus
L-MFELt Femoral Medial Epicn
L-MMLt Medial Malleolus
L-TROLt Trochanterion
L-TOE3Lt Digit II
R-ASISRt ASIS
R-CALRt Calcaneous Post
R-LFERt Femoral Lateral Epicn
R-LMRt Lateral Malleolus
R-MFERt Femoral Medial Epicn
R-MMRt Medial Malleolus
R- TRORt Trochanterion
R-TOE3Rt Digit II
SACRThis point is midpoint between Lt PSIS and Rt PSIS
The subjects in the test set were the following (the rest remaining in the training set): TDB_004_F, TDB_011_M, TDB_028_M, TDB_032_F, TDB_035_F, TDB_037_M, TDB_038_M, TDB_041_F, TDB_042_F, TDB_049_F, TDB_053_M, TDB_055_M, TDB_061_M, and TDB_071_F.

References

  1. Marras, W.S.; Ferguson, S.A.; Gupta, P.; Bose, S.; Parnianpour, M.; Kim, J.Y.; Crowell, R.R. The Quantification of Low Back Disorder Using Motion Measures. Methodology and Validation. Spine 1999, 24, 2091–2100. [Google Scholar] [CrossRef]
  2. Ferrario, V.F.; Sforza, C.; Poggio, C.E.; Schmitz, J.H.; Tartaglia, G. A Three-dimensional Non-invasive Study of Head Flexion and Extension in Young Non-patient Subjects. J. Oral Rehabil. 1997, 24, 361–368. [Google Scholar] [CrossRef]
  3. Bulgheroni, M.V.; Antonaci, F.; Sandrini, G.; Ghirmai, S.; Nappi, G.; Pedotti, A. A 3D Kinematic Method to Evaluate Cervical Spine Voluntary Movements in Humans. Funct. Neurol. 1998, 3, 239–245. [Google Scholar]
  4. Lu, T.-W.; Chang, C.-F. Biomechanics of Human Movement and Its Clinical Applications. Kaohsiung J. Med. Sci. 2012, 28, S13–S25. [Google Scholar] [CrossRef]
  5. Whittle, M.W. Clinical Gait Analysis: A Review. Hum. Mov. Sci. 1996, 15, 369–387. [Google Scholar] [CrossRef]
  6. Best, R.; Begg, R. Overview of Movement Analysis and Gait Features. In Computational Intelligence for Movement Sciences: Neural Networks and Other Emerging Techniques; IGI Global: Hershey, PA, USA, 2006; pp. 1–69. [Google Scholar]
  7. Wu, G.; Siegler, S.; Allard, P.; Kirtley, C.; Leardini, A.; Rosenbaum, D.; Whittle, M.; D’Lima, D.D.; Cristofolini, L.; Witte, H.; et al. ISB Recommendation on Definitions of Joint Coordinate System of Various Joints for the Reporting of Human Joint Motion—Part I: Ankle, Hip, and Spine. J. Biomech. 2002, 35, 543–548. [Google Scholar] [CrossRef] [PubMed]
  8. Wu, G.; van der Helm, F.C.T.; Veeger, H.E.J.; Makhsous, M.; Van Roy, P.; Anglin, C.; Nagels, J.; Karduna, A.R.; McQuade, K.; Wang, X.; et al. ISB Recommendation on Definitions of Joint Coordinate Systems of Various Joints for the Reporting of Human Joint Motion—Part II: Shoulder, Elbow, Wrist and Hand. J. Biomech. 2005, 38, 981–992. [Google Scholar] [CrossRef]
  9. Della Croce, U.; Leardini, A.; Chiari, L.; Cappozzo, A. Human Movement Analysis Using Stereophotogrammetry: Part 4: Assessment of Anatomical Landmark Misplacement and Its Effects on Joint Kinematics. Gait Posture 2005, 21, 226–237. [Google Scholar] [CrossRef] [PubMed]
  10. Chiari, L.; Croce, U.D.; Leardini, A.; Cappozzo, A. Human Movement Analysis Using Stereophotogrammetry: Part 2: Instrumental Errors. Gait Posture 2005, 21, 197–211. [Google Scholar] [CrossRef]
  11. Cappozzo, A.; Della Croce, U.; Leardini, A.; Chiari, L. Human Movement Analysis Using Stereophotogrammetry: Part 1: Theoretical Background. Gait Posture 2005, 21, 186–196. [Google Scholar]
  12. Kanko, R.M.; Laende, E.K.; Davis, E.M.; Selbie, W.S.; Deluzio, K.J. Concurrent Assessment of Gait Kinematics Using Marker-Based and Markerless Motion Capture. J. Biomech. 2021, 127, 110665. [Google Scholar] [CrossRef]
  13. Captury—Markerless Motion Capture Technology. Available online: https://captury.com/ (accessed on 18 January 2024).
  14. Theia Markerless—Markerless Motion Capture Redefined. Available online: https://www.theiamarkerless.ca/ (accessed on 18 January 2024).
  15. Iosa, M.; Picerno, P.; Paolucci, S.; Morone, G. Wearable Inertial Sensors for Human Movement Analysis. Expert Rev. Med. Devices 2016, 13, 641–659. [Google Scholar] [CrossRef]
  16. Picerno, P.; Iosa, M.; D’Souza, C.; Benedetti, M.G.; Paolucci, S.; Morone, G. Wearable Inertial Sensors for Human Movement Analysis: A Five-Year Update. Expert Rev. Med. Devices 2021, 18, 79–94. [Google Scholar] [CrossRef]
  17. RajKumar, A.; Vulpi, F.; Bethi, S.R.; Wazir, H.K.; Raghavan, P.; Kapila, V. Wearable Inertial Sensors for Range of Motion Assessment. IEEE Sens. J. 2020, 20, 3777–3787. [Google Scholar] [CrossRef]
  18. Kendall, A.; Grimes, M.; Cipolla, R. Posenet: A Convolutional Network for Real-Time 6-Dof Camera Relocalization. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2938–2946. [Google Scholar]
  19. Insafutdinov, E.; Pishchulin, L.; Andres, B.; Andriluka, M.; Schiele, B. DeeperCut: A Deeper, Stronger, and Faster Multi-Person Pose Estimation Model. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 34–50. [Google Scholar]
  20. Xiao, B.; Wu, H.; Wei, Y. Simple Baselines for Human Pose Estimation and Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 472–487. [Google Scholar]
  21. Mathis, A.; Mamidanna, P.; Cury, K.M.; Abe, T.; Murthy, V.N.; Mathis, M.W.; Bethge, M. DeepLabCut: Markerless Pose Estimation of User-Defined Body Parts with Deep Learning. Nat. Neurosci. 2018, 21, 1281–1289. [Google Scholar] [CrossRef] [PubMed]
  22. Martınez Hidalgo, G. OpenPose: Whole-Body Pose Estimation; Carnegie Mellon University: Pittsburgh, PA, USA, 2019. [Google Scholar]
  23. Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.-L.; Yong, M.G.; Lee, J.; et al. MediaPipe: A Framework for Building Perception Pipelines. arXiv 2019, arXiv:1906.08172. [Google Scholar]
  24. He, Y.; Yan, R.; Fragkiadaki, K.; Yu, S.-I. Epipolar Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 7776–7785. [Google Scholar]
  25. Uhlrich, S.D.; Falisse, A.; Kidziński, Ł.; Muccini, J.; Ko, M.; Chaudhari, A.S.; Hicks, J.L.; Delp, S.L. OpenCap: Human Movement Dynamics from Smartphone Videos. PLoS Comput. Biol. 2023, 19, e1011462. [Google Scholar] [CrossRef] [PubMed]
  26. Wade, L.; Needham, L.; McGuigan, P.; Bilzon, J. Applications and Limitations of Current Markerless Motion Capture Methods for Clinical Gait Biomechanics. PeerJ 2022, 10, e12995. [Google Scholar] [CrossRef] [PubMed]
  27. Sigal, L.; Balan, A.O.; Black, M.J. HumanEva: Synchronized Video and Motion Capture Dataset and Baseline Algorithm for Evaluation of Articulated Human Motion. Int. J. Comput. Vis. 2010, 87, 4–27. [Google Scholar] [CrossRef]
  28. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1325–1339. [Google Scholar] [CrossRef]
  29. Carnegie Mellon University—CMU Graphics Lab—Motion Capture Library. Available online: http://mocap.cs.cmu.edu/ (accessed on 15 February 2024).
  30. Kwolek, B.; Michalczuk, A.; Krzeszowski, T.; Switonski, A.; Josinski, H.; Wojciechowski, K. Calibrated and Synchronized Multi-View Video and Motion Capture Dataset for Evaluation of Gait Recognition. Multimed. Tools Appl. 2019, 78, 32437–32465. [Google Scholar] [CrossRef]
  31. Rosskamp, J.; Weller, R.; Zachmann, G. Effects of Markers in Training Datasets on the Accuracy of 6D Pose Estimation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 4–8 January 2024; pp. 4457–4466. [Google Scholar]
  32. Kanko, R.M.; Outerleys, J.B.; Laende, E.K.; Selbie, W.S.; Deluzio, K.J. Comparison of Concurrent and Asynchronous Running Kinematics and Kinetics from Marker-Based Motion Capture and Markerless Motion Capture under Two Clothing Conditions. bioRxiv 2023. [Google Scholar] [CrossRef]
  33. Kanko, R.M.; Laende, E.K.; Strutzenberger, G.; Brown, M.; Selbie, W.S.; DePaul, V.; Scott, S.H.; Deluzio, K.J. Assessment of Spatiotemporal Gait Parameters Using a Deep Learning Algorithm-Based Markerless Motion Capture System. J. Biomech. 2021, 122, 110414. [Google Scholar] [CrossRef] [PubMed]
  34. Kanko, R.M.; Laende, E.; Selbie, W.S.; Deluzio, K.J. Inter-Session Repeatability of Markerless Motion Capture Gait Kinematics. J. Biomech. 2021, 121, 110422. [Google Scholar] [CrossRef] [PubMed]
  35. Vafadar, S.; Skalli, W.; Bonnet-Lebrun, A.; Khalifé, M.; Renaudin, M.; Hamza, A.; Gajny, L. A Novel Dataset and Deep Learning-Based Approach for Marker-Less Motion Capture during Gait. Gait Posture 2021, 86, 70–76. [Google Scholar] [CrossRef] [PubMed]
  36. Vafadar, S.; Skalli, W.; Bonnet-Lebrun, A.; Assi, A.; Gajny, L. Assessment of a Novel Deep Learning-Based Marker-Less Motion Capture System for Gait Study. Gait Posture 2022, 94, 138–143. [Google Scholar] [CrossRef] [PubMed]
  37. Needham, L.; Evans, M.; Wade, L.; Cosker, D.P.; McGuigan, M.P.; Bilzon, J.L.; Colyer, S.L. The Development and Evaluation of a Fully Automated Markerless Motion Capture Workflow. J. Biomech. 2022, 144, 111338. [Google Scholar] [CrossRef]
  38. Delp, S.L.; Anderson, F.C.; Arnold, A.S.; Loan, P.; Habib, A.; John, C.T.; Guendelman, E.; Thelen, D.G. OpenSim: Open-Source Software to Create and Analyze Dynamic Simulations of Movement. IEEE Trans. Biomed. Eng. 2007, 54, 1940–1950. [Google Scholar] [CrossRef]
  39. Ruescas Nicolau, A.V.; Medina-Ripoll, E.; Parrilla Bernabé, E.; De Rosario Martínez, H. Human Tracking Dataset of 3D Anatomical Landmarks and Pose Keypoints. 2024. Available online: https://doi.org/10.17632/493s6f753v.2 (accessed on 1 January 2024).
  40. Parrilla, E.; Ballester, A.; Parra, P.; Ruescas, A.; Uriel, J.; Garrido, D.; Alemany, S. MOVE 4D: Accurate High-Speed 3D Body Models in Motion. In Proceedings of the 3DBODY.TECH 2019, Lugano, Switzerland, 22–23 October 2019; pp. 30–32. [Google Scholar]
  41. Ruescas Nicolau, A.V.; De Rosario, H.; Della-Vedova, F.B.; Bernabé, E.P.; Juan, M.-C.; López-Pascual, J. Accuracy of a 3D Temporal Scanning System for Gait Analysis: Comparative With a Marker-Based Photogrammetry System. Gait Posture 2022, 97, 28–34. [Google Scholar] [CrossRef]
  42. Ruescas-Nicolau, A.V.; De Rosario, H.; Bernabé, E.P.; Juan, M.-C. Positioning Errors of Anatomical Landmarks Identified by Fixed Vertices in Homologous Meshes. Gait Posture 2024, 108, 215–221. [Google Scholar] [CrossRef]
  43. Ruescas Nicolau, A.V.; Medina Ripoll, E.J.; Parrilla Bernabé, E.; De Rosario, H. Multimodal Human Motion Dataset of 3d Anatomical Landmarks and Pose Keypoints. Data Brief 2024, 53, 110157. [Google Scholar] [CrossRef]
  44. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  45. Chollet, F. Others Keras. 2015. Available online: https://github.com/fchollet/keras (accessed on 1 January 2024).
  46. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  47. O’Malley, T.; Bursztein, E.; Long, J.; Chollet, F.; Jin, H.; Invernizzi, L. Others KerasTuner. 2019. [Google Scholar]
  48. Yang, S.; Quan, Z.; Nie, M.; Yang, W. TransPose: Keypoint Localization via Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision 2021, Montreal, BC, Canada, 11–17 October 2021; pp. 11802–11812. [Google Scholar]
  49. Song, K.; Hullfish, T.J.; Scattone Silva, R.; Silbernagel, K.G.; Baxter, J.R. Markerless Motion Capture Estimates of Lower Extremity Kinematics and Kinetics Are Comparable to Marker-Based across 8 Movements. J. Biomech. 2023, 157, 111751. [Google Scholar] [CrossRef]
  50. Horsak, B.; Eichmann, A.; Lauer, K.; Prock, K.; Krondorfer, P.; Siragy, T.; Dumphart, B. Concurrent Validity of Smartphone-Based Markerless Motion Capturing to Quantify Lower-Limb Joint Kinematics in Healthy and Pathological Gait. J. Biomech. 2023, 159, 111801. [Google Scholar] [CrossRef] [PubMed]
  51. Grood, E.S.; Suntay, W.J. A Joint Coordinate System for the Clinical Description of Three-Dimensional Motions: Application to the Knee. J. Biomech. Eng. 1983, 105, 136–144. [Google Scholar] [CrossRef] [PubMed]
  52. Harrington, M.E.; Zavatsky, A.B.; Lawson, S.E.M.; Yuan, Z.; Theologis, T.N. Prediction of the Hip Joint Centre in Adults, Children, and Patients with Cerebral Palsy Based on Magnetic Resonance Imaging. J. Biomech. 2007, 40, 595–602. [Google Scholar] [CrossRef]
  53. Team, R.D.C. R: A Language and Environment for Statistical Computing; R Core Team: Vienna, Austria, 2010. [Google Scholar]
  54. Kuznetsova, A.; Brockhoff, P.B.; Christensen, R.H.B. lmerTest Package: Tests in Linear Mixed Effects Models. J. Stat. Softw. 2017, 82, 1–26. [Google Scholar] [CrossRef]
  55. Lüdecke, D.; Ben-Shachar, M.S.; Patil, I.; Waggoner, P.; Makowski, D. Performance: An R Package for Assessment, Comparison and Testing of Statistical Models. J. Open Source Softw. 2021, 6, 3139. [Google Scholar] [CrossRef]
  56. De Rosario, H. “Phia: Post-Hoc Interaction Analysis”. R Package Version 0.3-1. 2024. Available online: https://CRAN.R-project.org/package=phia (accessed on 13 February 2024).
  57. Nakagawa, S.; Schielzeth, H. A General and Simple Method for Obtaining R2 from Generalized Linear Mixed-effects Models. Methods Ecol. Evol. 2013, 4, 133–142. [Google Scholar] [CrossRef]
  58. Della Croce, U.; Cappozzo, A.; Kerrigan, D.C. Pelvis and Lower Limb Anatomical Landmark Calibration Precision and Its Propagation to Bone Geometry and Joint Angles. Med. Biol. Eng. Comput. 1999, 37, 155–161. [Google Scholar] [CrossRef]
  59. Ripic, Z.; Nienhuis, M.; Signorile, J.F.; Best, T.M.; Jacobs, K.A.; Eltoukhy, M. A Comparison of Three-Dimensional Kinematics between Markerless and Marker-Based Motion Capture in Overground Gait. J. Biomech. 2023, 159, 111793. [Google Scholar] [CrossRef] [PubMed]
  60. Kanko, R.M.; Outerleys, J.B.; Laende, E.K.; Selbie, W.S.; Deluzio, K.J. Comparison of Concurrent and Asynchronous Running Kinematics and Kinetics From Marker-Based and Markerless Motion Capture Under Varying Clothing Conditions. J. Appl. Biomech. 2024, 1, 1–9. [Google Scholar] [CrossRef] [PubMed]
  61. Wren, T.A.L.; Isakov, P.; Rethlefsen, S.A. Comparison of Kinematics between Theia Markerless and Conventional Marker-Based Gait Analysis in Clinical Patients. Gait Posture 2023, 104, 9–14. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Vertices that correspond to anatomical landmarks on homologous meshes showing the vertex-to-vertex correspondence between subjects (left) and throughout the sequence (right). The same landmarks correspond to the same vertex ID in all the meshes. Details on the left Femoral Lateral Epicondyle are shown.
Figure 1. Vertices that correspond to anatomical landmarks on homologous meshes showing the vertex-to-vertex correspondence between subjects (left) and throughout the sequence (right). The same landmarks correspond to the same vertex ID in all the meshes. Details on the left Femoral Lateral Epicondyle are shown.
Sensors 24 01923 g001
Figure 2. Selected keypoints estimated from MediaPipe pose network (blue) and 3D anatomical landmarks (green) extracted from a homologous mesh.
Figure 2. Selected keypoints estimated from MediaPipe pose network (blue) and 3D anatomical landmarks (green) extracted from a homologous mesh.
Sensors 24 01923 g002
Figure 3. MLP model architecture diagram.
Figure 3. MLP model architecture diagram.
Sensors 24 01923 g003
Figure 4. LSTM model architecture diagram.
Figure 4. LSTM model architecture diagram.
Sensors 24 01923 g004
Figure 5. Transformer model architecture diagram.
Figure 5. Transformer model architecture diagram.
Sensors 24 01923 g005
Figure 6. Errors in anatomical landmark positions depending on the movement and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Figure 6. Errors in anatomical landmark positions depending on the movement and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Sensors 24 01923 g006
Figure 7. Errors in anatomical landmark positions depending on the anatomical landmark and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Figure 7. Errors in anatomical landmark positions depending on the anatomical landmark and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Sensors 24 01923 g007
Figure 8. Errors in joint angles depending on the movement and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Figure 8. Errors in joint angles depending on the movement and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Sensors 24 01923 g008
Figure 9. Errors in joint angles depending on the joint, rotation axis and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Figure 9. Errors in joint angles depending on the joint, rotation axis and the model across all subjects in the test subset. (Left): observed distributions. (Right): marginal means of the LMM plus/minus their standard errors.
Sensors 24 01923 g009
Table 1. Anthropometric characteristics of the dataset and training and test subsets.
Table 1. Anthropometric characteristics of the dataset and training and test subsets.
SubsetSex (Size)Height (m)
Mean (std)
Weight (kg)
Mean (std)
Age (years)
Mean (std)
TrainingFemale (N = 29)1.64 (0.08)70.89 (18.19)38.21 (13.27)
Male (N = 28)1.78 (0.11)79.77 (17.83)37.04 (11.27)
Total (N = 57)1.71 (0.12)75.25 (18.41)37.63 (12.23)
TestFemale (N = 7)1.58 (0.08)55.57 (11.40)34.86 (11.36)
Male (N = 7)1.75 (0.08)78.54 (11.49)31.14 (9.41)
Total (N = 14)1.67 (0.12)67.06 (16.22)33.00 (10.21)
DatasetTotal (N = 71)1.70 (0.12)73.64 (18.18)36.72 (11.94)
Table 2. Errors (mean Euclidean distance in cm) for each anatomical landmark and movement across all subjects in the test subset with the MLP model.
Table 2. Errors (mean Euclidean distance in cm) for each anatomical landmark and movement across all subjects in the test subset with the MLP model.
Anatomical
Landmark (AL)
A-PoseGaitF-JumpJ-JacksJumpRunningSquatsT-JumpAcross
Movement
L-ASIS1.492.182.181.961.972.962.092.212.13
L-CAL0.981.841.371.911.362.651.391.551.63
L-LFE1.371.581.601.851.502.531.811.721.75
L-LM0.821.391.271.601.242.211.081.421.38
L-MFE0.991.851.701.911.592.931.801.691.81
L-MM0.821.601.451.681.332.421.181.351.48
L-TRO1.371.781.911.951.972.821.932.201.99
L-TOE31.192.051.901.961.653.441.512.221.99
R-ASIS1.432.302.271.801.962.561.991.972.03
R-CAL1.041.571.521.821.372.531.221.751.60
R-LFE1.301.661.621.801.533.121.611.641.79
R-LM0.771.401.511.431.112.091.151.441.36
R-MFE1.371.961.641.681.553.271.561.741.85
R-MM1.001.311.471.411.242.181.181.411.40
R-TRO1.301.861.911.771.762.622.011.991.90
R-TOE31.262.071.641.631.533.751.472.051.93
SACR1.271.532.031.711.882.281.861.721.78
Across
AL
1.161.761.701.761.562.731.581.771.75
Table 3. Errors (mean Euclidean distance in cm) for each anatomical landmark and movement across all subjects in the test subset with the LSTM model.
Table 3. Errors (mean Euclidean distance in cm) for each anatomical landmark and movement across all subjects in the test subset with the LSTM model.
Anatomical
Landmark (AL)
A-PoseGaitF-JumpJ-JacksJumpRunningSquatsT-JumpAcross
Movement
L-ASIS1.682.382.741.912.232.212.392.312.23
L-CAL0.641.000.921.200.831.340.690.920.94
L-LFE1.001.421.461.181.271.721.301.261.33
L-LM0.630.900.820.780.801.430.860.840.88
L-MFE0.952.061.641.431.262.411.331.391.56
L-MM0.651.130.971.060.881.630.800.861.00
L-TRO1.321.852.241.691.931.962.061.971.88
L-TOE31.031.481.371.311.031.900.911.191.28
R-ASIS1.842.422.632.052.262.122.262.292.23
R-CAL0.930.961.010.991.031.520.971.041.05
R-LFE0.821.501.131.121.231.981.231.241.28
R-LM0.881.190.900.981.011.780.960.981.08
R-MFE1.101.941.401.591.242.571.181.371.55
R-MM0.761.170.881.050.851.630.780.901.00
R-TRO1.621.792.101.741.941.672.102.061.88
R-TOE30.841.410.991.261.031.790.821.201.17
SACR1.291.601.751.551.661.661.621.631.60
Across
AL
1.061.541.471.351.321.841.311.381.41
Table 4. Errors (mean Euclidean distance in cm) for each anatomical landmark and movement across all subjects in the test subset with the Transformer model.
Table 4. Errors (mean Euclidean distance in cm) for each anatomical landmark and movement across all subjects in the test subset with the Transformer model.
Anatomical
Landmark (AL)
A-PoseGaitF-JumpJ-JacksJumpRunningSquatsT-JumpAcross
Movement
L-ASIS1.761.802.011.702.022.112.032.131.94
L-CAL0.861.050.991.000.961.530.900.991.03
L-LFE1.471.631.511.581.581.631.501.641.57
L-LM0.740.850.850.970.921.430.890.950.95
L-MFE1.561.731.661.721.611.691.541.661.65
L-MM0.771.041.001.160.941.460.981.001.04
L-TRO1.761.731.901.632.032.002.082.131.91
L-TOE31.131.521.591.511.252.221.161.651.50
R-ASIS1.731.832.021.802.071.931.972.191.94
R-CAL0.670.960.931.060.831.750.871.041.01
R-LFE1.201.421.581.391.331.891.401.461.46
R-LM0.550.980.910.930.731.840.780.910.95
R-MFE1.301.441.511.441.331.611.381.431.43
R-MM0.720.950.951.000.831.600.860.960.98
R-TRO1.721.811.901.681.951.941.972.091.88
R-TOE30.811.551.411.421.162.310.931.481.38
SACR1.781.771.921.761.921.841.782.021.85
Across
AL
1.211.421.451.401.381.811.351.511.44
Table 5. Average of RMSD (degrees) for each joint, axis, and movement across all subjects in the test subset with the MLP model.
Table 5. Average of RMSD (degrees) for each joint, axis, and movement across all subjects in the test subset with the MLP model.
MovementHipKneeAnkleAcross
Joint and Axis
FEAB-ADROTFEAB-ADROTFEAB-ADROT
Running7.395.219.044.676.5915.3510.112.8612.449.29
Gait5.133.136.713.553.648.085.338.736.775.68
F-Jump7.622.976.243.493.9410.585.829.018.996.52
J-Jacks5.252.527.963.893.777.737.067.746.635.84
T-Jump5.013.489.513.074.78.685.319.027.456.25
Jump6.082.86.063.063.855.934.176.925.484.93
Squats6.43.163.373.97.093.097.35.895.13
Across
movement
6.123.327.363.594.349.065.848.87.666.23
Table 6. Average of RMSD (degrees) for each joint, axis, and movement across all subjects in the test subset with the LSTM model.
Table 6. Average of RMSD (degrees) for each joint, axis, and movement across all subjects in the test subset with the LSTM model.
MovementHipKneeAnkleAcross
Joint and Axis
FEAB-ADROTFEAB-ADROTFEAB-ADROT
Running7.622.96.524.544.125.524.35.636.445.29
Gait9.322.046.534.292.963.732.935.645.484.77
F-Jump8.631.965.53.532.993.892.24.683.844.14
J-Jacks5.132.17.242.822.693.23.794.764.934.07
T-Jump7.152.316.653.242.673.372.9244.084.04
Jump6.821.785.23.42.423.192.293.943.53.61
Squats6.171.864.783.882.63.031.993.683.473.5
Across
movement
7.262.136.063.672.923.72.924.624.534.2
Table 7. Average of RMSD (degrees) for each joint, axis, and movement across all subjects in the test subset with the Transformer model.
Table 7. Average of RMSD (degrees) for each joint, axis, and movement across all subjects in the test subset with the Transformer model.
MovementHipKneeAnkleAcross
Joint and Axis
FEAB-ADROTFEAB-ADROTFEAB-ADROT
Running4.942.635.343.473.234.315.74.984.334.32
Gait3.991.914.782.922.23.333.494.173.523.37
F-Jump5.262.165.662.722.743.83.063.273.643.59
J-Jacks3.522.035.62.212.413.573.733.973.843.43
T-Jump4.912.437.132.762.983.843.23.314.023.84
Jump4.781.955.672.62.833.342.633.163.553.39
Squats4.072.075.092.742.993.31.992.633.853.19
Across
movement
4.492.175.612.782.773.643.43.643.823.59
Table 8. ANOVA table for anatomical landmark position errors (SS: sum of squares, MS: mean squares, DoF: degrees of freedom).
Table 8. ANOVA table for anatomical landmark position errors (SS: sum of squares, MS: mean squares, DoF: degrees of freedom).
SSMSDoF
Model137.8868.942
Movement381.4154.497
Anatomical landmark617.8938.6216
Model:Movement78.225.5914
Model:Anatomical landmark88.552.7732
Conditional R2 = 0.509, Marginal R2 = 0.431.
Table 9. ANOVA table for joint angle errors (SS: sum of squares, MS: mean squares, DoF: degrees of freedom).
Table 9. ANOVA table for joint angle errors (SS: sum of squares, MS: mean squares, DoF: degrees of freedom).
SSMSDoF
Model6745.123372.562
Movement2828.23471.376
Joint1032.79516.402
Axis3186.661593.332
Side67.5367.531
Model:Movement1233.55102.804
Model:Joint1454.16363.5412
Model:Axis1195.82298.954
Joint:Axis4614.301153.584
Model:Joint:Axis1593.14199.148
Conditional R2 = 0.538, Marginal R2 = 0.516.
Table 10. Marginal errors for the different models according to the fitted LMMs.
Table 10. Marginal errors for the different models according to the fitted LMMs.
MLPLSTMTransformer
Anatomical landmark distance (cm)1.751.411.44
RMSD (degrees)6.234.203.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ruescas-Nicolau, A.V.; Medina-Ripoll, E.; de Rosario, H.; Sanchiz Navarro, J.; Parrilla, E.; Juan Lizandra, M.C. A Deep Learning Model for Markerless Pose Estimation Based on Keypoint Augmentation: What Factors Influence Errors in Biomechanical Applications? Sensors 2024, 24, 1923. https://doi.org/10.3390/s24061923

AMA Style

Ruescas-Nicolau AV, Medina-Ripoll E, de Rosario H, Sanchiz Navarro J, Parrilla E, Juan Lizandra MC. A Deep Learning Model for Markerless Pose Estimation Based on Keypoint Augmentation: What Factors Influence Errors in Biomechanical Applications? Sensors. 2024; 24(6):1923. https://doi.org/10.3390/s24061923

Chicago/Turabian Style

Ruescas-Nicolau, Ana V., Enrique Medina-Ripoll, Helios de Rosario, Joaquín Sanchiz Navarro, Eduardo Parrilla, and María Carmen Juan Lizandra. 2024. "A Deep Learning Model for Markerless Pose Estimation Based on Keypoint Augmentation: What Factors Influence Errors in Biomechanical Applications?" Sensors 24, no. 6: 1923. https://doi.org/10.3390/s24061923

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop