Next Article in Journal
Fuzzy Cognitive Map Applications in Medicine over the Last Two Decades: A Review Study
Previous Article in Journal
Advances and Challenges in Wearable Glaucoma Diagnostics and Therapeutics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Gait Analysis Techniques with Markerless Pose Estimation Based on Smartphone Location

Department of Mechatronics Engineering, Incheon National University, Incheon 22012, Republic of Korea
*
Author to whom correspondence should be addressed.
Bioengineering 2024, 11(2), 141; https://doi.org/10.3390/bioengineering11020141
Submission received: 4 January 2024 / Revised: 25 January 2024 / Accepted: 29 January 2024 / Published: 30 January 2024

Abstract

:
Marker-based 3D motion capture systems, widely used for gait analysis, are accurate but have disadvantages such as cost and accessibility. Whereas markerless pose estimation has emerged as a convenient and cost-effective alternative for gait analysis, challenges remain in achieving optimal accuracy. Given the limited research on the effects of camera location and orientation on data collection accuracy, this study investigates how camera placement affects gait assessment accuracy utilizing five smartphones. This study aimed to explore the differences in data collection accuracy between marker-based systems and pose estimation, as well as to assess the impact of camera location and orientation on accuracy in pose estimation. The results showed that the differences in joint angles between pose estimation and marker-based systems are below 5°, an acceptable level for gait analysis, with a strong correlation between the two datasets supporting the effectiveness of pose estimation in gait analysis. In addition, hip and knee angles were accurately measured at the front diagonal of the subject and ankle angle at the lateral side. This research highlights the significance of careful camera placement for reliable gait analysis using pose estimation, serving as a concise reference to guide future efforts in enhancing the quantitative accuracy of gait analysis.

1. Introduction

Accurate measurement and analysis of human gait plays a critical role in gaining insight into the intricate mechanics of human movement, identifying abnormal gait patterns, and developing effective treatments for gait-related disorders. To objectively analyze human gait, various kinematic and kinetic parameters such as stride length, walking speed, joint angles, and ground reaction forces are measured. These parameters serve as valuable indicators of gait quality, symmetry, and efficiency, providing a comprehensive assessment of an individual’s gait pattern [1,2,3,4,5,6].
In the past, observational gait analysis was the primary method for quantitatively evaluating joint movement during gait. However, this method is subjective, and the data obtained may not be entirely reliable [7]. To obtain more accurate and detailed gait analysis, reflective marker-based motion capture systems have been developed. In these systems, reflective markers are attached to the subject’s body and tracked with an infrared camera to obtain marker coordinates. Human motion is then analyzed using the coordinates of these markers. Marker-based 3D motion capture is a widely used measurement tool for accurately quantifying human motion parameters [8].
However, the marker-based 3D motion capture system is very expensive and requires specialized personnel to handle the equipment or operate the system. Due to the need for specialized equipment and controlled environments, it can be challenging to create a dedicated experimental space for analyzing human motion analysis. In addition, placing markers on the subject’s body for motion analysis is time-consuming. As a result, both people with disabilities and the general population may have difficulty accessing such equipment and conducting experiments in dedicated experimental spaces [9].
In recent years, markerless human pose estimation (HPE) systems have gained significant attention to overcome the limitations of marker-based motion capture systems. HPE is a system that uses machine learning and deep learning algorithms to detect human joint positions in images or videos. A large amount of image data is used to build a dataset that labels human joint key points, and this dataset is trained with a deep learning model to accurately detect human joint key points. In this process, a convolutional neural network (CNN), one of the deep learning models, is mainly used to train the datasets. A CNN model is a neural network composed of multiple layers, and input images are passed through this neural network. Features are extracted from images input through CNN, and these features are used to detect human joint key points [10]. The HPE system can automatically detect the position of human joints using only images recorded with a smartphone, making it easier, simpler, and cheaper to use for various applications such as motion analysis, gesture recognition, and detection. With the advancements in computing technology, the utilization of HPE systems is becoming increasingly widespread. In addition, the barriers to analyzing human motion using traditional motion capture methods are rapidly decreasing, leading to a growing body of research exploring the use of webcam or smartphone-based HPE systems [11,12,13,14,15,16,17,18,19].
Human pose estimation (HPE) systems can be broadly categorized into two types: 2D and 3D HPE. The 2D HPE system tracks human joints or body parts on the entire surface of an image or video, whereas the 3D HPE system estimates the depth of human joints or body parts in an image or video [20,21]. With the development of computing technology and the collection of large-scale datasets, HPE has made remarkable progress, particularly in 2D images and videos. The 3D HPE system, which provides depth information from images and videos, is also advancing, but its performance is not yet satisfactory, and there is ample room for improvement [22]. The 3D HPE system is particularly challenging due to various difficulties, such as depth ambiguity, complex background, lack of training data, occlusion, and scale deformation [23]. For gait analysis, 2D analysis in the side view can quantify motion and address specific clinical problems [24].
To date, there have been many studies using 2D HPE to analyze gait kinematics and temporal and spatial gait parameters in normal subjects or patients with gait disorders, and the broader topic of estimating gait variables or identifying gait features has been addressed. Lv et al. [25] analyzed the walking characteristics of patients with knee joint disorders using sample entropy fused with AlphaPose human pose estimation, and proposed an algorithm capable of discriminating between healthy individuals and those with knee joint disorders. Sato et al. [26] proposed a method using OpenPose to estimate cadence from frontal plane videos of Parkinson’s disease patients using sagittal and frontal plane gait video data from normal subjects. Stenum et al. [27] compared spatiotemporal and sagittal kinematic gait parameters measured with OpenPose to 3D motion capture and proposed a reliable analysis workflow to accurately estimate various gait parameters and detect changes in the gait patterns of healthy individuals. Viswakumar et al. [28] evaluated the performance of the markerless systems OpenPose and MS Kinect using lower extremity joint angle variables during gait and investigated the impact of different lighting and clothing conditions. Rohan et al. [29] proposed an efficient mechanism for gait analysis by combining human pose estimation and convolutional neural networks, achieving a 97.3% accuracy in classifying normal and abnormal gait. Takeda et al. [30] compared lower limb joint angles obtained through video analysis using OpenPose and Kinovea, a markerless motion capture method, with 3D motion capture data and suggested that OpenPose, one of the human pose estimators, could be a suitable substitute for 3D marker-based motion capture. Ota et al. [31] measured lower extremity joint angles during treadmill walking and running using OpenPose and Vicon and suggested that an OpenPose-based motion analysis system can measure the lower extremity range of motion in the sagittal plane. Hii et al. [32] proposed an automated method for temporal gait analysis using MediaPipe and validated this approach against the Vicon motion capture system, demonstrating excellent agreement in most temporal gait parameters.
Previous studies have focused on using 2D HPE systems to measure gait kinematics and temporal and spatial gait parameters, and compare them to 3D marker-based motion capture for effectiveness. In a study investigating technical issues of HPE when performing motion analysis using HPE, Viswakumar et al. [28] investigated the effects of clothing and ambient light intensity on gait analysis results using HPE. Takeichi et al. [33] investigated the effect of the frame rate of video recorded by a smartphone camera on running form analysis using HPE. However, the crucial aspect of camera placement for accurate calculation of gait variables using a single camera to investigate technical issues in HPE has often been overlooked. Recognizing these challenges, this paper addresses the issue by investigating the effect of camera positions during treadmill walking using five smartphone cameras.
Building on the existing body of knowledge, this study aimed to not only compare HPE with marker-based systems but also to enhance the performance and accuracy of HPE in gait analysis. The importance of minimizing occlusions and obtaining reliable joint information during walking has led to the investigation of the specific influence of camera positions on treadmill walking using five smartphone cameras. In contrast to previous research that addressed issues such as clothing effects and ambient light intensity, this research focused on refining the methodology by strategically placing cameras to mitigate challenges associated with leg occlusions during the gait cycle. Investigating the effect of camera placement on the joint detection rate of the HPE system will contribute to a broader understanding of the factors that influence the accuracy and efficiency of gait analysis in clinical and research settings. This effort has the potential to advance the current state of the field and pave the way for further research. By highlighting the importance of proper camera placement in 2D HPE-based gait analysis, this research aimed to fill a gap in the existing literature and explore methods to improve accuracy and efficiency in clinical and research settings.

2. Methods

2.1. Participants

A total of 20 healthy volunteers (10 males and 10 females; mean age: 24.15 ± 1.8 years; mean height: 169.4 ± 8.3 cm; mean mass: 64.93 ± 15.3 kg) from the normal population participated in this experiment. Individuals with a history of serious injuries such as musculoskeletal injuries, neurological disorders, fractures, or those who have undergone surgery were excluded. The participants provided written informed consent before the start of this study, and the Ethics Committee of the authors’ affiliated university approved the experimental procedure used in this study.

2.2. Data Collection

Nine motion capture cameras (NaturalPoint, Corvallis, OR, USA, Optitrack Prime 41 Motion capture system, 4.1 MP, 120 fps) were used to measure gait motion. Each motion capture camera was strategically placed at an optimal distance to accurately capture the subject’s movements within the experimental space. The height of each infrared camera was standardized to 2.3 m to ensure consistent measurement conditions. A total of 39 reflective spherical markers were attached to anatomical landmarks for motion analysis: 4 on the head, 7 on each arm, 6 on each leg, 5 on the torso, and 4 on the pelvis. To prevent the markers from shaking at each joint attachment point, subjects were asked to wear shorts, short sleeves, or tight-fitting suits. Nevertheless, due to the possibility of distortion of the markers, we used Motive Optitrack software (version 3.0.2) to track the positions of the markers in real time, and manual editing is performed to minimize distortion. In addition, despite the participant wearing a T-shirt, we have implemented the preventive measures mentioned in the paper to ensure reliable data acquisition. Before this experiment, each subject adjusted the treadmill speed to their or her preferred speed, and a 3 min preliminary walking process was performed to adapt to walking on the treadmill. After the preliminary walking process, a 5 min break was taken, and each subject walked on the treadmill for 1 min at their preferred speed, and at the same time, the gait motion was recorded using 9 motion capture cameras and 5 smartphone cameras (Samsung Galaxy S22, 2340 × 1080 pixels, 30 fps). Preliminary experiments with smartphone cameras in front of and behind the subject showed that it was difficult to obtain reliable datasets for gait analysis. Since most gait analysis is performed from the side view, the experiment was conducted in five different positions to see if the most acceptable datasets could be obtained from a specific position in the side view. The smartphone cameras were positioned so that the angle between the subject and the smartphone was 90° from the subject’s side and 45° and 30°, respectively, (Figure 1). The distance between the smartphone camera and the subject was set to 3 m to ensure the subject’s whole body was included in the recording screen, and the height was set to 0.8 m. Based on several preliminary experiments, it was confirmed that distance and height did not critically affect the detection rate of pose estimation, as long as the subject’s whole body was captured on the camera screen.

2.3. Human Pose Estimation (HPE)

Compared to other HPE open-source libraries, Mediapipe is relatively fast to process and shows relatively high accuracy in video-based analysis methods. In addition, the pipeline’s use of pose estimation for a single person makes it highly personalized for gait analysis [32,34,35,36]. Therefore, in this study, customized HPE based on Mediapipe (version 0.9.3), an open-source software provided by Google as a cross-platform pipeline processing tool [37], was utilized to perform gait analysis. Mediapipe is a real-time system that detects feature points in a single image captured using a digital camera and provides trained models for pose estimation of the face, hands, and body. BlazePose is the HPE model configured in CNN-based architecture, and it is Mediapipe’s pose detection model that can track human movement from 33 landmark key points. The pose of a person can be identified from the landmark key points(x-y-z coordinates) detected by this model. BlazePose combines an encoder–decoder network architecture, which predicts both heatmaps of the joints and the coordinates of the joints simultaneously. The initial pose alignment provided by the BlazePose detector is complemented by a tracker, effectively detecting key points, and the presence of a person, and refining the regions of interest. This sophisticated model was trained on a substantial dataset, comprising 60,000 images containing a single person or few people, and an additional 25,000 images with only a single person performing various exercises [38]. In this study, the gait motions of each subject were recorded with smartphone cameras, and the x, y coordinates of the shoulder, hip, knee, ankle, toe, and heel of the BlazePose model were used to perform 2D gait analysis (Figure 2).

2.4. Data Processing

Following the acquisition of motion capture data, marker data from the marker-based motion capture system was carefully examined using Motive Optitrack software for accurate detection and quality improvement. Editing procedures were performed to enhance data reliability, including gap-filling interpolation techniques to address missing marker data points or gaps in the motion capture sequence. The Motive Optitrack software’s interactive visualization tool was utilized to verify and adjust marker positions, minimizing inaccuracies caused by noise or tracking errors.
Simultaneously, videos taken by the smartphone camera underwent processing using the Mediapipe algorithm executed with Jupyter notebook. A total of 33 anatomical landmarks were extracted from each frame, and the resulting data were converted into a CSV file format containing landmark coordinates for each frame. To reduce jitter and noise effects, a moving average filter in MATLAB (version R2021 a) was applied to the coordinates of each joint obtained from Mediapipe.
To address the difference in frame rates between the two methods, motion capture data were downsampled and rescaled to match the frame rate of the smartphone camera using MATLAB. Any remaining temporal discrepancies between the two methods were treated with manual adjustments to the kinematic data to ensure the same time instant of heel strike as a reference point for initiation of gait cycle, with approximately 50 gait cycles of 1 min duration analyzed per participant (Figure 3).
The MATLAB software (version R2021 a) was used to import CSV files containing coordinate values for individual joints to facilitate gait analysis. Specifically, within the side view, the hip, knee, and ankle joint angles were calculated based on the coordinates of the shoulder, hip, knee, ankle, and toe joints from each CSV file. The gait cycle was determined by identifying the heel strike and toe-off points within the coordinate data. To calculate the joint angles, it was imperative to ascertain the dimensions of the trunk, thigh, shank, and foot segments. As shown in Figure 4, the following equations were used to obtain measurements for each segment and then calculate the corresponding joint angles, with joint angles calculated in degrees:
ϕ s e g m e n t = t a n 1 ( Δ y Δ x )
θ h i p = ϕ t h i g h ϕ t r u n k
θ k n e e = ϕ t h i g h ϕ s h a n k
θ a n k l e = ϕ f o o t ϕ s h a n k 90 °

2.5. Error Calculation and Statistics

In this study, the reliability of the lower extremity joint angles obtained by the HPE system was evaluated using the mean absolute error (MAE), a frequently used metric when comparing kinematics between markerless and marker-based motion capture systems [27,31]. The MAE was calculated by subtracting the values obtained from the HPE system from those obtained by the marker-based motion capture system. The resulting differences were subjected to the absolute value function and divided by the total number of frames, thus considering only the magnitude of the errors. This approach allows a comprehensive evaluation of the overall reliability and effectiveness of the HPE system by comparing the results obtained from gait analysis using the HPE system with those obtained from the marker-based motion capture system. Lower MAE indicates a smaller difference in joint angles between the two methods, indicating the feasibility of using HPE in treadmill gait analysis. A higher correlation coefficient indicates a stronger association between the datasets, supporting the reliability of gait analysis using HPE.
M e a n a b s o l u t e e r r o r ( D e g r e e ) = 1 N J o i n t a n g l e i ( M ) J o i n t a n g l e i ( H ) i = 1 , 2 , , N i s t h e n u m b e r o f f r a m e M : m o t i o n c a p t u r e ,   H : h u m a n p o s e e s t i m a t i o n
The Pearson correlation coefficient was calculated to evaluate the relationship between lower extremity joint angles calculated from HPE data and lower extremity joint angles calculated from motion capture data.
ρ M , H = c o r r ( M , H ) = c o v ( M , H ) σ M σ H c o v : c o v a r i a n c e ,   σ : s t a n d a r d d e v i a t i o n s

3. Results

The majority of participants showed MAE of less than 5° between HPE and motion capture for hip, knee, and ankle joint angles (Table 1 and Figure 5). In addition, the correlation coefficient for lower extremity joint angles between HPE and motion capture was high, indicating that all joint angles were closely related to the true value (Table 2). In the case of hip joint angles, the lowest MAE of 1.92 ± 0.41 and high correlation coefficient of 0.97 ± 0.015 were observed at P2. Similarly, for knee joint angles, the lowest MAE of 2.4 ± 0.66 and the highest correlation coefficient of 0.98 ± 0.014 were observed at P2. For ankle joint angles, the lowest MAE of 2.29 ± 0.5 and the highest correlation coefficient of 0.84 ± 0.081 were observed at P3. Based on the average MAE and correlation coefficient of the subjects, the hip and knee joint angles were relatively accurately measured at P2, and the ankle joint angle was accurately measured at P3. Additionally, a difference was observed when comparing the left and right lower extremities. For hip joint angles, the average MAE on the left was 2.16 ± 0.63, whereas on the right, it was 2.54 ± 0.95. The average correlation coefficient was 0.96 ± 0.04 on the left and 0.94 ± 0.06 on the right. In the case of knee joint angles, the average MAE for the left was 2.68 ± 0.71, whereas for the right, it was 2.96 ± 0.9. The average correlation coefficients were 0.98 ± 0.01 on the left and 0.97 ± 0.02 on the right. As for ankle joint angles, the left exhibited an average MAE of 2.62 ± 0.66, whereas the right side had an MAE of 3.5 ± 1.17. The average correlation coefficients were 0.8 ± 0.11 on the left and 0.63 ± 0.2 on the right. Regarding the ankle joint angle, a notable observation was observed when comparing the results obtained from different camera positions. Specifically, at P5, most participants showed a relatively high MAE of 3.97 ± 1.35 and a comparatively low correlation coefficient of 0.63 ± 0.23.

4. Discussion

This study aimed to assess the reliability and effectiveness of the HPE system in gait analysis by comparing it with a marker-based motion capture system. When comparing previous gait analysis studies that used HPE with marker-based motion capture, Stenum et al. [27] reported MAE values of 4.0°, 5.6°, and 7.4° for hip, knee, and ankle joint angles, respectively. Viswakumar et al. [28] also reported MAE values of 7.73°, 5.82°, and 7.13° for hip, knee, and ankle joint angles between the two methods. In this study, the MAE values for hip, knee, and ankle joint angles were measured as 2.35°, 2.82°, and 3.06°, respectively. The results indicate a higher level of reliability compared to previous studies. However, it is important to note that our analysis was conducted on a treadmill, whereas the referenced studies were conducted in an overground walking setting. Therefore, it is crucial to consider the differences between these environments.
In addition, using five smartphone cameras, it was investigated whether the position of the smartphone camera affects the degree of joint detection associated with gait analysis results using the HPE system. The reliability and effectiveness of gait analysis using the HPE system were evaluated by calculating the MAE and correlation coefficient between the two methods based on hip, knee, and ankle joint angles. Overall, the results show that the MAE between HPE and marker-based motion capture for the lower extremity joint angles of most participants at all positions of the smartphone camera was less than 5° (Figure 5). This level of error is considered an acceptable margin of error in gait kinematic analysis [39]. The low MAE value was supported by a high correlation coefficient between HPE and marker-based motion capture for lower extremity joint angles (Table 2). Additionally, this study found that the position of the smartphone camera affected the average MAE and correlation coefficient.
Examining the difference between the lower extremity joint angles of individual participants, which represent the most closely resembling data to the average dataset of all participants, in terms of HPE and motion capture, reveals that the differences in angles are not substantial, and similar trends can be identified (Figure 6). In addition, the values of the two methods were compared based on the lower extremity joint angles calculated using HPE and motion capture, and the relationship was confirmed (Figure 7). The results of the comparative study with motion capture demonstrate that the HPE approach is reliable and effective for treadmill gait analysis, as evidenced by the low MAE and high correlation coefficient. These results suggest the feasibility of using HPE for precise analysis of treadmill gait without requiring specialized equipment beyond a smartphone camera.
The MAE and correlation coefficient between motion capture and HPE for lower extremity joint angles vary depending on the position of the smartphone camera. Identifying the position where the lower extremity joint angle difference has the lowest MAE and high correlation coefficient implies that the respective joint angle is calculated most accurately at that position. The results of this study indicate that the hip and knee joint angles are most accurately calculated when the smartphone camera is positioned at P2, with the lowest MAE and high correlation coefficient. Similarly, the ankle joint angle is accurately calculated when the smartphone camera is positioned at P3, with the lowest MAE and the high correlation coefficient. At position P2, the detection of joints necessary for calculating hip and knee joint angles is superior compared to other positions, resulting in more accurate measurements. At position P3, the detection of joints relevant to ankle joint angle calculation surpasses other positions, leading to enhanced accuracy. The significance of specific camera positions in optimizing the precision of joint angle calculations using the HPE system is highlighted by this observation. These findings suggest that adjusting the position of the smartphone camera can enhance the performance of gait analysis using the HPE system. This adjustment addresses issues such as joint visibility, potential occlusions, and detection reliability, thereby improving the overall accuracy and effectiveness of gait analysis conducted with HPE systems. This highlights the critical role of optimal camera placement in refining gait analysis precision within HPE systems.
In addition, the MAE for the left lower extremity joint angles was relatively lower compared to the right, and the correlation coefficient was relatively higher. This means that the left lower extremity joint angle results are more reliable than the right lower extremity joint angle results from the gait analysis using HPE, suggesting that the right lower extremity joints are obscured by the left joints due to the positioning of the smartphone cameras on the left side of the subject. Based on these findings, it suggests that the positioning of the smartphone camera, whether on the left or right side of the participant, could potentially influence the results of gait analysis using HPE [27].
In the case of the ankle joint angle, the correlation coefficient appears to be relatively low compared to the consistently high coefficients observed for the hip and knee joint angles. Previous studies also showed results showing that the error in ankle joint angle was large or that the joint detection rate was relatively low in the area below the ankle [40,41,42]. Therefore, further investigation is necessary to identify the underlying factors contributing to this variance. A plausible explanation for this discrepancy lies in the intricate nature of ankle joint angle estimation within the HPE framework. Unlike the hip and knee joints, where key points are relatively more prominent and reliably detectable, the estimation of the ankle joint angle heavily relies on accurate detection of the toe key points. The complexity of toe detection, which can be susceptible to occlusion and variability in footwear, might introduce inherent challenges that impact the precision of ankle joint angle calculations. Excluding P3, the correlation coefficients for the right ankle joint at the remaining positions are relatively low. This indicates that toe detection for calculating ankle joint angles was not accurately performed, particularly during the crossover of the left and right legs when the right toe was obscured by the left lower limb joint, resulting in suboptimal joint detection rates. In particular, the correlation coefficient for the right ankle joint angle at P5 was the lowest at 0.45 ± 0.149. This suggests that when the camera is positioned behind the subject for HPE to calculate joint angles, the toe of the right ankle is often obscured or inaccurately detected by other lower limb joints. Nevertheless, despite the challenges at P5, the higher correlation coefficient for the left ankle joint angle compared to the right indicates that, due to the camera being placed on the left side of the subject, the detection of the left toe was more accurate than that of the right toe. Given these findings, it becomes evident that prudent caution is essential when relying solely on ankle joint angle measurements derived from HPE. This is especially important in scenarios where accurate ankle joint angle information is critical, such as in evaluating specific pathologies or athletic performance. The observed limitation emphasizes the necessity for additional research and improvement of techniques for estimating ankle joint angles in the wider context of gait analysis based on HPE.
Nevertheless, for most subjects, the MAE between motion capture and HPE for all joint angles tend to be less than 5°, which is an acceptable error range for gait kinematics analysis, and the correlation coefficients also show high values [39]. These results indicate that the use of HPE provides joint detection results that closely resemble the actual joints detected using motion capture. These results are significant in both academic and clinical contexts. They highlight the capability of HPE to produce joint detection results that closely mirror the actual joints identified by motion capture methods. This alignment emphasizes the feasibility of utilizing HPE for treadmill-based gait analysis. The use of a smartphone camera as the primary device is a notable advantage, eliminating the need for specialized machinery. The implications of this extend beyond convenience, catering to individuals with mobility limitations, as well as paving the way for gait analysis to be performed in domestic settings, increasing accessibility to the general population.
This study addresses a notable research gap by examining the impact of smartphone camera positions on the accuracy of joint detection in 2D HPE-based gait analysis. This gap is crucial as it pertains to the practical implementation of such systems in real-world scenarios. Through a thorough evaluation of various camera positions, we aim to offer valuable insights for enhancing the utilization of HPE systems in gait analysis, a crucial aspect of biomechanics and clinical assessments. The significance of our findings extends to settings where accommodating specialized equipment may be challenging. In contexts where dedicated gait analysis facilities or advanced motion capture systems are not feasible, our study suggests that 2D HPE-based methods, informed by the insights gained from diverse smartphone camera positions, can provide a reliable alternative. This study enhances the reliability and applicability of 2D HPE-based methods in the field of gait analysis and biomechanics, offering practical implications for researchers and practitioners in both well-equipped and resource-constrained environments.

5. Limitations

It is imperative to recognize the influence of various environmental and technological factors on the accuracy and consistency of HPE results. These include fluctuations in lighting conditions, variations in ambient light intensity, differences in the texture and color of subjects’ clothing, and even factors such as frames per second and the performance capabilities of the smartphone camera performance. These complex variables underscore the need to recognize the limitations and potential sources of variability inherent in the HPE process. The interplay of these factors highlights the importance of cautious interpretation of results, particularly in scenarios where precision is paramount. In light of these considerations, future research efforts could focus on understanding the intricate interplay between environmental factors, technological specifications, and resultant HPE accuracy. Such a comprehensive investigation has the potential to refine the application of HPE and enhance its utility in academic, clinical, and broader contexts. Furthermore, it could contribute to the establishment of standardized protocols for optimized HPE-based gait analysis.
Our study is limited by the fact that we primarily focused on estimating flexion–extension (2D) during treadmill walking, and whereas this provides valuable insights, it is crucial to acknowledge that other joint movements and additional planes of motion, particularly relevant in clinical applications, were not explored in this study. Future research should address these limitations by expanding the scope to include a broader range of joint motions and dimensions.
Although our reported results indicate that the differences in joint angles between pose estimation and a marker-based system are below the commonly accepted threshold of 5° for gait analysis [39], it is imperative to recognize the potential significance of this discrepancy. A 5° difference, although deemed acceptable in general gait analysis, may be non-negligible in certain kinematic or biomechanical assessments where precision is critical. Researchers and practitioners should exercise caution when applying our findings in contexts where higher accuracy is required.
This study specifically analyzed lower extremity joint angles during treadmill walking. To gain a more comprehensive understanding of the influence of camera positions, it is necessary to extend our investigation to overground walking scenarios. Additionally, exploring the impact of camera positions on temporal walking variables would further contribute to the field’s knowledge. Future research endeavors should consider these aspects to provide a more holistic view of the relationship between camera placement and gait analysis outcomes.

6. Conclusions

This study presents the effectiveness of a customized human pose estimation (HPE) system in analyzing treadmill gait, which is compared with marker-based motion capture. The analysis of lower extremity joint angles shows an acceptable level of MAE below 5° and high correlation coefficients between HPE and motion capture. It is also important to note that the accuracy of the HPE system is affected by the position of the smartphone camera, considering that the hip and knee joint angles are accurately calculated at P2 and the ankle joint angle is accurately calculated at P3. This study highlights the potential of HPE adoption in various fields, such as rehabilitation and sports science. Future research will investigate the factors that affect HPE accuracy and establish standardized protocols, contributing to advancements in biomechanics.

Author Contributions

Conceptualization, J.Y. and K.P.; methodology, J.Y.; software, J.Y.; validation, J.Y. and K.P.; formal analysis, J.Y.; investigation, J.Y.; resources, J.Y.; data curation, J.Y.; writing—original draft preparation, J.Y.; writing—review and editing, K.P.; visualization, J.Y.; supervision, K.P.; project administration, K.P.; funding acquisition, K.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Research Foundation (NRF) grant funded by the Korean government (MSIT) (No. RS-2023-00222406).

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Ethics Committee of Incheon National University (7007971-202301-004A, 21 March 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Akhtaruzzaman, M.D.; Shafie, A.A.; Khan, M.D.R. Gait analysis: Systems, technologies, and importance. J. Mech. Med. Biol. 2016, 16, 1630003. [Google Scholar] [CrossRef]
  2. Roberts, M.; Mongeon, D.; Prince, F. Biomechanical parameters for gait analysis: A systematic review of healthy human gait. Phys. Ther. Rehabil. 2017, 4, 6. [Google Scholar] [CrossRef]
  3. Muro-de-la Herran, A.; Garcia-Zapirain, B.; Mendez-Zorrilla, A. Gait analysis methods: An overview of wearable and non-wearable systems, highlighting clinical applications. Sensors 2014, 14, 3362–3394. [Google Scholar] [CrossRef]
  4. Prakash, C.; Kumar, R.; Mittal, N. Recent developments in human gait research: Parameters, approaches, applications, machine learning techniques, datasets and challenges. Artif. Intell. Rev. 2018, 49, 1–40. [Google Scholar] [CrossRef]
  5. Ferreira, J.P.; Crisostomo, M.M.; Coimbra, A.P. Human gait acquisition and characterization. IEEE Trans. Instrum. Meas. 2009, 58, 2979–2988. [Google Scholar] [CrossRef]
  6. Caldas, R.; Mundt, M.; Potthast, W.; Buarque de Lima Neto, F.; Markert, B. A systematic review of gait analysis methods based on inertial sensors and adaptive algorithms. Gait Posture 2017, 57, 204–210. [Google Scholar] [CrossRef] [PubMed]
  7. Krebs, D.E.; Edelstein, J.E.; Fishman, S. Reliability of observational kinematic gait analysis. Phys. Ther. 1985, 65, 1027–1033. [Google Scholar] [CrossRef] [PubMed]
  8. Ziegler, J.; Reiter, A.; Gattringer, H.; Müller, A. Simultaneous identification of human body model parameters and gait trajectory from 3D motion capture data. Med. Eng. Phys. 2020, 84, 193–202. [Google Scholar] [CrossRef]
  9. Krishnan, C.; Washabaugh, E.P.; Seetharaman, Y. A low cost real-time motion tracking approach using webcam technology. J. Biomech. 2015, 48, 544–548. [Google Scholar] [CrossRef]
  10. Bulat, A.; Tzimiropoulos, G. Human Pose Estimation via Convolutional Part Heatmap Regression. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 717–732. [Google Scholar]
  11. Calvache, D.A.; Bernal, H.A.; Guarín, J.F.; Aguía, K.; Orjuela-Cañón, A.D.; Perdomo, O.J. Automatic estimation of pose and falls in videos using computer vision model. In Proceedings of the 16th International Symposium on Medical Information Processing and Analysis, Lima, Peru, 3 November 2020; Brieva, J., Lepore, N., Romero Castro, E., Linguraru, M.G., Eds.; SPIE: Bellingham, WA, USA, 2020. [Google Scholar]
  12. Güney, G.; Jansen, T.S.; Dill, S.; Schulz, J.B.; Dafotakis, M.; Hoog Antink, C.; Braczynski, A.K. Video-based hand movement analysis of Parkinson patients before and after medication using high-frame-rate videos and MediaPipe. Sensors 2022, 22, 7992. [Google Scholar] [CrossRef]
  13. Garg, S.; Saxena, A.; Gupta, R. Yoga pose classification: A CNN and MediaPipe inspired deep learning approach for real-world application. J. Ambient Intell. Humaniz. Comput. 2022, 14, 16551–16562. [Google Scholar] [CrossRef]
  14. Thaman, B.; Cao, T.; Caporusso, N. Face Mask Detection using MediaPipe Facemesh. In Proceedings of the 2022 45th Jubilee International Convention on Information, Communication and Electronic Technology (MIPRO), Opatija, Croatia, 23–27 May 2022; IEEE: Piscataway, NJ, USA, 2022. [Google Scholar]
  15. Khaleghi, L.; Artan, U.; Etemad, A.; Marshall, J.A. Touchless control of heavy equipment using low-cost hand gesture recognition. IEEE Internet Things Mag. 2022, 5, 54–57. [Google Scholar] [CrossRef]
  16. Moryossef, A.; Tsochantaridis, I.; Dinn, J.; Camgoz, N.C.; Bowden, R.; Jiang, T.; Rios, A.; Muller, M.; Ebling, S. Evaluating the immediate applicability of pose estimation for sign language recognition. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Nashville, TN, USA, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021. [Google Scholar]
  17. Docekal, J.; Rozlivek, J.; Matas, J.; Hoffmann, M. Human keypoint detection for close proximity human–robot interaction. In Proceedings of the 2022 IEEE-RAS 21st International Conference on Humanoid Robots (Humanoids), Ginowan, Japan, 28–30 November 2022. [Google Scholar]
  18. Chen, W.; Jiang, Z.; Guo, H.; Ni, X. Fall detection based on key points of human-skeleton using OpenPose. Symmetry 2020, 12, 744. [Google Scholar] [CrossRef]
  19. Hernández, Ó.G.; Morell, V.; Ramon, J.L.; Jara, C.A. Human pose detection for robotic-assisted and rehabilitation environments. Appl. Sci. 2021, 11, 4183. [Google Scholar] [CrossRef]
  20. Ben Gamra, M.; Akhloufi, M.A. A review of deep learning techniques for 2D and 3D human pose estimation. Image Vis. Comput. 2021, 114, 104282. [Google Scholar] [CrossRef]
  21. Topham, L.K.; Khan, W.; Al-Jumeily, D.; Hussain, A. Human body pose estimation for gait identification: A comprehensive survey of datasets and models: A comprehensive survey of datasets and models. ACM Comput. Surv. 2023, 55, 1–42. [Google Scholar] [CrossRef]
  22. Wang, J.; Tan, S.; Zhen, X.; Xu, S.; Zheng, F.; He, Z.; Shao, L. Deep 3D human pose estimation: A review. Comput. Vis. Image Underst. 2021, 210, 103225. [Google Scholar] [CrossRef]
  23. Ji, X.; Fang, Q.; Dong, J.; Shuai, Q.; Jiang, W.; Zhou, X. A survey on monocular 3D human pose estimation. Virtual Real. Intell. Hardw. 2020, 2, 471–500. [Google Scholar] [CrossRef]
  24. Ugbolue, U.C.; Papi, E.; Kaliarntas, K.T.; Kerr, A.; Earl, L.; Pomeroy, V.M.; Rowe, P.J. The evaluation of an inexpensive, 2D, video based gait assessment system for clinical use. Gait Posture 2013, 38, 483–489. [Google Scholar] [CrossRef] [PubMed]
  25. Lv, X.; Ta, N.; Chen, T.; Zhao, J.; Wei, H. Analysis of gait characteristics of patients with knee arthritis based on human posture estimation. Biomed Res. Int. 2022, 2022, 7020804. [Google Scholar] [CrossRef] [PubMed]
  26. Sato, K.; Nagashima, Y.; Mano, T.; Iwata, A.; Toda, T. Quantifying normal and parkinsonian gait features from home movies: Practical application of a deep learning—Based 2D pose estimator. PLoS ONE 2019, 14, e0223549. [Google Scholar] [CrossRef] [PubMed]
  27. Stenum, J.; Rossi, C.; Roemmich, R.T. Two-dimensional video-based analysis of human gait using pose estimation. PLoS Comput. Biol. 2021, 17, e1008935. [Google Scholar] [CrossRef] [PubMed]
  28. Viswakumar, A.; Rajagopalan, V.; Ray, T.; Gottipati, P.; Parimi, C. Development of a robust, simple, and affordable human gait analysis system using bottom-up pose estimation with a smartphone camera. Front. Physiol. 2021, 12, 784865. [Google Scholar] [CrossRef]
  29. Rohan, A.; Rabah, M.; Hosny, T.; Kim, S.H. Human pose estimation-based real-time gait analysis using convolutional neural network. IEEE Access 2020, 8, 191542–191550. [Google Scholar] [CrossRef]
  30. Takeda, I.; Yamada, A.; Onodera, H. Artificial Intelligence-Assisted motion capture for medical applications: A comparative study between markerless and passive marker motion capture. Comput. Methods Biomech. Biomed. Engin. 2021, 24, 864–873. [Google Scholar] [CrossRef] [PubMed]
  31. Ota, M.; Tateuchi, H.; Hashiguchi, T.; Ichihashi, N. Verification of validity of gait analysis systems during treadmill walking and running using human pose tracking algorithm. Gait Posture 2021, 85, 290–297. [Google Scholar] [CrossRef] [PubMed]
  32. Hii, C.S.T.; Gan, K.B.; Zainal, N.; Mohamed Ibrahim, N.; Azmin, S.; Mat Desa, S.H.; van de Warrenburg, B.; You, H.W. Automated gait analysis based on a marker-free pose estimation model. Sensors 2023, 23, 6489. [Google Scholar] [CrossRef]
  33. Takeichi, K.; Ichikawa, M.; Shinayama, R.; Tagawa, T. A mobile application for running form analysis based on pose estimation technique. In Proceedings of the 2018 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), San Diego, CA, USA, 23–27 July 2018; IEEE: Piscataway, NJ, USA, 2018. [Google Scholar]
  34. Kim, J.W.; Choi, J.Y.; Ha, E.J.; Choi, J.H. Human pose estimation using MediaPipe pose and optimization method based on a humanoid model. Appl. Sci. 2023, 13, 2700. [Google Scholar] [CrossRef]
  35. Tony Hii, C.S.; Gan, K.B.; Zainal, N.; Ibrahim, N.M.; Rani, S.A.M.; Shattar, N.A. Marker free gait analysis using pose estimation model. In Proceedings of the 2022 IEEE 20th Student Conference on Research and Development (SCOReD), Bangi, Malaysia, 8–9 November 2022; IEEE: Piscataway, NJ, USA, 2022. [Google Scholar]
  36. Chung, J.L.; Ong, L.Y.; Leow, M.C. Comparative analysis of skeleton-based human pose estimation. Future Internet 2022, 14, 380. [Google Scholar] [CrossRef]
  37. Lugaresi, C.; Tang, J.; Nash, H.; McClanahan, C.; Uboweja, E.; Hays, M.; Zhang, F.; Chang, C.L.; Yong, M.G.; Lee, J.; et al. MediaPipe: A Framework for Building Perception Pipelines. arXiv 2019, arXiv:1906.08172. [Google Scholar]
  38. Bazarevsky, V.; Grishchenko, I.; Raveendran, K.; Zhu, T.; Zhang, F.; Grundmann, M. BlazePose: On-device Real-time Body Pose tracking. arXiv 2020, arXiv:2006.10204. [Google Scholar]
  39. McGinley, J.L.; Baker, R.; Wolfe, R.; Morris, M.E. The reliability of three-dimensional kinematic gait measurements: A systematic review. Gait Posture 2009, 29, 360–369. [Google Scholar] [CrossRef] [PubMed]
  40. Kanko, R.M.; Laende, E.K.; Davis, E.M.; Selbie, W.S.; Deluzio, K.J. Concurrent assessment of gait kinematics using marker-based and markerless motion capture. J. Biomech. 2021, 127, 110665. [Google Scholar] [CrossRef] [PubMed]
  41. Yamamoto, M.; Shimatani, K.; Hasegawa, M.; Kurita, Y.; Ishige, Y.; Takemura, H. Accuracy of temporo-spatial and lower limb joint kinematics parameters using OpenPose for various gait patterns with orthosis. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 2666–2675. [Google Scholar] [CrossRef] [PubMed]
  42. Moro, M.; Marchesi, G.; Hesse, F.; Odone, F.; Casadio, M. Markerless vs. Marker-based gait analysis: A proof of concept study. Sensors 2022, 22, 2011. [Google Scholar] [CrossRef]
Figure 1. Overview of experimental setup. (A) 9 motion capture cameras and 5 smartphone cameras. The smartphone cameras are set at a distance of 3 m from the subject and a height of 0.8 m. (B) The skeleton model estimated by HPE according to each smartphone position.
Figure 1. Overview of experimental setup. (A) 9 motion capture cameras and 5 smartphone cameras. The smartphone cameras are set at a distance of 3 m from the subject and a height of 0.8 m. (B) The skeleton model estimated by HPE according to each smartphone position.
Bioengineering 11 00141 g001
Figure 2. Comparison of marker positions between two devices required for kinematics gait analysis. The green circles represent the joints used to calculate the lower extremity joint angle among the 33 landmarks in Mediapipe.
Figure 2. Comparison of marker positions between two devices required for kinematics gait analysis. The green circles represent the joints used to calculate the lower extremity joint angle among the 33 landmarks in Mediapipe.
Bioengineering 11 00141 g002
Figure 3. The workflow of the proposed approach.
Figure 3. The workflow of the proposed approach.
Bioengineering 11 00141 g003
Figure 4. Representation of the kinematic model utilized for calculating hip, knee, and ankle joint angles.
Figure 4. Representation of the kinematic model utilized for calculating hip, knee, and ankle joint angles.
Bioengineering 11 00141 g004
Figure 5. Boxplot of the MAE between motion capture and HPE for lower extremity joint angles. Each panel shows the MAE between HPE and motion capture for left and right lower extremity joint angles: (A) hip joint angle, (B) knee joint angle, (C) ankle joint angle. Boxplot shows the mean and SD of all subjects and individual subject data.
Figure 5. Boxplot of the MAE between motion capture and HPE for lower extremity joint angles. Each panel shows the MAE between HPE and motion capture for left and right lower extremity joint angles: (A) hip joint angle, (B) knee joint angle, (C) ankle joint angle. Boxplot shows the mean and SD of all subjects and individual subject data.
Bioengineering 11 00141 g005
Figure 6. Example of one subject’s lower extremity joint angles between HPE and motion capture. Panels visually show the variation within the gait cycle to illustrate how the results of lower extremity joint angles using HPE for each smartphone position compare to motion capture. (A) Left and right hip joint angle, (B) knee joint angle, (C) ankle joint angle.
Figure 6. Example of one subject’s lower extremity joint angles between HPE and motion capture. Panels visually show the variation within the gait cycle to illustrate how the results of lower extremity joint angles using HPE for each smartphone position compare to motion capture. (A) Left and right hip joint angle, (B) knee joint angle, (C) ankle joint angle.
Bioengineering 11 00141 g006
Figure 7. Example of left lower extremity joint angles comparisons among the different measurement systems. Panels show the relationship for lower extremity joint angles calculated using the two methods. (A) shows the relationship when the smartphone camera position is at P1, (B) at P2, (C) at P3, (D) at P4, and (E) at P5.
Figure 7. Example of left lower extremity joint angles comparisons among the different measurement systems. Panels show the relationship for lower extremity joint angles calculated using the two methods. (A) shows the relationship when the smartphone camera position is at P1, (B) at P2, (C) at P3, (D) at P4, and (E) at P5.
Bioengineering 11 00141 g007
Table 1. MAE between motion capture and HPE for lower extremity joint angles.
Table 1. MAE between motion capture and HPE for lower extremity joint angles.
Joint AnglesNumber of SubjectsMean ± SDp
P1P2P3P4P5
Hip (°)Left202.16 ± 0.561.94 ± 0.421.95 ± 0.342.26 ± 0.862.51 ± 0.690.018
Right202.61 ± 0.691.9 ± 0.42.07 ± 0.552.87 ± 1.093.26 ± 1.14<0.001
Knee (°)Left202.71 ± 0.72.39 ± 0.642.58 ± 0.792.78 ± 0.572.96 ± 0.760.042
Right202.72 ± 0.612.41 ± 0.72.87 ± 0.962.94 ± 0.723.85 ± 0.83<0.001
Ankle (°)Left202.92 ± 0.622.83 ± 0.752.19 ± 0.42.32 ± 0.592.85 ± 0.59<0.001
Right203.74 ± 0.862.91 ± 0.712.39 ± 0.583.38 ± 0.575.09 ± 0.89<0.001
Table 2. Correlation coefficient between motion capture and HPE for lower extremity joint angles.
Table 2. Correlation coefficient between motion capture and HPE for lower extremity joint angles.
Joint AnglesNumber of SubjectsMean ± SD 
P1P2P3P4P5
HipLeft200.96 ± 0.0170.97 ± 0.0140.97 ± 0.0230.94 ± 0.0770.94 ± 0.044
Right200.95 ± 0.030.96 ± 0.0160.96 ± 0.0260.91 ± 0.1140.92 ± 0.04
KneeLeft200.98 ± 0.0140.98 ± 0.0170.98 ± 0.0180.98 ± 0.0110.98 ± 0.011
Right200.98 ± 0.010.98 ± 0.0120.97 ± 0.0250.97 ± 0.020.96 ± 0.021
AnkleLeft200.73 ± 0.1240.75 ± 0.1130.84 ± 0.0740.84 ± 0.0950.81 ± 0.124
Right200.53 ± 0.180.7 ± 0.1940.83 ± 0.0870.67 ± 0.10.45 ± 0.149
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, J.; Park, K. Improving Gait Analysis Techniques with Markerless Pose Estimation Based on Smartphone Location. Bioengineering 2024, 11, 141. https://doi.org/10.3390/bioengineering11020141

AMA Style

Yang J, Park K. Improving Gait Analysis Techniques with Markerless Pose Estimation Based on Smartphone Location. Bioengineering. 2024; 11(2):141. https://doi.org/10.3390/bioengineering11020141

Chicago/Turabian Style

Yang, Junhyuk, and Kiwon Park. 2024. "Improving Gait Analysis Techniques with Markerless Pose Estimation Based on Smartphone Location" Bioengineering 11, no. 2: 141. https://doi.org/10.3390/bioengineering11020141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop