Next Article in Journal
Enhancing Fetal Electrocardiogram Signal Extraction Accuracy through a CycleGAN Utilizing Combined CNN–BiLSTM Architecture
Previous Article in Journal
An Improved Initial Alignment Method Based on SE2(3)/EKF for SINS/GNSS Integrated Navigation System with Large Misalignment Angles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Motion Capture Technology in Sports Scenarios: A Survey

by
Xiang Suo
1,
Weidi Tang
2 and
Zhen Li
1,*
1
School of Athletic Performance, Shanghai University of Sport, Shanghai 200438, China
2
School of Exercise and Health, Shanghai University of Sport, Shanghai 200438, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2947; https://doi.org/10.3390/s24092947
Submission received: 11 March 2024 / Revised: 26 April 2024 / Accepted: 4 May 2024 / Published: 6 May 2024
(This article belongs to the Section Wearables)

Abstract

:
Motion capture technology plays a crucial role in optimizing athletes’ skills, techniques, and strategies by providing detailed feedback on motion data. This article presents a comprehensive survey aimed at guiding researchers in selecting the most suitable motion capture technology for sports science investigations. By comparing and analyzing the characters and applications of different motion capture technologies in sports scenarios, it is observed that cinematography motion capture technology remains the gold standard in biomechanical analysis and continues to dominate sports research applications. Wearable sensor-based motion capture technology has gained significant traction in specialized areas such as winter sports, owing to its reliable system performance. Computer vision-based motion capture technology has made significant advancements in recognition accuracy and system reliability, enabling its application in various sports scenarios, from single-person technique analysis to multi-person tactical analysis. Moreover, the emerging field of multimodal motion capture technology, which harmonizes data from various sources with the integration of artificial intelligence, has proven to be a robust research method for complex scenarios. A comprehensive review of the literature from the past 10 years underscores the increasing significance of motion capture technology in sports, with a notable shift from laboratory research to practical training applications on sports fields. Future developments in this field should prioritize research and technological advancements that cater to practical sports scenarios, addressing challenges such as occlusion, outdoor capture, and real-time feedback.

1. Introduction

In the realm of sports, the pursuit of optimal performance, injury prevention, and effective training strategies has driven the continuous evolution of technology and scientific research [1,2]. Among the various tools and methods employed in sports science, motion capture technology has emerged as a crucial component in understanding, analyzing, and enhancing athletic performance. Motion capture refers to the process of recording and translating the movement of objects or people into digital data that can be analyzed and manipulated [3]. Its origins can be traced back to the 19th century when pioneering work by Eadweard Muybridge and H. C. Adam laid the foundation for capturing and studying human and animal locomotion [4].
Over the past few decades, significant advancements in software and hardware technology have propelled motion capture to new heights, expanding its applications across various fields, including rehabilitation, sports training, and human movement biomechanics [5,6]. In particular, the field of sports biomechanics, which integrates principles and methods from mechanics, anatomy, physiology, and other related disciplines to study the structure and function of the human movement system, has greatly benefited from the integration of motion capture technology [7]. By providing detailed kinematic and kinetic data, motion capture enables researchers and practitioners to gain deeper insights into the complex interplay of the nervous system, muscles, bone marrow, and joints during athletic performance.
One of the key areas where motion capture technology has made significant contributions is in postural analysis and balance assessment. Postural control and balance are essential factors that influence an athlete’s performance and susceptibility to injuries [8]. Traditional methods for assessing balance, such as force plates and laboratory-based motion capture systems, often face limitations in terms of cost, portability, and ecological validity. However, the advent of inertial sensors has revolutionized the field, offering a cost-effective and versatile solution for collecting and processing large amounts of athlete balance data in various settings [9].
The application of motion capture technology extends beyond postural analysis, encompassing a wide range of sports scenarios and research questions. From analyzing the biomechanics of swimming strokes [10] to investigating the kinematic factors influencing soccer kicking performance [11,12], motion capture has become an indispensable tool in sports science. Its ability to provide objective, quantifiable data has enabled researchers and coaches to optimize training programs, prevent injuries, and enhance overall athletic performance [2,13,14,15].
However, the rapid growth and diversification of motion capture technologies have also presented challenges for researchers and practitioners in selecting the most appropriate tools for their specific needs. Different motion capture systems, such as cinematography capture, electromagnetic capture, and computer vision capture, offer unique advantages and limitations that must be carefully considered [16]. Furthermore, the increasing complexity and volume of data generated by motion capture technologies necessitate the development of advanced analysis techniques, such as machine learning and pattern recognition, to fully harness their potential [17].
This review aims to provide a comprehensive overview of the various motion capture technologies used in sports scenarios, bridging the gap between technological advancements and practical implementation. By comparing and analyzing the characteristics, advantages, and limitations of each technology, we seek to guide researchers and practitioners in selecting the most suitable motion capture methods for their specific applications. Moreover, we discuss the current state-of-the-art applications of motion capture technology in sports, highlighting its potential to revolutionize athletic performance analysis, injury prevention, and rehabilitation.

2. Classification of Motion Capture Technology

To provide a comprehensive overview of motion capture technologies in sports scenarios, we conducted a thorough literature search focusing on the development, validation, and application of these technologies. The inclusion criteria for the selected articles were as follows: (1) the study involved the use of motion capture technology in sports contexts; (2) the article provided detailed information on the motion capture system, its validation, or its application; and (3) the study presented original research findings or technical advancements. Conference abstracts and studies focusing solely on non-sports applications of motion capture were excluded. The search results covered various motion capture systems, including traditional optical systems, wearable sensor-based systems, and computer vision-based approaches.
Among these technologies, traditional motion capture methods have long been considered the gold standard, typically relying on optical image analysis with markers as the benchmark. However, in recent years, the rapid development and application of deep learning have paved the way for wearable sensor-based motion capture and computer vision-based motion capture techniques [18]. Currently, motion capture technology can be broadly classified into two major categories: traditional optical systems [19] and computer vision-based motion capture technology, which has seen significant advancements in recent years. Additionally, wearable sensor technology and other types of technologies are also prominent in this field, as shown in Table 1.

2.1. Cinematography Capture Systems

Cinematography capture systems are generally considered the gold standard in the field of motion capture [34]. They utilize optical markers to calibrate the spatial coordinate system and accurately determine the real 3D coordinates of the subject being measured. The accuracy of optical triangulation in motion capture depends on several factors, including the relative positions of the cameras, the distance between the cameras and the object, the number and quality of optical markers, and the movement of the markers within the tracking space [35]. The trade-off between the sampling frequency and spatial resolution is also an important factor that affects the overall performance of the system. Typical application scenarios of modern optical motion capture technology are shown in Figure 1. The pipeline of cinematography motion capture involves the following key steps: synchronized multi-camera setup, marker placement on the subject, cameras capturing marker trajectory data, data processing and gap-filling, and mapping marker data onto a digital character to reconstruct high-fidelity 3D motion.
Cinematography capture systems can be divided into two types: active marker systems and passive marker systems. In passive marker systems, the markers used do not emit light, and the sensors rely on receiving reflected light from the markers for data recording. Representative products of passive marker systems include Vicon equipment. Compared to passive markers, active marker systems such as Optotrak3020 (Northern Digital, Waterloo, Canada) have stronger resistance to interference and greater stability [36]. However, the additional power supply and cables in active marker systems can interfere with the subject’s movement and actions [37]. Relying on fixed cameras for data collection, cinematography capture systems suffer from limitations such as a relatively fixed sampling area and a sampling range influenced by the number of cameras and their field of view. The maximum working range of optical sampling systems documented so far does not exceed 1000 m2 [38]. To meet the recording requirements of large-scale scenes, researchers have deployed over 20 cameras, significantly increasing the difficulty and cost of sampling. Furthermore, the calibration and synchronization efforts in post-processing also multiply. Another major bottleneck that is hard to overcome in image-based motion capture systems is occlusion. Some motion capture scenarios require restrictions on equipment placement, and the complexity of the subject’s movements leads to many gaps and blind spots in the collected data. This not only increases the workload of data processing but also reduces the precision of system output [39]. To meet the demands of testing in larger areas, an optical capture algorithm using a moving sampling camera has emerged. By mounting the camera on a sliding base that moves parallel to the direction of motion and using fixed markers on the ground for coordinate transformation, optical motion capture under linear motion conditions has been successfully achieved [40]. Although it comes at the cost of reduced accuracy and increased data processing time, this approach effectively reduces the complexity of motion capture systems for long-distance scenes. It has certain application value in motion capture scenarios such as short-distance sprints.
In addition, the iGPS system can achieve millimeter-level spatial positioning accuracy in a relatively large spatial scale [41]. However, such devices have limited performance in dynamic scenarios, as slightly higher speeds than walking can result in significantly increased positioning errors [42]. To meet the requirements of larger testing areas, an optical capture algorithm that uses moving cameras for tracking and sampling has been developed. By mounting the cameras on a sliding base parallel to the direction of motion and using fixed markers on the ground for coordinate transformation, optical motion capture under linear motion conditions has been successfully achieved [40]. Although this method sacrifices accuracy and increases the data processing time, it effectively reduces the complexity of motion capture systems for action acquisition at longer distances. This approach has potential applications in motion capture scenarios involving straight-line events such as sprints.

2.2. Electromagnetic Capture Systems

In addition to image capture systems, electromagnetic tracking systems are also an important component of motion capture technology. Wearable motion capture based on electromagnetic systems utilizes the calculation of the time difference between the reflection of electromagnetic waves from the base station to the target object to determine the target distance [43]. The main advantage of electromagnetic-based motion capture systems lies in their high tolerance for occlusion compared to optical motion capture systems. The positioning method, which does not rely on visible light, makes the impact of occlusion on such systems negligible [44]. A typical pipeline of wearable sensors in motion capture is shown in Figure 2. The wearable sensor pose estimation pipeline starts with calibrating the sensors to the body segments. Filtering is then applied to fuse the raw sensor data into cleaner orientation estimates. Linear accelerations are calculated by removing gravity, and double integrated to estimate segment position trajectories. Static periods are detected to apply zero-velocity updates, mitigating integration drift. Finally, a kinematic model combines the individual segment poses to reconstruct the full-body motion.
There are IMU (Inertial Measurement Unit) wearable sensors, which typically integrate accelerometers and gyroscopes, and some products also integrate magnetometers. Wearable technologies based on IMU sensors have gained increasing prominence in the field of motion analysis due to their simple and reliable communication protocols, compact device size, and flexible application scenarios [45]. Additionally, there are UWB (Ultra-Wide Band) wearable sensors. Commercial UWB systems have demonstrated stable performance in tasks such as indoor motion tracking, system deployment, and sensor localization. They can achieve dynamic tracking in indoor sports scenarios with measurement errors below 0.5% of the working range, and the positioning data accuracy can be improved by adjusting the sampling frequency [46]. This technology holds greater potential in sports with larger motion ranges and higher speeds, such as the comprehensive speed and posture tracking of ski jumpers. In addition to IMU and UWB, the LPM (Local Position Measurement) system is another typical electromagnetic positioning system, which utilizes radar wave reflections between base stations and reference points for positioning [47]. LPM offers notable advantages in terms of the size of the positioning area, system deployment complexity, and post-processing of data [48]. In addition to IMU, UWB, and LPM systems, GNSS (Global Navigation Satellite System) wearable sensors utilize satellite signals for precise positioning and motion data. They offer advantages in sports with larger motion ranges, providing comprehensive speed and posture tracking. While they may face limitations in challenging environments, combining GNSS with other sensor technologies enhances accuracy [49].

2.3. Computer Vision Capture Systems

In recent years, the advancement of hardware computing power in parallel computing technology and the progress in deep learning have mutually benefited each other [50,51,52,53]. Deep learning-based computer vision techniques have been widely applied in specialized fields such as medicine, sports, and industrial monitoring [54,55,56]. Motion capture based on computer vision offers advantages such as higher accuracy, faster speed, and reduced workload [57,58].
Human Pose Estimation (HPE) plays a crucial role in sports scenarios, which involves analyzing motion action videos for single-person or multi-person to obtain two-dimensional or three-dimensional coordinate information about body poses, based on the number of synchronized input videos, and it can be categorized into monocular pose estimation and multi-view pose estimation. As shown in Figure 3, typical human pose estimation involves frame acquisition and preprocessing to prepare the input image, which is then passed through a convolutional backbone network and upsampling network to predict the locations of key body joints via heatmaps and silhouettes. The model architecture is selected based on requirements, and the chosen model is trained on a dataset using backpropagation. The trained model then undergoes inference, post-processing, evaluation, and fine-tuning to optimize its performance on localizing body joints in new images.
With the improved speed and reliability of algorithmic recognition, computer vision-based motion capture has started to be used in various fields, including rehabilitation training, sports performance enhancement, and biomechanical analysis [59,60]. Despite the advantages of higher accuracy and stronger robustness, there are limitations in the application of multi-view pose estimation algorithms, such as the inability to extract information from regular motion broadcast videos. Currently, in the field of motion and vision related to computer vision capture applications, most use pose estimation algorithms based on single-view visual recognition for 2D or 3D athlete pose estimation.
Firstly, in practical application scenarios, the motion recognition algorithm is the core of computer vision capture systems, and the dataset forms the foundation of algorithm training. Various visual recognition models trained based on datasets can convert human motion in video inputs into spatial coordinates or vectors. The subsequent analysis and inference system work based on the output of the visual recognition system for training and adjustment. In 2021, Xin Chen et al. released the SportsCap system, which achieved real-time 3D motion capture for most sports scenes. It significantly improved the accuracy of motion capture in similar task scenarios compared to traditional methods, and achieved satisfactory levels of action classification capability [61].
Secondly, due to the rapid relative motion between the target individuals and the fixed coordinate system in sports, severe motion blur is one of the obstacles hindering the further application of computer vision capture. To address this, researchers like Hong Guo developed PhyCoVIS, which combines motion capture technology and visualization techniques to transform athlete motion data into easily understandable and analyzable visual charts and images. Through the PhyCoVIS system, coaches and athletes can assess the athlete’s body coordination, identify potential issues, formulate corresponding training plans, and compare the coordination of different athletes and analyze the coordination of specific motion sequences [62].
The bottleneck of computer vision capture in sports and sports-related fields lies in the lack of specialized visual recognition datasets. In addition to the scarcity of datasets suitable for outdoor scenes, most of the datasets are general datasets established by computer science researchers for human daily actions, rather than specialized datasets for sports scenes [63]. Using general pose estimation datasets for computer vision capture training can lead to problems such as decreased recognition accuracy and difficulties in recognition. To address this, besides establishing specialized datasets for sports scenes, another possible solution is to utilize few-shot learning techniques. Traditional machine learning algorithms require large amounts of data to train models, but few-shot learning techniques can accomplish learning tasks with very limited data available [64].
Currently, computer vision capture has not yet produced widely recognized and widely used pose estimation algorithms for high-precision capture scenarios. However, in applications where there is a higher tolerance for motion localization errors, such as football tactical analysis research, computer vision capture technology has irreplaceable advantages such as non-interference, ease of deployment, and fewer limitations. Some mature specialized applications already have considerable value.

2.4. Other Motion Capture Systems

In 2018, Dawei Liang et al. successfully extracted contextual information from the audio portion of videos, surpassing the context information that can be obtained from image-based methods [30]. Similar systems that utilize deep learning-based audio classifiers provide a new dimension of information for distinguishing complex actions.
In addition to sound waves, motion capture techniques utilizing radar (Radar Modality) also hold significant practical value. Sevgi Z. Gurbuz et al. utilized continuous radar RF data for action recognition and employed Generative Adversarial Networks (GANs) for sample synthesis, achieving comparable accuracy to Vicon systems in terms of output precision [31]. Compared to video-based motion capture systems, radar-based motion capture not only offers the same advantages but also enables data collection while ensuring privacy protection. This unique technological advantage makes radar-based motion capture highly promising with great technical potential [65].
Beyond the range of motion capture methods that use specialized equipment, there is the Wi-Fi Modality, which utilizes the echoes of Wi-Fi signals for motion capture. By extracting spatial–temporal information from the Channel State Information (CSI), and using Convolutional Neural Networks (CNNs) as the model classifier, it becomes possible to achieve cross-scene motion capture and recognition [32]. The greatest advantage of this technology is the freedom from the constraints of dedicated devices for motion capture tasks. In an ideal scenario, it would be possible to estimate human motion within the working range using Wi-Fi signals. This holds significant importance for low-cost and lightweight research projects, as it provides an irreplaceable solution for motion capture without the need for specialized equipment.
Fusion Motion Capture (FMC), which combines multiple data acquisition techniques, is one feasible solution for motion recognition in sports scenarios. Timo Von Marcard et al. integrated visual motion capture with wearable IMU sensors to improve and stabilize whole-body motion capture techniques [66]. Brodie et al. used a similar multimodal motion capture system to extract key performance indicators related to athletes’ body and physiological limits, such as average drag coefficient and maximum tilt angle during turns, which help optimize game strategies [33]. Hui-Min Shen et al. proposed a Magnetic Measurement Unit (MMU) approach for multisensor fusion in motion capture, enhancing system output accuracy and stability [67]. Additionally, Hasegawa et al. developed an action-assisted monitoring system that utilizes sound wave information to monitor and provide feedback on the skier’s center of gravity, assisting skiers in overcoming instinctive reactions of leaning back [68]. This helps guide specific technical optimizations for athletes, improving performance and reducing the risk of injury. The SmartSki system, designed by Anton Kos et al., underwent functional testing and validation by several professional alpine skiers during a year of snowfield testing [69]. However, its lack of portability hindered its practical implementation, keeping the system mostly in the experimental phase.
The main advantage of multimodal motion capture technology lies in the use of different types of data acquisition methods to simultaneously measure the pose and motion information of the moving target in different environments, achieving precision and reliability beyond the reach of a single technology [70,71]. This also implies broader potential application scenarios, faster data analysis capabilities, and better data visualization potential. At the same time, multimodal motion capture technology requires relatively more complex system integration, larger data processing capabilities, more precise data monitoring capabilities, and higher hardware costs. Overall, multimodal human motion capture technology has significant advantages in motion capture and analysis, but it needs to overcome technical challenges such as high equipment costs, high complexity, and difficulty in algorithm design. In the future, with the continuous development of technology, multimodal techniques are expected to become important means for human motion analysis and virtual reality, bringing more convenience and assurance to human life and health.

3. Application of Motion Capture in the Field of Sports

Motion capture technology exhibits different forms of expression and usage scenarios depending on the research objectives and topics. Specific application areas include human kinematics, biomechanics, motion technique analysis, athletic performance, and more [72]. In particular, motion capture technology can be used for topics such as technical analysis, training, athletic assessment, and sports injury medicine for athletes. For example, analyzing a soccer player’s running routes, shooting angles, and power can provide coaches and athletes with reliable quantitative metrics to improve team tactics and skills [73]. In swimming competitions, motion capture technology can measure swimmers’ speed, angular velocity, and other technical motion kinematics information, enabling the analysis of swimming postures and fluid resistance effects to assist athletes in improving training efficiency [74,75]. This article focuses on the typical research application scenarios of motion capture technology in the field of sports; Figure 1 illustrates the priority indicators and typical application scenarios of different motion capture techniques.

3.1. Construction of Athlete Performance Datasets

Xiong Zhao et al. utilized the Raptor-E motion capture system to collect and analyze three-dimensional motion data from nearly 200 athletes, establishing a database of motion and performance [76]. This enables researchers and system users to utilize advanced analysis methods such as pattern recognition and machine learning to study athletes’ movement patterns and gain new insights into the potential correlations between movement patterns and injury history, sport specificity, competition level, and demographics. Raptor-E is a passive marker-based optical motion capture system that offers advantages such as high accuracy, real-time performance, and strong outdoor interference resistance [77]. When combined with dedicated visualization systems designed for specific sports, the management and application of athlete performance datasets become more reliable and convenient [78].
Optical-based motion capture systems perform better in projects with simpler limb movements. However, they have limitations in terms of a relatively fixed sampling domain, where the sampling range is influenced by the number of cameras and their field of view. The maximum working range typically does not exceed 1000 square meters [38]. Additionally, the systems face occlusion issues. The mainstream solution currently involves using multiple cameras for video signal capture, but this increases the workload of post-processing and reduces the accuracy of system output [39].

3.2. Real-Time Assistance for Athlete Training and Competition

Compared to optical-based methods, motion capture using wearable sensors imposes less interference on athletes. To obtain more natural human gait kinematic information, Julie Rekant et al. installed 24 wearable sensors on 10 healthy males while using the Vicon system as a ground truth reference [79]. The study found significant differences between the measurements obtained from the wearable sensors and the Vicon system across different subjects, with the largest errors occurring in the sagittal plane, where the mean angular error was close to 5°. This error was considered relatively noticeable.
Although wearable motion capture solutions based on electromagnetic systems have advantages such as strong interference resistance and simple device setup, the positioning principles of such solutions limit their performance in terms of angular measurement accuracy. Most of these solutions rely on time-of-flight methods for distance measurement and positioning, resulting in significant cumulative errors during continuous positioning operations. Compared to optical-based motion capture solutions, commercially available products currently achieve lower tracking and positioning accuracy. Therefore, in typical motion capture tasks, electromagnetic-based motion capture solutions are not the preferred choice. However, in some special scenarios, such as when the motion range is too large, lighting conditions in the environment are not ideal, or occlusion issues cannot be resolved, specific electromagnetic-based motion capture solutions can be considered, and the system’s positioning accuracy needs to be validated to meet the research requirements.

3.3. Multi-Camera Motion Capture Technology for Training

In 2019, Eduard Pons et al. conducted a validation study on the effectiveness of a multi-camera motion capture system. The research team tracked 38 official matches of male athletes in the Spanish second division league and collected thousands of valid data points. The team validated the statistical indicators of standardized mean deviation and coefficient of variation for the system data, demonstrating that multi-camera motion capture technology can meet the research and training requirements in sports scenarios, with high repeatability and stability [80]. The bottleneck of computer vision-based motion capture in sports and specific fields lies in the lack of dedicated visual recognition datasets [63,81]. Apart from establishing specialized datasets for sports scenarios [82,83], another possible solution is to utilize few-shot learning techniques, which enable training with a small amount of data [64].
Currently, there is no widely recognized and widely used pose estimation algorithm in computer vision-based motion capture technology for scenarios with high accuracy requirements. However, in applications where the tolerance for motion positioning errors is higher, such as football tactical analysis and other research topics, vision-based motion capture technology offers irreplaceable advantages such as no interference, easy deployment, and fewer limitations. Some mature specialized applications already have considerable practical value.

4. Discussion

In the context of sports, the application of motion capture technology in sports scenarios requires careful consideration of the following factors compared to technical development environments:
(1)
Indoor laboratory settings offer controlled environments with better control over factors like lighting, temperature, and humidity, reducing noise and interference. However, outdoor sports scenarios present complex and variable conditions, including wind, sunlight, shadows, and different surfaces (e.g., plastic, grass, snow, ice, water), which can impact the accuracy and stability of motion capture systems. Outdoor conditions also experience changes in temperature, humidity, and visibility, while non-isolated environments introduce sound, lighting, and electromagnetic interference, further affecting motion capture tasks. Overcoming these environmental limitations is crucial for ensuring accurate and reliable motion capture in sports.
(2)
System setup: Indoor laboratory settings offer easier installation and calibration of motion capture systems due to the controlled environment. However, outdoor sports scenarios present more complex conditions, requiring additional effort for system setup and adjustment to achieve higher accuracy and precision. The precision of the measurement method is inversely proportional to the effective working range of the system. In sports scenarios, where large-scale motion scenes are common, capturing human kinematic information requires additional software and hardware optimization methods. These challenges necessitate careful system setup and optimization to ensure accurate motion capture in outdoor sports environments [84].
(3)
Motion characteristics: The movements in indoor laboratory settings are usually simple and single, such as gait analysis or arm movements. For these types of movements, indoor motion capture systems can provide high-quality data. In outdoor sports scenarios, there may be more complex movements, such as aerial rotations or limb flips, which can be distorted or inaccurate due to environmental factors. Motion capture for sports scenarios often requires capturing fast movements. In specific research scenarios, such as ballistic analysis in shooting or instant analysis of baseball swings, the sampling frequency may exceed 1000 Hz [16].
(4)
Data processing: Data processing is relatively straightforward in indoor laboratory settings because the data in controlled environments are usually stable and accurate [85]. In outdoor sports scenarios, data processing needs to consider more factors, such as lighting and environmental noise issues, occlusion and penetration issues, motion model establishment, data noise, and filtering issues [86].
When selecting the appropriate motion capture technology for specific sports applications, it is crucial to comprehensively consider factors such as the size of the capture space, the complexity of the athlete’s movements, the real-time performance requirements of the system, and the acceptability of the equipment’s impact on the athlete’s movements.
Cinematography capture systems offer high accuracy and are considered the gold standard in motion capture. However, they face limitations in terms of fixed capture areas, occlusion issues, and the impact of markers on natural movements. These systems are more suitable for biomechanical analysis in controlled environments.
Electromagnetic capture systems, including IMUs, UWB, and LPM, provide advantages in terms of occlusion resistance, flexible scenarios, and real-time performance. IMU-based wearable sensors are compact and widely used in motion analysis, while UWB systems enable high-precision indoor tracking, and LPM offers benefits in positioning area and deployment complexity. However, electromagnetic systems may face limitations in accuracy compared to optical systems and require careful sensor placement.
Computer vision capture systems have made significant advancements with deep learning techniques. Pose estimation algorithms enable markerless motion capture from video inputs, offering high accuracy, fast speed, and reduced interference with natural movements. These systems have been applied in rehabilitation, sports performance analysis, and biomechanical research. However, computer vision systems currently lack large-scale sport-specific datasets, which limits their performance in complex sports scenarios. Few-shot learning techniques are a potential solution to address this limitation.
Multimodal motion capture systems, which fuse data from multiple sensors, provide a promising approach for complex sports scenarios. By combining the strengths of different capture technologies, such as vision and wearable sensors, multimodal systems can achieve high accuracy, robustness, and adaptability to various environments. However, multimodal systems also face challenges in terms of system complexity, data synchronization, and higher costs.
The application of motion capture technology in sports scenarios faces various challenges, including environmental limitations, system setup, motion characteristics, and data processing. In this paper, a comparison and evaluation of the application of motion capture technology across different sports scenarios were conducted, with primary consideration given to factors such as repeatability, robustness, accuracy, and reliability. Due to differences in application scenarios, testing standards, and testing conditions among various motion capture systems, the assessment of reliability and repeatability may not be directly comparable, even for the same motion capture method. To address this issue, this paper employed a normalized evaluation approach by analyzing the relevant literature covered. The performance metrics from different studies were converted into a unified evaluation criteria by comparing the reported values within each specific metric and assigning them to a normalized scale. In cases where explicit performance metrics were not provided in the reviewed literature, this paper made qualitative assessments based on the overall conclusions and discussions presented in those studies, aligning them with the normalized scale.
The normalization process involved identifying the best-performing motion capture technology within each performance metric category and assigning it a score of 100%. The scores for the remaining technologies were then calculated as a percentage of the best-performing technology’s score within that category. This approach allowed for a standardized comparison of performance metrics across different motion capture systems, despite variations in testing conditions and reporting methods. The normalized scale was then divided into four ranges: 0–40%, 40–60%, 60–80%, and 80–100%, corresponding to the terms “low”, “medium”, “medium to high”, and “high”, respectively. By comparing the normalized performance metrics of each motion capture technology against these percentage ranges, this paper was able to assign the appropriate qualitative descriptor to each technology’s performance in terms of repeatability, robustness, accuracy, and reliability. Table 2 summarizes the performance metrics of various motion capture technologies, offering preliminary references for technology selection in subsequent research and applications.
In summary, the choice of motion capture technology depends on the specific requirements and constraints of the sports application. Researchers and practitioners should carefully evaluate factors such as capture space, motion complexity, and equipment impact on athletes to select the most suitable motion capture technology for their sports analysis tasks. As motion capture technology continues to evolve, addressing the challenges posed by outdoor sports scenarios and leveraging advancements in deep learning and multimodal systems will be crucial for advancing the field of sports biomechanics and performance analysis.

5. Conclusions

Motion capture technology offers numerous advantages and applications in the field of sports, enabling the analysis of athletes’ technical performance, training effectiveness, competitive assessment, and sports medicine-related information. Computer vision-based motion capture technology exhibits high recognition accuracy and wide applicability in various sports scenarios, particularly in large-scale scenes like football matches. Wearable sensor-based motion capture technology has made significant advancements in accuracy and multifunctional monitoring, proving particularly effective in small-scale scenarios such as swimming competitions. Imaging-based motion recognition remains dominant, offering unparalleled accuracy advantages across most scenarios. As artificial intelligence continues to develop, motion capture technology in sports will become more extensive and profound, supporting healthier, more scientific, and efficient training methods.
The research outlook in this field focuses on the future development of motion capture technology in sports. With ongoing advancements in technology, motion capture will provide even more accurate and objective data analysis and optimization solutions for athletes and coaches. Computer vision-based motion capture will see broader applications as improvements in capture accuracy, real-time performance, and stability are achieved through the further development and optimization of computer vision technology. Wearable sensor-based motion capture will grow in areas such as health monitoring and sports training. Multimodal motion capture technology will become more accurate and comprehensive, serving as a reliable solution for capturing complex scene motions. Additionally, the integration of emerging technologies like artificial intelligence will optimize motion capture algorithms and models, enhancing the efficiency and accuracy of sports performance analysis. In conclusion, motion capture technology brings opportunities and challenges to the sports field, supporting the pursuit of healthier, more scientific, and efficient training methods.

Author Contributions

Conceptualization, X.S.; writing—review and editing, X.S. and W.T.; supervision, project administration, Z.L.; funding acquisition, X.S. and Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (grant number: 62077037) and Research and Innovation Grant for Graduate Students, Shanghai University of Sport (Project Number: YJSCX-2023-034).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This article contains no data or material other than the articles used for the review and referenced.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chambers, R.; Gabbett, T.J.; Cole, M.H.; Beard, A. The Use of Wearable Microsensors to Quantify Sport-Specific Movements. Sports Med. 2015, 45, 1065–1081. [Google Scholar] [CrossRef] [PubMed]
  2. Camomilla, V.; Bergamini, E.; Fantozzi, S.; Vannozzi, G. Trends Supporting the In-Field Use of Wearable Inertial Sensors for Sport Performance Evaluation: A Systematic Review. Sensors 2018, 18, 873. [Google Scholar] [CrossRef] [PubMed]
  3. Menache, A. Understanding Motion Capture for Computer Animation; Elsevier: Amsterdam, The Netherlands, 2011. [Google Scholar]
  4. Adam, H.C. Eadweard Muybridge: The Human and Animal Locomotion Photographs; Bibliotheca Universalis; Taschen: Hong Kong, China, 2014; Available online: https://books.google.co.jp/books?id=tjomnwEACAAJ (accessed on 27 November 2023).
  5. Kolykhalova, K.; Camurri, A.; Völpe, G.; Sanguineti, M.; Puppo, E.; Niewiadomski, R. A multimodal dataset for the analysis of movement qualities in karate martial art. In Proceedings of the 2015 7th International Conference on Intelligent Technologies for Interactive Entertainment (INTETAIN), Turin, Italy, 10–12 June 2015; IEEE: New York, NY, USA, 2015; pp. 74–78. [Google Scholar]
  6. Qian, B.; Chen, H.; Wang, X.; Guan, Z.; Li, T.; Jin, Y.; Wu, Y.; Wen, Y.; Che, H.; Kwon, G.; et al. DRAC 2022: A public benchmark for diabetic retinopathy analysis on ultra-wide optical coherence tomography angiography images. Patterns 2024, 5, 100929. [Google Scholar] [CrossRef]
  7. Colyer, S.L.; Evans, M.; Cosker, D.P.; Salo, A.I.T. A Review of the Evolution of Vision-Based Motion Analysis and the Integration of Advanced Computer Vision Methods Towards Developing a Markerless System. Sports Med.—Open 2018, 4, 24. [Google Scholar] [CrossRef] [PubMed]
  8. Brognara, L.; Mazzotti, A.; Rossi, F.; Lamia, F.; Artioli, E.; Faldini, C.; Traina, F. Using Wearable Inertial Sensors to Monitor Effectiveness of Different Types of Customized Orthoses during CrossFit® Training. Sensors 2023, 23, 1636. [Google Scholar] [CrossRef] [PubMed]
  9. Walgaard, S.; Faber, G.S.; van Lummel, R.C.; van Dieën, J.H.; Kingma, I. The validity of assessing temporal events, sub-phases and trunk kinematics of the sit-to-walk movement in older adults using a single inertial sensor. J. Biomech. 2016, 49, 1933–1937. [Google Scholar] [CrossRef]
  10. Mooney, R.; Corley, G.; Godfrey, A.; Osborough, C.; Newell, J.; Quinlan, L.R.; ÓLaighin, G. Analysis of swimming performance: Perceptions and practices of US-based swimming coaches. J. Sports Sci. 2016, 34, 997–1005. [Google Scholar] [CrossRef]
  11. Inoue, K.; Nunome, H.; Sterzing, T.; Shinkai, H.; Ikegami, Y. Dynamics of the support leg in soccer instep kicking. J. Sports Sci. 2014, 32, 1023–1032. [Google Scholar] [CrossRef]
  12. Augustus, S.; Mundy, P.; Smith, N. Support leg action can contribute to maximal instep soccer kick performance: An intervention study. J. Sports Sci. 2017, 35, 89–98. [Google Scholar] [CrossRef]
  13. Philippaerts, R.M.; Vaeyens, R.; Janssens, M.; Van Renterghem, B.; Matthys, D.; Craen, R.; Bourgois, J.; Vrijens, J.; Beunen, G.; Malina, R.M. The relationship between peak height velocity and physical performance in youth soccer players. J. Sports Sci. 2006, 24, 221–230. [Google Scholar] [CrossRef]
  14. Gómez-Carmona, C.D.; Bastida-Castillo, A.; Ibáñez, S.J.; Pino-Ortega, J. Accelerometry as a method for external workload monitoring in invasion team sports. A systematic review. PLoS ONE 2020, 15, e0236643. [Google Scholar] [CrossRef] [PubMed]
  15. Ekegren, C.L.; Gabbe, B.J.; Finch, C.F. Sports Injury Surveillance Systems: A Review of Methods and Data Quality. Sports Med. 2016, 46, 49–65. [Google Scholar] [CrossRef]
  16. van der Kruk, E.; Reijne, M.M. Accuracy of human motion capture systems for sport applications; state-of-the-art review. Eur. J. Sport Sci. 2018, 18, 806–819. [Google Scholar] [CrossRef] [PubMed]
  17. Cust, E.E.; Sweeting, A.J.; Ball, K.; Robertson, S. Machine and deep learning for sport-specific movement recognition: A systematic review of model development and performance. J. Sports Sci. 2019, 37, 568–600. [Google Scholar] [CrossRef] [PubMed]
  18. Li, H.; Cui, C.; Jiang, S. Strategy for improving the football teaching quality by AI and metaverse-empowered in mobile internet environment. Wireless Netw. 2022. [Google Scholar] [CrossRef]
  19. Robertson, D.G.E.; Caldwell, G.E.; Hamill, J.; Kamen, G.; Whittlesey, S. Research Methods in Biomechanics; Human kinetics: Champaign, IL, USA, 2013. [Google Scholar]
  20. Whitting, J.; Steele, J.; McGhee, D.; Munro, B. Different measures of plantar-flexor flexibility and their effects on landing technique: Implications for injury. J. Sci. Med. Sport 2010, 13, e48. [Google Scholar] [CrossRef]
  21. Rizaldy, N.; Ferryanto, F.; Sugiharto, A.; Mahyuddin, A.I. Evaluation of action sport camera optical motion capture system for 3D gait analysis. IOP Conf. Ser. Mater. Sci. Eng. 2021, 1109, 012024. [Google Scholar] [CrossRef]
  22. Schmidt, M.; Nolte, K.; Kolodziej, M.; Ulbricht, A.; Jaitner, T. Accuracy of Three Global Positioning Systems for Determining Speed and Distance Parameters in Professional Soccer. In 13th World Congress of Performance Analysis of Sport and 13th International Symposium on Computer Science in Sport; Baca, A., Exel, J., Eds.; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2023; pp. 174–177. [Google Scholar] [CrossRef]
  23. Benjaminse, A.; Bolt, R.; Gokeler, A.; Otten, B. A validity study comparing xsens with vicon. ISBS Proc. Arch. 2020, 38, 752. [Google Scholar]
  24. Umek, A.; Kos, A. Validation of UWB positioning systems for player tracking in tennis. Pers. Ubiquit. Comput. 2022, 26, 1023–1033. [Google Scholar] [CrossRef]
  25. Houtmeyers, K.C.; Helsen, W.F.; Jaspers, A.; Nanne, S.; McLaren, S.; Vanrenterghem, J.; Brink, M.S. Monitoring Elite Youth Football Players’ Physiological State Using a Small-Sided Game: Associations With a Submaximal Running Test. Int. J. Sports Physiol. Perform. 2022, 17, 1439–1447. [Google Scholar] [CrossRef]
  26. Groos, D.; Ramampiaro, H.; Ihlen, E.A. EfficientPose: Scalable single-person pose estimation. Appl. Intell. 2021, 51, 2518–2533. [Google Scholar] [CrossRef]
  27. Aouaidjia, K.; Sheng, B.; Li, P.; Kim, J.; Feng, D.D. Efficient Body Motion Quantification and Similarity Evaluation Using 3-D Joints Skeleton Coordinates. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 2774–2788. [Google Scholar] [CrossRef]
  28. Li, Y.-C.; Chang, C.-T.; Cheng, C.-C.; Huang, Y.-L. Baseball Swing Pose Estimation Using OpenPose. In Proceedings of the 2021 IEEE International Conference on Robotics, Automation and Artificial Intelligence (RAAI), Hong Kong, China, 21–23 April 2021; pp. 6–9. [Google Scholar] [CrossRef]
  29. Nguyen, T.D.; Kresovic, M. A survey of top-down approaches for human pose estimation. arXiv 2022, arXiv:2202.02656. [Google Scholar] [CrossRef]
  30. Liang, D.; Thomaz, E. Audio-Based Activities of Daily Living (ADL) Recognition with Large-Scale Acoustic Embeddings from Online Videos. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 17. [Google Scholar] [CrossRef]
  31. Gurbuz, S.Z.; Rahman, M.M.; Kurtoglu, E.; Martelli, D. Continuous Human Activity Recognition and Step-Time Variability Analysis with FMCW Radar. In Proceedings of the 2022 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI), Ioannina, Greece, 27–30 September 2022; pp. 1–04. [Google Scholar] [CrossRef]
  32. Sheng, B.; Xiao, F.; Sha, L.; Sun, L. Deep Spatial–Temporal Model Based Cross-Scene Action Recognition Using Commodity WiFi. IEEE Internet Things J. 2020, 7, 3592–3601. [Google Scholar] [CrossRef]
  33. Brodie, M.; Walmsley, A.; Page, W. Fusion motion capture: A prototype system using inertial measurement units and GPS for the biomechanical analysis of ski racing. Sports Technol. 2008, 1, 17–28. [Google Scholar] [CrossRef]
  34. Corazza, S.; Mündermann, L.; Gambaretto, E.; Ferrigno, G.; Andriacchi, T.P. Markerless Motion Capture through Visual Hull, Articulated ICP and Subject Specific Model Generation. Int. J. Comput. Vis. 2009, 87, 156. [Google Scholar] [CrossRef]
  35. Maletsky, L.P.; Sun, J.; Morton, N.A. Accuracy of an optical active-marker system to track the relative motion of rigid bodies. J. Biomech. 2007, 40, 682–685. [Google Scholar] [CrossRef]
  36. Thewlis, D.; Bishop, C.; Daniell, N.; Paul, G. Next-Generation Low-Cost Motion Capture Systems Can Provide Comparable Spatial Accuracy to High-End Systems. J. Appl. Biomech. 2013, 29, 112–117. [Google Scholar] [CrossRef]
  37. Stancic, I.; Grujic Supuk, T.; Panjkota, A. Design, development and evaluation of optical motion-tracking system based on active white light markers. IET Sci. Meas. Technol. 2013, 7, 206–214. [Google Scholar] [CrossRef]
  38. Spörri, J.; Schiefermüller, C.; Müller, E. Collecting Kinematic Data on a Ski Track with Optoelectronic Stereophotogrammetry: A Methodological Study Assessing the Feasibility of Bringing the Biomechanics Lab to the Field. PLoS ONE 2016, 11, e0161757. [Google Scholar] [CrossRef]
  39. Panjkota, A. Outline of a Qualitative Analysis for the Human Motion in Case of Ergometer Rowing. In Proceedings of the 9th WSEAS International Conference on Simulation, Modelling and Optimization, Budapest, Hungary, 3–5 September 2009. [Google Scholar]
  40. Colloud, F.; Chèze, L.; André, N.; Bahuaud, P. An innovative solution for 3d kinematics measurement for large volumes. J. Biomech. 2008, 41 (Suppl. 1), S57. [Google Scholar] [CrossRef]
  41. Van der Kruk, E. Modelling and Measuring 3D Movements of a Speed Skater. 2013. Available online: https://repository.tudelft.nl/islandora/object/uuid%3A2a54e547-0a5a-468b-be80-a41a656cacc1 (accessed on 23 February 2023).
  42. Depenthal, C. iGPS used as kinematic measuring system. In Proceedings of the FIG Congress, Sydney, Australia, 11–16 April 2010; pp. 11–16. [Google Scholar]
  43. Stelzer, A.; Pourvoyeur, K.; Fischer, A. Concept and application of LPM—A novel 3-D local position measurement system. IEEE Trans. Microw. Theory Tech. 2004, 52, 2664–2669. [Google Scholar] [CrossRef]
  44. Pasku, V.; De Angelis, A.; De Angelis, G.; Arumugam, D.D.; Dionigi, M.; Carbone, P.; Moschitta, A.; Ricketts, D.S. Magnetic Field-Based Positioning Systems. IEEE Commun. Surv. Tutorials 2017, 19, 2003–2017. [Google Scholar] [CrossRef]
  45. Rana, M.; Mittal, V. Wearable Sensors for Real-Time Kinematics Analysis in Sports: A Review. IEEE Sensors J. 2021, 21, 1187–1207. [Google Scholar] [CrossRef]
  46. Perrat, B.; Smith, M.J.; Mason, B.S.; Rhodes, J.M.; Goosey-Tolfrey, V.L. Quality assessment of an Ultra-Wide Band positioning system for indoor wheelchair court sports. Proc. Inst. Mech. Eng. Part P J. Sports Eng. Technol. 2015, 229, 81–91. [Google Scholar] [CrossRef]
  47. Resch, A.; Pfeil, R.; Wegener, M.; Stelzer, A. Review of the LPM local positioning measurement system. In Proceedings of the 2012 International Conference on Localization and GNSS, Starnberg, Germany, 25–27 June 2012; pp. 1–5. [Google Scholar] [CrossRef]
  48. Stevens, T.G.A.; de Ruiter, C.J.; van Niel, C.; van de Rhee, R.; Beek, P.J.; Savelsbergh, G.J.P. Measuring Acceleration and Deceleration in Soccer-Specific Movements Using a Local Position Measurement (LPM) System. Int. J. Sports Physiol. Perform. 2014, 9, 446–456. [Google Scholar] [CrossRef]
  49. Gardner, C.; Navalta, J.W.; Carrier, B.; Aguilar, C.; Perdomo Rodriguez, J. Training Impulse and Its Impact on Load Management in Collegiate and Professional Soccer Players. Technologies 2023, 11, 79. [Google Scholar] [CrossRef]
  50. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  51. Pandey, M.; Fernandez, M.; Gentile, F.; Isayev, O.; Tropsha, A.; Stern, A.C.; Cherkasov, A. The transformational role of GPU computing and deep learning in drug discovery. Nat. Mach. Intell. 2022, 4, 211–221. [Google Scholar] [CrossRef]
  52. Papadrakakis, M.; Stavroulakis, G.; Karatarakis, A. A new era in scientific computing: Domain decomposition methods in hybrid CPU–GPU architectures. Comput. Methods Appl. Mech. Eng. 2011, 200, 1490–1508. [Google Scholar] [CrossRef]
  53. Abraham, M.J.; Murtola, T.; Schulz, R.; Páll, S.; Smith, J.C.; Hess, B.; Lindahl, E. GROMACS: High performance molecular simulations through multi-level parallelism from laptops to supercomputers. SoftwareX 2015, 1–2, 19–25. [Google Scholar] [CrossRef]
  54. Li, J.; Zhang, P.; Wang, T.; Zhu, L.; Liu, R.; Yang, X.; Wang, K.; Shen, D.; Sheng, B. DSMT-Net: Dual Self-Supervised Multi-Operator Transformation for Multi-Source Endoscopic Ultrasound Diagnosis. IEEE Trans. Med. Imaging 2024, 43, 64–75. [Google Scholar] [CrossRef]
  55. Dai, L.; Sheng, B.; Chen, T.; Wu, Q.; Liu, R.; Cai, C.; Wu, L.; Yang, D.; Hamzah, H.; Liu, Y.; et al. A deep learning system for predicting time to progression of diabetic retinopathy. Nat. Med. 2024, 30, 584–594. [Google Scholar] [CrossRef] [PubMed]
  56. Liu, R.; Ou, L.; Sheng, B.; Hao, P.; Li, P.; Yang, X.; Xue, G.; Zhu, L.; Luo, Y.; Zhang, P.; et al. Mixed-Weight Neural Bagging for Detecting m6A Modifications in SARS-CoV-2 RNA Sequencing. IEEE Trans. Biomed. Eng. 2022, 69, 2557–2568. [Google Scholar] [CrossRef]
  57. Thomas, G.; Gade, R.; Moeslund, T.B.; Carr, P.; Hilton, A. Computer vision for sports: Current applications and research topics. Comput. Vis. Image Underst. 2017, 159, 3–18. [Google Scholar] [CrossRef]
  58. Chen, Z.; Qiu, G.; Li, P.; Zhu, L.; Yang, X.; Sheng, B. MNGNAS: Distilling Adaptive Combination of Multiple Searched Networks for One-Shot Neural Architecture Search. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13489–13508. [Google Scholar] [CrossRef]
  59. Wang, J.; Tan, S.; Zhen, X.; Xu, S.; Zheng, F.; He, Z.; Shao, L. Deep 3D human pose estimation: A review. Comput. Vis. Image Underst. 2021, 210, 103225. [Google Scholar] [CrossRef]
  60. Sheng, B.; Li, P.; Ali, R.; Chen, C.L.P. Improving Video Temporal Consistency via Broad Learning System. IEEE Trans. Cybern. 2022, 52, 6662–6675. [Google Scholar] [CrossRef]
  61. Chen, X.; Pang, A.; Yang, W.; Ma, Y.; Xu, L.; Yu, J. SportsCap: Monocular 3D Human Motion Capture and Fine-Grained Understanding in Challenging Sports Videos. Int. J. Comput. Vis. 2021, 129, 2846–2864. [Google Scholar] [CrossRef]
  62. Guo, H.; Zou, S.; Lai, C.; Zhang, H. PhyCoVIS: A visual analytic tool of physical coordination for cheer and dance training. Comput. Animat. Virtual Worlds 2021, 32, e1975. [Google Scholar] [CrossRef]
  63. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; Theobalt, C. Monocular 3D Human Pose Estimation in the Wild Using Improved CNN Supervision. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; IEEE: New York, NY, USA, 2017; pp. 506–516. [Google Scholar]
  64. He, Y.; Wang, Y.; Fan, H.; Sun, J.; Chen, Q. FS6D: Few-Shot 6D Pose Estimation of Novel Objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 6814–6824. Available online: https://openaccess.thecvf.com/content/CVPR2022/html/He_FS6D_Few-Shot_6D_Pose_Estimation_of_Novel_Objects_CVPR_2022_paper.html (accessed on 13 March 2023).
  65. Li, X.; He, Y.; Jing, X. A Survey of Deep Learning-Based Human Activity Recognition in Radar. Remote Sens. 2019, 11, 1068. [Google Scholar] [CrossRef]
  66. von Marcard, T.; Pons-Moll, G.; Rosenhahn, B. Human Pose Estimation from Video and IMUs. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1533–1547. [Google Scholar] [CrossRef]
  67. Shen, H.-M.; Lian, C.; Wu, X.-W.; Bian, F.; Yu, P.; Yang, G. Full-pose estimation using inertial and magnetic sensor fusion in structurized magnetic field for hand motion tracking. Measurement 2021, 170, 108697. [Google Scholar] [CrossRef]
  68. Hasegawa, S.; Ishijima, S.; Kato, F.; Mitake, H.; Sato, M. Realtime sonification of the center of gravity for skiing. In Proceedings of the 3rd Augmented Human International Conference on—AH ’12, Megève, France, 8–9 March 2012; ACM Press: New York, NY, USA, 2012; pp. 1–4. [Google Scholar]
  69. Kos, A. Smart sport equipment: SmartSki prototype for biofeedback applications in skiing. Pers. Ubiquitous Comput. 2017, 22, 535–544. [Google Scholar] [CrossRef]
  70. Hwang, D.-H.; Aso, K.; Yuan, Y.; Kitani, K.; Koike, H. MonoEye: Multimodal Human Motion Capture System Using A Single Ultra-Wide Fisheye Camera. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology, UIST ’20, Virtual, 20–23 October 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 98–111. [Google Scholar] [CrossRef]
  71. Pons-Moll, G.; Baak, A.; Helten, T.; Müller, M.; Seidel, H.-P.; Rosenhahn, B. Multisensor-fusion for 3D full-body human motion capture. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 663–670. [Google Scholar] [CrossRef]
  72. Winter, D.A. Biomechanics and Motor Control of Human Movement; John Wiley & Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  73. Ángel López, J.; Segura-Giraldo, B.; Rodríguez-Sotelo, L.; García-Solano, K. Kinematic Soccer Kick Analysis Using a Motion Capture System. In Proceedings of the VII Latin American Congress on Biomedical Engineering CLAIB 2016, Bucaramanga, Colombia, 26–28 October 2016; pp. 682–685. [Google Scholar] [CrossRef]
  74. Yang, J.; Li, T.; Chen, Z.; Li, X. Research on the Method of Underwater Swimming Motion Capture. In Proceedings of the 2021 IEEE 7th International Conference on Virtual Reality (ICVR), Foshan, China, 20–22 May 2021; pp. 1–4. [Google Scholar] [CrossRef]
  75. Xu, S.; Lee, S. An Inertial Sensing-Based Approach to Swimming Pose Recognition and Data Analysis. J. Sens. 2022, 2022, e5151105. [Google Scholar] [CrossRef]
  76. Zhao, X.; Ross, G.; Dowling, B.; Graham, R.B. Three-Dimensional Motion Capture Data of a Movement Screen from 183 Athletes. Sci. Data 2023, 10, 235. [Google Scholar] [CrossRef]
  77. Noureen, S.; Shamim, N.; Roy, V.; Bayne, S. Real-Time Digital Simulators: A Comprehensive Study on System Overview, Application, and Importance. Int. J. Res. Eng. 2017, 4, 266–277. [Google Scholar] [CrossRef]
  78. Sheng, B.; Li, P.; Zhang, Y.; Mao, L.; Chen, C.L.P. GreenSea: Visual Soccer Analysis Using Broad Learning System. IEEE Trans. Cybern. 2021, 51, 1463–1477. [Google Scholar] [CrossRef]
  79. Rekant, J.; Rothenberger, S.; Chambers, A. Inertial measurement unit-based motion capture to replace camera-based systems for assessing gait in healthy young adults: Proceed with caution. Meas. Sens. 2022, 23, 100396. [Google Scholar] [CrossRef]
  80. Pons, E.; García-Calvo, T.; Resta, R.; Blanco, H.; del Campo, R.L.; García, J.D.; Pulido, J.J. A comparison of a GPS device and a multi-camera video technology during official soccer matches: Agreement between systems. PLoS ONE 2019, 14, e0220729. [Google Scholar] [CrossRef] [PubMed]
  81. Ostrek, M.; Rhodin, H.; Fua, P.; Müller, E.; Spörri, J. Are Existing Monocular Computer Vision-Based 3D Motion Capture Approaches Ready for Deployment? A Methodological Study on the Example of Alpine Skiing. Sensors 2019, 19, 4323. [Google Scholar] [CrossRef] [PubMed]
  82. Nibali, A.; Millward, J.; He, Z.; Morgan, S. ASPset: An outdoor sports pose video dataset with 3D keypoint annotations. Image Vis. Comput. 2021, 111, 104196. [Google Scholar] [CrossRef]
  83. Akada, H.; Wang, J.; Shimada, S.; Takahashi, M.; Theobalt, C.; Golyanik, V. UnrealEgo: A New Dataset for Robust Egocentric 3D Human Motion Capture. arXiv 2022, arXiv:2208.01633. [Google Scholar] [CrossRef]
  84. Begon, M.; Colloud, F.; Fohanno, V.; Bahuaud, P.; Monnet, T. Computation of the 3D kinematics in a global frame over a 40m-long pathway using a rolling motion analysis system. J. Biomech. 2009, 42, 2649–2653. [Google Scholar] [CrossRef]
  85. Wang, J.; Wang, S.; Wang, Y.; Hu, H.; Yu, J.; Zhao, X.; Liu, J.; Chen, X.; Li, Y. A data process of human knee joint kinematics obtained by motion-capture measurement. BMC Med. Inform. Decis. Mak. 2021, 21, 121. [Google Scholar] [CrossRef]
  86. Saini, N.; Price, E.; Tallamraju, R.; Enficiaud, R.; Ludwig, R.; Martinovic, I.; Ahmad, A.; Black, M.J. Markerless Outdoor Human Motion Capture Using Multiple Autonomous Micro Aerial Vehicles. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 823–832. Available online: https://openaccess.thecvf.com/content_ICCV_2019/html/Saini_Markerless_Outdoor_Human_Motion_Capture_Using_Multiple_Autonomous_Micro_Aerial_ICCV_2019_paper.html (accessed on 3 March 2023).
Figure 1. Pipeline of cinematography motion capture.
Figure 1. Pipeline of cinematography motion capture.
Sensors 24 02947 g001
Figure 2. Pipeline of wearable sensor motion capture.
Figure 2. Pipeline of wearable sensor motion capture.
Sensors 24 02947 g002
Figure 3. Typical pipeline of HPE.
Figure 3. Typical pipeline of HPE.
Sensors 24 02947 g003
Table 1. Classification of motion capture technology.
Table 1. Classification of motion capture technology.
CategoryApplication Example
Cinematography CaptureActive MarkerLanding Technique [20]
Passive MarkerGait Analysis [21]
Electromagnetic Capture SystemsGNSSSoccer Player Kinematic Data Acquisition [22]
IMUMotion Data Validation [23]
UWBTennis Player Positioning [24]
LPMYouth Soccer Performance [25]
Computer Vision CaptureSingle-Person2DGait Analysis [26]
3DHandball Action Analysis [27]
Multi-PersonBottom-UpBaseball Swing Assessment [28]
Top-DownGait Analysis [29]
OtherAudio ModalityActivity Recognition [30]
Radar ModalityActivity Recognition [31]
Wi-Fi ModalityCross-scene Action Recognition [32]
Fusion ModalitySki Racing Biomechanics [33]
Table 2. Comparative analysis of characteristic indicators of different motion capture methods.
Table 2. Comparative analysis of characteristic indicators of different motion capture methods.
Motion Capture TechnologyAccuracyAdvantagesConstraintsRobustnessRepeatabilityReliabilitySports ScenariosSports Applications
CinematographyHighHigh accuracy, suitable for complex movementsLimited capture volume, marker occlusionMediumHighHighLab-based analysis, technique evaluationBiomechanical analysis, technique optimization, injury prevention
Wearable Sensors Medium No marker occlusion, large capture volume, real-time tracking Prone to electromagnetic interference, lower accuracy than optical systemsHighHighMedium to HighIndoor and outdoor training, competition monitoringReal-time performance tracking, load monitoring, tactical analysis
Computer VisionMedium to HighMarkerless tracking, flexible setupLine of sight, lighting, computationally intensive, sensitive to lighting conditionsMedium to HighHighMediumLab-based analysis, technique evaluationBiomechanical analysis, technique optimization, movement pattern recognition
Others (e.g., Fusion Modality)HighIntegrating the advantages of multiple sensorsSensor synchronizationHighHighHighComprehensive performance analysisMultifaceted performance assessment, injury risk prediction
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Suo, X.; Tang, W.; Li, Z. Motion Capture Technology in Sports Scenarios: A Survey. Sensors 2024, 24, 2947. https://doi.org/10.3390/s24092947

AMA Style

Suo X, Tang W, Li Z. Motion Capture Technology in Sports Scenarios: A Survey. Sensors. 2024; 24(9):2947. https://doi.org/10.3390/s24092947

Chicago/Turabian Style

Suo, Xiang, Weidi Tang, and Zhen Li. 2024. "Motion Capture Technology in Sports Scenarios: A Survey" Sensors 24, no. 9: 2947. https://doi.org/10.3390/s24092947

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop