Next Article in Journal
Robust Localization and Tracking of VRUs with Radar and Ultra-Wideband Sensors for Traffic Safety
Previous Article in Journal
Assessing Time–Frequency Analysis Methods for Non-Stationary EMG Bursts: Application to an Animal Model of Parkinson’s Disease
Previous Article in Special Issue
DynamicVLN: Incorporating Dynamics into Vision-and-Language Navigation Scenarios
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating a Multi-Camera Markerless System for Capturing Basketball-Specific Movements: An Exploration Using 25 Hz Video Streams

1
Institute of Physical Education and Training, Capital University of Physical Education and Sports, Beijing 100191, China
2
School of Athletic Performance, Shanghai University of Sport, Shanghai 200438, China
3
College of Physical Education and Sports, Beijing Normal University, Beijing 100875, China
4
College of Physical Education, Dalian University, Dalian 116622, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work and share first authorship.
Sensors 2026, 26(5), 1689; https://doi.org/10.3390/s26051689
Submission received: 29 January 2026 / Revised: 23 February 2026 / Accepted: 5 March 2026 / Published: 7 March 2026
(This article belongs to the Special Issue Multi-Sensor Systems for Object Tracking—2nd Edition)

Abstract

Markerless motion capture (MMC) provides a non-invasive alternative for motion analysis; however, its validity at the standard frame rate of 25 Hz commonly used in broadcast and surveillance applications remains to be established. This study evaluated the performance of a 25 Hz multi-camera MMC workflow using consumer-grade cameras for capturing basketball-specific movements. Three highly trained male athletes completed seven tasks, including sprinting and simulated sport-specific skills, while being synchronously recorded by six MMC cameras (DJI Action 5 Pro, 25 fps) and a 10-camera Vicon system (25 Hz). Kinematic data were processed using an RTMDet–RTMPose pipeline and low-pass filtered at 6 Hz. Waveform validity was assessed using Pearson’s correlation coefficient (r) and the root mean square error (RMSE). The displacement magnitudes of 12 joints showed excellent agreement (r = 0.916–0.994; median nRMSE = 0.54–1.32%), indicating robust trajectory reconstruction. In contrast, agreement decreased for derivative variables: velocity (r = 0.583–0.867) and acceleration (r = 0.232–0.677) were highly sensitive to the low sampling rate and numerical differentiation. Although a 25 Hz configuration is insufficient for high-precision impact analysis, it provides acceptable accuracy for macroscopic displacement tracking and external-load quantification in resource-constrained training environments. Future optimization should prioritize temporal synchronization to improve the reliability of derivative variables.

1. Introduction

Basketball is characterized by rapid accelerations and decelerations, frequent changes of direction, explosive jump–landing cycles, and highly coordinated upper- and lower-limb actions [1]. Quantifying three-dimensional (3D) joint kinematics during these sport-specific actions is essential for evidence-based training, injury-risk screening, return-to-play decision-making, and long-term performance monitoring [2,3]. In recent years, optical marker-based motion-capture systems (e.g., Vicon) have been widely accepted as a reference method for 3D kinematics because of their high spatial accuracy and mature processing workflows [4]. However, marker-based systems remain difficult to deploy in real-world sport settings, as they require dedicated infrastructure, time-consuming marker placement, and a limited capture volume that can be exceeded by fast, large-displacement training actions [5,6]. In addition, markers may detach, become occluded, or introduce soft-tissue artifacts, all of which can bias kinematic outputs, particularly during high-speed movements [7].
Recent advances in deep-learning-based pose estimation have accelerated the development of markerless motion capture (MMC) systems [8,9]. These approaches identify anatomical keypoints directly from video and reconstruct human motion trajectories, offering non-invasive, low-barrier, and scalable measurement solutions [9,10,11]. Importantly, multi-camera MMC provides redundancy to mitigate partial occlusion and can preserve trajectory continuity even when athletes move near the boundary of the capture volume [12]. Compared with Vicon [4], it is particularly suitable for continuous acquisition and analysis of natural movement patterns in training environments [13,14]. Prior studies have shown that MMC can achieve accuracy comparable to marker-based systems for lower-limb joint kinematics and selected kinetic outcomes. Across multi-task, non-sport-specific settings (e.g., running, jumping, squatting, and cutting), MMC has generally demonstrated acceptable accuracy and reliability for kinematic assessment of the lower limbs, trunk, and upper limbs [4,15,16,17]. Nevertheless, most validation studies rely on expensive high-frame-rate industrial cameras (>60 Hz, and often 100 Hz), which constrain broader adoption in large-scale sport programs and resource-limited settings. In practice, 25 or 30 Hz video streams remain the most common data source in game broadcasts, routine training recordings, and standard venue surveillance systems.
Despite these advances, several challenges must be addressed before multi-camera MMC can be considered a reliable alternative for sport-specific biomechanical measurement. First, most validation work has focused on walking or generic functional tasks, whereas basketball imposes more complex demands on coordination, limb speed, and self-occlusion [9,12]. Second, agreement in positional signals does not necessarily translate to accurate derivatives such as velocity and acceleration. This is particularly critical at low frame rates, where numerical differentiation amplifies small spatial noise and attenuates high-frequency impact-related content [18]. Accordingly, defining the performance boundaries of low-frame-rate MMC—between macroscopic external-load metrics (e.g., displacement and mean velocity) and microscopic kinematic features (e.g., instantaneous velocity)—is crucial for real-world deployment. Third, beyond cross-system validity, practical applications require test–retest reliability, because the usefulness of MMC for training monitoring depends on whether observed changes exceed measurement noise [19,20].
Therefore, this study evaluated the concurrent validity and test–retest reliability of a consumer-grade 25 Hz multi-camera markerless motion capture (MMC) workflow for basketball-specific movements, using a laboratory-grade 10-camera Vicon system as the reference standard. The primary research question was: to what extent does this low-frame-rate MMC approach agree with Vicon when quantifying 3D joint displacement, velocity, and acceleration across major upper- and lower-limb joints during representative basketball tasks? We hypothesized that MMC would show excellent waveform agreement for displacement across joints and tasks (r > 0.90), whereas agreement for velocity and acceleration would be substantially lower. In addition, we assessed test–retest reliability across two testing days to determine whether trial-level MMC metrics are sufficiently repeatable for practical training monitoring applications.

2. Materials and Methods

2.1. Participants

Three experienced male basketball athletes were recruited (age: 25.00 ± 4.58 years; height: 189.33 ± 4.04 cm; body mass: 91.00 ± 17.69 kg; training experience: 7.00 ± 1.00 years). Inclusion criteria were: (i) ability to proficiently perform all experimental tasks; (ii) no history of severe lower-limb injury within the previous two years; and (iii) absence of acute conditions during testing that could affect performance. All participants provided written informed consent. The protocol was approved by the Scientific Research Ethics Committee of Shanghai University of Sport.

2.2. Experimental Setup and Data Acquisition

Experiments were conducted in the Biomechanics Laboratory of Shanghai University of Sport. The overall experimental capture setup is shown in Figure 1. The reference system was a 10-camera Vicon optical motion-capture setup (Vicon Vero) surrounding an approximately 5.33 m × 5.33 m capture area (camera height ~4 m), recorded in Vicon Nexus. This configuration represents a standard laboratory reference setup with high camera redundancy to minimize reconstruction uncertainty. The MMC system used six consumer-grade action cameras (DJI Action 5 Pro, SZ DJI Technology Co., Ltd., Shenzhen, China) positioned around the same capture area (approximately 1 m outside the boundary; height ~2 m), recording at 1920 × 1080 resolution and 25 fps. Both systems were intentionally configured at the same nominal sampling rate (25 Hz) to enable a like-for-like comparison. The MMC camera count was deliberately limited to six to reflect realistic deployment constraints for 3 × 3 basketball training/competition monitoring, where space, budget, setup time, and interference with play make larger portable arrays difficult; consequently, lower redundancy may reduce keypoint visibility and triangulation robustness during partially occluded, high-speed movements [21]. Camera calibration followed a fixed-camera, moving-checkerboard multi-view procedure; intrinsic and extrinsic parameters were estimated to define a unified world coordinate system. Calibration quality was assessed using mean reprojection error (~2 pixels), which was deemed acceptable for subsequent triangulation.

2.3. MMC Pipeline and Marker Configuration

To obtain 3D joint trajectories, we implemented a top-down multi-view 3D reconstruction workflow. Multi-view videos were first temporally synchronized using Mel-frequency cepstral coefficients (MFCC) extracted from audio signals to achieve frame-level alignment across views [22]. For each view, person detection was performed using RTMDet, followed by RTMPose to output COCO-style 2D keypoint coordinates and confidence scores [23,24]. The MMC 2D pose estimation employed the official RTMPose model (OpenMMLab MMPose framework) with COCO-pretrained weights further trained on general human pose datasets, without basketball-specific fine-tuning. This generic, out-of-the-box configuration was deliberately chosen to evaluate the realistic performance boundary of a consumer-grade pipeline readily deployable by coaches and researchers without domain-specific labelled data. Compared with earlier models, RTMPose has demonstrated improved pose-estimation accuracy in prior work [25].
For trajectory tracking, we used the OCSORT tracker combined with SOLIDER re-identification, and introduced a 3D-matching penalty during association to improve robustness [26,27]. After obtaining 2D pose trajectories for each view, 3D pose sequences were reconstructed using triangulation [28]. The MMC model outputs 17 COCO body keypoints (nose; left/right eyes; left/right ears; left/right shoulders; left/right elbows; left/right wrists; left/right hips; left/right knees; left/right ankles). For cross-system comparison, we selected 12 limb-related keypoints (shoulders, elbows, wrists, hips, knees, and ankles; left/right) that have direct anatomical relevance to our basketball-specific movements and can be matched to Vicon markers or marker-derived joint centers. The 5 facial keypoints (nose, eyes, and ears) were excluded because they have no corresponding reflective markers in our Vicon setup, are prone to occlusion and less stable detection in sport-like trials, and are not required for the joint-level kinematic comparisons targeted in this study.
Vicon marker placement followed a modified anatomical-landmark protocol commonly used in optical motion capture [29,30], with minor adaptations to enhance anatomical comparability with the selected MMC COCO keypoints and to improve the stability of marker-derived joint-center estimation. The marker placement configuration and participant standing posture are illustrated in Figure 2. Specifically, 22 reflective markers used for cross-system kinematic analyses were placed on major anatomical landmarks: bilateral shoulders (acromion region; 2), elbows (medial and lateral epicondyles; 4), wrists (radial and ulnar styloid processes; 4), hips/pelvis (bony prominences near the anterior and posterior superior iliac spines; 4), knees (medial and lateral femoral condyles; 4), and ankles (medial and lateral malleoli; 4). In addition, two shoe-mounted markers located near the toe region were used solely to facilitate take-off event identification for cross-system temporal alignment (MMC vs. Vicon); these markers were not included in any subsequent trajectory-based computations, kinematic derivations, or cross-system comparisons.
Accordingly, our primary validation targets were kinematic quantities derived from these joint trajectories—displacement, velocity, and acceleration—because they are directly obtainable from both systems and are most relevant to practical training/competition monitoring. In particular, magnitude-based kinematics are less sensitive to constant spatial offsets that may arise from different anatomical definitions, while numerical differentiation still reflects the real-world propagation of micro-errors into first- and second-order derivatives. By contrast, estimating whole-body center of mass would require additional biomechanical modeling assumptions, which could confound an “out-of-the-box” system-to-system comparison. Center-of-mass metrics can be derived post hoc from the same 3D trajectories and will be explored in future work.

2.4. Task Protocol and Synchronization

Before each testing session, participants completed a ~10 min standardized warm-up (3 min light jogging, 4 min dynamic stretching, and three sets of low-intensity lateral shuffles and small hops) to ensure adequate range of motion and neuromuscular activation. Participants then stood naturally at the center of the capture area for ~3 s to check marker visibility and static posture quality. Prior to data collection, the Vicon system was calibrated following the standard procedure, including camera calibration and wand/plate-based volume definition. An operator moved a T-shaped calibration wand throughout the capture volume with wide-range, multi-directional motions to ensure visibility across cameras, and then placed the wand flat on the floor at the center of the testing area to complete calibration. Reflective markers were placed on four floor reference landmarks, and their 3D coordinates were recorded to establish a unified world coordinate system for that session.
To synchronize Vicon and MMC, before each task trial the researcher gave a verbal cue and the participant performed a single vertical jump in place, then stood still before initiating the task. This event was clearly identifiable in the multi-view videos and produced an abrupt change in Vicon marker trajectories; the corresponding frame was used as the synchronization reference (tsync) to align both systems at the frame level. At least 1 min rest was provided between high-intensity tasks and extended when necessary based on perceived fatigue to minimize fatigue-related degradation in movement quality and trajectory stability. Seven task conditions were tested (see Figure 3 for an overview of the experimental procedure and task set): three key-area locomotion conditions (walking/running/sprinting), a step-in vertical jump, a free-throw posture simulation, a three-step layup, and a crossover step-back jump shot. For the key-area locomotion task, the four corners of the capture area (A, B, C, D) served as path markers; participants performed shuttle movements along the prescribed route at walking, running, and sprinting speeds. To reduce interference from the ball on keypoint detection and marker visibility, all three basketball skill tasks were performed without a ball (simulated movements). This choice minimized occlusion and improved tracking stability for cross-system validation; ball-present conditions will be examined in future work. Each participant performed one trial per task on each testing day. Across two days, this yielded 3 (participants) × 2 (days) × 7 (tasks) = 42 paired synchronized trials for analysis.

2.5. Data Processing and Outcome Variables

Vicon and MMC outputs were exported as time series of 3D joint-center coordinates in the world coordinate system, with the same sampling frequency (25 Hz). The two systems were time-aligned using the synchronization event (a vertical jump performed immediately before each trial). After synchronization, paired trials were merged frame-by-frame using a shared frame index, and only the overlapping frame range present in both systems was retained for subsequent analyses. To reduce the influence of between-trial initial position offsets, joint coordinates were converted to relative displacements with respect to the first retained frame. For each joint, raw coordinates were denoted as X(t), Y(t), and Z(t), and relative displacement components were computed as follows:
x t = X t X t 0 ,   y t = Y t Y t 0 ,   z ( t ) = Z ( t ) Z ( t 0 )
Here, t 0 denotes the first frame of the overlapped segment after frame-wise merging. Each component signal was then filtered using a 4th-order low-pass Butterworth filter with a 6 Hz cutoff, implemented with zero-phase forward–backward filtering (filtfilt) to avoid phase delay [31]. Filtering was implemented in Python 3.10 using scipy.signal.filtfilt. Given the 25 Hz sampling rate and the amplification of high-frequency noise in derivative signals, a conservative 6 Hz cutoff was selected to suppress 2D/3D reconstruction jitter and differentiation noise in velocity/acceleration estimates, aligning with the intended use of consumer devices in resource-limited settings. Velocity and acceleration were computed via numerical differentiation of the filtered displacement signals. The sampling interval was Δ t   = 1 25 S, and the first derivative was estimated using a central-difference scheme. A central-difference scheme was chosen because it is symmetric (time-centered) and second-order accurate, which reduces numerical bias compared with one-sided differences when estimating derivatives from discretely sampled kinematic time series [32].
v x t = x t + 1 x t 1 2 Δ t ,   v y t = y t + 1 y t 1 2 Δ t ,   v z t = z t + 1 z t 1 2 Δ t
Similarly, acceleration components were computed using second-order central differences.
a x t = x t + 1 2 x t + x t 1 ( Δ t ) 2 ,   a y t = y t + 1 2 y t + y t 1 ( Δ t ) 2 , a z t = z t + 1 2 z t + z t 1 ( Δ t ) 2
Because central differences require neighboring frames, the first and last frames of each trial were excluded from derivative-based analyses. In addition to axis-specific components, magnitude (vector-norm) variables were calculated to provide direction-independent summaries of movement intensity.
p o s m a g t = x t 2 + y t 2 + z t 2
v m a g t = v x t 2 + v y t 2 + v z t 2
a m a g t = a x t 2 + a y t 2 + a z t 2
These magnitude variables—displacement magnitude (posmag), velocity magnitude (vmag), and acceleration magnitude (amag)—served as primary outcomes for waveform validity/agreement and inter-system agreement analyses. For reliability analyses, trial-level features were extracted from each magnitude time series, including the mean, peak, and 95th percentile (P95).

2.6. Statistical Analysis

All data processing and statistical analyses were performed in Python 3.10 (NumPy, pandas and SciPy). After frame-level alignment and 6 Hz low-pass filtering, displacement magnitude ( p o s m a g ), velocity magnitude ( v m a g ), and acceleration magnitude ( a m a g ) were computed for 12 joints, using paired trials as the statistical unit [33]. Waveform validity/agreement was assessed using Pearson’s correlation coefficient (r) to quantify shape similarity, and RMSE and nRMSE to quantify amplitude error. nRMSE% was normalized by the full-sample amplitude range (max–min) of the corresponding Vicon waveform for the same joint and variable and expressed as a percentage. Because nRMSE% can be inflated when the reference range is small (e.g., near-static segments), it was interpreted alongside RMSE. Inter-system agreement was evaluated using Bland–Altman analysis, reporting mean bias and 95% limits of agreement (LoA = bias ± 1.96 × SD) [34]. To avoid treating frame-level observations as independent samples, repeated-measures Bland–Altman methods were applied to obtain more robust LoA estimates. Lin’s concordance correlation coefficient (CCC) was reported as a complementary agreement metric [35]. Test–retest reliability across two testing days was assessed using trial-level features (mean, peak, P95) extracted from the magnitude series. Reliability was quantified using the ICC (two-way random-effects, absolute agreement, single measurement; ICC(A,1)) with 95% confidence intervals, and absolute error was summarized using CV%, SEM, and MDC95 ( M D C 95   =   1.96   ×   2   ×   S E M ) [36,37]. For reporting, r and RMSE are presented as mean ± SD, whereas nRMSE% is presented as median [IQR] to reduce denominator effects in cross-condition comparisons. The main text reports test–retest reliability results for the mean feature only, whereas reliability results for peak and P95 are provided in the Supplementary Materials. The same filtering and the same numerical differentiation scheme were applied to both MMC and Vicon signals to ensure a like-for-like comparison. During pipeline development, alternative cutoff settings were evaluated, and a conservative 6 Hz cutoff was adopted to suppress frame-to-frame jitter prior to numerical differentiation.

3. Results

3.1. Data Quality Control and Paired-Trial Overview

At the trial level, MMC and Vicon data were successfully paired in a 1:1 manner, yielding 42 paired trials included for subsequent analyses. All paired trials were frame-aligned and cropped to the overlapping segment using a shared frameindex. Data integrity checks identified one paired trial in which the Vicon right-shoulder trajectory was missing the z-axis component. Accordingly, for trial-level waveform validity analyses (r/RMSE/nRMSE%), this issue affected only the right-shoulder statistics (right-shoulder: n = 41; all other joints: n = 42). For frame-level agreement analyses requiring complete 12-joint vectors (pooled Bland–Altman and CCC), the affected paired trial was excluded a priori based on predefined criteria; thus, frame-level analyses were conducted on 41 paired trials.

3.2. Overall Waveform Validity Across 12 Joints

Table 1, Table 2 and Table 3 summarize the overall waveform validity between MMC and Vicon across 12 joints for three kinematic magnitude variables: displacement magnitude (posmag), velocity magnitude (vmag), and acceleration magnitude (amag). Waveform shape similarity was quantified using Pearson’s correlation coefficient (r), whereas amplitude error was quantified using RMSE and normalized RMSE (nRMSE%, reported as median [IQR]).
As shown in Table 1, displacement magnitude (posmag) demonstrated high waveform agreement across joints (rmean = 0.916–0.994) with small amplitude errors (RMSEmean = 0.075–0.176 m; median nRMSE% = 0.54–1.32%). In Table 2, waveform agreement for velocity magnitude (vmag) decreased relative to displacement (rmean = 0.583–0.867), accompanied by larger errors (RMSEmean = 0.670–1.285 m/s; median nRMSE% = 0.62–3.76%). Table 3 further indicates that acceleration magnitude (amag) exhibited the lowest agreement (rmean = 0.232–0.677) and markedly higher RMSE values (14.294–31.906 m/s2; median nRMSE% = 0.30–8.45%). Overall, waveform validity followed a consistent pattern across variables, with posmag outperforming vmag, and vmag outperforming amag. Joint-wise distributions of waveform correlations are shown in the Supplementary Materials (Figure S6).

3.3. Task-Stratified Waveform Validity Across Seven Movement Tasks

To compare task-dependent differences in waveform agreement between MMC and Vicon, trial-level validity metrics were stratified and summarized by task condition. Table 4 reports, for each of the seven tasks, the across-joint mean correlation coefficient ( r m e a n ) and mean RMSE ( R M S E m e a n ) for three magnitude variables (posmag, vmag, and amag). Because one trial in Tri01 had a missing right-shoulder field, the number of observations was n = 71 for Tri01, whereas all other tasks had n = 72.
For displacement magnitude (posmag), all tasks showed consistently high agreement ( r m e a n = 0.939–0.999). Key-area locomotion (Tri01–Tri03) and the step-in vertical jump (Tri04) approached ceiling-level correlations ( r m e a n = 0.996–0.999) with small errors ( R M S E m e a n = 0.029–0.086 m). Among the basketball skill tasks, the free-throw simulation (Tri05) exhibited the lowest displacement agreement ( r m e a n = 0.939, R M S E m e a n = 0.045 m). The three-step layup (Tri06) and crossover step-back jump shot (Tri07) maintained high correlations ( r m e a n = 0.972–0.978), but with larger errors (Tri06: R M S E m e a n = 0.366 m; Tri07: R M S E m e a n = 0.124 m).
For velocity magnitude (vmag), between-task variability increased markedly ( r m e a n = 0.181–0.866). Key-area locomotion and the step-in vertical jump (Tri01–Tri04) showed moderate-to-high agreement ( r m e a n = 0.778–0.824) with R M S E m e a n = 0.294–0.704 m/s, whereas the free-throw simulation (Tri05) yielded the highest velocity agreement ( r m e a n = 0.866, R M S E m e a n = 0.199 m/s). In contrast, the three-step layup (Tri06) showed the lowest velocity agreement and the largest error ( r m e a n = 0.181, R M S E m e a n = 4.423 m/s).
For acceleration magnitude (amag), task dependence became even more pronounced ( r m e a n = 0.061–0.595). Key-area locomotion tasks showed relatively low correlations (Tri01–Tri03: r m e a n = 0.402–0.434), and the sprinting condition exhibited higher error (Tri03: R M S E m e a n = 17.902 m/s2). The step-in vertical jump and free-throw simulation (Tri04–Tri05) showed higher correlations ( r m e a n = 0.574–0.593) with R M S E m e a n = 5.228–10.321 m/s2, and the crossover step-back jump shot (Tri07) also performed relatively well ( r m e a n = 0.595, R M S E m e a n = 6.002 m/s2). The three-step layup (Tri06) exhibited the lowest acceleration agreement and the largest error ( r m e a n = 0.061, R M S E m e a n = 102.351 m/s2). Records of the anomalous trials in Tri06, supplementary visualizations, and sensitivity analyses are provided in the Supplementary Materials.

3.4. Waveform Visualization

Figure 4 presents a time-series comparison of displacement magnitude (posmag) waveforms across 12 joints in a representative paired trial. Overall, MMC and Vicon showed highly consistent peak–trough timing and overall waveform trends; for a few joints, local amplitude deviations or slight temporal offsets were observed during rapidly changing segments. Representative waveform comparisons for velocity and acceleration magnitudes (vmag and amag) are provided in the Supplementary Materials (Figures S1 and S2).

3.5. Frame-Level Agreement Between MMC and Vicon

To evaluate frame-level agreement, pooled Bland–Altman plots were used for visualization, and repeated-measures Bland–Altman analyses were performed to estimate the overall bias and 95% limits of agreement (LoA). Because one paired trial contained incomplete right-shoulder 3-axis data, this trial was excluded from frame-level analyses, leaving 41 valid paired trials. The resulting numbers of pooled “joint × frame” observations were 510,540 for posmag, 510,048 for vmag, and 509,556 for amag (Table 5). To facilitate interpretation of the frame-level agreement results, it should be noted that each point in the pooled Bland–Altman plot represents a paired MMC–Vicon observation at a specific frame for a specific joint (i.e., a joint–frame pair). The mean difference (bias) reflects the systematic offset between systems, whereas the 95% limits of agreement (LoA = bias ± 1.96 × SD) quantify the expected range within which most frame-level differences fall. Because frames are nested within trials and subjects (i.e., repeated measurements), treating all frames as independent would underestimate variability. Therefore, we report repeated-measures Bland–Altman estimates, which provide more robust LoA by accounting for within-trial/within-subject dependence.
As shown in Figure 5, the posmag differences were centered around zero with relatively small dispersion. Table 5 indicates that the posmag bias was close to zero (bias = 0.0029 m), with LoA ranging from −0.4252 to 0.4310 m. In contrast, differences for vmag and amag exhibited greater dispersion. Table 5 shows slight negative biases for vmag and amag (bias = −0.0163 m/s and −1.0451 m/s2, respectively), accompanied by wider LoA. Results for Lin’s concordance correlation coefficient (CCC) are reported in the Supplementary Materials. The corresponding pooled Bland–Altman plots for vmag and amag are provided in the Supplementary Materials (Figures S3 and S4).

3.6. Test–Retest Reliability and Measurement Error

Test–retest reliability and measurement error were further evaluated across two testing days. For each trial, mean, peak, and 95th percentile (P95) features were extracted from the posmag, vmag, and amag magnitude series, and ICC(A,1), CV%, SEM, and MDC95 were computed. To limit manuscript length, only the mean-feature results are summarized in the main text (Table 6), whereas results for peak and P95 are provided in the Supplementary Materials (Tables S3 and S4).
Table 6 suggests that day-to-day stability for the mean feature was broadly comparable between the two systems. The median CV% values for MMC were 3.97% (posmag), 3.93% (vmag), and 6.22% (amag), while the corresponding values for Vicon were 4.00%, 4.52%, and 12.99%, respectively. The median MDC95 values were 0.111 (posmag), 0.073 (vmag), and 0.504 (amag) for MMC, and 0.113, 0.075, and 0.787 for Vicon, respectively (Table 6). ICC(A,1) estimates are also reported in Table 6.

4. Discussion

This study evaluated the 3D kinematic performance of a markerless motion capture (MMC) approach based on deep-learning pose estimation and multi-view 3D reconstruction for basketball-specific movements, using a laboratory-grade optical motion capture system (Vicon) as the reference. Overall, MMC stably reproduced the major spatiotemporal patterns of joint trajectories at the displacement-magnitude level (posmag): waveform correlations across 12 joints remained high (r = 0.916–0.994), with relatively small displacement errors (RMSE = 0.075–0.176 m). Agreement analyses further supported this finding: repeated-measures Bland–Altman results showed a negligible displacement bias (bias ≈ 0.0029 m) with limits of agreement of approximately −0.4252 to 0.4310 m. These findings suggest that, without reflective markers or complex instrumentation, a multi-camera video–based MMC workflow is practically feasible for training monitoring and movement assessment at the level of displacement trajectory reconstruction and other low-order kinematic features, consistent with prior markerless validation studies reporting high agreement for displacement-related measures [38,39]. Importantly, this displacement-dominant agreement pattern is also aligned with broader markerless validation evidence in functional and sport-like tasks, where position/displacement signals tend to be substantially more robust than numerically differentiated variables [31].
To mitigate the large discrepancies in derivative variables, we focused on suppressing high-frequency jitter before numerical differentiation. In particular, we evaluated alternative low-pass cutoff settings during development and adopted a conservative 6 Hz, zero-phase filter for both systems to improve derivative interpretability, while acknowledging that extreme transients are attenuated under this setting. We did not switch pose networks or perform basketball-specific fine-tuning in the present study because our aim was to benchmark an out-of-the-box MMC workflow, and the dataset did not provide task-specific labeled 2D keypoints or 3D ground truth for supervised adaptation. Nevertheless, domain-specific fine-tuning on basketball imagery (with 2D labels) may further reduce keypoint jitter and could be a practical route to improving velocity and acceleration agreement in future work.
Importantly, agreement decreased as the kinematic order increased, which is attributable to the amplification of subtle temporal offsets and high-frequency jitter during numerical differentiation. Compared with displacement, derivative variables showed reduced agreement: joint-level correlations for velocity magnitude (vmag) were r = 0.583–0.867 with larger errors (RMSE = 0.670–1.285 m/s), while acceleration magnitude (amag) exhibited greater joint-dependent variability (r = 0.232–0.677; RMSE = 14.294–31.906 m/s2). This monotonic degradation from displacement to velocity and then acceleration reflects both aliasing under low sampling rates and noise amplification inherent to numerical differentiation [40]. Under the 6 Hz filtering setting used here, the system captured the macroscopic movement trends but inevitably smoothed high-frequency transient features common in basketball, such as take-off and abrupt deceleration [41]. This observation is consistent with established motion-capture considerations: even when displacement reconstruction is highly consistent, small timing offsets or differences in inherent smoothing strategies can produce substantial deviations in derivative waveforms [42]. Similar conclusions have been reported in other markerless validation efforts, where position-level tracking was generally acceptable but derivative- or high-dynamic outcomes were more error-prone, especially under rapid motions and partial occlusions [43,44]. Accordingly, derivative variables from the 25 Hz MMC workflow are more appropriate for relative, low-frequency/aggregate intensity descriptors (e.g., trial-level mean or percentile-based metrics such as P95) rather than for quantifying instantaneous peak values in absolute terms.
At the task level, MMC yielded higher derivative agreement in relatively controlled movements with minimal occlusion (e.g., free-throw simulation), whereas discrepancies became more pronounced in tasks involving larger displacements, sharper tempo changes, or more complex postural dynamics. For example, in the three-step layup (Tri06), correlations for velocity and acceleration were markedly lower and RMSE values were substantially larger, indicating the weakest inter-system agreement for derivative variables under this task condition. This pattern may be attributable to the task’s rapid translation and posture changes and may also be jointly influenced by capture-volume boundaries, occlusion, and 3D reconstruction stability. In addition, the lower camera redundancy of the six-camera markerless setup may have further amplified these effects during high-speed actions by reducing multi-view keypoint visibility and triangulation robustness when occlusions occur; therefore, the observed task-dependent discrepancies should be interpreted as the result of multiple interacting factors, rather than being attributed to any single cause.
From a joint-specific perspective, proximal joints (e.g., shoulders and hips) generally exhibited higher displacement-level agreement, whereas distal joints—particularly the wrist—were more challenging, likely due to the sensitivity of vision-based tracking to small, highly dynamic segments and the higher probability of occlusion for distal joints during sport-specific actions [44,45]. In addition, COCO keypoint definitions used by MMC are not anatomically identical to joint centers inferred by Vicon from the geometric centers of reflective markers. Such model-definition differences are a known source of systematic offsets; even with highly consistent waveform trends, they can lead to larger dispersion in the magnitude of distal segments [38]. Nevertheless, the mean displacement correlation for the wrist remained above 0.90, indicating that the system can still provide practically useful displacement-trajectory information for most basketball-specific movements.
Regarding test–retest reliability, the two systems showed broadly similar reliability patterns. Displacement metrics were the most repeatable (median CV for posmag ≈ 3%), whereas velocity and acceleration metrics showed greater variability (median CV ≈ 20% and 40%, respectively). Although the study is exploratory with only three participants, the analysis of 42 tightly paired trials and more than 43,000 synchronized frames was sufficient to reveal performance patterns of a 25 Hz system at the algorithm-validation level.
Limitations and practical implications. This study has several limitations. First, the sample size was small (n = 3), and tasks were performed in a controlled laboratory setting without a ball, which may limit generalizability to game-like contexts involving contact, ball-handling, and multi-athlete occlusions. Second, the MMC workflow used six consumer-grade cameras to reflect field-deployment constraints; reduced camera redundancy may decrease multi-view keypoint visibility and triangulation robustness during high-speed, partially occluded actions, thereby disproportionately degrading numerically differentiated velocity and acceleration. Third, we used an out-of-the-box RTMPose model (COCO-style keypoints) without basketball-specific fine-tuning because task-specific labeled 2D keypoints or 3D ground truth were not available for supervised adaptation in the present dataset; domain adaptation may reduce keypoint jitter and improve derivative agreement. Fourth, higher acquisition rates (e.g., 50–100 Hz) were not evaluated, and the conservative 6 Hz low-pass filtering required at 25 Hz to suppress jitter inevitably attenuates high-frequency transients and may underestimate instantaneous derivative peaks. Taken together, these constraints clarify the intended positioning of this work in the trade-off between sampling rate and deployability: the 25 Hz setting represents an engineering compromise aligned with low-barrier training and sideline monitoring, where rapid feedback and feasible deployment often outweigh maximal precision for every instantaneous metric. Under these practical conditions, our findings delineate a “performance boundary” for consumer devices—displacement-dominant outputs can achieve high agreement, whereas velocity and acceleration derived by numerical differentiation remain the primary accuracy bottleneck. Future work should systematically quantify the benefits of higher frame rates under otherwise identical camera layout and processing, and, without substantially increasing deployment cost, prioritize improving synchronization accuracy and calibration stability, increasing camera redundancy where feasible, refining anatomical correspondence of keypoints, and exploring basketball-specific model adaptation and data augmentation, with validation extended to more ecologically complex settings involving multiple athletes, greater occlusion, and game-like contact and constraints.

5. Conclusions

This study validated the practical performance of an MMC workflow based on 25 Hz consumer-grade video streams for basketball-specific movements. The results demonstrate that, under the tested conditions requiring rapid feedback and non-instrumented data collection, this approach can provide accurate displacement-related 3D trajectories (r > 0.91) across joints and tasks. These displacement trajectories were analyzed as relative displacement magnitudes and can directly support displacement-based external-load descriptors such as distance covered, trajectory/path features, and mean speed derived from the 3D trajectories. We note that the present study did not evaluate higher-level biomechanical variables that require additional modeling assumptions, such as joint angles or center-of-mass metrics, which should be examined in dedicated follow-up work. In contrast, agreement for velocity and acceleration was substantially lower, reflecting an observed limitation under the current 25 Hz setting and processing pipeline, where micro timing/phase deviations and keypoint jitter introduced by markerless detection, tracking, and triangulation can be amplified by numerical differentiation. Because sampling rate was not manipulated in this study, isolating frame-rate effects will require future multi-rate experiments. Finally, given that the tasks were performed in a controlled laboratory environment without ball-handling or game-like interactions, we refrain from claiming ecological validity; instead, these findings delineate practical performance boundaries and support the workflow’s deployability potential for low-burden training monitoring, while field validation in more representative settings remains necessary. Future work should prioritize improving multi-camera temporal synchronization and incorporating basketball-specific pose priors to further mitigate the accuracy limitations of derivative variables.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s26051689/s1. SA. Supplementary Methods; SB. Supplementary Figures. Figure S1. Waveforms across 12 joints for velocity magnitude (vmag). (a) Full-scale visualization preserving extreme transient peaks. (b) Zoomed-in visualization where y-axis limits were set for each joint separately to the 1st–99th percentiles, computed from values pooled across MMC and Vicon within that joint, to highlight the bulk waveform patterns. Figure S2. Waveforms across 12 joints for acceleration magnitude (amag). (a) Full-scale visualization preserving extreme transient peaks. (b) Zoomed-in visualization where y-axis limits were set for each joint separately to the 1st–99th percentiles, computed from values pooled across MMC and Vicon within that joint, to highlight the bulk waveform patterns. Figure S3. Pooled Bland–Altman plot of frame-level agreement between MMC and Vicon for velocity magnitude (vmag) across 12 joints (25 Hz; 6 Hz filtering). The center line denotes the mean bias and dashed lines denote the 95% limits of agreement. Figure S4. Pooled Bland–Altman plot of frame-level agreement between MMC and Vicon for acceleration magnitude (amag) across 12 joints (25 Hz; 6 Hz filtering). The center line denotes the mean bias and dashed lines denote the 95% limits of agreement. Figure S5. Task-level waveform validity across seven basketball-specific tasks. Figure S6. Joint-wise distributions of waveform correlation (r) across all paired trials for displacement (posmag), velocity (vmag), and acceleration (amag) magnitudes. Boxplots summarize the median and interquartile range, with whiskers indicating dispersion across trials. Figure S7. Power spectral density (PSD) of displacement magnitude (posmag), velocity magnitude (vmag), and acceleration magnitude (amag) for MMC and Vicon in a representative paired trial, shown before (raw) and after 6 Hz zero-phase low-pass filtering (LPF). PSD curves are summarized as the median across the 12 joints; SC. Supplementary Tables. Table S1. Sensitivity analysis of task-level waveform validity estimates with and without Tri06 (Layup, 3-step). Table S2. Frame-level concordance between MMC and Vicon across 12 joints (pooled joint–frame points), reporting Pearson’s r, Lin’s concordance correlation coefficient (CCC), and mean difference (MMC–Vicon). Table S3. Summary of test–retest reliability for the peak feature (pooled across 12 joints × 7 tasks), including ICC(A,1), CV%, and MDC95. Table S4. Summary of test–retest reliability for the P95 feature (pooled across 12 joints × 7 tasks), including ICC(A,1), CV%, and MDC95. Table S5. Fraction of total power spectral density (PSD) below 6 Hz for displacement, velocity, and acceleration magnitudes (median across 12 joints) for MMC and Vicon, before and after 6 Hz low-pass filtering.

Author Contributions

Conceptualization, Z.L. and M.Z.; methodology, Z.L. and Z.T.; software, Z.L.; validation, Z.T., W.Z. and G.Y.; formal analysis, Z.L. and J.T.; investigation, Z.L., W.Z. and G.Y.; resources, W.Z., G.Y. and M.Z.; data curation, Z.L. and Z.T.; writing—original draft preparation, Z.L., Z.T. and J.T.; writing—review and editing, W.Z., G.Y., M.Z. and X.X.; visualization, Z.L. and W.Z.; supervision, M.Z. and X.X.; project administration, M.Z. and X.X.; funding acquisition, M.Z. and X.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shanghai Sports Bureau, grant number 25J022, and the Shanghai Municipal Education Commission, grant number A1-N3017-24-0013-3.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Shanghai University of Sport (protocol code: 102772025RT095 and date of approval: 25 March 2025).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Due to privacy and ethical restrictions, the raw multi-view video recordings collected for this study are not publicly available. The processed kinematic datasets exported from the reference optical motion-capture system (Vicon) and the markerless motion-capture pipeline (MMC) (i.e., 3D joint-center trajectories in the global coordinate system) will be made publicly available in an open repository upon acceptance of this manuscript; prior to that, these data are available from the corresponding authors upon reasonable request. The source code of the MMC pipeline cannot be shared publicly because it contains proprietary components protected by intellectual property/patent restrictions. Key processing parameters and outcome-variable definitions are provided in the manuscript and Supplementary Materials.

Acknowledgments

The authors wish to express their sincere gratitude to Wei Wang and Yuchen Yang from Shanghai Artificial Intelligence Laboratory for their significant technical support throughout this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Harper, D.J.; McBurnie, A.J.; Santos, T.D.; Eriksrud, O.; Evans, M.; Cohen, D.D.; Rhodes, D.; Carling, C.; Kiely, J. Biomechanical and Neuromuscular Performance Requirements of Horizontal Deceleration: A Review with Implications for Random Intermittent Multi-Directional Sports. Sports Med. 2022, 52, 2321–2354. [Google Scholar] [CrossRef]
  2. Di Paolo, S.; Lopomo, N.F.; Della Villa, F.; Paolini, G.; Figari, G.; Bragonzoni, L.; Grassi, A.; Zaffagnini, S. Rehabilitation and Return to Sport Assessment after Anterior Cruciate Ligament Injury: Quantifying Joint Kinematics during Complex High-Speed Tasks through Wearable Sensors. Sensors 2021, 21, 2331. [Google Scholar] [CrossRef]
  3. Straub, R.K.; Powers, C.M. Biomechanical Predictors of Primary ACL Injury: A Scoping Review of Prospective Studies. Gait Posture 2025, 116, 22–29. [Google Scholar] [CrossRef]
  4. Turner, J.A.; Chaaban, C.R.; Padua, D.A. Validation of OpenCap: A Low-Cost Markerless Motion Capture System for Lower-Extremity Kinematics during Return-to-Sport Tasks. J. Biomech. 2024, 171, 112200. [Google Scholar] [CrossRef] [PubMed]
  5. Brambilla, C.; Marani, R.; Romeo, L.; Lavit Nicora, M.; Storm, F.A.; Reni, G.; Malosio, M.; D’Orazio, T.; Scano, A. Azure Kinect Performance Evaluation for Human Motion and Upper Limb Biomechanical Analysis. Heliyon 2023, 9, e21606. [Google Scholar] [CrossRef] [PubMed]
  6. Yang, C.; Wei, L.; Huang, X.; Tu, L.; Xu, Y.; Li, X.; Hu, Z. Comparison of Lower Limb Kinematic and Kinetic Estimation during Athlete Jumping between Markerless and Marker-Based Motion Capture Systems. Sci. Rep. 2025, 15, 18552. [Google Scholar] [CrossRef] [PubMed]
  7. Ancillao, A.; Aertbelien, E.; De Schutter, J. Effect of the Soft Tissue Artifact on Marker Measurements and on the Calculation of the Helical Axis of the Knee during a Squat Movement: A Study on the CAMS-Knee Dataset. Med. Eng. Phys. 2022, 110, 103915. [Google Scholar] [CrossRef]
  8. Colyer, S.L.; Evans, M.; Cosker, D.P.; Salo, A.I.T. A Review of the Evolution of Vision-Based Motion Analysis and the Integration of Advanced Computer Vision Methods towards Developing a Markerless System. Sports Med.—Open 2018, 4, 24. [Google Scholar] [CrossRef]
  9. Edriss, S.; Romagnoli, C.; Caprioli, L.; Bonaiuto, V.; Padua, E.; Annino, G. Commercial Vision Sensors and AI-Based Pose Estimation Frameworks for Markerless Motion Analysis in Sports and Exercises: A Mini Review. Front. Physiol. 2025, 16, 1649330. [Google Scholar] [CrossRef]
  10. Moran, M.F.; Rogler, I.C.; Wager, J.C. Inter-Session Repeatability of Marker-Less Motion Capture of Treadmill Running Gait. Appl. Sci. 2023, 13, 1702. [Google Scholar] [CrossRef]
  11. Qian, C.; Li, D.; Yu, X.; Yang, Z.; Ma, Q. OpenMoCap: Rethinking Optical Motion Capture under Real-World Occlusion. In Proceedings of the 33rd ACM International Conference on Multimedia, Dublin, Ireland, 27–31 October 2025; pp. 7529–7537. [Google Scholar]
  12. Scataglini, S.; Abts, E.; Van Bocxlaer, C.; Van den Bussche, M.; Meletani, S.; Truijen, S. Accuracy, Validity, and Reliability of Markerless Camera-Based 3D Motion Capture Systems versus Marker-Based 3D Motion Capture Systems in Gait Analysis: A Systematic Review and Meta-Analysis. Sensors 2024, 24, 3686. [Google Scholar] [CrossRef] [PubMed]
  13. El Kaid, A.; Baïna, K. A Systematic Review of Recent Deep Learning Approaches for 3D Human Pose Estimation. J. Imaging 2023, 9, 275. [Google Scholar] [CrossRef]
  14. Needham, L.; Evans, M.; Wade, L.; Cosker, D.P.; McGuigan, M.P.; Bilzon, J.L.; Colyer, S.L. The Development and Evaluation of a Fully Automated Markerless Motion Capture Workflow. J. Biomech. 2022, 144, 111338. [Google Scholar] [CrossRef] [PubMed]
  15. Lam, W.W.T.; Fong, K.N.K. Validity and Reliability of Upper Limb Kinematic Assessment Using a Markerless Motion Capture (MMC) System: A Pilot Study. Arch. Phys. Med. Rehabil. 2024, 105, 673–681.e2. [Google Scholar] [CrossRef]
  16. Verheul, J.; Hughes, O.; Hitchens, L.; Atherton, T.; Sauter, T.; Radwan, A.; Pryce, E.; Mitchell, A.; Burton, S. Markerless Motion Capture for Running: Validity and Reliability of Whole-Body, Joint, and Muscle Kinematics. J. Biomech. 2025, 195, 113133. [Google Scholar] [CrossRef] [PubMed]
  17. Lima, Y.L.; Collings, T.; Hall, M.; Bourne, M.N.; Diamond, L.E. Validity and Reliability of Trunk and Lower-Limb Kinematics during Squatting, Hopping, Jumping and Side-Stepping Using OpenCap Markerless Motion Capture Application. J. Sports Sci. 2024, 42, 1847–1858. [Google Scholar] [CrossRef]
  18. Chartrand, R. Numerical Differentiation of Noisy, Nonsmooth Data. Int. Sch. Res. Not. 2011, 2011, 164564. [Google Scholar] [CrossRef]
  19. Bland, J.M.; Altman, D.G. Statistical Methods for Assessing Agreement between Two Methods of Clinical Measurement. Int. J. Nurs. Stud. 2010, 47, 931–936. [Google Scholar] [CrossRef]
  20. Das, K.; de Paula Oliveira, T.; Newell, J. Comparison of Markerless and Marker-Based Motion Capture Systems Using 95% Functional Limits of Agreement in a Linear Mixed-Effects Modelling Framework. Sci. Rep. 2023, 13, 22880. [Google Scholar] [CrossRef]
  21. Zheng, W.; Zhang, M.; Dong, R.; Qiu, M.; Wang, W. Feasibility and Accuracy of an RTMPose-Based Markerless Motion Capture System for Single-Player Tasks in 3x3 Basketball. Sensors 2025, 25, 4003. [Google Scholar] [CrossRef]
  22. Muda, L.; Begam, M.; Elamvazuthi, I. Voice Recognition Algorithms Using Mel Frequency Cepstral Coefficient (MFCC) and Dynamic Time Warping (DTW) Techniques. arXiv 2010, arXiv:1003.4083. [Google Scholar] [CrossRef]
  23. Lyu, C.; Zhang, W.; Huang, H.; Zhou, Y.; Wang, Y.; Liu, Y.; Zhang, S.; Chen, K. RTMDet: An Empirical Study of Designing Real-Time Object Detectors. arXiv 2022, arXiv:2212.07784. [Google Scholar] [CrossRef]
  24. Jiang, T.; Lu, P.; Zhang, L.; Ma, N.; Han, R.; Lyu, C.; Li, Y.; Chen, K. Rtmpose: Real-Time Multi-Person Pose Estimation Based on Mmpose. arXiv 2023, arXiv:2303.07399. [Google Scholar]
  25. Perez, M.; Liu, J.; Kot, A.C. Skeleton-Based Relational Reasoning for Group Activity Analysis. Pattern Recognit. 2022, 122, 108360. [Google Scholar] [CrossRef]
  26. Cao, J.; Pang, J.; Weng, X.; Khirodkar, R.; Kitani, K. Observation-Centric SORT: Rethinking SORT for Robust Multi-Object Tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 9686–9696. [Google Scholar]
  27. Chen, W.; Xu, X.; Jia, J.; Luo, H.; Wang, Y.; Wang, F.; Jin, R.; Sun, X. Beyond Appearance: A Semantic Controllable Self-Supervised Learning Framework for Human-Centric Visual Tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 15050–15061. [Google Scholar]
  28. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  29. Wu, G.; Helm, F.C.T.; Van der Helm, F.C.; Veeger, H.D.; Makhsous, M.; Roy, P.V.; Anglin, C.; Nagels, J.; Karduna, A.R.; McQuade, K.; et al. ISB Recommendation on Definitions of Joint Coordinate Systems of Various Joints for the Reporting of Human Joint Motion—Part II: Shoulder, Elbow, Wrist and Hand. J. Biomech. 2005, 38, 981–992. [Google Scholar] [CrossRef] [PubMed]
  30. Wu, G.; Siegler, S.; Allard, P.; Kirtley, C.; Leardini, A.; Rosenbaum, D.; Whittle, M.; D’Lima, D.D.; Cristofolini, L.; Witte, H.; et al. ISB Recommendation on Definitions of Joint Coordinate System of Various Joints for the Reporting of Human Joint Motion—Part I: Ankle, Hip, and Spine. J. Biomech. 2002, 35, 543–548. [Google Scholar] [CrossRef]
  31. Song, K.; Hullfish, T.J.; Scattone Silva, R.; Silbernagel, K.G.; Baxter, J.R. Markerless Motion Capture Estimates of Lower Extremity Kinematics and Kinetics Are Comparable to Marker-Based across 8 Movements. J. Biomech. 2023, 157, 111751. [Google Scholar] [CrossRef]
  32. Edwards, W.B.; Derrick, T.R.; Hamill, J. Time Series Analysis in Biomechanics. In Handbook of Human Motion; Springer Nature: Berlin/Heidelberg, Germany, 2017; pp. 1–24. [Google Scholar]
  33. Winter, D.A. Biomechanics and Motor Control of Human Movement; John Wiley & Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  34. Bland, J.M.; Altman, D.G. Agreement between Methods of Measurement with Multiple Observations per Individual. J. Biopharm. Stat. 2007, 17, 571–582. [Google Scholar] [CrossRef]
  35. Lawrence, I.; Lin, K. A Concordance Correlation Coefficient to Evaluate Reproducibility. Biometrics 1989, 45, 255–268. [Google Scholar]
  36. Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef]
  37. Weir, J.P. Quantifying Test-Retest Reliability Using the Intraclass Correlation Coefficient and the SEM. J. Strength Cond. Res. 2005, 19, 231–240. [Google Scholar]
  38. Oonk, G.; Kempe, M.; Lemmink, K.; Buurke, T. Examining the Concurrent Validity of Markerless Motion Capture in Dual-Athlete Team Sports Movements. J. Sports Sci. 2025, 1–12. [Google Scholar] [CrossRef] [PubMed]
  39. Bae, K.; Lee, S.; Bak, S.-Y.; Kim, H.S.; Ha, Y.; You, J.H. Concurrent Validity and Test Reliability of the Deep Learning Markerless Motion Capture System during the Overhead Squat. Sci. Rep. 2024, 14, 29462. [Google Scholar] [CrossRef] [PubMed]
  40. Hindle, B.R.; Keogh, J.W.; Lorimer, A.V. Inertial-Based Human Motion Capture: A Technical Summary of Current Processing Methodologies for Spatiotemporal and Kinematic Measures. Appl. Bionics Biomech. 2021, 2021, 6628320. [Google Scholar] [CrossRef]
  41. Crenna, F.; Rossi, G.B.; Berardengo, M. Filtering Biomechanical Signals in Movement Analysis. Sensors 2021, 21, 4580. [Google Scholar] [CrossRef]
  42. Needham, L.; Evans, M.; Cosker, D.P.; Colyer, S.L. Can Markerless Pose Estimation Algorithms Estimate 3D Mass Centre Positions and Velocities during Linear Sprinting Activities? Sensors 2021, 21, 2889. [Google Scholar] [CrossRef]
  43. D’Haene, M.; Chorin, F.; Colson, S.S.; Guérin, O.; Zory, R.; Piche, E. Validation of a 3D Markerless Motion Capture Tool Using Multiple Pose and Depth Estimations for Quantitative Gait Analysis. Sensors 2024, 24, 7105. [Google Scholar] [CrossRef]
  44. Lahkar, B.K.; Muller, A.; Dumas, R.; Reveret, L.; Robert, T. Accuracy of a Markerless Motion Capture System in Estimating Upper Extremity Kinematics during Boxing. Front. Sports Act. Living 2022, 4, 939980. [Google Scholar] [CrossRef] [PubMed]
  45. Needham, L.; Evans, M.; Cosker, D.P.; Wade, L.; McGuigan, P.M.; Bilzon, J.L.; Colyer, S.L. The Accuracy of Several Pose Estimation Methods for 3D Joint Centre Localisation. Sci. Rep. 2021, 11, 20673. [Google Scholar] [CrossRef]
Figure 1. Experimental site for the dual-system synchronized validation.
Figure 1. Experimental site for the dual-system synchronized validation.
Sensors 26 01689 g001
Figure 2. Schematic of marker placement.
Figure 2. Schematic of marker placement.
Sensors 26 01689 g002
Figure 3. Schematic of the experimental procedure.
Figure 3. Schematic of the experimental procedure.
Sensors 26 01689 g003
Figure 4. MMC vs. Vicon waveforms across 12 joints for displacement magnitude (posmag) in a representative trial. Vicon is shown in orange and MMC in blue (both solid lines).
Figure 4. MMC vs. Vicon waveforms across 12 joints for displacement magnitude (posmag) in a representative trial. Vicon is shown in orange and MMC in blue (both solid lines).
Sensors 26 01689 g004
Figure 5. Pooled Bland–Altman plot of frame-level agreement between MMC and Vicon for displacement magnitude (posmag) across 12 joints.
Figure 5. Pooled Bland–Altman plot of frame-level agreement between MMC and Vicon for displacement magnitude (posmag) across 12 joints.
Sensors 26 01689 g005
Table 1. Overall validity summary for displacement magnitude (posmag) across 12 joints.
Table 1. Overall validity summary for displacement magnitude (posmag) across 12 joints.
Jointr (Mean ± SD)RMSE (Mean ± SD)nRMSE% (Median [Q1, Q3])n
L Shoulder0.994 ± 0.0080.079 ± 0.1060.59 [0.46, 1.23]42
R Shoulder0.993 ± 0.0090.079 ± 0.1100.63 [0.51, 1.13]41
L Elbow0.994 ± 0.0090.087 ± 0.1020.65 [0.40, 2.56]42
R Elbow0.994 ± 0.0090.088 ± 0.1180.92 [0.44, 1.17]42
L Wrist0.916 ± 0.2010.176 ± 0.1951.32 [0.82, 7.09]42
R Wrist0.959 ± 0.1330.160 ± 0.3030.57 [0.46, 0.99]42
L Hip0.989 ± 0.0250.075 ± 0.1050.74 [0.45, 0.98]42
R Hip0.990 ± 0.0240.086 ± 0.1060.81 [0.68, 1.46]42
L Knee0.992 ± 0.0120.118 ± 0.1530.63 [0.31, 3.95]42
R Knee0.994 ± 0.0080.110 ± 0.1360.72 [0.36, 3.77]42
L Ankle0.992 ± 0.0140.081 ± 0.1140.58 [0.24, 1.16]42
R Ankle0.989 ± 0.0350.101 ± 0.1670.54 [0.22, 1.85]42
Note: Values are reported as r (mean ± SD), RMSE (mean ± SD), and nRMSE% (median [Q1, Q3]) for each joint.
Table 2. Overall validity summary for velocity magnitude (vmag) across 12 joints.
Table 2. Overall validity summary for velocity magnitude (vmag) across 12 joints.
Jointr (Mean ± SD)RMSE (Mean ± SD)nRMSE% (Median [Q1, Q3])n
L Shoulder0.841 ± 0.2860.670 ± 1.4393.06 [2.04, 4.53]42
R Shoulder0.867 ± 0.2430.711 ± 1.5503.32 [2.24, 4.14]41
L Elbow0.694 ± 0.3210.955 ± 1.3201.46 [0.82, 8.67]42
R Elbow0.640 ± 0.2621.067 ± 1.4543.21 [1.16, 5.50]42
L Wrist0.616 ± 0.3701.285 ± 1.7673.76 [0.70, 5.55]42
R Wrist0.799 ± 0.3150.807 ± 1.4832.04 [1.61, 2.97]42
L Hip0.712 ± 0.2810.862 ± 1.4211.69 [0.35, 3.95]42
R Hip0.583 ± 0.2941.031 ± 1.4532.74 [0.70, 5.42]42
L Knee0.612 ± 0.3081.231 ± 1.5512.15 [1.19, 10.10]42
R Knee0.658 ± 0.3141.224 ± 1.5532.55 [0.43, 10.22]42
L Ankle0.806 ± 0.3130.838 ± 1.4640.62 [0.44, 3.58]42
R Ankle0.836 ± 0.2850.809 ± 1.5960.70 [0.51, 1.87]42
Note: Values are reported as r (mean ± SD), RMSE (mean ± SD), and nRMSE% (median [Q1, Q3]) for each joint.
Table 3. Overall validity summary for acceleration magnitude (amag) across 12 joints.
Table 3. Overall validity summary for acceleration magnitude (amag) across 12 joints.
Jointr (Mean ± SD)RMSE (Mean ± SD)nRMSE% (Median [Q1, Q3])n
L Shoulder0.584 ± 0.23814.294 ± 32.7566.75 [4.31, 8.31]42
R Shoulder0.573 ± 0.20415.600 ± 35.4488.45 [5.54, 10.92]41
L Elbow0.346 ± 0.30423.986 ± 30.7381.70 [0.43, 7.80]42
R Elbow0.257 ± 0.27727.185 ± 32.7123.05 [1.30, 5.92]42
L Wrist0.386 ± 0.32931.906 ± 44.4542.97 [0.33, 4.56]42
R Wrist0.557 ± 0.30116.715 ± 33.9643.16 [2.35, 4.82]42
L Hip0.346 ± 0.26420.082 ± 32.9311.07 [0.20, 4.31]42
R Hip0.232 ± 0.27925.378 ± 33.1082.85 [0.28, 4.61]42
L Knee0.292 ± 0.28930.656 ± 36.3662.08 [1.15, 8.16]42
R Knee0.388 ± 0.29430.330 ± 37.0412.86 [0.25, 9.09]42
L Ankle0.615 ± 0.36919.171 ± 34.3150.30 [0.22, 3.58]42
R Ankle0.677 ± 0.30017.219 ± 37.3500.30 [0.20, 0.84]42
Note: Values are reported as r (mean ± SD), RMSE (mean ± SD), and nRMSE% (median [Q1, Q3]) for each joint.
Table 4. Waveform validity results under different task conditions.
Table 4. Waveform validity results under different task conditions.
Task IDTask Nameposmag rmean (posmag)RMSE (posmag, m)rmean (vmag)RMSE (vmag, m/s)rmean (amag)RMSE (amag, m/s2)
Tri01Lane Drill Walk0.9990.0350.7840.2940.4047.633
Tri02Lane Drill Run0.9990.0390.8000.3570.4349.425
Tri03Lane Drill Sprint0.9960.0860.7780.7040.40217.902
Tri04Drop-step vertical jump0.9980.0290.8240.3890.57410.321
Tri05Free-throw simulation0.9390.0450.8660.1990.5935.228
Tri06Layup (3-step)0.9780.3660.1814.4230.061102.351
Tri07Crossover + step-back jump shot0.9720.1240.8200.3310.5956.002
Table 5. Repeated-measures Bland–Altman summary of frame-level agreement between MMC and Vicon across 12 joints. Bias indicates the mean MMC–Vicon difference; LoA denotes the 95% limits of agreement. N represents pooled joint–frame observations (paired values at each frame for each joint) after excluding one trial with incomplete joint coordinates.
Table 5. Repeated-measures Bland–Altman summary of frame-level agreement between MMC and Vicon across 12 joints. Bias indicates the mean MMC–Vicon difference; LoA denotes the 95% limits of agreement. N represents pooled joint–frame observations (paired values at each frame for each joint) after excluding one trial with incomplete joint coordinates.
VariableBiasLoA (Lower)LoA (Upper)N (Joint–Frame Points)Subjects
posmag (m)0.0029−0.42520.4310510,5403
vmag (m/s)−0.0163−6.32456.2919510,0483
amag (m/s2)−1.0451−194.4521192.3620509,5563
Note: N represents joint–frame observations pooled across 12 joints and all valid frames.
Table 6. Test–retest reliability summary (mean feature) for MMC and Vicon across 12 joints and 7 tasks.
Table 6. Test–retest reliability summary (mean feature) for MMC and Vicon across 12 joints and 7 tasks.
SystemVariableICC(A,1) (Median, Range)CV% (Median, Range)MDC95 (Median)
MMCposmag0.00 (−2.01–1.00)3.97 (0.00–39.22)0.111
MMCvmag0.52 (−0.25–0.99)3.93 (0.09–16.45)0.073
MMCamag0.44 (−0.34–0.99)6.22 (0.00–23.07)0.504
Viconposmag0.29 (−0.33–1.00)4.00 (0.00–39.23)0.113
Viconvmag0.37 (−0.20–0.99)4.52 (0.05–15.85)0.075
Viconamag0.28 (−0.31–0.99)12.99 (0.00–51.05)0.787
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Tan, Z.; Zheng, W.; Yang, G.; Tao, J.; Zhang, M.; Xu, X. Evaluating a Multi-Camera Markerless System for Capturing Basketball-Specific Movements: An Exploration Using 25 Hz Video Streams. Sensors 2026, 26, 1689. https://doi.org/10.3390/s26051689

AMA Style

Li Z, Tan Z, Zheng W, Yang G, Tao J, Zhang M, Xu X. Evaluating a Multi-Camera Markerless System for Capturing Basketball-Specific Movements: An Exploration Using 25 Hz Video Streams. Sensors. 2026; 26(5):1689. https://doi.org/10.3390/s26051689

Chicago/Turabian Style

Li, Zhaoyu, Zhenbin Tan, Wen Zheng, Ganling Yang, Junye Tao, Mingxin Zhang, and Xiao Xu. 2026. "Evaluating a Multi-Camera Markerless System for Capturing Basketball-Specific Movements: An Exploration Using 25 Hz Video Streams" Sensors 26, no. 5: 1689. https://doi.org/10.3390/s26051689

APA Style

Li, Z., Tan, Z., Zheng, W., Yang, G., Tao, J., Zhang, M., & Xu, X. (2026). Evaluating a Multi-Camera Markerless System for Capturing Basketball-Specific Movements: An Exploration Using 25 Hz Video Streams. Sensors, 26(5), 1689. https://doi.org/10.3390/s26051689

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop