Next Article in Journal
Risk Assessment Method for CPS-Based Distributed Generation Cluster Control in Active Distribution Networks Under Cyber Attacks
Previous Article in Journal
Monocular Visual/IMU/GNSS Integration System Using Deep Learning-Based Optical Flow for Intelligent Vehicle Localization
Previous Article in Special Issue
Analysis of AWS Rekognition and Azure Custom Vision Performance in Parking Sign Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CGFusionFormer: Exploring Compact Spatial Representation for Robust 3D Human Pose Estimation with Low Computation Complexity

College of Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(19), 6052; https://doi.org/10.3390/s25196052
Submission received: 29 August 2025 / Revised: 25 September 2025 / Accepted: 29 September 2025 / Published: 1 October 2025

Abstract

Transformer-based 2D-to-3D lifting methods have demonstrated outstanding performance in 3D human pose estimation from 2D pose sequences. However, they still encounter challenges with the relatively poor quality of 2D joints and substantial computational costs. In this paper, we propose a CGFusionFormer to address these problems. We propose a compact spatial representation (CSR) to robustly generate local spatial multihypothesis features from part of the 2D pose sequence. Specifically, CSR models spatial constraints based on body parts and incorporates 2D Gaussian filters and nonparametric reduction to improve spatial features against low-quality 2D poses and reduce the computational cost of subsequent temporal encoding. We design a residual-based Hybrid Adaptive Fusion module that combines multihypothesis features with global frequency domain features to accurately estimate the 3D human pose with minimal computational cost. We realize CGFusionFormer with a PoseFormer-like transformer backbone. Extensive experiments on the challenging Human3.6M and MPI-INF-3DHP benchmarks show that our method outperforms prior transformer-based variants in short receptive fields and achieves a superior accuracy–efficiency trade-off. On Human3.6M (sequence length 27, 3 input frames), it achieves 47.6 mm Mean Per Joint Position Error (MPJPE) at only 71.3 MFLOPs, representing about a 40 percent reduction in computation compared with PoseFormerV2 while attaining better accuracy. On MPI-INF-3DHP (81-frame sequences), it reaches 97.9 Percentage of Correct Keypoints (PCK), 78.5 Area Under the Curve (AUC), and 27.2 mm MPJPE, matching the best PCK and achieving the lowest MPJPE among the compared methods under the same setting.

1. Introduction

3D human pose estimation (HPE) is an active research topic in computer vision that restores key point (such as joints, limbs, etc.) positions of the human body in three-dimensional space from camera images (monocular or multi-view) or from video sequences and then reconstructs the human pose. It is one of the fundamental techniques used in human behavior analysis and serves as the foundation for many fields of sensor perception application, such as motion analysis, medical rehabilitation, human–computer interaction, and virtual reality. Many of the latest methods choose 2D HPE as an intermediate to estimate 3D pose based on 2D poses (or 3D pose sequence based on 2D pose sequence), called 2D-to-3D lifting methods, which generally involves the following steps: 2D pose exploration, 2D-to-3D lifting, and feature fusion. Due to the fact that the above algorithms only take 2D poses as input, it is highly dependent on the estimation accuracy of 2D poses. Once 2D HPE fails, it will seriously affect subsequent 2D-to-3D lifting. Therefore, exploring the 2D HPE produced by offline 2D pose estimators is still a fundamental prerequisite to achieve robust 3D HPE.
The 2D-to-3D lifting methods [1,2,3,4,5] are gaining prominence in 3D human pose estimation over data-heavy approaches [6,7,8,9,10] of directly extracting 3D poses from videos. However, the effectiveness of lifting-based methods is constrained by the quality of 2D pose detection. In addition, handling a large number of temporal features also incurs significant computational costs. In practice, 2D detections are frequently degraded by occlusion, truncation, fast motion, and extreme viewpoints, producing temporally jittery joint trajectories and biased limb lengths that propagate to the lifting stage and exacerbate depth ambiguity. From an efficiency perspective, the token complexity of spatio-temporal attention scales with sequence length and joint/channel width, while frequency–domain pipelines often require dimensional expansion for alignment, increasing memory traffic and multiply–accumulate operations. Consequently, naively enlarging temporal context or stacking multiple hypotheses tends to trade robustness for computational cost, motivating compact, noise-aware spatial encodings before temporal modeling. Recently, the requirement to model long-range temporal dependencies in 2D human pose sequences makes transformers an excellent fit for 2D-to-3D lifting methods [11,12,13,14,15,16]. PoseFormer [17] introduces a widely adopted architecture comprising spatial transformer encoders and temporal transformer encoders. To counteract unreliable 2D joint detection, MHFormer [18] enhances robustness by repeatedly encoding to generate diverse pose hypotheses. This approach directly applies temporal self-attention to spatial representations, incurring substantial computational costs. PoseFormerV2 [19] adopts a computationally efficient frequency domain approach based on Discrete Cosine Transform (DCT) coefficients and low-pass filters to model skeletal sequences. However, to ensure feature alignment, it expands the dimensionality of frequency–domain features before temporal encoding, which leads to an increased computational cost.
In this paper, we propose a transformer-based CGFusionFormer to robustly estimate the three-dimensional human pose from a 2D pose sequence in a low computational complexity way, as shown in Figure 1. Our key insight is how to make full use of local and global correlations in the 2D pose sequence and to effectively extract both features, even if the 2D pose is not very accurate, and then efficiently fuse and utilize. First, we focus on exploring spatial correlation among joints and propose a compact spatial representation (CSR), which generates local spatial multihypothesis features (called CSR domain features) from part of the 2D pose sequence. Using a Body-Part Spatial Transformer Encoder (BPS) and Filter-based Multihypothesis Compression (FMC), CSR improves the robustness of spatial feature representation against low-quality 2D pose detection and reduces the computational costs of subsequent temporal encoding. Then, we design a residual network structure, hybrid adaptive fusion (HAF), which effectively integrates CSR domain features and frequency domain features to accurately estimate the 3D human pose. The frequency domain features are global frequency features of the entire 2D pose sequence obtained in the same way as in PoseFormerV2 [19]. The experimental results on the Human3.6M [20] and MPI-INF-3DHP [21] datasets demonstrate that our work achieves an outstanding trade-off between accuracy and speed with promising 3D HPE. Our main contributions are as follows:
  • We propose a compact spatial representation (CSR) that reduces subsequent computational costs while enhancing the capability against unreliable 2D pose detection. CSR is the first attempt to explore 2D pose sequences locally while reducing subsequent 2D-to-3D lifting computational cost at the same time.
  • We design an effective hybrid adaptive fusion (HAF) to integrate features in both the CSR domain and the frequency domain, thereby improving the accuracy of 3D pose estimation at the expense of low computational overhead. HAF enriches the residual network structure and can adapt to the general fusion of multiple features.
  • Our CGFusionFormer achieves comparable results on two challenging datasets for 3D HPE, demonstrating a superior speed–accuracy trade-off compared to previous methods.

2. Related Work

Our work emphasizes robustness and explores a compact spatial representation (CSR) to achieve model lightweighting. This section delves into transformer-based techniques in 2D–3D lifting model design and analyzes and summarizes strategies for enhancing visual processing robustness, specifically in 3D human pose estimation.

2.1. Transformer-Based Methods for 3D Human Pose Estimation

PoseFormer is the inaugural fully transformer-based network in pose estimation, with its spatial–temporal architecture becoming a widely adopted standard in subsequent research. CrossFormer [22] refines the internal organization of spatial and temporal transformers by introducing cross-frame interaction modules and cross-joint feature interactions, thereby strengthening spatio-temporal coupling. StriderFormer [23] redesigns the temporal branch with a strided transformer encoder to mine inter-frame dependencies efficiently, while P-STMO [12] integrates multi-frame fusion into a unified spatio-temporal transformer. Beyond architecture, P-STMO [12] also adopts an MAE-style pre-training strategy [24] with random masking in space and time to let the encoder learn the joint correlation between frames and within frames.
A parallel line of work, inspired by the seq2seq paradigm [25], performs a direct sequence-to-sequence lifting from 2D to 3D. MixSTE [26] employs a recurrent spatio-temporal encoding scheme built upon PoseFormer [17], enhancing temporal attention at the joint-trajectory level so that joint-specific temporal relationships can be captured. MixSTE [26] also shows that spatial and temporal interactions are fused rather than isolated. HSTFormer [27] divides 17 joints into 6 body parts, facilitating the modeling of temporal correlation between parts. MotionBert [28] and other multitask designs further boost accuracy by leveraging temporal signals, albeit at increased computational cost.
Taken together, most transformer-based approaches chiefly advance temporal modeling—via cross-frame coupling, sequence-to-sequence lifting, body-part grouping, and pretraining—whereas compact and informative spatial encodings remain comparatively underexplored. Addressing this gap requires lightweight spatial priors that can be cleanly integrated into existing pipelines without incurring substantial computational overhead.

2.2. Strategies to Enhance Robustness in Computer Vision

2.2.1. Multi-Hypothesis Methods

Wehrbein [29] introduces a method inspired by Normalizing Flows [30] that leverages the concept of bijectivity to map precise 3D human poses to the 2D domain. By combining heatmaps from 2D pose detectors with latent vectors preserved during the 2D–3D process, the proposed approach proves remarkably effective in handling occlusion scenarios. MHFormer [18] merges multihypothesis and transformer concepts, using stacked spatial transformers to generate varied pose hypotheses, which are then enriched through self-encoding and cross-encoding, achieving notable results with just three pose hypotheses. This method confirms that the interaction and fusion of multiple representations can significantly enhance the robustness of the model.

2.2.2. Noise-Based Processing Methods

MULLER [31] applies Gaussian filtering to suppress noise while preserving key features. PoseFormerV2 [19] models skeleton sequences with low-pass filtering, removing high-frequency noise from 2D pose sequences and extracting frequency-domain cues, which markedly improves robustness. Inspired by DDPM [32], Shan et al. [33] propose a diffusion-based pipeline: Gaussian noise is progressively injected into ground-truth 3D poses, and the model iteratively denoises together with 2D inputs to restore accurate 3D poses, which highlights the value of Gaussian noise processing in robust estimation. Simo-Serra et al. [34] propagate noise from the image plane to the shape space, further corroborating the benefit of noise modeling for robustness.
Overall, noise-aware processing complements architectural advances and offers a practical route to robust 3D lifting. A promising direction is to combine compact spatial representations with simple multi-hypothesis consolidation and lightweight spatial–frequency cues, thereby reducing the burden on temporal encoders and improving resilience to inaccuracies in 2D detections under constrained computing.

3. Method

3.1. Overview

As shown in Figure 1, CGFusionFormer takes a sequence of 2D poses as input and outputs a high-quality 3D pose estimation with low computational cost in a way of combining features of local and global receptive fields.
Formally, let P 2 d = { p i | i = K , , 1 , 0 , 1 , , K } represent the set of 2D poses in a sequence, where p i is the pose in frame i, 0 represents the current frame, and the objective of CGFusionFormer is to predict the 3D pose of the current frame 0:
p 3 d = D e c ( H a f ( E n c c s r ( P 2 d , f ) , E n c f r e q ( P 2 d , F ) ) ) ,
where f is the number of frames (3, as in Figure 1) around the current frame 0 and serves as the local receptive field of E n c c s r ( · ) , and F is the entire sequence length (9, as in Figure 1, F = 2 K + 1 ) and serves as the global receptive field of E n c f r e q ( · ) . E n c c s r ( · ) and E n c f r e q ( · ) represent modules to generate local spatial multi-hypothesis features and global frequency features, respectively. H a f ( · ) is the hybrid feature fusion and D e c ( · ) represents the regression operation to output the estimated 3D pose p 3 d .
The CGFusionFormer focuses on spatial representation and hybrid feature fusion. The proposed Compact Spatial Representation (CSR) module first utilizes the Body-Part Spatial Transformer Encoder (BPS) module to obtain local spatial features from part of the 2D pose sequence (f frames) and then feeds them into the Filter-Based Multi-Hypothesis Compression (FMC) module. In the FMC, we incorporate multiple Gaussian filters to improve robustness and perform feature compression in the spatial domain to reduce the computational cost associated with subsequent temporal encoding. Furthermore, we integrate advanced designs such as DCT coefficients and low-pass filters to transform poses in the total sequence from the temporal domain to the frequency domain globally. The Hybrid Adaptive Fusion (HAF) module, a novel residual network, fully utilizes the features of CSR, frequency, and CSR–frequency to enable CGFusionFormer to produce accurate 3D HPE.
Although CGFusionFormer is designed to predict a 3D pose from a sequence of 2D poses, it can serve as a basic architecture for 2D-to-3D lifting methods that choose 2D HPE as an intermediate to estimate 3D pose sequences based on 2D pose sequences from monocular or multi-view camera images or videos.

3.2. Compact Spatial Representation (CSR)

We describe the modules used in CSR and explain how CSR reduces computational cost while ensuring feature quality and improving robustness.

3.2.1. Body-Part Spatial Transformer Encoder (BPS)

Human motion can be achieved through local motions of different body parts. Therefore, accurate and independent modeling of local body parts effectively establishes the internal connections of the overall pose.
The input to BPS is indicated by P 0 R f × J × C (the subset of P 2 d ), where f represents the pose frames in the receptive field, J represents the number of joints in each frame, and C represents the high-dimensional representation of each joint.
Here, we add constraints based on body part segmentation, typically by grouping the 17 joints into 5 sets according to the dynamic chain structure of a human body:
B t r u n k = { h e a d , n e c k , t h o r a x , s p i n e , h i p } , B r _ l e g = { r _ f e e t , r _ k n e e , r _ h i p } , B l _ l e g = { l _ f e e t , l _ k n e e , l _ h i p } , B r _ a r m = { r _ s h o u l d e r , r _ w r i s t , r _ e l b o w } , B l _ a r m = { l _ s h o u l d e r , l _ w r i s t , l _ e l b o w } .
We encode these five sets of joints ( B t r u n k , B r _ l e g , B l _ l e g , B r _ a r m , and B l _ a r m ) using a learnable dictionary B R 5 × C and transform them into body embedding vectors, which is a process termed Body-Part Partitioning (BPP). We apply BPP in two components based on the vanilla spatial encoder. First, before the integration of a learnable spatial positional embedding E S p o s into the spatial encoder (Figure 2), the introduction of BPP is essential to impose constraints on human motion representations based on body parts external to the encoder. Secondly, BPP is integrated within a multi-head self-attention mechanism, named BpMHSA (Figure 2), to enhance connections among fine-grained features. The output of the spatial encoder with L 1 layers is Z S p a t i a l R f × J × C .

3.2.2. Filter-Based Multi-Hypothesis Compression (FMC)

In scenarios of low-quality 2D pose detection, Gaussian filters effectively attenuate high-frequency noise but risk over-smoothing the signal and suppressing fine-grained motion cues (e.g., rapid local motions). As discussed above, our spatial representation Z S p a t i a l strengthens local correlations and preserves the granularity of local motion, partly mitigating this drawback. To further reduce the model’s sensitivity to 2D detection quality and improve robustness, we insert a set of 2D Gaussian filters that operate on Z S p a t i a l within the 3D HPE pipeline. The module is implemented as a multi-branch block (Figure 3): each branch B l ( l 3 ) applies a distinct Gaussian filter to Z S p a t i a l and outputs an initial hypothesis feature.
For the upper two branches, we build a hierarchy of 2D Gaussian filters by iteratively applying the kernel g times to obtain W ( g ) ( g < 3 ), where a larger g yields stronger smoothing. The bottom branch employs an identity filter W G . Following [31], we construct inter-branch difference filters by subtracting the outputs from parallel branches, to stabilize the denoising process and preserve the fidelity of the post-smoothing feature. As the dimensionality of the feature increases, the computational cost in subsequent encoders inevitably increases. Therefore, we apply a parameter-free reduction along the channel dimension C to compress the hypothesis of each branch to C C . In this way, FMC produces multiple hypotheses with varying levels of smoothness, which are finally fused and unfolded to form Z CSR R F × ( J · C ) .

3.3. Hybrid Adaptive Fusion (HAF)

As shown in Figure 4, to address the loss of detailed features of human motion due to feature compression, we design a residual structure to fuse features before and after processing through the CSR-Frequency Feature Fusion module.
In the beginning, we describe how these features are obtained. To maintain consistency with the dimensions of Z C S R , we employ DCT coefficients and low-pass filters from [19] to transform the 2D pose sequence into the frequency domain, resulting in Z F r e q R F × ( J · C ) , which represents features in the global frequency domain because the total pose frames F of the sequence are used.
Subsequently, the CSR-Frequency Feature Fusion module fuses Z C S R and Z F r e q to produce Z C S R _ F r e q R ( f + F ) × ( J · C ) , where Z C S R and Z F r e q are concatenated with different learnable positional embeddings E C p o s , E F p o s , respectively.
In addition, to further enhance both local and global correlations, we merge features before and after the CSR-frequency Feature Fusion module, with an increase in computational cost. Firstly, features of the CSR domain Z C S R and those of the frequency domain Z F r e q are concatenated along the temporal dimension to match the size of Z C S R _ F r e q , resulting in Z C S R _ F r e q R ( f + F ) × ( J · C ) . Secondly, Z C S R _ F r e q and Z C S R _ F r e q are spatially merged to form the integrated feature Z H A C o n c a t R ( f + F ) × ( 2 · J · C ) . To adaptively adjust the impact of local and global correlations during the fusion process, we introduce a balancing parameter λ into Z C S R _ F r e q . Ultimately, the final representation Z H A F u s i o n R ( f + F ) × ( J · C ) is derived using a fully connected layer with learnable weights W F R ( 2 · J · C ) × ( J · C ) . The above process can be summarized as follows:
Z C S R _ F r e q = C o n c a t ( Z C S R , Z F r e q ) , Z H A C o n c a t = C o n c a t ( λ Z C S R _ F r e q , Z C S R _ F r e q ) , Z H A F u s i o n = Z H A F W F .

3.4. Regression Head

We employ a linear regression head, composed of a 1 D convolution layer and linear mapping, to aggregate temporal information and derive the final 3 D pose y R 1 × ( J · 3 ) for the middle frames of the sequence, where J denotes the number of joints. Specifically, the 1 D convolution layer effectively captures the temporal features within the sequence, while the linear mapping transforms these features into the final 3 D joint positions.

4. Experiment

4.1. Datasets and Evaluation Metrics

  • Human3.6M [20] is a widely used benchmark for 3D human pose estimation. It was recorded in a controlled indoor studio with synchronized and calibrated multi-view RGB cameras (typically four views). The dataset contains over 3.6 million video frames of 11 subjects performing 15 everyday actions (e.g., walking, phoning) and provides accurate 3D joint annotations captured by a motion-capture system. Following the standard subject split, we use S1, S5, S6, S7, and S8 for training and S9 and S11 for testing. We adopt the Human3.6M-17 joint convention with pelvis as the root and report errors in millimeters in the camera coordinate system. The results are reported with two common metrics: MPJPE (Mean Per-Joint Position Error; Protocol #1) and P-MPJPE (Protocol #2), which computes MPJPE after a rigid Procrustes alignment [35].
  • MPI-INF-3DHP [21] is also captured by camera. Compared to Human3.6M, it covers more diverse motions and scenes, including indoor green screen setups with varied lighting and a portion of outdoor environments, resulting in larger viewpoint changes, more occlusions, and more diverse backgrounds. The 3D labels are obtained via multi-view reconstruction and motion-capture pipelines with manual verification, making the dataset suitable for evaluating generalization. Following [4], we report MPJPE, PCK (Percentage of Correct Keypoints), and AUC (Area Under the Curve).

4.2. Implementation Details and Analysis

4.2.1. Hyperparameter Settings

In our design, there are three critical hyperparameters: ksize, which denotes the kernel size of Gaussian filters W 1 and W 2 ; std, the standard deviation of Gaussian filters; C , the feature dimension C after compression in the FMC. We implement a Gaussian filter with k s i z e = 5 and s t d = 1 and set both L 1 and L 2 to 4 layers (Figure 1). We set C and C to 32 and 24, respectively, to adequately preserve spatial features. We also employ three critical hyperparameters: the number of input frames (f) for the spatial encoder, the input sequence length (F) as the global (or expanded) receptive field, and the number of retained DCT coefficients (n) for preserved long-range temporal information. We set n = f when f is 3 or 9. When f = 1 , we set n to 3.

4.2.2. Experimental Settings

Our experiments are carried out on a Quadro RTX 6000 GPU, with the model implemented in PyTorch [36] version 1.12.0, under the Ubuntu 22.04 operating system. During training, we set the initial learning rate at 7 × 10−5, incorporating an exponential learning rate decay schedule with a decay factor of 0.99. We employ AdamW [37] as the optimizer, training for 120 epochs, each with a weight decay of 0.1. To ensure fairness, following [17], we implement horizontal flip augmentation. We use CPN [38] 2D pose detection for Human3.6M following [17] and ground truth 2D detection for MPI-INF-3DHP following [39].

4.3. Comparison with State-of-the-Art Methods

We compare our method with state-of-the-art transformer-based approaches on two datasets, Human3.6M and MPI-INF-3DHP, demonstrating that our method achieves superior accuracy while maintaining lower computational costs.

4.3.1. Results on Human3.6M

We compare our approach with representative transformer-based baselines on Human3.6M and report the results in Table 1. We vary the total sequence length F and the number of input frames fed into the model (denoted by f). We define the expansion ratio r = F / f (i.e., the ratio of the total sequence length to the number of input frames).
Our method is highly effective when the value of f is small and the expansion ratio r is large. For example, at r = 81 (F = 81, f = 1>), CGFusionFormer achieves an MPJPE of 47.3 mm using only 46.7 MFLOPs; by comparison, MHFormer [18] requires 342.9 MFLOPs to reach 47.8 mm MPJPE, and PoseFormerV2 [19] needs 77.2 MFLOPs for 47.6 mm MPJPE.
In addition, our proposed method achieves a superior speed–accuracy trade-off. As shown in Table 1, under similar computational budgets (approximately 75 MFLOPs) and the same sequence length (81), our method achieves an MPJPE of 47.1 mm, which is better than that of 47.6 mm obtained by PoseFormerV2 [19].

4.3.2. Results on MPI-INF-3DHP

To evaluate the generality of the model, we conduct comparative experiments on the much more complex MPI-INF-3DHP dataset. The results are listed in Table 2. Our model uses an 81-frame sequence as input with n = f = 9 settings, and our implementation follows [12]. The parameter settings and experimental results of the compared methods are from their original papers. Our proposed method achieves the best performance on several metrics. The results further demonstrate that our method can fully explore the features in short receptive fields and efficiently combine local and global receptive field features, making it adaptable for much more challenging 3D HPE from the 2D poses sequence.

4.3.3. Robustness and Qualitative Comparisons

To evaluate the robustness of our method, we follow the experimental setup of [12] by injecting zero-mean Gaussian noise into the ground-truth 2D detections on the Human3.6M dataset. As illustrated in Figure 5, we report the performance degradation (in MPJPE) under increasing noise levels. For fair comparison, all models are evaluated using 27-frame input sequences with f = n = 3 . The results show that our approach not only exhibits stronger robustness under mild noise but also achieves a greater accuracy improvement than other methods even with slight noise (e.g., ( s i g m a = 3 )).
Complementing these quantitative findings, Figure 6 presents qualitative comparisons to assess the robustness of 3D pose predictions under noisy 2D detections. We inject Gaussian noise into the 2D joints (with the same protocol as in Figure 5) to mimic detector drift. For fairness, all methods take 81-frame input sequences. We observe that CGFusionFormer produces more accurate and physically plausible poses, as emphasized by the annotated arrows, especially for peripheral joints (hands and feet) and the head. Compared with PoseFormer [17] and PoseFormerV2 [19], these improvements are consistent with the gains attributed to our CSR and upport the robustness indicated in Figure 5.

4.3.4. Ablation Study

Table 3 shows the results of our method under different components and structures on the Human3.6M dataset. It is evident that the BPS module effectively improves accuracy with almost no increase in MFLOPs. The addition of the FMC module significantly reduces MFLOPs with only a 0.9 mm MPJPE loss compared to the baseline (1.7 × slower). Subsequently, the HAF module achieves 47.6 mm MPJPE at the cost of only four MFLOPs. The results indicate the following: (i) The substantial reduction in MFLOPs after incorporating the FMC reflects the effectiveness of decreasing computational costs. (ii) the BPS and HAF, requiring minimal computational costs, play a crucial role in enhancing performance. Notably, the baseline model is PoseFormerV2 [19], with all added modules built upon it. The inclusion of the FMC module resulted in a slight performance drop because it compressed and lost some information, as reflected in the comparison of MPJPE scores.

4.3.5. Computational Complexity Analysis

To enable real-time applications of lifting-based methods, we evaluate computational complexity (Table 4). In real-time detection, because future frames are unavailable, we fill future inputs with the current frame, which leads to some performance degradation. We compare against models with nearly identical MPJPE; despite using only 46.7 MFLOPs, CGFusionFormer exhibits the smallest accuracy drop (3.7 mm MPJPE). Our method demonstrates superior stability and reliability under frame-loss scenarios compared with other robustness-oriented approaches. We also run real-time detection on an NVIDIA GeForce RTX 3050 and an NVIDIA Jetson AGX Xavier using HRNet [41] for 2D pose detection and feed the resulting sequences into subsequent 3D pose estimators. We then compare models in terms of frames per second (FPS) and MFLOPs. Our method delivers strong performance with a minimal computational budget.

5. Conclusions

This paper addresses core challenges such as the relatively low quality of 2D joints and significant computational costs in transformer-based 2D-to-3D lifting 3D human pose estimation methods. We propose a CGFusionFormer to estimate the 3D human pose robustly with low computational complexity from a sequence of 2D poses. To enhance the robustness of the model against unreliable 2D pose detection and reduce the computational costs associated with subsequent temporal encoding, the compact spatial representation module is proposed, which concentrates on exploring local spatial correlation in 2D pose sequences with a Body-Part Spatial Transformer Encoder and multichannel Gaussian filters-based multihypothesis compression. To further improve the robustness of the model, the Hybrid Adaptive Fusion module is designed to fully fuse features from CSR and those from the global correlation explored in 2D poses sequence. Quantitatively, CGFusionFormer attains 47.6 mm MPJPE on Human3.6M with only 71.3 MFLOPs when using 3 input frames over 27-frame sequences and with longer context reaches 47.1 mm MPJPE at similar computing. On MPI-INF-3DHP, it achieves 97.9 PCK, 78.5 AUC, and 27.2 mm MPJPE, matching the best PCK and delivering the lowest MPJPE among compared methods under the same setting. Qualitatively, under synthetic Gaussian noise, our predictions remain more physically plausible—especially for peripheral joints—than compared baselines. These results substantiate the claimed accuracy–efficiency trade-off and the robustness benefits brought by CSR and HAF. CGFusionFormer is currently designed to predict a 3D pose from a sequence of 2D poses; it can be easily extended to estimate a 3D pose sequence based on 2D pose sequences from monocular or multiview camera images or videos. Enabling real-time usage of CGFusionFormer is also part of our future work.

Author Contributions

Conceptualization, T.L.; methodology, H.W.; software, H.W.; validation, H.W.; formal analysis, T.L.; resources, T.L.; writing—original draft preparation, T.L. and H.W.; writing—review and editing, D.X.; supervision, D.X.; funding acquisition, D.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China grant number 62172150.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rommel, C.; Letzelter, V.; Samet, N.; Marlet, R.; Cord, M.; Pérez, P.; Valle, E. ManiPose: Manifold-Constrained Multi-Hypothesis 3D Human Pose Estimation. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 14 December 2024; Globerson, A., Mackey, L., Belgrave, D., Fan, A., Paquet, U., Tomczak, J., Zhang, C., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2024; Volume 37, pp. 107350–107378. [Google Scholar]
  2. Xu, J.; Guo, Y.; Peng, Y. FinePOSE: Fine-Grained Prompt-Driven 3D Human Pose Estimation via Diffusion Models. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 561–570. [Google Scholar] [CrossRef]
  3. He, R.; Xiang, S.; Tao, P.; Yu, Y. Monocular 3D Human Pose Estimation Based on Global Temporal-Attentive and Joints-Attention In Video. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–9 June 2023; pp. 1–5. [Google Scholar]
  4. Hu, W.; Zhang, C.; Zhan, F.; Zhang, L.; Wong, T.T. Conditional Directed Graph Convolution for 3D Human Pose Estimation. In Proceedings of the 29th ACM International Conference on Multimedia, New York, NY, USA, 20–24 October 2021; MM ’21. pp. 602–611. [Google Scholar] [CrossRef]
  5. Liu, K.; Zou, Z.; Tang, W. Learning Global Pose Features in Graph Convolutional Networks for 3D Human Pose Estimation. In Proceedings of the Computer Vision—ACCV 2020, Kyoto, Japan, 30 November–4 December 2020; Ishikawa, H., Liu, C.L., Pajdla, T., Shi, J., Eds.; ACCV: Cham, Switzerland, 2021; pp. 89–105. [Google Scholar] [CrossRef]
  6. Shin, S.; Kim, J.; Halilaj, E.; Black, M.J. WHAM: Reconstructing World-Grounded Humans with Accurate 3D Motion. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 2070–2080. [Google Scholar] [CrossRef]
  7. Pavlakos, G.; Zhou, X.; Daniilidis, K. Ordinal Depth Supervision for 3D Human Pose Estimation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7307–7316. [Google Scholar] [CrossRef]
  8. Pavlakos, G.; Zhou, X.; Derpanis, K.G.; Daniilidis, K. Coarse-to-Fine Volumetric Prediction for Single-Image 3D Human Pose. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1263–1272. [Google Scholar] [CrossRef]
  9. Li, S.; Chan, A.B. 3D Human Pose Estimation from Monocular Images with Deep Convolutional Neural Network. In Proceedings of the Computer Vision—ACCV 2014, Singapore, 1–5 November 2014; Cremers, D., Reid, I., Saito, H., Yang, M.H., Eds.; ACCV: Cham, Switzerland, 2015; pp. 332–347. [Google Scholar] [CrossRef]
  10. Lie, W.N.; Vann, V. Estimating a 3D Human Skeleton from a Single RGB Image by Fusing Predicted Depths from Multiple Virtual Viewpoints. Sensors 2024, 24, 8017. [Google Scholar] [CrossRef] [PubMed]
  11. Zheng, H.; Li, H.; Dai, W.; Zheng, Z.; Li, C.; Zou, J.; Xiong, H. HiPART: Hierarchical Pose AutoRegressive Transformer for Occluded 3D Human Pose Estimation. In Proceedings of the 2025 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 10–17 June 2025; pp. 16807–16817. [Google Scholar] [CrossRef]
  12. Shan, W.; Liu, Z.; Zhang, X.; Wang, S.; Ma, S.; Gao, W. P-STMO: Pre-trained Spatial Temporal Many-to-One Model for 3D Human Pose Estimation. In Proceedings of the Computer Vision—ECCV 2022, Tel Aviv, Israel, 23–27 October 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; ACM: Cham, Switzerland, 2022; pp. 461–478. [Google Scholar] [CrossRef]
  13. Chen, H.; He, J.Y.; Xiang, W.; Cheng, Z.Q.; Liu, W.; Liu, H.; Luo, B.; Geng, Y.; Xie, X. HDFormer: High-order Directed Transformer for 3D Human Pose Estimation. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, Macao, China, 19–25 August 2023; pp. 581–589. [Google Scholar] [CrossRef]
  14. Tang, Z.; Qiu, Z.; Hao, Y.; Hong, R.; Yao, T. 3D Human Pose Estimation with Spatio-Temporal Criss-Cross Attention. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 4790–4799. [Google Scholar] [CrossRef]
  15. Wei, F.; Xu, G.; Wu, Q.; Qin, P.; Pan, L.; Zhao, Y. Whole-Body 3D Pose Estimation Based on Body Mass Distribution and Center of Gravity Constraints. Sensors 2025, 25, 3944. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, H.; Quan, W.; Zhao, R.; Zhang, M.; Jiang, N. Learning Temporal–Spatial Contextual Adaptation for Three-Dimensional Human Pose Estimation. Sensors 2024, 24, 4422. [Google Scholar] [CrossRef] [PubMed]
  17. Zheng, C.; Zhu, S.; Mendieta, M.; Yang, T.; Chen, C.; Ding, Z. 3D Human Pose Estimation with Spatial and Temporal Transformers. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 11636–11645. [Google Scholar] [CrossRef]
  18. Li, W.; Liu, H.; Tang, H.; Wang, P.; Van Gool, L. MHFormer: Multi-Hypothesis Transformer for 3D Human Pose Estimation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13137–13146. [Google Scholar] [CrossRef]
  19. Zhao, Q.; Zheng, C.; Liu, M.; Wang, P.; Chen, C. PoseFormerV2: Exploring Frequency Domain for Efficient and Robust 3D Human Pose Estimation. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 8877–8886. [Google Scholar] [CrossRef]
  20. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1325–1339. [Google Scholar] [CrossRef]
  21. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; Theobalt, C. Monocular 3D Human Pose Estimation in the Wild Using Improved CNN Supervision. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 506–516. [Google Scholar] [CrossRef]
  22. Hassanin, M.; Khamis, A.; Bennamoun, M.; Boussaïd, F.; Radwan, I. CrossFormer: Cross Spatio-Temporal Transformer for 3D Human Pose Estimation. arXiv 2022, arXiv:2203.13387. [Google Scholar] [CrossRef]
  23. Li, W.; Liu, H.; Ding, R.; Liu, M.; Wang, P.; Yang, W. Exploiting Temporal Contexts With Strided Transformer for 3D Human Pose Estimation. IEEE Trans. Multim. 2023, 25, 1282–1293. [Google Scholar] [CrossRef]
  24. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked Autoencoders Are Scalable Vision Learners. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 15979–15988. [Google Scholar] [CrossRef]
  25. Pavllo, D.; Feichtenhofer, C.; Grangier, D.; Auli, M. 3D Human Pose Estimation in Video With Temporal Convolutions and Semi-Supervised Training. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7745–7754. [Google Scholar] [CrossRef]
  26. Zhang, J.; Tu, Z.; Yang, J.; Chen, Y.; Yuan, J. MixSTE: Seq2seq Mixed Spatio-Temporal Encoder for 3D Human Pose Estimation in Video. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13222–13232. [Google Scholar] [CrossRef]
  27. Qian, X.; Tang, Y.; Zhang, N.; Han, M.; Xiao, J.; Huang, M.; Lin, R. HSTFormer: Hierarchical Spatial-Temporal Transformers for 3D Human Pose Estimation. arXiv 2023, arXiv:2301.07322. [Google Scholar] [CrossRef]
  28. Zhu, W.; Ma, X.; Liu, Z.; Liu, L.; Wu, W.; Wang, Y. MotionBERT: Unified Pretraining for Human Motion Analysis. arXiv 2022, arXiv:2210.06551. [Google Scholar] [CrossRef]
  29. Wehrbein, T.; Rudolph, M.; Rosenhahn, B.; Wandt, B. Probabilistic Monocular 3D Human Pose Estimation with Normalizing Flows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 11179–11188. [Google Scholar] [CrossRef]
  30. Rezende, D.J.; Mohamed, S. Variational inference with normalizing flows. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 7–9 July 2015; Volume 37, pp. 1530–1538. [Google Scholar]
  31. Tu, Z.; Milanfar, P.; Talebi, H. MULLER: Multilayer Laplacian Resizer for Vision. arXiv 2023, arXiv:2304.02859. [Google Scholar] [CrossRef]
  32. Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 6840–6851. [Google Scholar]
  33. Shan, W.; Liu, Z.; Zhang, X.; Wang, Z.; Han, K.; Wang, S.; Ma, S.; Gao, W. Diffusion-Based 3D Human Pose Estimation with Multi-Hypothesis Aggregation. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 4–6 October 2023; pp. 14715–14725. [Google Scholar] [CrossRef]
  34. Simo-Serra, E.; Ramisa, A.; Alenyà, G.; Torras, C.; Moreno-Noguer, F. Single image 3D human pose estimation from noisy observations. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2673–2680. [Google Scholar] [CrossRef]
  35. Zheng, C.; Wu, W.; Chen, C.; Yang, T.; Zhu, S.; Shen, J.; Kehtarnavaz, N.; Shah, M. Deep Learning-based Human Pose Estimation: A Survey. ACM Comput. Surv. 2024, 56, 1–37. [Google Scholar] [CrossRef]
  36. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; pp. 8024–8035. [Google Scholar]
  37. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  38. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Yu, G.; Sun, J. Cascaded pyramid network for multi-person pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7103–7112. [Google Scholar] [CrossRef]
  39. Lin, J.; Lee, G.H. Trajectory Space Factorization for Deep Video-Based 3D Human Pose Estimation. In Proceedings of the 30th British Machine Vision Conference, Cardiff, UK, 9–12 September 2019; p. 101. [Google Scholar]
  40. Chen, T.; Fang, C.; Shen, X.; Zhu, Y.; Chen, Z.; Luo, J. Anatomy-Aware 3D Human Pose Estimation With Bone-Based Pose Decomposition. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 198–209. [Google Scholar] [CrossRef]
  41. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5686–5696. [Google Scholar]
Figure 1. Overview of CGFusionFormer. To exemplify, given a sequence of 2D poses as input, CGFusionFormer uses 3 frames (index −1, 0, 1) to extract local CSR domain features and an entire 9 frames to capture global frequency domain features and outputs the estimated 3D human pose of frame 0 after effective and efficient hybrid feature fusion.
Figure 1. Overview of CGFusionFormer. To exemplify, given a sequence of 2D poses as input, CGFusionFormer uses 3 frames (index −1, 0, 1) to extract local CSR domain features and an entire 9 frames to capture global frequency domain features and outputs the estimated 3D human pose of frame 0 after effective and efficient hybrid feature fusion.
Sensors 25 06052 g001
Figure 2. Body-Part Spatial Transformer and Body-Part Multi-head Attention.
Figure 2. Body-Part Spatial Transformer and Body-Part Multi-head Attention.
Sensors 25 06052 g002
Figure 3. Filter-Based Multi-Hypothesis Compression.
Figure 3. Filter-Based Multi-Hypothesis Compression.
Sensors 25 06052 g003
Figure 4. Hybrid Adaptive Fusion.
Figure 4. Hybrid Adaptive Fusion.
Sensors 25 06052 g004
Figure 5. Comparison of robustness between CGFusionFormer and other methods on Human3.6M. We add zero-mean Gaussian noise with varying standard deviations to ground truth 2D detection, comparing the performance drop ( MPJPE , in mm) as s i g m a increases.
Figure 5. Comparison of robustness between CGFusionFormer and other methods on Human3.6M. We add zero-mean Gaussian noise with varying standard deviations to ground truth 2D detection, comparing the performance drop ( MPJPE , in mm) as s i g m a increases.
Sensors 25 06052 g005
Figure 6. Qualitative comparison of the proposed method with PoseFormer and PoseFormerV2. We introduce Gaussian noise to the 2D detection and compare the model’s performance under conditions of 2D joint drift. Orange arrows emphasize the more accurate and realistic joint estimations in 3D poses by CGFusionFormer compared to other methods.
Figure 6. Qualitative comparison of the proposed method with PoseFormer and PoseFormerV2. We introduce Gaussian noise to the 2D detection and compare the model’s performance under conditions of 2D joint drift. Orange arrows emphasize the more accurate and realistic joint estimations in 3D poses by CGFusionFormer compared to other methods.
Sensors 25 06052 g006
Table 1. Quantitative comparison on Human3.6M (in mm). f and Seq.Len. denote the local and global (or expanded) receptive field, respectively, as stated before. (*) indicates the use of an additional pre-training stage. Bolded scores represent the best outcomes.
Table 1. Quantitative comparison on Human3.6M (in mm). f and Seq.Len. denote the local and global (or expanded) receptive field, respectively, as stated before. (*) indicates the use of an additional pre-training stage. Bolded scores represent the best outcomes.
MethodfSeq. Len.MFLOPsMPJPE↓/P-MPJPE↓
PoseFormer [17]2727542.147.0/-
StridedTrans [23]8181342.547.5/-
MHFormer [18]99342.947.8/-
MHFormer [18]27271031.845.9/-
P-STMO [12] (*)818149345.6/-
STCFormer [14]2727217344.1/-
PoseFormerV2 [19]12777.248.7/37.8
CGFusionFormer (ours)12746.748.5/37.8
PoseFormerV2 [19]18177.247.6/37.3
CGFusionFormer (ours)18146.747.3/37.4
PoseFormerV2 [19]327117.347.9/37.4
CGFusionFormer (ours)32771.347.6/37.3
PoseFormerV2 [19]381117.347.1/37.3
CGFusionFormer (ours)38171.347.1/37.4
PoseFormerV2 [19]981351.746.0/36.1
CGFusionFormer (ours)98121546.5/36.5
Table 2. Quantitative comparison with the state-of-the-art methods on MPI-INF-3DHP. (*) indicates using an additional pretraining stage. Higher PCK , higher AUC , and lower MPJPE indicate better performance. Bolded scores represent the best outcomes.
Table 2. Quantitative comparison with the state-of-the-art methods on MPI-INF-3DHP. (*) indicates using an additional pretraining stage. Higher PCK , higher AUC , and lower MPJPE indicate better performance. Bolded scores represent the best outcomes.
MethodSeq. Len.PCK↑AUC↑MPJPE↓
Pavllo et al. [25]8186.051.984.0
Pavllo et al. [25]24385.551.584.8
Lin et al. [39]2583.651.479.8
Chen et al. [40]8187.954.078.8
PoseFormer [17]995.463.257.7
MHFormer [18]993.863.358.0
MixSTE [26]2794.466.554.9
P-STMO [12] (*)8197.975.832.2
PoseFormerV2 [19]8197.978.827.8
CGFusionFormer (ours)8197.978.527.2
Table 3. Ablation study on the baseline involving module replacement or addition. BPS represents the Body-Part Spatial Transformer Encoder, FMC denotes the Filter-Based Multi-Hypothesis Compression, and HAF stands for the Hybrid Adaptive Fusion.
Table 3. Ablation study on the baseline involving module replacement or addition. BPS represents the Body-Part Spatial Transformer Encoder, FMC denotes the Filter-Based Multi-Hypothesis Compression, and HAF stands for the Hybrid Adaptive Fusion.
BPSFMCHAFMFLOPsMPJPE↓
Baseline#1 117.347.9
+BPS#2 117.347.5
+FMC#3 67.748.4
+HAF#471.747.6
The symbol ‘#’ denotes the experiment configuration numbers. ‘✓’ indicates that the corresponding module is included in the model variant.
Table 4. Comparison of computational complexity ( MFLOPs ), accuracy ( MPJPE ), and inference speed. f denotes the number of input frames per inference; Seq .   Len . is the global receptive field. FPS   ( l ) is measured on a GeForce RTX 3050 device and FPS   ( e ) on a Jetson AGX Xavier device, and Perform .   Drop   ( mm ) refers to the performance decrease ( MPJPE ) caused by using the current frame in place of future frames in the input sequence. Bolded scores represent the best outcomes.
Table 4. Comparison of computational complexity ( MFLOPs ), accuracy ( MPJPE ), and inference speed. f denotes the number of input frames per inference; Seq .   Len . is the global receptive field. FPS   ( l ) is measured on a GeForce RTX 3050 device and FPS   ( e ) on a Jetson AGX Xavier device, and Perform .   Drop   ( mm ) refers to the performance decrease ( MPJPE ) caused by using the current frame in place of future frames in the input sequence. Bolded scores represent the best outcomes.
MethodfSeq.
Len.
MFLOPsFPS (l)FPS (e)Perform.
Drop (mm)
MHFormer [18] CVPR’2281813132.282.5714.595.25
PoseFormerV2 [19] CVPR’23381117.389.9317.744.3
CGFusionFormer18146.797.9518.213.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, T.; Wang, H.; Xiao, D. CGFusionFormer: Exploring Compact Spatial Representation for Robust 3D Human Pose Estimation with Low Computation Complexity. Sensors 2025, 25, 6052. https://doi.org/10.3390/s25196052

AMA Style

Lu T, Wang H, Xiao D. CGFusionFormer: Exploring Compact Spatial Representation for Robust 3D Human Pose Estimation with Low Computation Complexity. Sensors. 2025; 25(19):6052. https://doi.org/10.3390/s25196052

Chicago/Turabian Style

Lu, Tao, Hongtao Wang, and Degui Xiao. 2025. "CGFusionFormer: Exploring Compact Spatial Representation for Robust 3D Human Pose Estimation with Low Computation Complexity" Sensors 25, no. 19: 6052. https://doi.org/10.3390/s25196052

APA Style

Lu, T., Wang, H., & Xiao, D. (2025). CGFusionFormer: Exploring Compact Spatial Representation for Robust 3D Human Pose Estimation with Low Computation Complexity. Sensors, 25(19), 6052. https://doi.org/10.3390/s25196052

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop