Next Article in Journal
Supplement Use Among Athletes: Insights from Gyms in Sarajevo
Previous Article in Journal
Verifiable Threshold Multi-Party Fully Homomorphic Encryption from Share Resharing
Previous Article in Special Issue
Infrared Multi-Scale Small-Target Detection Algorithm Based on Feature Pyramid Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Axial Compression Transformer for Efficient Human Pose Estimation

School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(9), 4746; https://doi.org/10.3390/app15094746
Submission received: 22 March 2025 / Revised: 16 April 2025 / Accepted: 23 April 2025 / Published: 24 April 2025

Abstract

:
Transformer has a wide range of applications in human posture estimation. It can model the global dependence relationship of images through the self-attention mechanism to obtain key human body information. However, Transformer consumes a lot of computation. An axial compression pose transformer (ACPose) method is proposed to reduce part of the computational cost of Transformer by the axial compression of the input matrix, while maintaining the global receptive field by feature fusion. A Local Enhancement Module is constructed to avoid the loss of too much feature information in the compression process. In the COCO dataset experiment, there was a significant reduction in computational cost compared to those of state-of-the-art transformer-based algorithms.

1. Introduction

Two-dimensional (2D) human pose estimation is a crucial task in computer vision, aiming to detect and localize the keypoints of the human body from a given image. With the advancement of deep learning methods, 2D human pose estimation has achieved significant improvements in both accuracy and real-time performance, enabling its wide application in various real-world scenarios, such as motion analysis [1,2,3,4,5] and human–computer interaction [6,7,8,9,10,11].
In the field of motion analysis, human pose estimation technology provides technical support for athletes. By capturing an athlete’s posture, their technique can be analyzed to adjust its standardization and accuracy, facilitating coaches in refining technical details and customizing training plans. In virtual reality and augmented reality fields, human pose estimation technology is a crucial component for human–computer interaction. By recognizing human postures and mapping them to the virtual world, real-time human–computer interaction is enabled, enhancing immersion and interactivity.
Over the last decade, Convolutional Neural Networks (CNNs) have been widely adopted for human pose estimation due to their ability to capture local features through local receptive fields, allowing them to focus on keypoint information [12,13,14]. However, CNNs’ strong reliance on local features limits their capacity to capture global contextual information, which is crucial for accurate pose estimation as it requires a comprehensive understanding of the image content.
The introduction of Transformer architectures has brought about significant breakthroughs in this field. Compared to the traditional methods, Transformers leverage their self-attention mechanism to effectively capture the relationships between different parts of an input sequence, leading to a superior performance in keypoint detection tasks (as shown in [15,16,17,18,19,20]). Furthermore, the multi-layer self-attention mechanism enhances the model’s ability to recognize complex poses. To combine the advantages of CNNs and Transformers, other researchers have proposed various hybrid architectures [21]. For instance, TransPose [22] utilizes CNNs for feature extraction and Transformers to capture global dependencies, while TokenPose [23] introduces keypoint tokens to further improve performance. However, these methods often face a high computational cost, which hinders their deployment in real-time applications.
Addressing this challenge, we propose an axial compression module designed to simplify the network structure within a Transformer, while preserving crucial information, enabling efficient global modeling. Our method integrates the axial compression module into the Transformer architecture. The input data undergo lightweight processing, followed by attention calculations. Specifically, we employ average pooling operations to compress the input along both the horizontal and vertical axes, generating two matrices. These are then processed through a multi-head attention mechanism to capture the relative positional information of the keypoints. Additionally, our unique design for splitting attention heads allows for each head to focus more intently on specific directional position information (horizontal or vertical), thereby enhancing the model’s understanding of the global features associated with keypoints. To recover the potentially lost local visual features during the axial compression process, we introduce a Local Enhancement Module. Specifically, after obtaining the output from the axial compression module, we calculate the residual difference between the original input and the compressed output. This residual is then subjected to average pooling dimensionality reduction, followed by multi-head self-attention operations, effectively extracting local information missed by the axial compression module. By fusing this extracted local information with the output of the axial compression module, we achieve a synergistic combination of global features and local details. The basic principle is as follows:
(1)
Text is human-generated, highly condensed information, while images are natural information containing a lot of redundancy and noise. Compression operations can reduce the useless information in images.
(2)
For human pose estimation tasks, compression operations do not significantly lose global semantic information. CNNs initially extract the local features from images, while Transformer structures tend to establish global dependencies.
(3)
From a biological perspective, human vision can partially fill in missing parts of an image. Similarly, the performance loss caused by missing some visual information in Transformers is relatively low.
Compared to other lightweight methods, such as pruning [24,25], upsampling [26], and depthwise convolution [27], our axial compression module significantly reduces the computational cost, while retaining keypoint global feature information. The experimental results demonstrate that this method effectively decreases the model parameter count and the computational cost, while maintaining high detection accuracy, providing a novel solution for lightweight human pose estimation tasks.
Building upon this idea, we propose a new and efficient attitude estimation method based on axial compression Transformer, named ACPose. Figure 1 illustrates the ACPose architecture, we utilizes TokenPose [23] as the baseline network. It first extracts feature maps using CNNs, flattens them into visual tokens, and adds randomly initialized keypoint tokens to represent human keypoints. These tokens are then fed into Transformer. Subsequently, the input Transformer’s feature matrix undergoes horizontal and vertical average pooling. Multi-head attention is applied to the compressed matrices, and the results are fused. The difference matrix obtained by subtracting the fused matrix from the original Transformer input is average-pooled and subjected to self-attention. This result is then merged with the output of the axial compression stage. Finally, we use the traditional heatmap methods to predict the keypoints.
In summary, this article has the following three contributions:
  • Novel Compress–Merge Strategy: This paper introduces a novel compress–merge strategy, implemented in the proposed axial compression pose Transformer (ACPose) network for efficient human pose estimation. By employing compression operations to reduce the computational cost, while utilizing feature fusion to maintain Transformer’s global receptive field, this strategy mitigates performance degradation caused by irrelevant information in human pose estimation tasks.
  • Local Enhancement Method: This work proposes a local enhancement method that leverages the difference between the compressed and original inputs to amplify the lost local details. This technique effectively enhances the extraction of local features, leading to a significant improvement in prediction accuracy.
  • Competitive Performance with Reduced Complexity: Extensive experiments conducted on prominent public datasets demonstrate that our approach achieves a substantial 81.2% reduction in computational cost of the Transformer part, while maintaining competitive accuracy. These results highlight the effectiveness of our proposed method in achieving lightweight, yet high-performing human pose estimation.

2. Related Work

2.1. Efficient Vision Transformer

In recent years, we have witnessed significant advancements in vision Transformers, with notable contributions in areas such as image classification, object detection, and image segmentation. While some methods have achieved impressive accuracy at a high computational cost, their resource demands often hinder practical applications. Consequently, some researchers have explored techniques to streamline Transformer architectures and design efficient vision Transformer models. For instance, BEiT [28] employs a masked learning strategy by randomly masking portions of an image and training the model to predict the occluded content. This approach enables the model to learn global contextual information, while reducing the computational costs. SeaFormer [29] utilizes a combination of axial compression and convolutional fusion to simultaneously capture both global and local image information. DilateFormer [30] focuses on the self-attention mechanism, selecting a limited number of patches surrounding each query patch for attention computation. By employing different dilation rates across different heads, it effectively reduces the computational cost of the Transformer layers.

2.2. Transformer-Based Pose Estimation Methods

The multi-head self-attention mechanism within Transformers effectively captures the long-range dependencies between human keypoints, which is crucial for accurately predicting their relationships. Recognizing this, some researchers have applied vision Transformers to the field of human pose estimation, resulting in the emergence of numerous Transformer-based human pose estimation networks.
Several notable works have applied Transformers to human pose estimation. TFPose [31] recasts the task as a sequence prediction problem utilizing Transformer encoder-decoder architecture for keypoint regression. VITpose [32], based on the ViT [12] backbone, employs a simple Transformer baseline and enhances accuracy by increasing the model size. HRFormer [33] leverages parallel multi-resolution streams to jointly learn features and integrates them through a multi-scale fusion module, achieving the fusion of different distance attention mechanisms. TokenPose [18] extracts features using CNNs and introduces the concept of keypoint tokens, enabling Transformer to learn the global constraints. These methods demonstrate the effectiveness of Transformers in human pose estimation, with TokenPose [18] achieving competitive accuracy, while maintaining a relatively small model size. Consequently, this paper adopts TokenPose [18] as the baseline network for experimentation.

3. Methods

Our objective is to develop a model capable of efficiently performing human pose estimation. The proposed approach first extracts features from input human images using CNNs. These feature maps are then partitioned into patches and flattened into one-dimensional vectors, which are combined with randomly initialized keypoint tokens and fed into Transformer. Within Transformer, we replace the standard multi-head attention operation with axial compression and Local Enhancement Modules. Finally, the output is mapped to a heatmap through an MLP, and then decoded to obtain the keypoint coordinates, yielding the final prediction results. The following five sections present the details of our approach.

3.1. Tokenization

Feature extraction and tokenization are performed using a CNN. An input image is fed into the CNN to extract features. To enhance model efficiency, we employ the first three stages of HRNet [34], which have a parameter count one-fourth that of the complete HRNet. Following convolutional network feature extraction, the output feature map has dimensions 1/4 the size of the original input image (denoted as x H × W × C ). This output feature map is then divided into a grid along both the height (H) and width (W) dimensions, resulting in H P h × W P w grid cells. Each grid cell has a size of x p P h × P w × C . Each grid cell is flattened into a one-dimensional sequence of size P h × P w × C , which is then mapped to a visual token through a linear layer. Since accurate keypoint localization is crucial for human pose estimation, two-dimensional positional embedding p e i is incorporated into each specific visual token v i , resulting in a visual token of v 1 + p e 1 ,   v 2 + p e 2 , ,   v L + p e L , where L represents the total number of visual tokens. Concurrently, N keypoint tokens are randomly initialized to represent the human keypoints, with N determined by the number of keypoints in the dataset. Finally, these two types of token are connected and input into Transformer for training.

3.2. Axial Compression Module

This paper proposes an axial compression module for lightweight network architectures. Figure 2a depicts the traditional approach, while Figure 2b shows a schematic diagram of the method proposed in this paper. Within Transformer, the input data are first processed by this axial compression module for lightweight computation and attention calculation. Specifically, assuming the input is X n × d , where n denotes the number of sequences and d denotes the sequence length, we employ average pooling to simplify the network structure, while emphasizing the relative positions of the keypoints. X is averaged along both the horizontal and vertical axes, resulting in two compressed matrices, X h n × h and X v h × d , respectively. Here, h represents the number of heads in the multi-head attention mechanism. This effectively compresses the global information into these two matrices. Subsequently, X h and X v are linearly mapped to generate two sets of matrices: the query ( Q n , Q d ), the key ( K n , K d ), and the value ( V n , V d ). These qkv matrix pairs are then fed into a multi-head attention module to compute the attention scores.
y h = p = 1 n s o f t m a x p q ( h ) p × k ( h ) T d k v ( h ) p
y v = p = 1 d s o f t m a x p q ( v ) p × k ( v ) T d k v ( v ) p
In this context, d k represents the key dimension. The attention score determines the attention level that the current query token should give to each key. Furthermore, when splitting the attention heads, q ( h ) , k ( h ) , and v ( h ) are divided into h heads of size n × 1, and q ( v ) , k ( v ) , and v ( v ) are divided into h heads of size 1 × d. This operation allows for multiple attention heads to capture richer horizontal and vertical positional information. The two output matrices, y h and y v , are then fused to form a global positional feature representation. To ensure that the fused representation fully establishes global dependencies, the matrices y h and y v , which participate in fusion, must acquire comprehensive information along each individual axis, both horizontally and vertically. Therefore, after obtaining the feature matrices y h n × h and y v h × d through the self-attention mechanism, they are subjected to average pooling operations along both the horizontal and vertical axes, and further compressed into y h 1 n × 1 and y v 1 1 × d , respectively. This aggregates the multi-dimensional information obtained from the multi-head attention mechanism onto a single axis. Subsequently, to fuse these two axial feature matrices, we use the repeat operation to expand the matrices y h 2 n × d and y v 2 n × d into y h 2 n × d and y v 2 n × d , resulting in two feature matrices with the same shape. Finally, by summing y h 2 and y v 2 , we obtain the feature matrix y o . Each point in this matrix encompasses both the horizontal positional features and the vertical positional features, indicating that the keypoint tokens have acquired comprehensive global positional information and established long-range dependencies with visual features.

3.3. Detail Augmentation Module

To recover local visual features that may be lost during the compression process, we propose a Local Enhancement Module. Specifically, after obtaining the output y o from the axial compression module, we subtract the output y o from the original Transformer input X. The resulting difference is then subjected to average pooling with a kernel size of k   ×   k (where k takes on values of 2, 3, and 4) to reduce dimensionality and maintain computational efficiency. Subsequently, multi-head self-attention is applied to the pooled matrix to effectively extract local information that may have been overlooked by the axial compression module. Finally, the output is fused with y o . Since the dimension of the pooled matrix is reduced, we employ bilinear interpolation to expand it back to its original dimension before matching y o ’s dimensions. The final result, obtained by adding this expanded output to y o , represents the output of the Local Enhancement Module.

3.4. Transformer Architecture

Depending on the version setup, we use different Transformer layers as the encoder, as illustrated in Figure 1. These Transformer layer primarily consists of an axial compression module, a feature fusion module, and a feed-forward network. Furthermore, both the axial compression module and the feed-forward network are preceded by regularization and residual connection mechanisms. The computational formula for this is as follows:
L a y e r N o r m X + M u l t i H e a d A t t e n t i o n X
L a y e r N o r m X + F e e d F o r w a r d X
where M u l t i H e a d A t t e n t i o n X represents the output of the multi-head attention layer, and F e e d F o r w a r d X represents the output of the feed-forward network layer; their output shapes are consistent with the original input X, allowing for direct summation.
Human keypoint detection is a complex task primarily focused on obtaining keypoint coordinates. While the standard Transformer methods utilize self-attention mechanisms to acquire global receptive fields for extracting positional information, this often comes with a significant computational burden. Therefore, addressing these concerns, we compress the matrices during self-attention operations, streamlining the network structure while simultaneously achieving a global receptive field comparable to that of standard Transformers through the fusion of information along two axes. Unlike other prevalent Pose Transformer architectures, our proposed Transformer presents a more efficient and lightweight network structure.

3.5. Heatmap

During the output stage of Transformer, we select only N keypoint tokens as the output and process them through a multi-layer perceptron (MLP). This mapping results in a two-dimensional heatmap P N × H × W , where H and W represent 1/4 the size of the original image dimensions. Subsequently, a reshape operation transforms this heatmap into P N × H * × W * , matching the dimensions of the original image. We can then locate the human keypoints by identifying the position of the maximum response value on this heatmap, as illustrated in Figure 3. The location corresponding to the maximum response represents our predicted human keypoint. During model training, we employ the Mean Squared Error (MSE) loss function to compare the predicted heatmap with the ground truth heatmap. The mathematical formula for the MSE loss function is as follows:
L M S E = 1 2 i = 1 N P i P i ^ 2
where N represents the number of human keypoints, which varies depending on the dataset; P i denotes the ground truth heatmap, and P i ^ represents the predicted heatmap.

4. Results

4.1. Setup

In this paper, we designed three versions of ACPose, as shown in Table 1. The base utilizes the first three stages of HRNet-W32 as the backbone for the CNN stage, large employs the first three stages of HRNet-W48 as the backbone, and D24 denotes Transformer with 24 layers.
During training, the images from the COCO dataset are cropped to 256 × 192 dimensions, while those from the MPII dataset are cropped to 256 × 256. We employ the Adam optimizer with a total of 300 epochs and an initial learning rate of 1 × 10−3. The learning rate is adjusted to 1 × 10−4 and 1 × 10−5 at epochs 200 and 260, respectively. To address the uncertainty caused by insufficient model training and limited training data, we adjusted the batch size during model training, conducting multiple trainings with settings of 32, 64, and 128. The experiments are conducted on a hardware environment consisting of two NVIDIA RTX 4090 24G GPUs with a software environment comprising PyTorch 2.3.0 and CUDA 12.1. The manufacturer is ASUS and the location is Shanghai, China.

4.2. Quantitative Results

The COCO dataset comprises over 200,000 images and 250,000 person instances; each instance is annotated with 17 keypoints. The COCO dataset is divided into train, validation, and test-dev sets, containing 57k, 5k, and 20k images, respectively. We trained our model on the COCO train2017 set and evaluated it on the COCO validation and COCO test-dev sets. A commonly used evaluation metric for the COCO dataset is Average Precision (AP), which is calculated based on Object Keypoint Similarity (OKS). The formula for calculating OKS is as follows:
O K S = i e d i 2 / 2 s 2 k i 2 δ v i > 0 i δ v i > 0
In this case, d i represents the Euclidean distance between the predicted keypoint coordinates and their corresponding ground truth values. v i indicates whether each keypoint is visible. s represents the area of the human part of it, and k i represents the weight assigned to each keypoint.
The formula for calculating AP is as follows:
A P t = p δ O K S > t p 1
When t = 0.5 and 0.75, AP is represented as A P 50 and A P 75 , while A P L and A P M represent evaluation metrics for large-scale and medium-scale objects, respectively. Additionally, we also employ Average Recall (AR) as an evaluation metric.
We use GFLOPs as the evaluation standard for model computational cost, where FLOP stands for floating-point operations, s represents seconds, and FLOPs represent the number of floating-point operations per second. Therefore, GFLOPs represent billions of operations per second. We utilize the thop library to output the GFLOPs in the model network.
As shown in Table 2, we evaluated the performance of ACPose on the mainstream COCO dataset. The results demonstrate that our ACPose model achieves a more significant reduction in computational cost compared to those of the other prevailing models, while maintaining competitive AP scores. Specifically, compared to TokenPose-L/D6, our ACPose-L/D6 reduces the Transformer-related computational cost by 81.2% with only a 0.4% decrease in accuracy; the results show that the original method, TokenPose, generates a large number of matrix operations due to the excessive data volume in the Transformer layers, resulting in high GFLOPs. Our method significantly reduces the data volume through compression operations, thereby substantially lowering the GFLOPs, primarily within the Transformer structure. Furthermore, our ACPose-B achieves a 47.3% reduction in the Transformer-related computational cost, while improving accuracy by 0.1%.
Compared to TransPose-H-A6, our method exhibits substantial reductions in computational cost and demonstrates superior accuracy. Additionally, as evident from the table, our approach out-performs the ViTPose and SHaRPose methods by reducing the computational cost without compromising the performance.
The MPII dataset comprises over 25,000 images and 40,000 person instances, each instance annotated with 16 human keypoints. In the MPII dataset, we primarily utilize the Percentage of Correct Keypoints (PCKs) as the evaluation metric, specifically PCK@0.5 as our final experimental result.
As shown in Table 3, we evaluated our model on the MPII dataset. Compared to the other mainstream methods, our model demonstrates a significant reduction in the computational cost. Compared to the TokenPose [18] method, G F L O P s T is reduced by 39.0%, with only a 0.2% loss in accuracy. Specifically, ACPose-L/D6 achieves 96.1% and 86.1% prediction accuracy for the shoulder joint (Sho) and wrist joint (Wri), respectively, surpassing the TokenPose-L/D6 method. This indicates that our approach does not compromise overall prediction accuracy, while achieving a well-balanced efficient human pose estimation network in terms of both precision and computational cost.

4.3. Qualitative Results

Figure 4 presents the comparative analysis of our method, ACPose, and the TokenPose [18] method on the COCO dataset. As illustrated in the figure, ACPose demonstrates a more accurate inference of the human keypoint locations. The first image reveals that ACPose excels at identifying the keypoints in overlapping images and exhibits heightened sensitivity in recognizing smaller human targets within an image. This suggests that ACPose’s Local Enhancement Module effectively enhances the recognition capability for fine-grained details within images. The second and third images showcase ACPose’s superior accuracy in predicting the human keypoints even under occlusion conditions. In conclusion, ACPose consistently out-performs the other methods in terms of pose estimation accuracy.

4.4. Ablation Study

To investigate the impact of the pooling kernel size k in the Local Enhancement Module on both prediction accuracy and computational cost, we conducted ablation experiments on the MPII dataset. The aim was to identify the optimal balance between resource consumption and model performance. As shown in Table 4, we varied the matrix compression size during the pooling stage within the Local Enhancement Module. The results demonstrate that as the pooling kernel size k decreases, both the computational cost and the parameter count gradually increase, accompanied by a corresponding improvement in the performance metrics. Specifically, when k = 4, G F L O P s T is reduced by 45.3% with a 0.9% decrease in accuracy. When k = 3, G F L O P s T is reduced by 43.8% with a 0.4% decrease in accuracy. When k = 2, G F L O P s T is reduced by 39.0% with a 0.2% decrease in accuracy. Based on this analysis, we conclude that k = 2 achieves the optimal balance between computational cost and accuracy, representing the most favorable combination.
To verify the influence of different compression degrees in axial compression on the results and obtain the optimal compression size, ablation experiments were conducted on the MPII dataset. The experiments mainly focused on adjusting the parameter h in the axial compression module to investigate the impact of the compression degree on the performance metrics. As shown in Table 5, as h decreases, the compression degree gradually increases, the computational cost and the parameter amount gradually decrease, and the accuracy metrics remain basically consistent. Specifically, when h is 36, G F L O P s T is reduced by 10.9%; when h is 24, G F L O P s T is reduced by 29.7%; and when h is 12, G F L O P s T is reduced by 39.0%. Based on the above analysis, we believe that when h is 12, a balance between computational cost and performance is achieved, making it the optimal compression choice.

5. Discussion

This paper demonstrates the higher efficiency of Transformer structures in focusing on global feature extraction. We propose ACPose, an efficient human pose estimation method that achieves significant reductions in model consumption without sacrificing accuracy, thereby addressing the balance between accuracy and speed in human pose estimation. Through the careful design of axial compression and Local Enhancement Modules, ACPose achieves competitive results on both the COCO and MPII datasets. The qualitative experiments show that our method can more accurately predict keypoint locations compared to other lightweight models when applied to human pose estimation. Furthermore, ablation studies were conducted to determine the optimal parameter settings for our model.
Despite achieving outstanding performance on the public datasets, we encountered certain challenges during our research. Specifically, while reducing dimensionality within Transformer, effectively lowering complexity, the current method does not differentiate the contribution of various tokens to human keypoint prediction during compression. This results in a degree of accuracy loss. Therefore, addressing this issue and exploring more nuanced token-specific compression strategies will be the focus of our future work.

Author Contributions

Conceptualization, W.T. and H.Z.; methodology, W.T. and H.Z.; software, W.T.; validation, W.T. and H.Z.; formal analysis, W.T.; investigation, W.T.; resources, W.T.; data curation, W.T.; writing—original draft preparation, W.T.; writing—review and editing, W.T.; visualization, W.T.; supervision, H.Z.; project administration, H.Z. and X.S.; funding acquisition, H.Z. and X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Zhejiang Sci-Tech University Research Startup Funding Project, grant number 11121731282202-01.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study are openly available in Github at https://github.com/KKTYZ/ACPose (accessed on 12 March 2025).

Acknowledgments

We would like to express our sincere gratitude to Haixiang Zhang for his invaluable guidance and support. We also thank the editor and the reviewers for their insightful comments and contributions to this paper.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Luvizon, D.C.; Picard, D.; Tabia, H. 2D/3D Pose Estimation and Action Recognition Using Multitask Deep Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 5137–5146. [Google Scholar]
  2. Geng, Z.; Wang, C.; Wei, Y.; Liu, Z.; Li, H.; Hu, H. Human Pose as Compositional Tokens. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 18–22 June 2023; pp. 660–671. [Google Scholar]
  3. Wang, Y.; Xia, Y.; Liu, S. BCCLR: A Skeleton-Based Action Recognition with Graph Convolutional Network Combining Behavior Dependence and Context Clues. Comput. Mater. Contin. 2024, 78, 4489–4507. [Google Scholar] [CrossRef]
  4. Xu, X.; Zhang, Y. High-resolution multi-scale feature fusion network for running posture estimation. Appl. Sci. 2024, 14, 3065. [Google Scholar] [CrossRef]
  5. Jiang, J.H.; Xia, N. PCNet: A Human Pose Compensation Network Based on Incremental Learning for Sports Actions Estimation. Complex Intell. Syst. 2025, 11, 17. [Google Scholar] [CrossRef]
  6. Luvizon, D.C.; Picard, D.; Tabia, H. Multi-task deep learning for real-time 3D human pose estimation and action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2752–2764. [Google Scholar] [CrossRef] [PubMed]
  7. Jha, P.; Yadav, G.P.K.; Bandhu, D.; Hemalatha, N.; Mandava, R.K.; Adin, M.Ş.; Saxena, K.K.; Patel, M. Human–machine interaction and implementation on the upper extremities of a humanoid robot. Discov. Appl. Sci. 2024, 6, 152. [Google Scholar] [CrossRef]
  8. Azhar, M.H.; Jalal, A. Human-Human Interaction Recognition Using Mask R-CNN and Multi-Class SVM. In Proceedings of the 2024 3rd International Conference on Emerging Trends in Electrical, Control, and Telecommunication Engineering (ETECTE), Lahore, Pakistan, 26–27 November 2024; IEEE: New York, NY, USA, 2024; pp. 1–6. [Google Scholar]
  9. Hernández, Ó.G.; Morell, V.; Ramon, J.L.; Jara, C.A. Human pose detection for robotic-assisted and rehabilitation environments. Appl. Sci. 2021, 11, 4183. [Google Scholar] [CrossRef]
  10. Wang, B.; Song, C.; Li, X.; Zhou, H.; Yang, H.; Wang, L. A deep learning-enabled visual-inertial fusion method for human pose estimation in occluded human-robot collaborative assembly scenarios. Robot. Comput. Integr. Manuf. 2025, 93, 102906. [Google Scholar] [CrossRef]
  11. Atari, R.; Bamani, E.; Sintov, A. Human Arm Pose Estimation with a Shoulder-worn Force-Myography Device for Human-Robot Interaction. IEEE Robot. Autom. Lett. 2025, 10, 2974–2981. [Google Scholar] [CrossRef]
  12. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16×16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations ICLR 2021, Vienna, Austria, 4 May 2021. [Google Scholar] [CrossRef]
  13. Dong, C.; Du, G. An enhanced real-time human pose estimation method based on modified YOLOv8 framework. Sci. Rep. 2024, 14, 8012. [Google Scholar] [CrossRef]
  14. Xia, H.; Zhang, T. Self-attention network for human pose estimation. Appl. Sci. 2021, 11, 1826. [Google Scholar] [CrossRef]
  15. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. RMPE: Regional Multi-Person Pose Estimation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar]
  16. Xiao, B.; Wu, H.; Wei, Y. Simple Baselines for Human Pose Estimation and Tracking. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 466–481. [Google Scholar]
  17. Liu, Z.; Lin, W.; Shi, Y.; Zhao, J. A Robustly Optimized BERT Pre-Training Approach with Post-Training. In Proceedings of the China National Conference on Chinese Computational Linguistics (CNCL), Hohhot, China, 13–15 August 2021; Springer: Cham, Switzerland, 2021; pp. 471–484. [Google Scholar]
  18. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 5485–5551. [Google Scholar] [CrossRef]
  19. Xu, Z.; Dai, M.; Zhang, Q.; Jiang, X. HRPVT: High-Resolution Pyramid Vision Transformer for Medium and Small-Scale Human Pose Estimation. Neurocomputing 2025, 619, 129154. [Google Scholar] [CrossRef]
  20. Ji, A.; Fan, H.; Xue, X. Vision-Based Body Pose Estimation of Excavator Using a Transformer-Based Deep-Learning Model. J. Comput. Civ. Eng. 2025, 39, 04024064. [Google Scholar] [CrossRef]
  21. Tian, Z.; Fu, W.; Woźniak, M.; Liu, S. PCDPose: Enhancing the lightweight 2D human pose estimation model with pose-enhancing attention and context broadcasting. Pattern Anal. Appl. 2025, 28, 59. [Google Scholar] [CrossRef]
  22. Yang, S.; Quan, Z.; Nie, M.; Yang, W. Transpose: Keypoint Localization via Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 11802–11812. [Google Scholar]
  23. Li, Y.; Zhang, S.; Wang, Z.; Yang, S.; Yang, W.; Xia, S.T.; Zhou, E. TokenPose: Learning Keypoint Tokens for Human Pose Estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 11313–11322. [Google Scholar]
  24. An, X.; Zhao, L.; Gong, C.; Wang, N.; Wang, D.; Yang, J. SharpPose: Sparse High-Resolution Representation for Human Pose Estimation. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Palo Alto, CA, USA, 25–27 March 2024; AAAI Press: Palo Alto, CA, USA, 2024; Volume 38, pp. 691–699. [Google Scholar]
  25. Tu, H.; Qiu, Z.; Yang, K.; Tan, X.; Zheng, X. HP-YOLO: A Lightweight Real-Time Human Pose Estimation Method. Appl. Sci. 2025, 15, 3025. [Google Scholar] [CrossRef]
  26. Cai, M.; Jeon, W.S.; Rhee, S.Y. LW-FastPose: A Lightweight Network for Human Pose Estimation Based on Improvements to FastPose. In Proceedings of the 2025 International Conference on Electronics, Information, and Communication (ICEIC), Rome, Italy, 16–17 January 2025; IEEE: New York, NY, USA, 2025; pp. 1–5. [Google Scholar]
  27. Noh, W.J.; Moon, K.R.; Lee, B.D. SMS-Net: Bridging the Gap Between High Accuracy and Low Computational Cost in Pose Estimation. Appl. Sci. 2024, 14, 10143. [Google Scholar] [CrossRef]
  28. Bao, H.; Dong, L.; Piao, S.; Wei, F. Beit: Bert pre-training of image transformers. arXiv 2021, arXiv:2106.08254. [Google Scholar] [CrossRef]
  29. Wan, Q.; Huang, Z.; Lu, J.; Yu, G.; Zhang, L. SeaFormer: Squeeze-Enhanced Axial Transformer for Mobile Semantic Segmentation. In Proceedings of the Eleventh International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 February 2023; ICLR: New Orleans, LA, USA, 2023. [Google Scholar]
  30. Jiao, J.; Tang, Y.M.; Lin, K.Y.; Gao, Y.; Ma, A.J.; Wang, Y.; Zheng, W.S. Dilateformer: Multi-scale dilated transformer for visual recognition. IEEE Trans. Multimed. 2023, 25, 8906–8919. [Google Scholar] [CrossRef]
  31. Mao, W.; Ge, Y.; Shen, C.; Tian, Z.; Wang, X.; Wang, Z. Tfpose: Direct human pose estimation with transformers. arXiv 2021, arXiv:2103.15320. [Google Scholar] [CrossRef]
  32. Xu, Y.; Zhang, J.; Zhang, Q.; Tao, D. Vitpose: Simple vision transformer baselines for human pose estimation. Adv. Neural Inf. Process. Syst. 2022, 35, 38571–38584. [Google Scholar] [CrossRef]
  33. Yuan, Y.; Fu, R.; Huang, L.; Lin, W.; Zhang, C.; Chen, X.; Wang, J. Hrformer: High-resolution vision transformer for dense predict. Adv. Neural Inf. Process. Syst. 2021, 34, 7281–7293. [Google Scholar] [CrossRef]
  34. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
Figure 1. The main structure of ACPose.
Figure 1. The main structure of ACPose.
Applsci 15 04746 g001
Figure 2. A core idea map of the axial compression module. (a) depicts the traditional approach; (b) shows a schematic diagram of the method proposed in this paper.
Figure 2. A core idea map of the axial compression module. (a) depicts the traditional approach; (b) shows a schematic diagram of the method proposed in this paper.
Applsci 15 04746 g002
Figure 3. Heatmap of human pose predicted by ACPose. Each row in this figure represents heatmap showing responses to different human keypoints. It can be seen that network effectively distinguishes between various keypoints.
Figure 3. Heatmap of human pose predicted by ACPose. Each row in this figure represents heatmap showing responses to different human keypoints. It can be seen that network effectively distinguishes between various keypoints.
Applsci 15 04746 g003
Figure 4. Experimental comparison of ACPose and TokenPose methods.
Figure 4. Experimental comparison of ACPose and TokenPose methods.
Applsci 15 04746 g004
Table 1. Structure configuration table of different ACPose models.
Table 1. Structure configuration table of different ACPose models.
Model#ParamsCNN BackboneHeadsLayersGFLOPs
ACPose-Base 13.9M HRNetW32-stage38125.1
ACPose-Large/D620.8MHRNetW48-stage3869.1
ACPose-Large/D2427.5MHRNetW48-stage312249.8
Table 2. Performance comparison between ACPose and mainstream models on COCO dataset.
Table 2. Performance comparison between ACPose and mainstream models on COCO dataset.
Method#ParamsGFLOPs G F L O P s T AP A P 50 A P 75 A P L A P M AR
TransPose-R-A46.0M8.93.3872.689.179.968.879.878.0
TransPose-H-S8.0M10.24.8874.289.680.870.681.079.5
TransPose-H-A617.5M21.811.475.890.182.171.982.880.8
TokenPose-B *13.5M5.71.2975.689.881.471.381.480.0
TokenPose-L/D6 *20.8M10.31.9777.790.081.871.882.480.4
ViTPose-Base86.0M18.6-75.890.783.278.468.781.1
SHaRPose-Base93.9M17.1-75.590.682.382.272.280.8
ACPose-B *13.9M5.1 (−11%)0.68 (−47.3%)75.7 (+0.1%)92.582.772.880.078.2
ACPose-L/D6 *21.0M8.8 (−15%)0.37 (−81.2%)77.4 (−0.4%)93.684.874.381.879.9
ACPose-L/D24 *27.5M9.91.5277.493.683.874.481.979.9
* The asterisk indicates the usage of GTBox; G F L O P s T represents the computational cost of the Transformer part.
Table 3. Performance comparison between ACPose and mainstream models on MPII dataset.
Table 3. Performance comparison between ACPose and mainstream models on MPII dataset.
Method#Params G F L O P s T HeaShoElbWriHipKneAnkMean
SimpleBaseline-R5034.0M-96.495.389.083.288.484.079.688.5
SimpleBaseline-R10153.0M-96.995.995.984.488.484.580.789.1
SimpleBaseline-R15253.0M-97.095.990.085.089.285.381.389.6
HRNetW3228.5M-96.996.090.685.888.786.682.690.1
TokenPose-L/D621.4M0.6497.195.991.085.889.586.182.790.2
ACPose-L/D6 (Ours)21.66M0.39 (−39.0%)97.196.190.986.188.986.181.690.0 (−0.2%)
G F L O P s T represents the GFLOPs of the Transformer part. Bold indicates the optimal indicator.
Table 4. Performance comparison of different size pooled cores on MPII datasets.
Table 4. Performance comparison of different size pooled cores on MPII datasets.
Modelk#ParamsGFLOPs G F L O P s T Mean
TokenPose-L/D6-21.49 M11.850.6490.2
ACPose-L/D6421.49 M11.560.35 (−45.3%)89.4 (−0.9%)
ACPose-L/D6321.53 M11.570.36 (−43.8%)89.8 (−0.4%)
ACPose-L/D6221.66 M11.600.39 (−39.0%)90.0 (−0.2%)
Table 5. Performance comparison of different compression degrees h on MPII datasets.
Table 5. Performance comparison of different compression degrees h on MPII datasets.
Modelh#ParamsGFLOPs G F L O P s T Mean
TokenPose-L/D6-21.49 M11.850.6490.2
ACPose-L/D63621.96 M11.780.57 (−10.9%)90.1 (−0.1%)
ACPose-L/D62421.73 M11.660.45 (−29.7%)90.0 (−0.2%)
ACPose-L/D61221.66 M11.600.39 (−39.0%)90.0 (−0.2%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, W.; Zhang, H.; Song, X. An Axial Compression Transformer for Efficient Human Pose Estimation. Appl. Sci. 2025, 15, 4746. https://doi.org/10.3390/app15094746

AMA Style

Tan W, Zhang H, Song X. An Axial Compression Transformer for Efficient Human Pose Estimation. Applied Sciences. 2025; 15(9):4746. https://doi.org/10.3390/app15094746

Chicago/Turabian Style

Tan, Wen, Haixiang Zhang, and Xinyi Song. 2025. "An Axial Compression Transformer for Efficient Human Pose Estimation" Applied Sciences 15, no. 9: 4746. https://doi.org/10.3390/app15094746

APA Style

Tan, W., Zhang, H., & Song, X. (2025). An Axial Compression Transformer for Efficient Human Pose Estimation. Applied Sciences, 15(9), 4746. https://doi.org/10.3390/app15094746

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop