Next Article in Journal
The Applicability of a Complete Archive of Keyhole Imagery for Land-Use Change Detection in China (1960–1984)
Previous Article in Journal
Quantitative Evaluation of Focus Measure Operators in Optical Microscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial–Temporal Heatmap Masked Autoencoder for Skeleton-Based Action Recognition

1
Department of Education, Ocean University of China, Qingdao 266100, China
2
Department of Campus Security, Ocean University of China, Qingdao 266100, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3146; https://doi.org/10.3390/s25103146
Submission received: 3 April 2025 / Revised: 9 May 2025 / Accepted: 12 May 2025 / Published: 16 May 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

Skeleton representation learning offers substantial advantages for action recognition by encoding intricate motion details and spatial–temporal dependencies among joints. However, fully supervised approaches necessitate large amounts of annotated data, which are often labor-intensive and costly to acquire. In this work, we propose the Spatial–Temporal Heatmap Masked Autoencoder (STH-MAE), a novel self-supervised framework tailored for skeleton-based action recognition. Unlike coordinate-based methods, STH-MAE adopts heatmap volumes as its primary representation, mitigating noise inherent in pose estimation while capitalizing on advances in Vision Transformers. The framework constructs a spatial–temporal heatmap (STH) by aggregating 2D joint heatmaps across both spatial and temporal axes. This STH is partitioned into non-overlapping patches to facilitate local feature learning, with a masking strategy applied to randomly conceal portions of the input. During pre-training, a Vision Transformer-based autoencoder equipped with a lightweight prediction head reconstructs the masked regions, fostering the extraction of robust and transferable skeletal representations. Comprehensive experiments on the NTU RGB+D 60 and NTU RGB+D 120 benchmarks demonstrate the superiority of STH-MAE, achieving state-of-the-art performance under multiple evaluation protocols.

1. Introduction

Human action recognition poses a critical challenge in computer vision, involving the interpretation of human behaviors and discrimination between distinct actions. Prior research has explored diverse modalities for feature extraction, including RGB images [1,2], optical flow [3,4], audio signals [5], and skeletal data [6,7,8]. Advances in deep learning and pose estimation have significantly elevated the prominence of skeleton-based approaches. Skeletal data, typically represented as sequences of joint coordinates, encode pose information while remaining invariant to background clutter and lighting variations [9]. In skeleton-based action recognition, predominant methods rely on fully supervised learning frameworks employing Convolutional Neural Networks (CNNs) [9,10], Recurrent Neural Networks (RNNs) [11,12], Graph Convolutional Networks (GCNs) [13,14], and transformers [7,15]. Despite achieving notable performance, these methods are prone to overfitting and require labor-intensive annotated datasets.
To mitigate these challenges, self-supervised learning methodologies have gained traction in the field of skeleton action recognition [16,17]. These approaches utilize unlabeled data to derive meaningful representations. Some self-supervised strategies are centered on pretext tasks, such as reconstructing motions or solving jigsaw puzzles, to learn local features. However, these tasks may fall short in capturing comprehensive temporal dynamics. Conversely, contrastive learning techniques have emerged, training models to differentiate between positive and negative skeleton sequence pairs from varying perspectives [18,19,20]. These approaches prioritize high-level contextual understanding but may overly rely on the quantity of contrastive pairs, potentially overlooking inter-frame joint correlations.
A recent innovation in self-supervised learning is the introduction of masked autoencoders, which have demonstrated exceptional generalization capabilities and robust performance across diverse computer vision tasks [21,22]. This method involves obscuring substantial portions of the input image and training the model to reconstruct the original image using the visible portions. The original MAE approach employs an asymmetric encoder–decoder architecture, masking 75% of the input image to create a nontrivial self-supervisory task. However, while MAE excels in image-based tasks, applying this technique to human skeleton sequences presents distinct challenges due to the rich semantic content and complex temporal nature of these sequences.
To address the existing challenges in skeleton-based action recognition, we introduce an innovative self-supervised masked autoencoder framework known as the Spatial–Temporal Heatmap Masked Autoencoder (STH-MAE). Our STH-MAE framework extends the core MAE principles to spatio-temporal data by introducing a hierarchical masking strategy. Unlike MAE, which focuses solely on random patch masking, STH-MAE incorporates both spatial and temporal masking, capturing the intricate dynamics of skeleton sequences. Additionally, STH-MAE introduces a lightweight temporal decoder module that leverages the sequential nature of skeleton data, enabling more accurate reconstruction of both spatial relationships and temporal transitions. Effective masking strategies are thus critical to adequately capture the temporal dynamics and spatial relationships inherent in skeleton data, distinguishing STH-MAE’s approach from the original MAE framework.
The core structure of the STH-MAE pipeline is built on several fundamental principles. Initially, a spatial–temporal heatmap is created by integrating heatmaps across both spatial and temporal dimensions. During the pre-training phase, a unique spatial–temporal masking approach is employed, which segments the heatmap into uniform, non-overlapping patches and masks selected portions. To optimize spatial–temporal representation learning, we investigate various factors such as the masking ratio, heatmap resolution, training epochs, and the depth of the decoder to find an ideal balance between computational cost and performance. The encoder is designed to extract generalized feature representations from the input, while the decoder is responsible for reconstructing the masked heatmap patches. By exploiting the spatial–temporal properties of the heatmap, the autoencoder’s backbone is constructed based on the Vision Transformer (ViT) architecture, incorporating standard transformer blocks within an asymmetric encoder–decoder configuration. In the fine-tuning phase, only the encoder is employed, complemented by a simple output layer, to predict actions. Our method demonstrates superior performance compared to existing self-supervised learning techniques in skeleton-based action recognition, achieving competitive results without the need for additional data. This approach democratizes access to advanced action recognition capabilities, especially in scenarios where labeled data are scarce or expensive to obtain. Our key contributions can be summarized as follows:
  • We introduce a streamlined and effective method for masked autoencoder in skeleton-based action recognition, emphasizing the acquisition of comprehensive and adaptable representations through the masking and reconstruction of skeleton sequences.
  • We present a spatial–temporal heatmap as the primary representation of skeletons, enhancing robustness in pose estimation and capitalizing on recent advancements in Visual Transformers.
  • Our model underwent extensive evaluation on the NTU RGB+D 60 and NTU RGB+D 120 datasets. Experimental results indicate that STH-MAE achieves state-of-the-art performance in self-supervised settings.

2. Related Works

2.1. Self-Supervised Skeleton-Based Action Recognition

Self-supervised skeleton-based action recognition has garnered considerable attention in research, with the primary objective of learning action feature representations from unlabeled skeleton sequences [23]. The advancements in this field can be broadly categorized into two approaches: those based on pretext tasks [16,24,25,26] and those utilizing contrastive learning techniques [27,28,29,30]. For instance, Zheng et al. [16] introduced a skeleton inpainting framework designed to capture long-term motion dynamics. Similarly, Lin et al. [24] employed multiple tasks, such as solving jigsaw puzzles, to learn more generalized skeleton features. Xu et al. [25] proposed a method involving reverse sequential predictions within an encoder–decoder structure to extract motion patterns. However, the feature representations derived from these methods might be overly specialized to the specific pre-designed tasks. In recent years, the incorporation of contrastive learning has led to significant progress in self-supervised skeleton-based action recognition. Contrastive learning has been effective in enhancing the learning of action representations by employing carefully crafted augmentations to generate varied movement patterns with consistent semantics. The primary challenge lies in applying robust augmentations that do not distort the skeletal structure to the point of semantic loss, which can result in unstable training. To mitigate this issue, Zhang et al. [29] introduced the Hierarchical Consistent Contrastive Learning framework, which integrates strong augmentations with a directional clustering mechanism to ensure hierarchical consistency in the learned representations. This framework gradually increases the complexity of augmentations and employs an asymmetric loss function. Additionally, Chen et al. [31] explored the significance of data augmentation and contrastive pair generation within the context of self-supervised skeleton-based action recognition. They developed SkeleMixCLR, a framework that utilizes spatio-temporal skeleton mixing augmentations to create challenging contrastive samples, thereby improving representation learning. This approach involves mixing two skeleton sequences and extending contrastive pairs with trimmed and truncated views. Research directions also focus on creating more robust and efficient methods for recognizing actions from skeleton sequences. Paoletti et al. [32] and Hua et al. [30] have highlighted the potential of unsupervised and part-aware contrastive learning in this domain. These methods emphasize high-level contextual information but often rely heavily on the quantity of contrastive pairs, potentially overlooking the correlation information between joints across frames. Leveraging advanced machine learning techniques, such as transformers, could further improve the ability of models to capture complex spatial and temporal relationships within skeleton sequences. In summary, learning task-specific representations that may not generalize to downstream action recognition, while contrastive learning based approaches require careful design of augmentations to avoid semantic distortion. This paper uses masked autoencoding, a task-agnostic framework, where reconstructing masked heatmap patches forces the model to learn comprehensive spatial–temporal dependencies.

2.2. Masked Autoencoder

In the realm of self-supervised learning, the masked autoencoder has emerged as a significant tool, allowing models to acquire meaningful data representations by inferring masked or omitted portions of the input. By leveraging the information from the visible parts of the input, this method enhances the model’s comprehension of the inherent data structure. Germain et al. [33] pioneered the Masked Autoencoder for Distribution Estimation, a technique that enables autoencoder outputs to be interpreted as conditional probabilities, significantly improving the generation of models with complete joint probability distributions. Zhou et al. [34] investigated the application of masked autoencoders for pre-training Vision Transformers in the context of medical image analysis, showing their efficacy in pre-training tasks involving natural image analysis. Geng et al. [35] introduced the Multimodal Masked Autoencoder, which facilitates the learning of transferable representations across visual and linguistic data without the need for modality-specific encoders. Zhang et al. [36] provided a comprehensive survey that underscored the increasing interest in masked autoencoders for self-supervised learning in vision and other domains, highlighting its potential as a promising avenue for self-supervised learning research. Further expanding on these concepts, Radosavovic et al. [37] explored the use of masked autoencoders in practical robot learning scenarios, while Yang et al. [38] examined their utility in domain-invariant self-supervised learning. Liu et al. [39] proposed Masked Decision Prediction as a self-supervised pre-training technique for reinforcement learning and behavioral cloning, utilizing masked autoencoders to reconstruct absent data in state–action trajectories. Reed et al. [40] introduced Scale-MAE, a scale-aware masked autoencoder designed for multiscale geospatial representation learning, emphasizing the importance of understanding relationships between data at various scales. Additionally, Woo et al. [41] and Wang et al. [42] extended the applicability of masked autoencoders to Convolutional Neural Networks and video pre-training, respectively, demonstrating the scalability and efficiency of masked autoencoders in developing foundational models for diverse data types. Overall, masked autoencoders have demonstrated versatility and effectiveness as tools for self-supervised learning and pre-training across various domains. Methods such as SkeletonMAE [43] apply masking to coordinate sequences, struggle with pose estimation noise and lack explicit spatial–temporal structure. Therefore, this paper introduces spatial–temporal heatmaps as the input representation, masking entire joints over time intervals, preserving structural relationships in spatial–temporal dimension.

2.3. Pose Estimation Heatmap

Human pose estimation represents a pivotal task within the domain of computer vision, aiming to accurately detect and pinpoint a person’s key points to comprehend posture and movement. In recent years, heatmap-based methodologies, which encode keypoint locations as heatmaps, have emerged as the leading approach in this field [44] and have proven particularly effective. These methods generate a probability distribution over the image, indicating the likelihood of each pixel being the keypoint. The key advantages of heatmap-based approaches include their ability to handle multiple keypoints simultaneously and their robustness to small perturbations in the image [45,46]. Despite the widespread adoption of heatmap-based regression in keypoint estimation tasks, including human pose estimation, this method exhibits significant limitations [47]. Comparative analyses between heatmap-based detection and integral regression for 2D pose estimation have shed light on crucial distinctions between these techniques [48]. Although heatmap regression predicts the likelihood of keypoint presence within a 2D space, it is critical to address the biases and inherent limitations of this approach [49]. Recent advancements in human pose estimation have also investigated sophisticated heatmap estimation structures, such as the MH Pose method, to improve both accuracy and performance [50].
While spatial–temporal data are widely applied across fields such as environmental monitoring [51], industrial systems [52], and causal discovery [52], our research targets skeleton-based action recognition, a domain marked by unique challenges including pose estimation noise and limited labeled data. Converting spatial–temporal data into heatmaps presents an engaging strategy, particularly for action recognition, as it simplifies complex data while preserving interpretability and minimizing information loss. Unlike Hu et al., who used spatial–temporal heatmaps in autonomous driving planning [53], our approach leverages heatmaps to encode joint probability distributions across spatial–temporal dimensions, thereby explicitly modeling human motion dynamics. Whereas Hu et al. applied heatmaps within a supervised learning framework for motion planning, our method, STH-MAE, employs a self-supervised masked autoencoder approach. By reconstructing masked heatmap patches, STH-MAE learns generalized representations without relying on labeled data, addressing the scarcity of annotated skeleton action sequences.

3. Proposed Method

We aim to establish an efficient and streamlined framework tailored for self-supervised skeleton-based action classification representation learning via masked autoencoders. The proposed method, referred to as STH-MAE, is visually summarized in Figure 1. Initially, the input action sequence is processed through a Spatial–Temporal Heatmap Masking module. Following this, a Visual Transformer-based autoencoder, equipped with a single-layer prediction head, is utilized to reconstruct the masked segments of the spatial–temporal heatmap. Detailed explanations of STH-MAE are provided in subsequent sections.

3.1. Spatial–Temporal Heatmap Masking

Spatial–Temporal Heatmap. The 2D poses extracted from the video frames are transformed for integration into STH-MAE by reshaping them into a spatial–temporal heatmap volume. Formally, a 2D pose is represented as a heatmap of dimensions K × H × W , where T denotes the temporal dimension representing the action sequence length, K corresponds to the anatomical joints in the human skeletal representation, and H × W specifies the spatial resolution of joint-specific confidence maps, as shown in Figure 2. Each heatmap encodes probabilistic joint localization through spatial coordinate likelihood estimation. The construction process involves two systematic phases. First, per-frame heatmap aggregation generates K distinct confidence maps, each of dimension H × W for every temporal instance t T , forming a 3D tensor K × H × W through channel-wise concatenation. Second, temporal integration stacks these frame-level tensors along the temporal axis via tensor concatenation, yielding the complete 4D STH representation T × K × H × W .
To ensure a consistent number of input patches for the encoder–decoder, the spatial–temporal heatmap is ultimately resized to T H W / K × K × K , conforming it to an image-like format. The following steps are performed to reshape STH: Flatten Spatial-Temporal Dimensions combines T, H, and W into a single dimension; Partition into Tokens divides the flattened dimension by K to create T H W / K dimensional vectors; Reshape into 3D Tensor then rearrange the flattened sequence into a 3D tensor with dimensions
ST H reshaped R K × K × T H W K
where K × K is grid of tokens, equivalent to ViT’s patch grid and T H W / K is the feature vector per token. The process transforms a 4D volume of sequential 2D poses into a 3D spatial–temporal heatmap volume, enabling the capture of both spatial and temporal dynamics of human actions. This representation is crucial for understanding complex motion patterns and interactions between different joints over time. Additionally, the heatmap volume can be easily integrated into transformer-based architectures, which have shown superior performance in capturing long-range dependencies in both spatial and temporal dimensions.
Masking. Following the method outlined in ViT [54], the spatial–temporal heatmap is divided into regular, non-overlapping K × K patches based on the joints and time intervals of the original 2D heatmap. Each patch comprises normalized 2D heatmaps of the same joint over a fixed time interval, which aids in convergence. A random sampling strategy is then employed to select a subset of these patches for masking, ensuring that patches are selected without replacement and follow a uniform distribution. Through experimentation, we have found that maintaining a relatively high, yet not extreme, masking ratio is most effective for our method. The set of masked tokens, denoted as P g t , with a masking ratio m, is used as the ground truth for calculating the reconstruction loss. Our masking strategy draws inspiration from Vision Transformers but is tailored to the specific needs of skeleton-based action recognition. By partitioning the spatial–temporal heatmap into joint-time patches and employing a high masking ratio, we encourage the model to focus on learning comprehensive representations from limited visible information. This not only improves the model’s robustness to missing data but also enhances its ability to infer complex action patterns from partial observations.

3.2. Autoencoder Pre-Training

In our framework, the autoencoder is designed with a Vision Transformer (ViT) backbone, incorporating traditional transformer blocks and an asymmetric encoder–decoder architecture. The autoencoder’s final layer features a single-layer prediction head that aims to reconstruct the input.
Encoder–Decoder. Our approach applies the encoder only to unmasked patches, processing solely the visible tokens T v and excluding the mask tokens T m . By focusing on a subset of the total tokens, we enable the training of considerably larger encoders with reduced computational and memory requirements. The unmasked tokens undergo linear projection embedding, with positional embeddings added, and are then processed through a sequence of transformer blocks, resulting in encoded tokens represented as T e .
The decoder in our architecture mirrors the encoder but with a fewer number of transformer blocks. It accepts both encoded visible tokens T e and mask tokens T m as inputs. Each mask token, representing a missing patch to be predicted, is denoted by a shared, learned vector. Positional embeddings are incorporated for all tokens, including mask tokens, to maintain spatial location information. The integration of positional embeddings ensures that spatial information is preserved throughout the encoding and decoding processes, which is crucial for accurate reconstruction. The decoder processes these inputs to produce decoded mask tokens H m , which are then forwarded to the prediction head. The encoder–decoder process can be summarized as follows:
T e = Encoder ( T v ) , H m = Decoder ( concat ( T e , T m ) ) .
During pre-training, the decoder is exclusively tasked with the reconstruction of masked tokens, while the encoder generates skeleton representations for action recognition. This separation allows for a flexible design of the decoder, independent of the encoder’s configuration. Our experiments explore various decoder sizes. The significance of our autoencoder’s architecture lies in its capacity to handle the complexity of visual data while optimizing computational efficiency. By exclusively focusing on visible tokens during encoding, we can allocate resources more effectively, enabling the training of larger models. Additionally, the flexibility in decoder design allows us to experiment with various configurations to determine the optimal balance between accuracy and computational overhead.
Prediction Head. Our approach, named STH-MAE, reconstructs the input image by predicting the pixel values for each masked patch. As the concluding layer of the autoencoder’s backbone, the prediction head’s purpose is to reconstruct the masked tokens. We utilize a fully connected layer as the prediction head. The output H m from the decoder is projected by the prediction head to a vector with dimensions matching the total number of pixels in a patch, followed by a reshape operation to generate the predicted masked tokens P p r e .
P p r e = Reshape ( FC ( H m ) ) .
The fully connected layer within the prediction head is meticulously designed to handle high-dimensional data effectively, ensuring precise projection and reshaping operations.
Reconstruction Target. To compute the reconstruction loss, we compare the predicted masked tokens P p r e with the actual masked tokens P g t using Mean Squared Error (MSE):
L = 1 N i = 1 N | | P p r e i P g t i | | 2 ,
where N represents the number of pixels in a patch, and P p r e i and P g t i are the predicted and ground-truth values for the ith pixel, respectively. The use of MSE as the loss function is crucial in minimizing the difference between the predicted and actual masked tokens, thereby optimizing the model’s performance.

4. Experiments

4.1. Datasets

The NTU-RGB+D 60 and NTU-RGB+D 120 datasets are the most prevalent datasets utilized in contemporary research.
  • NTU RGB+D 60 (NTU 60) is an extensive dataset designed specifically for the recognition of human actions based on skeletal data. It consists of 56,578 video sequences covering 60 different action categories, with each human figure represented by 25 joints. This dataset includes two standard benchmark protocols: Cross-Subject (X-Sub) and Cross-View (X-View).
  • NTU RGB+D 120 (NTU 120), an extension of NTU 60, provides a larger dataset with 113,945 sequences encompassing 120 action labels. It also incorporates two benchmark protocols: Cross-Subject (X-Sub) and Cross-Set (X-Set). For these datasets, the estimation of 2D heatmaps for human joints is performed using HRNet [55] on sequences of RGB frames.

4.2. Evaluation Protocol

In our study, we employed the widely adopted evaluation protocols for these datasets.
  • Linear Evaluation Protocol: The models were evaluated using a linear evaluation approach for the action recognition task. This involves training a linear classifier (comprising a fully-connected layer followed by a softmax layer) with the encoder weights fixed.
  • Semi-supervised Evaluation Protocol: The encoder is pre-trained using the entire dataset and then the whole model is finetuned with only 1% or 10% randomly selected labeled data.
  • Finetune Evaluation Protocol: In this protocol, a linear classifier is appended to the pre-trained encoder, and the entire model is fine-tuned to compare its performance with fully supervised methods.

4.3. Implementation Details

All experiments were conducted using the PyTorch 1.9.0 deep learning framework on two RTX 3090 GPUs. Training employed the AdamW optimizer with a base learning rate of 0.001 and a weight decay of 0.05. The default mask ratio was set to 0.75, and the resolution was set to 8 × 8 . The autoencoder’s backbone consisted of 24 transformer blocks for the encoder and 8 transformer blocks for the decoder. For the NTU 60 and NTU 120 datasets, all samples were reshaped to 85 frames. Additionally, sine–cosine positional embedding was added to both encoder and decoder inputs. During self-supervised representation learning, the autoencoder model was trained on the training set without ground-truth labels for 20 epochs. For supervised learning, the encoder model and linear classifier were randomly initialized and jointly trained on the training set with ground-truth action labels.

4.4. Ablation Study

This section delves into the analysis of key components and hyperparameters of STH-MAE. By systematically varying the masking ratio, heatmap resolution, decoder depth, and pre-training epochs, we can better understand the trade-offs and performance impacts of each parameter. Unless specified otherwise, all experiments are carried out on the NTU 60 dataset utilizing the Cross-Subject benchmarks and a linear evaluation protocol.
Ablation study on masking ratio. We investigated the effects of varying masking ratios on STH-MAE’s performance. Employing a random strategy, we tested patch masking ratios of 0.55, 0.75, and 0.95. As shown in Table 1, a masking ratio of 0.75 provides the best results among the tested ratios. Higher ratios foster exploration beneficial for tasks requiring extensive comprehension, whereas lower ratios emphasize exploitation, advantageous for detailed tasks. Achieving a balanced trade-off in the masking ratio enhances overall performance effectively.
Ablation study on heatmap resolution. The influence of 2D heatmap resolution on STH-MAE is examined, as illustrated in Table 2. Results demonstrate that the highest accuracy is obtained at a resolution of 8 × 8 , consistent with expectations. Resolutions that are too low fail to capture features from ambiguous joints adequately, while excessively high resolutions struggle to discern complex spatial–temporal joint relationships. This balance is particularly critical for tasks involving complex spatial–temporal relationships, where both joint ambiguity and joint relationship complexity need to be effectively managed.
Ablation study on decoder depth. The decoder depth in our model refers to the number of transformer blocks employed. We tested different decoder depths, specifically 12, 8, and 4 blocks. Our experiments (refer to Table 3) reveal that STH-MAE achieves the highest performance with a Top1 accuracy of 84.31% when the decoder depth is set to 8 blocks. In contrast, deeper (12 blocks) and shallower (4 blocks) configurations result in lower performance, with Top1 accuracies of 83.47% and 82.90%, respectively. This difference can be attributed to the nature of the tasks involved—heatmap patch reconstruction versus recognition. In an autoencoder, the final layers focus primarily on reconstruction and are less critical for recognition. A deeper decoder captures these reconstruction nuances better, leading to more abstract latent representations. Understanding this balance can inform the design of more effective models for various tasks in computer vision. Moreover, as our STH-MAE framework employs a hierarchical transformer architecture, the computational complexity scales linearly with the number of input patches and transformer layers. However, our pre-training masking strategy, 75% patch masking, reduces the encoder’s effective input to 0.25 patches, lowering pre-training complexity by 4× compared to full-input training.
Ablation study on pre-training epoch. In the realm of self-supervised learning, it is commonly observed that extending the duration of pre-training epochs tends to yield more favorable outcomes. In our investigation, we systematically increased the pre-training epochs from 1 to 20, assessing optimal linear outcomes every 5 epochs. The data presented in Table 4 highlight that the highest accuracy, reaching 84.31%, was achieved at the 20-epoch mark, establishing it as the default setting. Noteworthy is the substantial enhancement of 4.3% observed between epochs 5 and 10. However, the subsequent progression from epochs 15 to 20 saw a marginal 0.1% improvement, signaling diminishing returns in efficacy beyond the 20-epoch threshold.

4.5. Comparison with State-of-the-Art

Linear evaluation. When comparing with current methods, the performance of STH-MAE stands out prominently, as shown in Table 5. Remarkably, STH-MAE consistently outperforms other approaches across all datasets evaluated. On the NTU 60 CS and NTU 120 X-sub datasets, it surpasses the previous leading method, ActCLR, by 3.4% and 5.3% respectively. This significant improvement highlights the efficacy of our spatial–temporal prediction targets.
Fine-tuned evaluation. The fine-tuning results, depicted in Table 6, highlight the efficacy of STH-MAE when trained with full supervision. On both the NTU 60 and NTU 120 datasets, STH-MAE not only matches but often exceeds the performance of the current state-of-the-art methods. Notably, while HYSP [60] holds strong performance metrics, STH-MAE enhances performance across almost all metrics, except for the X-sub on NTU 120.
Semi-supervised evaluation. Table 7 showcases the evaluation of semi-supervised results on the NTU 60 datasets. Particularly noteworthy is the performance of STH-MAE, which achieves a recognition rate of 39.8% for 1% of the data and 74.5% for 10% of the data on the CS benchmark. Similarly, on the CV benchmark, it achieves a recognition rate of 39.4% for 1% of the data and 77.2% for 10% of the data. It is important to highlight that STH-MAE not only matches but often surpasses the performance of the current state-of-the-art methods. Despite having limited labeled data, STH-MAE demonstrates significant performance improvements, showcasing its ability to learn effectively even in scenarios with less supervision. This capability of STH-MAE to perform well under such conditions is critical for real-world applications, especially where labeled data are scarce or expensive to acquire. Consequently, STH-MAE’s adeptness in delivering high performance under these circumstances makes it exceptionally valuable for practical implementations, ensuring that optimal performance can be maintained even with limited resources.
STH-MAE achieves state-of-the-art performance in most benchmarks, though minor variations arise due to inherent trade-offs in self-supervised learning. As a reconstruction-based framework, STH-MAE, prioritizes generalizable spatial–temporal dependencies, which excel in robustness and label efficiency but may lag slightly in highly specialized tasks requiring fine-grained distinctions. This reflects the balance between generality and specificity in representation learning, suggesting future exploration of adaptive strategies or hybrid objectives to address such gaps while retaining computational advantages.

5. Conclusions

In this study, we propose STH-MAE, a novel self-supervised framework for skeleton-based action recognition that leverages spatial–temporal heatmap reconstruction to learn comprehensive skeletal representations. Our approach begins by aggregating 2D heatmaps across spatial and temporal dimensions to construct a unified spatial–temporal heatmap. A randomized masking strategy is then applied to occlude patches within this heatmap, followed by reconstruction using a Visual Transformer architecture with a single-layer prediction head. Through this process, the model learns transferable skeletal features without requiring labeled data. Comprehensive evaluations on two large-scale benchmarks validate the effectiveness of STH-MAE, achieving state-of-the-art results across multiple metrics and outperforming existing methods.
Future research could extend this work by integrating multi-modal data to enhance action recognition robustness, or exploring cross-domain transfer learning strategies through large-scale pre-training followed by task-specific fine-tuning. These directions may further advance the adaptability and performance of skeleton-based models in real-world applications.

Author Contributions

Conceptualization, C.B. and W.L.; methodology, C.B. and T.W.; software, Y.Y.; formal analysis, Y.Y.; investigation, Y.Y.; writing—original draft preparation, T.W.; writing—review and editing, C.B. and W.L.; funding acquisition, C.B. and W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62407042, 62277045), Humanity and Social Science Foundation of Ministry of Education (No. 24YJC880004), and Natural Science Foundation of Shandong Province (No. ZR2024QF075).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from Rose Lab and are available https://rose1.ntu.edu.sg/dataset/actionRecognition/ (accessed on 2 April 2025) with the permission of Rose Lab.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  2. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning spatiotemporal features with 3d convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015. [Google Scholar]
  3. Simonyan, K.; Zisserman, A. Two-stream convolutional networks for action recognition in videos. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  4. Sun, S.; Kuang, Z.; Sheng, L.; Ouyang, W.; Zhang, W. Optical flow guided feature: A fast and robust motion representation for video action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018. [Google Scholar]
  5. Sun, Z.; Ke, Q.; Rahmani, H.; Bennamoun, M.; Wang, G.; Liu, J. Human action recognition from various data modalities: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3200–3225. [Google Scholar] [CrossRef] [PubMed]
  6. Zhou, H.; Liu, Q.; Wang, Y. Learning discriminative representations for skeleton based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 20–22 June 2023. [Google Scholar]
  7. Wang, L.; Koniusz, P. 3mformer: Multi-order multi-mode transformer for skeletal action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 20–22 June 2023. [Google Scholar]
  8. Bian, C.; Feng, W.; Wan, L.; Wang, S. Structural knowledge distillation for efficient skeleton-based action recognition. IEEE Trans. Image Process. 2021, 30, 2963–2976. [Google Scholar] [CrossRef] [PubMed]
  9. Duan, H.; Zhao, Y.; Chen, K.; Lin, D.; Dai, B. Revisiting skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022. [Google Scholar]
  10. Bavil, A.F.; Damirchi, H.; Taghirad, H.D. Action Capsules: Human skeleton action recognition. Comput. Vis. Image Underst. 2023, 233, 103722. [Google Scholar] [CrossRef]
  11. Li, S.; Li, W.; Cook, C.; Zhu, C.; Gao, Y. Independently recurrent neural network (indrnn): Building a longer and deeper rnn. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UR, USA, 19–21 June 2018. [Google Scholar]
  12. Wang, H.; Wang, L. Beyond joints: Learning representations from primitive geometries for skeleton-based action recognition and detection. IEEE Trans. Image Process. 2018, 27, 4382–4394. [Google Scholar] [CrossRef] [PubMed]
  13. Chen, Y.; Zhang, Z.; Yuan, C.; Li, B.; Deng, Y.; Hu, W. Channel-wise topology refinement graph convolution for skeleton-based action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021. [Google Scholar]
  14. Song, Y.F.; Zhang, Z.; Shan, C.; Wang, L. Constructing stronger and faster baselines for skeleton-based action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1474–1488. [Google Scholar] [CrossRef] [PubMed]
  15. Kong, J.; Bian, Y.; Jiang, M. MTT: Multi-scale temporal transformer for skeleton-based action recognition. IEEE Signal Process. Lett. 2022, 29, 528–532. [Google Scholar] [CrossRef]
  16. Zheng, N.; Wen, J.; Liu, R.; Long, L.; Dai, J.; Gong, Z. Unsupervised representation learning with long-term dynamics for skeleton based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  17. Su, K.; Liu, X.; Shlizerman, E. Predict & cluster: Unsupervised skeleton based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020. [Google Scholar]
  18. Bian, C.; Feng, W.; Wang, S. Self-supervised representation learning for skeleton-based group activity recognition. In Proceedings of the ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022. [Google Scholar]
  19. Bian, C.; Feng, W.; Meng, F.; Wang, S. Global–local contrastive multiview representation learning for skeleton-based action recognition. Comput. Vis. Image Underst. 2023, 229, 103655. [Google Scholar] [CrossRef]
  20. Shah, A.; Roy, A.; Shah, K.; Mishra, S.; Jacobs, D.; Cherian, A.; Chellappa, R. Halp: Hallucinating latent positives for skeleton-based self-supervised learning of actions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 20–22 June 2023. [Google Scholar]
  21. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022. [Google Scholar]
  22. Fang, Y.; Wang, W.; Xie, B.; Sun, Q.; Wu, L.; Wang, X.; Huang, T.; Wang, X.; Cao, Y. Eva: Exploring the limits of masked visual representation learning at scale. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 20–22 June 2023. [Google Scholar]
  23. Guan, S.; Yu, X.; Huang, W.; Fang, G.; Lu, H. DMMG: Dual min-max games for self-supervised skeleton-based action recognition. IEEE Trans. Image Process. 2023, 33, 395–407. [Google Scholar] [CrossRef] [PubMed]
  24. Lin, L.; Song, S.; Yang, W.; Liu, J. Ms2l: Multi-task self-supervised learning for skeleton based action recognition. In Proceedings of the ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020. [Google Scholar]
  25. Xu, S.; Rao, H.; Hu, X.; Cheng, J.; Hu, B. Prototypical contrast and reverse prediction: Unsupervised skeleton based action recognition. IEEE Trans. Multimed. 2021, 25, 624–634. [Google Scholar] [CrossRef]
  26. Yang, S.; Liu, J.; Lu, S.; Er, M.H.; Kot, A.C. Skeleton cloud colorization for unsupervised 3d action representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021. [Google Scholar]
  27. Li, L.; Wang, M.; Ni, B.; Wang, H.; Yang, J.; Zhang, W. 3d human action representation learning via cross-view consistency pursuit. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021. [Google Scholar]
  28. Guo, T.; Liu, H.; Chen, Z.; Liu, M.; Wang, T.; Ding, R. Contrastive learning from extremely augmented skeleton sequences for self-supervised action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22–28 February 2022. [Google Scholar]
  29. Zhang, J.; Lin, L.; Liu, J. Hierarchical consistent contrastive learning for skeleton-based action recognition with growing augmentations. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  30. Hua, Y.; Wu, W.; Zheng, C.; Lu, A.; Liu, M.; Chen, C.; Wu, S. Part aware contrastive learning for self-supervised action recognition. In Proceedings of the International Joint Conference on Artificial Intelligence, Macao, China, 19–25 August 2023. [Google Scholar]
  31. Chen, Z.; Liu, H.; Guo, T.; Chen, Z.; Song, P.; Tang, H. Contrastive learning from spatio-temporal mixed skeleton sequences for self-supervised skeleton-based action recognition. arXiv 2022, arXiv:2207.03065. [Google Scholar]
  32. Paoletti, G.; Cavazza, J.; Beyan, C.; Del Bue, A. Unsupervised human action recognition with skeletal graph laplacian and self-supervised viewpoints invariance. In Proceedings of the British Machine Vision Conference, Virtual, 22–15 November 2021. [Google Scholar]
  33. Germain, M.; Gregor, K.; Murray, I.; Larochelle, H. Made: Masked autoencoder for distribution estimation. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  34. Zhou, L.; Liu, H.; Bae, J.; He, J.; Samaras, D.; Prasanna, P. Self pre-training with masked autoencoders for medical image analysis. arXiv 2022, arXiv:2203.05573. [Google Scholar]
  35. Geng, X.; Liu, H.; Lee, L.; Schuurmans, D.; Levine, S.; Abbeel, P. Multimodal masked autoencoders learn transferable Representations. In Proceedings of the Workshop on Pre-Training: Perspectives, Pitfalls, and Paths Forward at ICML, Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  36. Zhang, C.; Zhang, C.; Song, J.; Yi, J.S.K.; Zhang, K.; Kweon, I.S. A survey on masked autoencoder for self-supervised learning in vision and beyond. arXiv 2022, arXiv:2208.00173. [Google Scholar]
  37. Radosavovic, I.; Xiao, T.; James, S.; Abbeel, P.; Malik, J.; Darrell, T. Real-world robot learning with masked visual pre-training. In Proceedings of the Conference on Robot Learning, Atlanta, GA, USA, 6–9 November 2023. [Google Scholar]
  38. Yang, H.; Tang, S.; Chen, M.; Wang, Y.; Zhu, F.; Bai, L.; Zhao, R.; Ouyang, W. Domain invariant masked autoencoders for self-supervised learning from multi-domains. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  39. Liu, F.; Liu, H.; Grover, A.; Abbeel, P. Masked autoencoding for scalable and generalizable decision making. In Proceedings of the Advances in Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
  40. Reed, C.J.; Gupta, R.; Li, S.; Brockman, S.; Funk, C.; Clipp, B.; Keutzer, K.; Candido, S.; Uyttendaele, M.; Darrell, T. Scale-mae: A scale-aware masked autoencoder for multiscale geospatial representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023. [Google Scholar]
  41. Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. Convnext v2: Co-designing and scaling convnets with masked autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  42. Wang, L.; Huang, B.; Zhao, Z.; Tong, Z.; He, Y.; Wang, Y.; Wang, Y.; Qiao, Y. Videomae v2: Scaling video masked autoencoders with dual masking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  43. Wu, W.; Hua, Y.; Zheng, C.; Wu, S.; Chen, C.; Lu, A. Skeletonmae: Spatial-temporal masked autoencoders for self-supervised skeleton action recognition. In Proceedings of the IEEE International Conference on Multimedia and Expo Workshops, Brisbane, Australia, 10–14 July 2023. [Google Scholar]
  44. Li, S.; Zhang, H.; Ma, H.; Feng, J.; Jiang, M. SSA Net: Small scale-aware enhancement network for human pose estimation. Sensors 2023, 23, 7299. [Google Scholar] [CrossRef] [PubMed]
  45. Zhang, F.; Zhu, X.; Dai, H.; Ye, M.; Zhu, C. Distribution-aware coordinate representation for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  46. Yang, F.; Song, Z.; Xiao, Z.; Mo, Y.; Chen, Y.; Pan, Z.; Zhang, M.; Zhang, Y.; Qian, B.; Jin, W. Error compensation heatmap decoding for human pose estimation. IEEE Access 2021, 9, 114514–114522. [Google Scholar]
  47. McNally, W.; Vats, K.; Wong, A.; McPhee, J. Rethinking keypoint representations: Modeling keypoints and poses as objects for multi-person human pose estimation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  48. Gu, K.; Yang, L.; Yao, A. Removing the bias of integral pose regression. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  49. Kim, D.; Wang, K.; Saenko, K.; Betke, M.; Sclaroff, S. A unified framework for domain adaptive pose estimation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  50. Zhou, H.; Hong, C.; Han, Y.; Huang, P.; Zhuang, Y. Mh pose: 3d human pose estimation based on high-quality heatmap. In Proceedings of the IEEE International Conference on Big Data, Virtual, 15–18 December 2021. [Google Scholar]
  51. Wang, Y.; Kong, Y.; Chen, H.; Ding, Y. Spatial-temporal characteristics of drought detected from meteorological data with high resolution in Shaanxi Province, China. J. Arid Land 2020, 12, 561–579. [Google Scholar] [CrossRef]
  52. Wang, Y.; Peng, H.; Wang, G.; Tang, X.; Wang, X.; Liu, C. Monitoring industrial control systems via spatio-temporal graph neural networks. Eng. Appl. Artif. Intell. 2023, 122, 106144. [Google Scholar] [CrossRef]
  53. Hu, Y.; Li, K.; Liang, P.; Qian, J.; Yang, Z.; Zhang, H.; Shao, W.; Ding, Z.; Xu, W.; Liu, Q. Imitation with spatial-temporal heatmap: 2nd place solution for nuplan challenge. arXiv 2023, arXiv:2306.15700. [Google Scholar]
  54. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, Virtual, 26 April–1 May 2020. [Google Scholar]
  55. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  56. Rao, H.; Xu, S.; Hu, X.; Cheng, J.; Hu, B. Augmented skeleton based contrastive action learning with momentum lstm for unsupervised action recognition. Inf. Sci. 2021, 569, 90–109. [Google Scholar] [CrossRef]
  57. Nie, Q.; Liu, Z.; Liu, Y. Unsupervised 3d human pose representation with viewpoint and pose disentanglement. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  58. Mao, Y.; Zhou, W.; Lu, Z.; Deng, J.; Li, H. Cmd: Self-supervised 3d action representation learning with cross-modal mutual distillation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  59. Zhou, Y.; Duan, H.; Rao, A.; Su, B.; Wang, J. Self-supervised action representation learning from partial spatio-temporal skeleton sequences. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  60. Franco, L.; Mandica, P.; Munjal, B.; Galasso, F. HYperbolic self-paced learning for self-supervised skeleton-based action representations. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  61. Lin, L.; Zhang, J.; Liu, J. Actionlet-dependent contrastive learning for unsupervised skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  62. Hu, J.; Hou, Y.; Guo, Z.; Gao, J. Global and local contrastive learning for self-supervised skeleton-based action recognition. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 10578–10589. [Google Scholar] [CrossRef]
  63. Si, C.; Nie, X.; Wang, W.; Wang, L.; Tan, T.; Feng, J. Adversarial self-supervised learning for semi-supervised 3d action recognition. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  64. Su, Y.; Lin, G.; Wu, Q. Self-supervised 3d skeleton action representation learning with motion consistency and continuity. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  65. Thoker, F.M.; Doughty, H.; Snoek, C.G. Skeleton-contrastive 3D action representation learning. In Proceedings of the ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021. [Google Scholar]
  66. Chen, Y.; Zhao, L.; Yuan, J.; Tian, Y.; Xia, Z.; Geng, S.; Han, L.; Metaxas, D.N. Hierarchically self-supervised transformer for human skeleton representation learning. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
Figure 1. The proposed STH-MAE scheme for self-supervised skeleton-based action classification is depicted in the overall pipeline. At the top, the STH masking process is illustrated. Two-dimensional heatmaps are extracted from video frames and reformulated into an STH, which is then divided into patches that are randomly masked. The result is a set of K × K patches, each containing a 1D vector of length T H W / K . The T H W / K calculation is derived from flattening the spatial–temporal ( T × H × W ) dimensions. At the bottom, the autoencoder pre-training is shown. The encoder processes only the visible tokens, while mask tokens are added to the input sequence of the decoder to reconstruct the masked patches.
Figure 1. The proposed STH-MAE scheme for self-supervised skeleton-based action classification is depicted in the overall pipeline. At the top, the STH masking process is illustrated. Two-dimensional heatmaps are extracted from video frames and reformulated into an STH, which is then divided into patches that are randomly masked. The result is a set of K × K patches, each containing a 1D vector of length T H W / K . The T H W / K calculation is derived from flattening the spatial–temporal ( T × H × W ) dimensions. At the bottom, the autoencoder pre-training is shown. The encoder processes only the visible tokens, while mask tokens are added to the input sequence of the decoder to reconstruct the masked patches.
Sensors 25 03146 g001
Figure 2. Heatmap structure for a single frame.
Figure 2. Heatmap structure for a single frame.
Sensors 25 03146 g002
Table 1. Ablation study on masking ratio.
Table 1. Ablation study on masking ratio.
Masking Ratio0.550.750.95
Top182.8484.3178.65
Top596.0898.3395.24
Bold numbers indicate the best results in each task.
Table 2. Ablation study on heatmap resolution.
Table 2. Ablation study on heatmap resolution.
Resolution4 × 48 × 816 × 1632 × 3248 × 48
Top182.6884.3181.9878.4271.03
Top597.9798.3397.8596.9794.60
Bold numbers indicate the best results in each task.
Table 3. Ablation study on decoder depth.
Table 3. Ablation study on decoder depth.
Depth4812
Top182.9084.3183.47
Top598.2398.3398.19
Bold numbers indicate the best results in each task.
Table 4. Ablation study on pre-training epoch.
Table 4. Ablation study on pre-training epoch.
Pre-Training Epoch15101520
Top170.0179.8283.2284.2184.31
Top594.4597.5398.1598.3698.33
Bold numbers indicate the best results in each task.
Table 5. Linear evaluation results with state-of-the-art methods on NTU 60 and NTU 120 datasets.
Table 5. Linear evaluation results with state-of-the-art methods on NTU 60 and NTU 120 datasets.
MethodYearNTU 60NTU120
CSCVX-SubX-Set
LongT GAN [16]AAAI1839.148.1--
MS2L [24]ACM MM2052.6---
AS-CAL [56]IS2158.564.848.649.2
P&C [17]CVPR2050.776.342.741.7
SeBiReNet [57]ECCV20-79.7--
SkeletonCLR [27]CVPR2175.079.860.762.6
Colorization [26]ICCV2175.283.1--
CrossSCLR [27]CVPR2177.883.467.967.1
AimCLR [28]AAAI2278.983.868.268.8
CMD [58]ECCV2279.886.970.371.5
PSTL [59]AAAI2379.183.869.270.3
HiCRL [29]AAAI2380.485.570.070.4
HaLP [20]CVPR2379.786.871.172.2
HYSP [60]ICLR2378.282.661.864.6
ActCLR [61]CVPR2380.986.769.070.5
skeleton-logoCLR [62]TSCVT2482.487.272.873.5
STH-MAE (Ours)-84.387.074.375.6
Bold numbers indicate the best results in each task.
Table 6. Fine-tuned evaluation results with state-of-the-art methods on NTU 60 and NTU 120 datasets.
Table 6. Fine-tuned evaluation results with state-of-the-art methods on NTU 60 and NTU 120 datasets.
MethodYearNTU 60NTU120
CSCVX-SubX-Set
SkeletonCLR [27]CVPR2182.288.973.675.3
AimCLR [28]AAAI2283.089.276.476.7
SkeletonMAE [43]ICMEW2386.692.976.879.1
HYSP [60]ICLR2386.593.581.482.0
ActCLR [61]CVPR2385.891.279.480.9
skeleton-logoCLR [62]TSCVT2486.193.679.280.0
STH-MAE (Ours)-89.894.980.183.5
Bold numbers indicate the best results in each task.
Table 7. Semi-supervised evaluation results with state-of-the-art methods on NTU 60 dataset.
Table 7. Semi-supervised evaluation results with state-of-the-art methods on NTU 60 dataset.
MethodYear1% Data10% Data
CSCVCSCV
LongT GAN [16]AAAI 201835.2-62.0-
ASSL [63]ECCV 2020--64.369.8
MS2L [24]ACM MM 202031.1-65.2-
MCC [64]ICCV 2021--60.865.8
Skeleton-Contrastive [65]ACM MM 202135.738.165.972.5
Hi-TRS [66]ECCV 202239.142.970.774.8
STH-MAE (Ours)-39.839.474.577.2
Bold numbers indicate the best results in each task.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bian, C.; Yang, Y.; Wang, T.; Lu, W. Spatial–Temporal Heatmap Masked Autoencoder for Skeleton-Based Action Recognition. Sensors 2025, 25, 3146. https://doi.org/10.3390/s25103146

AMA Style

Bian C, Yang Y, Wang T, Lu W. Spatial–Temporal Heatmap Masked Autoencoder for Skeleton-Based Action Recognition. Sensors. 2025; 25(10):3146. https://doi.org/10.3390/s25103146

Chicago/Turabian Style

Bian, Cunling, Yang Yang, Tao Wang, and Weigang Lu. 2025. "Spatial–Temporal Heatmap Masked Autoencoder for Skeleton-Based Action Recognition" Sensors 25, no. 10: 3146. https://doi.org/10.3390/s25103146

APA Style

Bian, C., Yang, Y., Wang, T., & Lu, W. (2025). Spatial–Temporal Heatmap Masked Autoencoder for Skeleton-Based Action Recognition. Sensors, 25(10), 3146. https://doi.org/10.3390/s25103146

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop