Next Article in Journal
Dynamic Task Scheduling Based on Greedy and Deep Reinforcement Learning Algorithms for Cloud–Edge Collaboration in Smart Buildings
Previous Article in Journal
Image Captioning Model Based on Multi-Step Cross-Attention Cross-Modal Alignment and External Commonsense Knowledge Augmentation
Previous Article in Special Issue
Lightweight Network for Spoof Fingerprint Detection by Attention-Aggregated Receptive Field-Wise Feature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TransMODAL: A Dual-Stream Transformer with Adaptive Co-Attention for Efficient Human Action Recognition

1
Department of Computer Engineering, Faculty of Engineering, Ayatollah Boroujerdi University, Boroujerd 69199-69737, Iran
2
Department of Computer and System Sciences, Stockholm University, 10691 Stockholm, Sweden
3
School of Computing, University of Georgia, Athens, GA 30602, USA
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(16), 3326; https://doi.org/10.3390/electronics14163326
Submission received: 28 July 2025 / Revised: 17 August 2025 / Accepted: 19 August 2025 / Published: 21 August 2025

Abstract

Human Action Recognition has seen significant advances through transformer-based architectures, yet achieving a nuanced understanding often requires fusing multiple data modalities. Standard models relying solely on RGB video can struggle with actions defined by subtle motion cues rather than appearance. This paper introduces TransMODAL, a novel dual-stream transformer that synergistically fuses spatiotemporal appearance features from a pre-trained VideoMAE(Video Masked AutoEncoders) backbone with explicit skeletal kinematics from a state-of-the-art pose estimation pipeline (RT-DETR(Real-Time DEtection Transformer) + ViTPose++). We propose two key architectural innovations to enable effective and efficient fusion: a CoAttentionFusion module that facilitates deep, iterative cross-modal feature exchange between the RGB and pose streams, and an efficient AdaptiveSelector mechanism that dynamically prunes less informative spatiotemporal tokens to reduce computational overhead. Evaluated on three challenging benchmarks, TransMODAL demonstrates robust generalization, achieving accuracies of 98.5% on KTH, 96.9% on UCF101, and 84.2% on HMDB51. These results significantly outperform a strong VideoMAE-only baseline and are competitive with state-of-the-art methods, demonstrating the profound impact of explicit pose guidance. TransMODAL presents a powerful and efficient paradigm for composing pre-trained foundation models to tackle complex video understanding tasks by providing a fully reproducible implementation and strong benchmark results.

1. Introduction

Human action recognition (HAR) has undergone a significant paradigm shift, evolving from early methods based on handcrafted features to deep learning models capable of learning hierarchical representations directly from data [1,2,3]. More recent methods leverage deep architectures and hybrid models for spatiotemporal representation learning. For instance, deeper two-stream architectures were developed to better integrate spatial and temporal cues, achieving improved performance in surveillance video settings [4], and refined spatiotemporal fusion strategies have further enhanced feature integration [5]. Additionally, hybrid models combining Conv-RBM(Convolution Restricted Boltzmann Machine) with LSTM(Long Short-Term Memory) and optimized frame selection have demonstrated the flexibility and effectiveness of temporal modeling techniques in HAR [6]. A recent work demonstrated that augmenting 3D skeleton inputs with spatial–temporal transformations significantly improves HAR robustness and model generalization [7]. Three-Dimensional Convolutional Neural Networks (3D-CNNs) such as the Inflated 3D ConvNet (I3D) [8] and factorized variants like R(2+1)D [9] set the standard by extending the success of 2D-CNNs into the temporal domain. More recently, vision transformers (ViT) have revolutionized the field, with models like VideoMAE [10] demonstrating that a self-supervised pre-training objective of masked autoencoding can produce powerful, data-efficient learners for video.
Despite these advances, a fundamental “modality gap” persists. Models relying solely on RGB video excel at capturing appearance, texture, and scene context but can be confounded by actions with subtle motion or those performed by individuals with similar appearances. For example, distinguishing “jogging” from “running” may depend more on the cadence and extension of limbs than on the person’s visual appearance or background. Conversely, skeletal pose data, which represents the body’s kinematics as a set of keypoint coordinates, offers an explicit, viewpoint-invariant representation of motion but lacks the rich contextual information present in RGB frames [11]. This suggests that the optimal approach lies in the effective fusion of these complementary modalities.
This realization has spurred a recent wave of research into pose-guided action recognition. Methods such as Baradel et al. 2018’s Pose-driven Attention for RGB [12] and Song et al. 2020’s Modality Compensation Network [13] leverage pose information to guide a model’s attention toward semantically salient spatiotemporal regions, thereby improving performance in fine-grained and complex action categories. Building on this momentum, our work addresses the critical challenge of how best to fuse appearance and pose information within a modern transformer architecture while maintaining computational efficiency. The core of our work is not just the development of a new model, but a demonstration of a compositional methodology: we leverage powerful, off-the-shelf foundation models—VideoMAE for appearance, RT-DETR for person detection [14], and ViTPose++ for pose estimation [15]—and focus our innovation on the crucial fusion and efficiency-enhancing layers that connect them. This pragmatic approach reflects a modern research paradigm that intelligently combines existing capabilities to solve new challenges.
This paper presents TransMODAL, a dual-stream architecture designed for effective and efficient pose-guided human action recognition. Our primary contributions are twofold. First, we propose a Novel Dual-Stream Architecture (TransMODAL), a compositional, end-to-end trainable model that synergistically fuses two powerful and complementary data streams. The first stream leverages a pre-trained VideoMAE backbone to extract rich, contextual appearance features from RGB video. The second, a dedicated PoseEncoder stream, processes explicit kinematic information from a sequence of 2D skeletal keypoints. This pose data is generated via a state-of-the-art pre-processing pipeline composed of the RT-DETR model for robust person detection and the ViTPose++ model for accurate keypoint estimation. By integrating these specialized foundation models, our architecture focuses its innovation directly on the crucial task of multi-modal fusion. Second, we introduce Adaptive Multi-Modal Fusion and Pruning, which consists of two novel modules that form the technical core of our architecture. The first, CoAttentionFusion, facilitates a deep, iterative dialogue between the two data streams. Through a symmetric cross-attention mechanism, the appearance stream queries the pose stream for kinematic context, while the pose stream simultaneously queries the appearance stream for visual evidence. This allows each modality to enrich its representation with context from the other. The second module, AdaptiveSelector, addresses the challenge of computational efficiency inherent in transformer models. It is a lightweight module with a learnable scoring mechanism that intelligently identifies and prunes redundant spatiotemporal tokens from the fused representation. This significantly reduces computational overhead and inference latency without compromising classification accuracy.

2. Related Work

Our work is situated at the confluence of three major trends in video understanding: spatiotemporal convolutional networks, transformer-based video models, and pose-guided recognition systems.

2.1. Spatiotemporal Convolutions for Action Recognition

Spatial–temporal fusion networks combining TSN and Bi-LSTM have been effectively used for HAR in dynamic rescue scenarios [16]. For several years, 3D-CNNs were the dominant architecture for HAR. A seminal work in this area is the Inflated 3D ConvNet (I3D) by Carreira and Zisserman 2017, which “inflates” 2D filters into 3D to process video volumes and bootstrap weights from ImageNet [8]. This work also introduced the large-scale Kinetics dataset, which is quickly becoming the de facto standard for pre-training action recognition models [17]. I3D achieved state-of-the-art performance on benchmarks like UCF-101 with Kinetics pre-training, reaching 97.9% Top-1 accuracy [8]. Building on the success of I3D but aiming for greater computational efficiency, [9] proposed R(2+1)D, which factorizes spatiotemporal convolutions. Instead of a full 3D convolution (e.g., with a kernel of size t × k × k), R(2+1)D performs a 2D spatial convolution (kernel 1 × k × k) followed by a 1D temporal convolution (kernel t × 1 × 1). This decomposition has two main benefits: it increases the number of nonlinearities for a fixed number of parameters, enhances representational capacity, and simplifies optimization. R(2+1)D also established strong baseline performance, achieving 97.3% Top-1 accuracy on UCF-101 with Kinetics pre-training. These CNN-based models represent the established “old guard” against which new architectures are often measured [9]. However, their reliance on fixed convolutional kernels inherently limits their ability to capture long-range spatiotemporal dependencies, a key challenge that transformer-based models are designed to overcome.

2.2. Transformer-Based Video Understanding

More recently, transformer architectures have emerged as the new state of the art. Our model’s appearance stream is built directly upon VideoMAE (Masked Autoencoders for Video) [10]. Inspired by its counterparts in natural language processing (BERT: Bidirectional Encoder Representations from Transformers) and image processing (MAE), VideoMAE is pre-trained using a self-supervised objective: a very high percentage (e.g., 90–95%) of spatiotemporal patches in a video are masked, and the model must reconstruct the missing pixels [10]. A key finding of the VideoMAE work is that these models are remarkably data-efficient learners; they can achieve high performance on downstream tasks even when pre-trained on relatively small datasets (a few thousand videos) without any external supervised data. This property makes VideoMAE an ideal backbone for our work, providing powerful generic features that can be effectively adapted. The original VideoMAE paper reported strong results of 91.3% on UCF101 and 62.6% on HMDB51 without extra data [10]. Other influential video transformers, such as TimeSformer by Bertasius et al. 2021 [18] and ViViT by Arnab et al. 2021 [19], have also demonstrated the power of attention mechanisms for modeling long-range spatiotemporal dependencies in video. While powerful, these models typically operate on RGB input alone, leaving them vulnerable to ambiguity in actions defined by subtle motion cues rather than distinct visual appearance, which highlights the need for multi-modal integration.

2.3. Pose-Guided Action Recognition

The third critical area of related work involves leveraging human pose to improve action recognition. The rationale is that pose provides a compact, high-level representation of human motion robust to variations in clothing, background clutter, and viewpoint [20]. AutoML-based skeleton pipelines have shown promise, such as ‘single-path one-shot’ neural architecture search optimized for depth-camera data [21]. While early works explored fusing pose with CNN features, recent efforts have focused on integrating pose within transformer frameworks, an active and promising research direction. However, the optimal method for fusing appearance and pose information to maximize performance while maintaining computational efficiency remains a critical and open question. For instance, the Pose-Guided Video Transformer (PGVT) uses 2D body joint coordinates to explicitly guide the spatial attention mechanism of a vision transformer, forcing the model to focus on pose-relevant appearance features [22]. Other approaches have explored multi-task learning, where action and pose are predicted simultaneously [23], or have used pose features as positional embedding for video tokens [24].
Our work contributes to this burgeoning area by proposing CoAttentionFusion, a novel and symmetric mechanism for iterative feature exchange between modalities, distinguishing it from prior fusion strategies. TransMODAL thus sits at the intersection of these three research thrusts: it leverages the powerful representations of VideoMAE, incorporates the efficiency lessons from models like R(2+1)D (via our AdaptiveSelector), and introduces a new fusion technique to the field of pose-guided recognition.

3. The TransMODAL Architecture

The TransMODAL architecture is fundamentally based on the transformer paradigm, chosen for its unique strengths in processing sequential and high-dimensional data like video. From a theoretical perspective, the core self-attention mechanism of transformers is highly suitable for HAR as it can model long-range spatiotemporal dependencies without the inductive biases of convolutional networks. This allows the model to capture complex relationships between body parts and their movements over extended time periods. From a technical standpoint, the multi-head attention mechanism provides a powerful and flexible framework for multi-modal fusion. Our novel CoAttentionFusion module leverages this by using the query–key–value (QKV) formulation to allow the pose and RGB streams to iteratively query one another, enabling a deep and synergistic feature integration that is central to our model’s performance.
The proposed architecture is a dual-stream network that processes and fuses RGB video and 2D pose sequences for robust action recognition. The overall pipeline, depicted in Figure 1, is a multi-stage process that leverages pre-trained models for initial feature extraction and introduces novel modules for efficient fusion and classification. The process begins with person detection and pose estimation to generate two parallel input streams. The RGB stream, consisting of person-centric video clips, is encoded by a frozen VideoMAE backbone [10]. The pose stream, a sequence of 2D skeletal coordinates, is processed by our lightweight PoseEncoder [20]. The core of our contribution lies in the subsequent stages, where the two feature streams are deeply fused by the CoAttentionFusion module, intelligently pruned by the AdaptiveSelector [25], and finally passed to a linear classifier to predict the action [8,9]. The following subsections detail each of these components.

3.1. Overall Pipeline

  • Person Detection and Tracking: We first employ an off-the-shelf, pre-trained RT-DETR model for each video to detect all persons in every frame [14]. A simple tracker associates detections across frames, yielding a set of person-centric video clips.
  • Pose Estimation: A pre-trained ViTPose++ model is used for each person clip to estimate a sequence of 2D human poses [15]. As specified in our data loader, we extract 17 keypoints corresponding to the COCO format for each frame. This results in a tensor of skeletal coordinates.
  • Dual-Stream Encoding: The pipeline then splits into two parallel streams. The cropped RGB person clip is fed into a frozen VideoMAE backbone, while the corresponding sequence of pose coordinates is passed to our lightweight PoseEncoder.
  • Fusion, Selection, and Classification: Visual and pose tokens undergo iterative cross-modal refinement via CoAttentionFusion (inspired by STAR-Transformer’s zigzag attention [26] and MM-ViT’s modality factorization [27]) and are then pruned by AdaptiveSelector using lightweight learnable scoring (echoing DynamicViT’s token sparsification [25]). The remaining tokens are averaged and classified.

Upstream Model Selection and Hyperparameter Justification

The selection of upstream models for person detection and pose estimation was guided by the principle of using state-of-the-art, publicly available components to ensure high-quality inputs for our novel fusion architecture. RT-DETR [14] was chosen for its excellent balance of high accuracy and real-time performance, surpassing many models in the YOLO (You Only Look Once) family in efficiency [28]. Similarly, ViTPose++ [15] represents a top-performing transformer-based approach for pose estimation. Crucially, our pipeline is modular; these components could be replaced with other high-performing models (e.g., different YOLO versions for detection or models like OpenPose for estimation [29]) without altering the core TransMODAL architecture, allowing for future improvements as new upstream models become available.
The key hyperparameters were selected to align with established standards from these foundational models and to accommodate hardware constraints. The input resolution ( H = 224 ,   W = 224 ), clip length ( T = 16 ), and embedding dimension ( D = 768 ) are standard for the ViT-Base architecture used in VideoMAE [10]. The number of tokens ( N = 196 ) is a direct result of dividing a 224 × 224 image into 16 × 16 patches. The number of keypoints ( J = 17 ) corresponds to the standard COCO format output by ViTPose++. Finally, the batch size ( B = 4 ) was the maximum feasible value given the memory constraints of the NVIDIA Tesla T4 GPUs (manufactured by NVIDIA Corporation, Santa Clara, CA, USA) provided through the Kaggle cloud platform, which were used for training.

3.2. Input Modalities and Encoders

3.2.1. RGB Appearance Stream

The primary visual feature extractor is a pre-trained VideoMAE model (MCG-NJU/videomae-base-finetuned-kinetics) with a vision transformer (ViT-Base) architecture. The model processes input clips of c l i p _ l e n = 16 frames. The ViT backbone, with an embedding dimension of D = 768 , converts the video clip into a sequence of spatiotemporal patch embeddings. The parameters of this backbone are frozen during training to leverage its powerful, pre-learned features and to maintain training efficiency.

3.2.2. Pose Kinematics Stream (PoseEncoder)

This stream encodes explicit skeletal motion into tokens that are aligned in dimension with the RGB stream. The input is a tensor P R B × T × J × 2 , where B is the batch size (we use B = 4 in Section 4.2), T is the clip length (default T = 16 ), J is the number of joints ( J = 17 ) , and the last dimension contains the ( x , y ) image coordinates for each joint per frame. For each frame t , we take the joint matrix P t R J × 2 , flatten it to p t = vec P t R 2 J , and project it into the model’s embedding space R D (shared with the RGB tokens). The pose token e t is obtained by (1):
e t = W pose p t + b pose
where W pose R D × 2 J and b pose R D are learnable parameters, and D denotes the embedding dimension (default D = 768 ). Stacking { e t } t = 1 T yields the sequence E R T × D . We then add learnable temporal position embeddings E temp R T × D to encode frame order, producing H 0 = E + E temp .
To capture temporal dynamics, H 0 is processed by a single transformer encoder layer comprising multi-head self-attention and a position-wise feed-forward network, each wrapped with residual connections and LayerNorm. This yields the pose token sequence X pose R T × D . Using the same embedding dimension D as the RGB stream enables the token-wise fusion described in Section 3.3 without additional projection. The temporal attention operates over T frames (here T = 16 ), making its computational cost modest relative to the visual backbone, while providing a dedicated pathway for modeling pose dynamics.

3.3. Core Fusion and Selection Modules

The primary innovations of TransMODAL lie in how it fuses the two modalities and maintains computational efficiency.

3.3.1. Co-Attention Fusion

This module performs deep, bidirectional cross-modal exchange so that the pose stream can query appearance context and, symmetrically, the appearance stream can query kinematic cues (see Figure 2a). Let the appearance tokens be A R N a × D (obtained by flattening T frames and P patches per frame so that N a = T × P ) and the pose tokens be P R N p × D (typically N p = T , one token per frame). We apply dual cross-attention as in (2) and (3).
P ~ = softmax P W Q A W K T d k A W V
A ~ = softmax A W Q P W K T d k P W V
where W Q , W K , W V R D × d k and W Q , W K , W V R D × d k are learnable projections for the two directions, and d k is the per-head key/query dimension (for h heads, d k = D ). We implement (2) and (3) with multi-head cross-attention, followed by Add&Norm residuals and position-wise feed-forward networks (FFN-Feed Forward Network) as shown in (4)–(6):
P + = LN P + MHA Q = P , K = A , V = A
A + = LN A + MHA Q = A , K = P , V = P
P = LN P + + FFN P + , A = LN A + + FFN A +
The block can be stacked L times to yield progressively fused representations. In practice, we concatenate A and P along the token axis (or equivalently inject P into the frame dimension of A ) and linearly project back to R D , producing fused tokens Z R N z × D that are consumed by the AdaptiveSelector described in Section 3.3.2. This design preserves the token dimensionality D shared with the RGB and pose streams, while allowing each modality to attend to the other at the token level, rather than relying on coarse late fusion.

3.3.2. Adaptive Feature Selector (AdaptiveSelector)

To curb the quadratic cost of downstream self-attention, the AdaptiveSelector, depicted in Figure 2b, prunes the fused tokens in two lightweight, learnable stages. Let the fused representation be reshaped to F R T × P × D for a single clip, where T is the number of frames, P the number of patch tokens per frame, and D the embedding dimension.
  • Stage 1—Frame Selection: For each frame t , we first aggregate its patch tokens with a permutation-invariant pooling operator (we use the mean over P tokens) and score the frame with a linear unit using (7):
    s t = w f pool F t , : 1 P p = 1 P F t , p , : + b f
    where w f R D and b f R are learnable parameters. We then select the index set T of the t o p k f frames (default k f = 8 ) by sorting { s t } t = 1 T .
  • Stage 2—Token Selection within selected frames: Within each high-salience frame t T , we compute a token-wise score with a second linear unit using (8):
    r t , p = w p F t , p , : + b p
    where w p R D and b p R are learnable. For each t T , we retain the t o p k t tokens according to r t , p (default k t = 12 ), yielding a fixed-size subset X sel R k f × k t × D . This two-stage selection reduces the effective token budget from T × P to k f × k t without introducing heavy computation in the scoring path. With our defaults ( T = 16 , P = 14 × 14 = 196 , k f = 8 , k t = 12 ), the token count drops from 3136 to 96 (≈33× reduction), while the learned scorers w f , b f , w p , b p are trained end-to-end via the classification loss. The selected tokens are then pooled and fed to the final classifier.

3.4. Implementation and Complexity

The architectural specifics and computational characteristics of the trainable components of TransMODAL are summarized in Table 1. The VideoMAE backbone parameters are frozen and thus do not contribute to the trainable parameter count.

4. Experimental Evaluation

We conducted a series of experiments to validate the effectiveness of the TransMODAL architecture, compare its performance against relevant baselines, and analyze the contribution of its core components through ablation studies.

4.1. Datasets and Protocol

Our primary evaluation benchmark was the KTH Action Recognition Dataset [30]. It contains 600 video clips across six action classes: boxing, handclapping, handwaving, jogging, running, and walking. These actions are performed by 25 subjects in four distinct scenarios. The videos are recorded at 25 fps. Following the standard evaluation protocol for this dataset, we performed a subject-specific split, which ensures 100% reproducibility: subjects 1–16 were used for training, subject 17 for validation, and subjects 18–25 for testing [30]. To test generalization, we used the UCF-101 dataset [31]. This is a more challenging “in-the-wild” dataset collected from YouTube, comprising 13,320 videos across 101 action classes, also at 25 fps. We also evaluated the HMDB51 dataset [32], which consists of 6766 clips from 51 action categories, sourced primarily from movies and web videos. It is known for its challenging camera motion and lower video quality, making it a robust test of model performance.

4.2. Implementation Details

The proposed model was implemented using the PyTorch version 2.5.1. deep learning framework. The experiments were conducted in a high-performance computing environment featuring two NVIDIA Tesla T4 GPUs, each equipped with 15,360 MiB of memory and running on CUDA version 12.2. The system was optimized for deep learning tasks, ensuring efficient GPU memory utilization, with both GPUs initialized with 0 MiB memory usage. This setup provided the computational resources required for training and inference on large-scale video datasets in a reasonable timeframe. We used the AdamW optimizer (torch.optim.AdamW) with a base learning rate of 1 × 10−4 and a weight decay 0.05 [33]. A cosine annealing schedule (torch.optim.lr_scheduler.CosineAnnealingLR) was used to decay the learning rate (SGDR: Stochastic Gradient Descent with Warm Restarts [34]) over the course of training. Models were trained for 100 epochs with a batch size of 4. For data processing, input videos were uniformly sampled to form clips of 16 frames (clip_len = 16) with a sampling stride of 2 (frame_rate = 2), ensuring consistent temporal coverage across all samples.

4.3. Main Results and Baselines

We compared TransMODAL against established CNN-based, transformer-based, and hybrid models, using the performance metrics reported in their original publications on the standard test splits. To demonstrate the generalizability of our approach, we report results across all three datasets. Table 2 shows our results on KTH, while Table 3 and Table 4 provide a comparative analysis against state-of-the-art methods on UCF101 and HMDB51, respectively.
The results on KTH clearly demonstrate the value of the proposed dual-modal fusion. Our full TransMODAL model achieves a Top-1 accuracy of 98.5%, which is a significant improvement of 1.7 percentage points over the strong unimodal baseline (TransMODAL—No Pose). This confirms that explicitly modeling skeletal kinematics provides complementary information that is crucial for disambiguating actions on this dataset. To further validate our method’s effectiveness, we evaluated it on the more diverse UCF101 and HMDB51 datasets, achieving strong Top-1 accuracies of 96.9% and 84.2%, respectively.

4.4. Ablation Studies

To dissect the architecture and validate our design choices on our primary benchmark, we conducted a series of ablation studies on the KTH validation set. The results are summarized in Table 5.
The ablation results yield several key findings:
  • Pose is Critical: Removing the pose stream entirely (Row 1) causes the largest drop in accuracy (−1.7%), confirming it as the most impactful component for performance.
  • AdaptiveSelector is Efficient: Removing the AdaptiveSelector (Row 2) results in a marginal 0.3% drop in accuracy but increases latency by over 46% (from 35.2 ms to 51.5 ms). This demonstrates that our token pruning strategy is highly effective at reducing computational cost with a negligible impact on performance.
  • top_k_frames Affects Sensitivity: Varying top_k_frames shows a clear trade-off. Reducing it to 4 (Row 3) harms accuracy, suggesting that important temporal information is lost. Increasing it to 12 (Row 4) provides no significant benefit over our default of 8, validating our hyperparameter choice.

4.5. Qualitative Analysis

To gain a deeper, qualitative understanding of the model’s behavior, we analyzed its predictions and failure modes. The confusion matrix in Figure 3 shows excellent performance and high discriminative power for the TransMODAL model on the KTH test set, achieving an overall accuracy of 98.50%. The matrix is strongly diagonal, indicating that the vast majority of samples for each class are correctly classified. Classes with highly distinct motion patterns, such as boxing, handclapping, and handwaving, are recognized with near-perfect precision (1.00, 0.99, and 0.99) and recall (1.00, 0.99, and 0.99). The most frequent confusion, as is common for this dataset, occurs between jogging and running due to their high visual and kinematic similarity. Even so, the model effectively distinguishes them, misclassifying only 3% of jogging instances as running and maintaining high F1-scores for both classes (0.97 and 0.97). This demonstrates that even when actions lie on a motion continuum, the dual-modal fusion of appearance and explicit pose data provides sufficient information for robust classification.
To further investigate the model’s behavior, particularly concerning the confusion between kinematically similar actions, we provide qualitative examples from the KTH test set in Figure 4. The top panel (a) shows a sequence correctly classified as “jogging.” In this example, the ViTPose++ pipeline produces a clean and consistent pose track, where the gait and limb movements are clearly defined. This stable pose sequence provides a strong, unambiguous signal to the PoseEncoder, allowing the model to make a confident and accurate prediction. Conversely, the bottom panel (b) illustrates a common failure case where a “running” sequence is misclassified as “jogging.” While the actions are visually similar, the pose estimations in this clip appear less stable, especially in the later frames where motion blur is more pronounced. The slight degradation in keypoint accuracy, combined with the inherent kinematic similarity between a slow run and a fast jog, likely introduces enough ambiguity for the model to err on the side of the more common “jogging” class. This highlights the model’s sensitivity to the quality of the upstream pose estimation and underscores the challenge of distinguishing actions that lie on a motion continuum.
To further probe the model’s internal reasoning and identify which spatiotemporal regions contribute most to its predictions, we visualized attention heatmaps for several action classes. As shown in Figure 5, these visualizations reveal that TransMODAL has successfully learned to focus on semantically relevant body parts and track their motion over time. For “handclapping” (a), the model’s attention is consistently concentrated on the hands and upper torso, correctly identifying the area of action. In the “boxing” sequence (b), the heatmaps highlight the upper body, with the most intense focus tracking the fists and arms as they perform the punching motion. For the full-body action of “running” (c), the model’s attention is appropriately distributed across the legs and torso, capturing the dynamics of the running gait. These heatmaps provide strong evidence that the dual-modal fusion mechanism effectively guides the model to learn class-discriminative and spatiotemporally relevant features, rather than relying on spurious background correlations.

5. Analysis and Discussion

The experimental results provide strong evidence for the effectiveness of the TransMODAL architecture. The quantitative lift of 1.7% in Top-1 accuracy from adding the pose stream on KTH (Table 2) underscores the value of multi-modal fusion. While the RGB stream captures appearance, the pose stream provides an explicit, disentangled representation of the human form’s dynamics, which is critical for distinguishing kinematically similar actions like “jogging” and “running.”
The results on UCF101 and HMDB51 (Table 3 and Table 4) demonstrate the model’s strong generalization capabilities. On UCF101, TransMODAL’s 96.9% accuracy is competitive, though it does not surpass two-stream models like I3D (97.9%) that utilize computationally expensive optical flow. However, our model significantly outperforms the RGB-only VideoMAE baseline (91.3%), showing the clear benefit of pose fusion over a single modality. Most notably, on the challenging HMDB51 dataset, TransMODAL achieves a highly competitive accuracy of 84.2%. While this does not surpass the larger, state-of-the-art VideoMAE V2-g model (88.7%), it is important to note that our model significantly outperforms the standard VideoMAE baseline (62.6%) and other strong methods like PERF-Net (83.2%). This demonstrates that for datasets with high intra-class variation and potential video quality issues, the explicit structural guidance from pose is particularly beneficial, allowing a more compact model like TransMODAL to achieve performance in the same tier as much larger, state-of-the-art architectures.
The ablation study on the AdaptiveSelector module (Table 5, Row 2) reveals a crucial insight into the trade-off between performance and efficiency. Removing the selector and processing all tokens results in a marginal 0.3% drop in accuracy (from 98.5% to 98.2%) but causes a significant 46% increase in inference latency (from 35.2 ms to 51.5 ms). This demonstrates that a large portion of the spatiotemporal tokens are redundant for the final classification task. The AdaptiveSelector’s lightweight, learnable scoring mechanism provides an effective method for identifying and pruning these less informative tokens, making the model substantially more efficient with a negligible impact on its predictive power.
To further contextualize the efficiency of our approach, Table 6 provides a direct comparison against prominent two-stream models that also leverage motion information (via optical flow). TransMODAL achieves a competitive accuracy on UCF101 while requiring significantly fewer trainable parameters and a lower computational load (FLOPs) than the classic I3D model. This highlights the effectiveness of using a lightweight pose stream as an efficient alternative to computationally expensive optical flow for capturing motion dynamics.
Our error analysis, informed by the high overall accuracy shown in the confusion matrix (Figure 3), indicates that the model’s primary challenge lies in distinguishing between actions with very similar kinematics. The qualitative example in Figure 4 provides a clear illustration of this limitation: a “running” sequence is misclassified as “jogging.” The attention heatmaps in Figure 5c confirm that the model is correctly focusing on the legs and torso, which are the most discriminative regions for this action. However, the failure case in Figure 4b suggests that even when the model attends to the correct body parts, severe motion blur can degrade the quality of the underlying keypoint estimates from the ViTPose++ pipeline. This introduces ambiguity, causing the model to default to the kinematically similar “jogging” class. This “garbage-in, garbage-out” phenomenon, where noisy input from an upstream module limits performance, points to the main weakness of a two-stage approach. A promising direction for future work is therefore the exploration of end-to-end trainable models, where the pose estimator is fine-tuned alongside the action recognition head to make it more robust to such real-world video artifacts.

6. Conclusions and Future Work

This paper introduced TransMODAL, a dual-stream transformer architecture for human action recognition that effectively fuses RGB appearance features with skeletal pose kinematics. We demonstrated that it is possible to build a highly performant and efficient system by composing powerful pre-trained foundation models (VideoMAE, RT-DETR, ViTPose++) and focusing innovation on the fusion and efficiency layers. Our key contributions, the CoAttentionFusion and AdaptiveSelector modules, enable deep cross-modal feature exchange while dynamically pruning redundant tokens to reduce computational cost. TransMODAL demonstrates strong performance and generalization across multiple benchmarks, achieving 98.5% on KTH, 96.9% on UCF101, and a competitive 84.2% on HMDB51. These results validate our design choices and establish a new, reproducible benchmark for efficient dual-modal action recognition.
Our work opens several avenues for future research. First, while our results are strong, evaluation on even-larger-scale datasets such as Kinetics 400 [17] and Something Something V2 [39] is needed to test its scalability fully. Second, exploring more-sophisticated fusion mechanisms and pose encoders—for instance, adopting Graph Convolutional Networks (GCNs) to better model the spatial relationships between joints [40]—could yield further improvements. Third, incorporating complementary modalities (e.g., sensor-based data) may enhance robustness; for example, recent studies using smartphone accelerometer and gyroscope signals processed via ensemble learning suggest valuable insights for multimodal HAR [41]. Finally, and perhaps most importantly, transitioning from the current two-stage pipeline to an end-to-end trainable system is a key future direction. Jointly fine-tuning the pose-estimation backbone with the action-recognition head could improve robustness to noisy inputs, such as the motion blur shown in Figure 4, and potentially lead to a more holistic spatiotemporal representation.

Author Contributions

Conceptualization, M.J.; methodology, M.J.; software, M.J. and M.I.; validation, M.J., M.I. and H.R.A.; formal analysis, M.J. and M.I.; investigation, M.J.; resources, M.J.; data curation, M.I.; writing—original draft preparation, M.J.; writing—review and editing, M.J. and H.R.A.; visualization, M.I.; supervision, H.R.A.; project administration, M.J.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original data presented in the study is openly available: KTH: https://www.csc.kth.se/cvap/actions/ (accessed on 31 January 2025), UCF101: https://www.crcv.ucf.edu/research/data-sets/ucf101/ (accessed on 30 March 2025), and HMDB51: https://serre-lab.clps.brown.edu/resource/hmdb-a-large-human-motion-database/ (accessed on 22 May 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dong, Y.; Zhou, R.; Zhu, C.; Cao, L.; Li, X. Hierarchical activity recognition based on belief functions theory in body sensor networks. IEEE Sens. J. 2022, 22, 15211–15221. [Google Scholar] [CrossRef]
  2. Joudaki, M.; Ebrahimpour Komleh, H. Introducing a new architecture of deep belief networks for action recognition in videos. JMVIP 2024, 11, 43–58. [Google Scholar]
  3. Teng, Q.; Wang, K.; Zhang, L.; He, J. The layer-wise training convolutional neural networks using local loss for sensor-based human activity recognition. IEEE Sens. J. 2020, 20, 7265–7274. [Google Scholar] [CrossRef]
  4. Han, Y.; Zhang, P.; Zhuo, T.; Huang, W.; Zhang, Y. Going Deeper with Two-Stream ConvNets for Action Recognition in Video Surveillance. Pattern Recognit. Lett. 2018, 107, 83–90. [Google Scholar] [CrossRef]
  5. Abdelbaky, A.; Aly, S. Two-Stream Spatiotemporal Feature Fusion for Human Action Recognition. Vis. Comput. 2021, 37, 1821–1835. [Google Scholar] [CrossRef]
  6. Joudaki, M.; Imani, M.; Arabnia, H.R. A New Efficient Hybrid Technique for Human Action Recognition Using 2D Conv-RBM and LSTM with Optimized Frame Selection. Technologies 2025, 13, 53. [Google Scholar] [CrossRef]
  7. Xin, C.; Kim, S.; Cho, Y.; Park, K.S. Enhancing Human Action Recognition with 3D Skeleton Data: A Comprehensive Study of Deep Learning and Data Augmentation. Electronics 2024, 13, 747. [Google Scholar] [CrossRef]
  8. Carreira, J.; Zisserman, A. Quo vadis, action recognition? A new model and the Kinetics dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6299–6308. [Google Scholar] [CrossRef]
  9. Tran, D.; Wang, H.; Torresani, L.; Ray, J.; LeCun, Y.; Paluri, M. A closer look at spatiotemporal convolutions for action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 6450–6459. [Google Scholar] [CrossRef]
  10. Tong, Z.; Song, Y.; Wang, J.; Wang, L. VideoMAE: Masked autoencoders are data-efficient learners for self-supervised video pre-training. Adv. Neural Inf. Process. Syst. 2022, 35, 10078–10093. [Google Scholar]
  11. Sun, Z.; Ke, Q.; Rahmani, H.; Bennamoun, M.; Wang, G.; Liu, J. Human action recognition from various data modalities: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3200–3225. [Google Scholar] [CrossRef]
  12. Baradel, F.; Wolf, C.; Mille, J. Human activity recognition with pose-driven attention to RGB. In Proceedings of the 29th British Machine Vision Conference (BMVC), Newcastle, UK, 3–6 September 2018; pp. 1–14. [Google Scholar]
  13. Song, S.; Liu, J.; Li, Y.; Guo, Z. Modality compensation network: Cross-modal adaptation for action recognition. IEEE Trans. Image Process. 2020, 29, 3957–3969. [Google Scholar] [CrossRef]
  14. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar] [CrossRef]
  15. Xu, Y.; Zhang, J.; Zhang, Q.; Tao, D. ViTPose++: Vision transformer for generic body pose estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 46, 1212–1230. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Guo, Q.; Du, Z.; Wu, A. Human Action Recognition for Dynamic Scenes of Emergency Rescue Based on Spatial-Temporal Fusion Network. Electronics 2023, 12, 538. [Google Scholar] [CrossRef]
  17. Kay, W.; Carreira, J.; Simonyan, K.; Zhang, B.; Hillier, C.; Vijayanarasimhan, S.; Viola, F.; Green, T.; Back, T.; Natsev, P.; et al. The Kinetics human action video dataset. arXiv 2017, arXiv:1705.06950. [Google Scholar] [CrossRef]
  18. Bertasius, G.; Wang, H.; Torresani, L. Is space-time attention all you need for video understanding? In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 813–824. [Google Scholar]
  19. Arnab, A.; Dehghani, M.; Heigold, G.; Sun, C.; Lučić, M.; Schmid, C. ViViT: A video vision transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 6836–6846. [Google Scholar] [CrossRef]
  20. Fang, H.-S.; Xie, S.; Tai, Y.-W.; Lu, C. RMPE: Regional multi-person pose estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar] [CrossRef]
  21. Jiang, Y.; Yu, S.; Wang, T.; Sun, Z.; Wang, S. Skeleton-Based Human Action Recognition Based on Single Path One-Shot Neural Architecture Search. Electronics 2023, 12, 3156. [Google Scholar] [CrossRef]
  22. Wang, J.; Tan, S.; Zhen, X.; Xu, S.; Zheng, F.; He, Z.; Shao, L. Deep 3D human pose estimation: A review. Comput. Vis. Image Underst. 2021, 210, 103225. [Google Scholar] [CrossRef]
  23. Bevilacqua, A.; MacDonald, K.; Rangarej, A.; Widjaya, V.; Caulfield, B.; Kechadi, T. Human activity recognition with convolutional neural networks. In Proceedings of the European Conference Machine Learning and Knowledge Discovery in Databases, Cham, Switzerland, 10–14 September 2018; pp. 541–552. [Google Scholar] [CrossRef]
  24. Reilly, D.; Chadha, A.; Das, S. Seeing the pose in the pixels: Learning pose-aware representations in vision transformers. arXiv 2023, arXiv:2306.09331. [Google Scholar] [CrossRef]
  25. Rao, Y.; Zhao, W.; Liu, B.; Lu, J.; Zhou, J.; Hsieh, C.J. DynamicViT: Efficient vision transformers with dynamic token sparsification. Adv. Neural Inf. Process. Syst. 2021, 34, 13937–13949. [Google Scholar]
  26. Ahn, D.; Kim, S.; Hong, H.; Ko, B.C. STAR-Transformer: A spatio-temporal cross attention transformer for human action recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 2–7 January 2023; pp. 3330–3339. [Google Scholar] [CrossRef]
  27. Chen, J.; Ho, C.M. MM-ViT: Multi-modal video transformer for compressed video action recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 1910–1921. [Google Scholar] [CrossRef]
  28. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7473. [Google Scholar] [CrossRef]
  29. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.E.; Sheikh, Y. OpenPose: Realtime multi-person 2D pose estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 172–186. [Google Scholar] [CrossRef] [PubMed]
  30. Schuldt, C.; Laptev, I.; Caputo, B. Recognizing human actions: A local SVM approach. In Proceedings of the 17th International Conference on Pattern Recognition (ICPR), Cambridge, UK, 26 August 2004; Volume 3, pp. 32–36. [Google Scholar] [CrossRef]
  31. Soomro, K.; Zamir, A.R.; Shah, M. Ucf101: A dataset of 101 human action classes from videos in the wild. arXiv 2012, arXiv:1212.0402. [Google Scholar] [CrossRef]
  32. Kuehne, H.; Jhuang, H.; Garrote, E.; Poggio, T.; Serre, T. HMDB: A large video database for human motion recognition. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 2556–2563. [Google Scholar] [CrossRef]
  33. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  34. Loshchilov, I.; Hutter, F. SGDR: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
  35. Li, Y.; Wu, C.Y.; Fan, H.; Mangalam, K.; Xiong, B.; Malik, J.; Feichtenhofer, C. MViTv2: Improved multiscale vision transformers for classification and detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4804–4814. [Google Scholar] [CrossRef]
  36. Li, K.; Wang, Y.; He, Y.; Li, Y.; Wang, Y.; Wang, L.; Qiao, Y. Uniformerv2: Spatiotemporal learning by arming image vits with video uniformer. arXiv 2022, arXiv:2211.09552. [Google Scholar]
  37. Li, Y.; Lu, Z.; Xiong, X.; Huang, J. PERF-Net: Pose empowered RGB-Flow Net. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 513–522. [Google Scholar] [CrossRef]
  38. Wang, L.; Huang, B.; Zhao, Z.; Tong, Z.; He, Y.; Wang, Y.; Wang, Y.; Qiao, Y. VideoMAE v2: Scaling video masked autoencoders with dual masking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 14549–14560. [Google Scholar] [CrossRef]
  39. Goyal, R.; Ebrahimi Kahou, S.; Michalski, V.; Materzynska, J.; Westphal, S.; Kim, H.; Haenel, V.; Freund, I.; Yianilos, P.; Mueller-Freitag, M.; et al. The “Something Something” video database for learning and evaluating visual common sense. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5842–5850. [Google Scholar] [CrossRef]
  40. Yan, S.; Xiong, Y.; Lin, D. Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 April 2018; Volume 32. [Google Scholar] [CrossRef]
  41. Tan, T.-H.; Wu, J.-Y.; Liu, S.-H.; Gochoo, M. Human Activity Recognition Using an Ensemble Learning Algorithm with Smartphone Sensor Data. Electronics 2022, 11, 322. [Google Scholar] [CrossRef]
Figure 1. The detailed architecture of the TransMODAL model. The network consists of two parallel streams for RGB and pose data. The outputs are merged via our proposed CoAttentionFusion module, which performs symmetric cross-attention. The subsequent AdaptiveSelector prunes the fused tokens before they are passed to a final classifier.
Figure 1. The detailed architecture of the TransMODAL model. The network consists of two parallel streams for RGB and pose data. The outputs are merged via our proposed CoAttentionFusion module, which performs symmetric cross-attention. The subsequent AdaptiveSelector prunes the fused tokens before they are passed to a final classifier.
Electronics 14 03326 g001
Figure 2. Detailed architecture of the proposed (a) CoAttentionFusion and (b) AdaptiveSelector modules. The fusion module uses two parallel cross-attention paths to refine the visual and pose features, which are then combined. The selector module uses a two-stage, learnable scoring process to prune the fused tokens first by frame and then by token, reducing the data dimensionality while preserving the most salient information.
Figure 2. Detailed architecture of the proposed (a) CoAttentionFusion and (b) AdaptiveSelector modules. The fusion module uses two parallel cross-attention paths to refine the visual and pose features, which are then combined. The selector module uses a two-stage, learnable scoring process to prune the fused tokens first by frame and then by token, reducing the data dimensionality while preserving the most salient information.
Electronics 14 03326 g002
Figure 3. Confusion matrix for TransMODAL on the KTH test set, with an overall accuracy of 98.50%. The strong diagonal indicates high classification accuracy across all classes. The model achieves perfect or near-perfect precision and recall for all classes.
Figure 3. Confusion matrix for TransMODAL on the KTH test set, with an overall accuracy of 98.50%. The strong diagonal indicates high classification accuracy across all classes. The model achieves perfect or near-perfect precision and recall for all classes.
Electronics 14 03326 g003
Figure 4. Qualitative examples from the KTH test set. (a) A “jogging” sequence that is correctly classified. The pose estimations are stable and provide a clear kinematic signal. (b) A failure case where a “running” sequence is misclassified as “jogging.” The kinematic similarity, potentially exacerbated by minor inaccuracies in the pose track due to motion blur, leads to the incorrect prediction.
Figure 4. Qualitative examples from the KTH test set. (a) A “jogging” sequence that is correctly classified. The pose estimations are stable and provide a clear kinematic signal. (b) A failure case where a “running” sequence is misclassified as “jogging.” The kinematic similarity, potentially exacerbated by minor inaccuracies in the pose track due to motion blur, leads to the incorrect prediction.
Electronics 14 03326 g004
Figure 5. Visualization of attention heatmaps over time for three action classes from the KTH dataset. The model demonstrates a clear ability to focus on the most discriminative body regions for each action: (a) the hands and torso for “handclapping,” (b) the fists and upper body for “boxing,” and (c) the legs and torso for “running.” This confirms that the model has learned to ground its predictions in semantically relevant motion patterns.
Figure 5. Visualization of attention heatmaps over time for three action classes from the KTH dataset. The model demonstrates a clear ability to focus on the most discriminative body regions for each action: (a) the hands and torso for “handclapping,” (b) the fists and upper body for “boxing,” and (c) the legs and torso for “running.” This confirms that the model has learned to ground its predictions in semantically relevant motion patterns.
Electronics 14 03326 g005
Table 1. Implementation and complexity details for the TransMODAL architecture. Parameter and FLOP counts are calculated for the trainable components of the model.
Table 1. Implementation and complexity details for the TransMODAL architecture. Parameter and FLOP counts are calculated for the trainable components of the model.
Module Name (Code)Key HyperparametersParameters (M)FLOPs (G)
VideoMAE Backboneembed_dim = 76887 (frozen)N/A
PoseEncoderembed_dim = 768, num_joints = 170.220.02
CoAttentionFusionnum_heads = 8, depth = 19.450.81
AdaptiveSelectortop_k_frames = 8, top_k_tokens = 120.05<0.01
ActionClassifiernum_classes = 60.60<0.01
Total Trainable 10.32~0.83
Table 2. Comparative performance on the KTH test set. The “No Pose” result is from our ablation study and represents a strong VideoMAE-only baseline.
Table 2. Comparative performance on the KTH test set. The “No Pose” result is from our ablation study and represents a strong VideoMAE-only baseline.
ModelInput ModalityPre-Trained onKTH Top-1 Acc (%)
Two-stream ConvNets [4]RGB + FlowImageNet93.1
ST-VLAD-PCANet [5]RGB-93.33
2D Conv-RBM + LSTM [6]RGB-97.3
VideoMAE(B/16) [10]RGBKinetics96.5
TransMODAL (Proposed method)—No PoseRGBKinetics96.8
TransMODAL (Proposed method)RGB + PoseKinetics98.5
Table 3. Comparative performance on the UCF101 test set (split 1). Our proposed model is competitive, with strong baselines without requiring optical flow.
Table 3. Comparative performance on the UCF101 test set (split 1). Our proposed model is competitive, with strong baselines without requiring optical flow.
ModelInput ModalityPre-Trained onUCF101 Top-1 Acc (%)
I3D (Two-Stream) [8]RGB + FlowKinetics97.9
R(2+1)D (Two-Stream) [9]RGB+ FlowKinetics97.3
VideoMAE [10]RGBKinetics91.3
MViTv2 [35]RGBKinetics-40098.6
UniFormerV2 [36]RGBKinetics-40098.9
PERF-Net [37]RGB + Flow + PoseS3D-G98.6
TransMODAL (Proposed method)RGB + PoseKinetics96.9
Table 4. Comparative performance on the HMDB51 test set (split 1). Our proposed model achieves highly competitive accuracy, demonstrating strong performance against other state-of-the-art methods.
Table 4. Comparative performance on the HMDB51 test set (split 1). Our proposed model achieves highly competitive accuracy, demonstrating strong performance against other state-of-the-art methods.
ModelInput ModalityPre-Trained onHMDB51 Top-1 Acc (%)
2D Conv-RBM + LSTM [6]RGB-81.5
I3D (Two-Stream) [8]RGB + FlowKinetics80.2
R(2+1)D (Two-Stream) [9]RGB+ FlowKinetics78.7
VideoMAE [10]RGBKinetics62.6
MViTv2 [33]RGBKinetics-40085.5
UniFormerV2 [34]RGBKinetics-40086.1
PERF-Net [37]RGB + Flow + PoseS3D-G83.2
VideoMAE V2-g [38]RGBUnlabeledHybrid88.7
TransMODAL (Proposed method)RGB + PoseKinetics84.2
Table 5. Ablation studies on the KTH validation set. The impact of key architectural components is analyzed by measuring changes in Top-1 accuracy and inference latency. The results validate the significant contributions of both the pose stream and the AdaptiveSelector module to the model’s final performance and efficiency.
Table 5. Ablation studies on the KTH validation set. The impact of key architectural components is analyzed by measuring changes in Top-1 accuracy and inference latency. The results validate the significant contributions of both the pose stream and the AdaptiveSelector module to the model’s final performance and efficiency.
ConfigurationTop-1 Acc (%)Δ vs. Full ModelLatency (ms/batch)
Full TransMODAL Model98.5-35.2
No Pose Stream (VideoMAE-only)96.8−1.729.8
No AdaptiveSelector (use all tokens)98.2−0.351.5
top_k_frames = 497.5−1.031.1
top_k_frames = 1298.4−0.139.4
Table 6. Efficiency and performance comparison on UCF101. TransMODAL achieves competitive accuracy with greater parameter value and computational efficiency compared to traditional flow-based two-stream models. Note: Table 1 reports FLOPs for trainable modules only (VideoMAE frozen), whereas Table 6 reports an end-to-end forward-pass estimate for the full TransMODAL pipeline.
Table 6. Efficiency and performance comparison on UCF101. TransMODAL achieves competitive accuracy with greater parameter value and computational efficiency compared to traditional flow-based two-stream models. Note: Table 1 reports FLOPs for trainable modules only (VideoMAE frozen), whereas Table 6 reports an end-to-end forward-pass estimate for the full TransMODAL pipeline.
ModelInput ModalityTrainable Params (M)FLOPs (G)Top-1 Acc (%)
I3D (Two-Stream) [8]RGB + Flow27.910897.9
R(2+1)D (Two-Stream) [9]RGB + Flow31.415297.3
TransMODAL (Proposed)RGB + Pose10.32~4596.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Joudaki, M.; Imani, M.; Arabnia, H.R. TransMODAL: A Dual-Stream Transformer with Adaptive Co-Attention for Efficient Human Action Recognition. Electronics 2025, 14, 3326. https://doi.org/10.3390/electronics14163326

AMA Style

Joudaki M, Imani M, Arabnia HR. TransMODAL: A Dual-Stream Transformer with Adaptive Co-Attention for Efficient Human Action Recognition. Electronics. 2025; 14(16):3326. https://doi.org/10.3390/electronics14163326

Chicago/Turabian Style

Joudaki, Majid, Mehdi Imani, and Hamid R. Arabnia. 2025. "TransMODAL: A Dual-Stream Transformer with Adaptive Co-Attention for Efficient Human Action Recognition" Electronics 14, no. 16: 3326. https://doi.org/10.3390/electronics14163326

APA Style

Joudaki, M., Imani, M., & Arabnia, H. R. (2025). TransMODAL: A Dual-Stream Transformer with Adaptive Co-Attention for Efficient Human Action Recognition. Electronics, 14(16), 3326. https://doi.org/10.3390/electronics14163326

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop