Next Article in Journal
A Hierarchical Prediction Method for Pedestrian Head Injury in Intelligent Vehicle with Combined Active and Passive Safety System
Previous Article in Journal
Facial Expression Realization of Humanoid Robot Head and Strain-Based Anthropomorphic Evaluation of Robot Facial Expressions
Previous Article in Special Issue
Complex-Exponential-Based Bio-Inspired Neuron Model Implementation in FPGA Using Xilinx System Generator and Vivado Design Suite
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Modal Enhancement Transformer Network for Skeleton-Based Human Interaction Recognition

School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
*
Author to whom correspondence should be addressed.
This work was done as an intern at the Institute of Science on Brain Inspired Intelligence, Chongqing University.
Biomimetics 2024, 9(3), 123; https://doi.org/10.3390/biomimetics9030123
Submission received: 15 January 2024 / Revised: 11 February 2024 / Accepted: 16 February 2024 / Published: 20 February 2024
(This article belongs to the Special Issue Biologically Inspired Vision and Image Processing)

Abstract

:
Skeleton-based human interaction recognition is a challenging task in the field of vision and image processing. Graph Convolutional Networks (GCNs) achieved remarkable performance by modeling the human skeleton as a topology. However, existing GCN-based methods have two problems: (1) Existing frameworks cannot effectively take advantage of the complementary features of different skeletal modalities. There is no information transfer channel between various specific modalities. (2) Limited by the structure of the skeleton topology, it is hard to capture and learn the information about two-person interactions. To solve these problems, inspired by the human visual neural network, we propose a multi-modal enhancement transformer (ME-Former) network for skeleton-based human interaction recognition. ME-Former includes a multi-modal enhancement module (ME) and a context progressive fusion block (CPF). More specifically, each ME module consists of a multi-head cross-modal attention block (MH-CA) and a two-person hypergraph self-attention block (TH-SA), which are responsible for enhancing the skeleton features of a specific modality from other skeletal modalities and modeling spatial dependencies between joints using the specific modality, respectively. In addition, we propose a two-person skeleton topology and a two-person hypergraph representation. The TH-SA block can embed their structural information into the self-attention to better learn two-person interaction. The CPF block is capable of progressively transforming the features of different skeletal modalities from low-level features to higher-order global contexts, making the enhancement process more efficient. Extensive experiments on benchmark NTU-RGB+D 60 and NTU-RGB+D 120 datasets consistently verify the effectiveness of our proposed ME-Former by outperforming state-of-the-art methods.

1. Introduction

Human action recognition technology has been widely used in the fields of bionic robots, medical care, video surveillance, intelligent transportation, and human–robot interaction [1,2,3,4,5,6,7,8]. Skeletons are made up of a variety of different shapes, with complex internal and external structures, which play a role in supporting and protecting the human body. As a biological feature of the human body, the skeleton data can fully represent the kinetic features of human actions. Compared to RGB and depth data, skeleton data are more robust than RGB and depth data to illumination change and viewpoint occlusion. Advances in depth cameras and pose estimation methods [9,10,11,12] have also made it easier for bionic robots to acquire skeleton data.
Previous studies [13,14,15] have confirmed that second-order information about the skeleton plays a complementary role in action recognition. The second-order information about the skeleton, which could also be called the skeletal modality, included joint coordinates, bone vector, joint coordinate motion, and bone vector motion. An early approach to multi-modal fusion strategy, known as back-end fusion, was extremely time-consuming and could not be trained in an end-to-end way. Diverse modalities of skeleton data were input into the model separately to obtain partial results. Then, the partial results were manually fused to obtain the final recognition result. Song et al. [16] proposed a training framework called an input branch network, which allowed the skeleton features of each modality to be trained separately in the branch network, and then concatenated the features from different modalities in a specific layer to continue training using the main stream network. Although this method improved the training efficiency to a certain extent, it had one disadvantage: in the input branch network, there is no information transfer between different skeletal modalities. The absence of information transfer meant that every specific skeletal modality could not learn complementary features from other skeletal modalities. Furthermore, the data distribution of skeleton features after concatenation was considerably different, which hindered the subsequent recognition.
In order to solve the problem of missing information transfer, we utilize the cross-attention mechanism to fuse skeleton features from different modalities. Through successive training iterations, the cross-attention mechanism can progressively enhance the skeleton features of a specific modality by learning the directional pairwise attention between the specific modality and other skeletal modalities. However, the conventional cross-attention mechanism can only handle the features of two modalities. Inspired by the human visual neural network, which consists of multiple subnets [17] where different neural subnets are connected by dendrites in the brain cortex [18], we propose to add a global context subnet to interact with other skeletal modalities. Specifically, our improvement is to build a context progressive fusion block to save and update the global context, and then use the global context to interact with each skeletal modality. The global context is gradually transformed from low-level to high-level with network training. During this process, the global context aggregated sufficient information from different skeletal modalities.
Although Transformer had shown advantages in multi-modal enhancement, it worked poorly with structured data, such as skeleton data. The main reason was that each joint in the skeleton topology contained particular spatial structural information, which is not perceived by the vanilla Transformer. Ying et al. [19] and Zhou et al. [20] proposed two methods for embedding the spatial structural information about the skeleton into the Transformer. They were relative position embedding (RPE) and hypergraph embedding, which embed the pairwise and higher-order relations of skeletal joints into the Transformer, respectively. However, the above methods were used for single-person action recognition and were not suitable for two-person interaction recognition.
To solve the limitations of the existing skeletal topology, we design a simple yet effective two-person skeleton topology and hypergraph. We first consider the skeleton topology of two persons as a whole and then connect the limbs and heads of two persons to obtain the two-person skeleton topology. Then, we use a learnable method to automatically convert the two-person skeleton topology into a two-person hypergraph representation. By embedding these two kinds of spatial structure information into the Transformer, the model can better recognize the interaction information about the two people.
In this paper, we propose a multi-modal enhancement transformer (ME-Former) network, which can enhance different skeletal modalities simultaneously and recognize two-person interaction. ME-Former includes a multi-modal enhancement module (ME) and a context progressive fusion block (CPF). More specifically, each ME module consists of a multi-head cross-modal attention block (MH-CA) and a two-person hypergraph self-attention block (TH-SA). The MH-CA block is responsible for utilizing the global context generated from the CPF block to enhance the skeleton features of specific modalities. The TH-SA block can encode the two-person skeleton topology and two-person hypergraph representation as structural information and embed them into self-attention to better learn two-person interaction. On the other hand, the CPF block performing the information interaction among all the skeletal modalities is proposed to generate and update the global context.
The strategy of our research is as follows: Section 1 introduces the research background of this work and our solutions to existing two problems. Section 2 describes the related work, including skeleton-based action recognition methods and human interaction recognition methods. Section 3 recaps the prior knowledge involved in this work. Section 4 states the network structure and implementation details of the ME-Former. Section 5 evaluates the validity of our proposed ME-Former and compares it with other advanced methods. Section 6 summarizes this paper and predicts the future development direction. The main contributions of our work are as follows:
  • A novel multi-modal enhancement transformer (ME-Former) network is proposed for skeleton-based human interaction recognition, overcoming the challenge of different skeletal modalities not being able to effectively utilize complementary features;
  • We propose a simple yet effective two-person skeleton topology and a two-person hypergraph representation, which can model the pairwise and higher-order relations between human joints, and use the TH-SA block to embed the two kinds of structural information into the ME-Former for better recognizing human interaction;
  • The extensive experimental results highlight the benefits of our multi-modal enhancement transformer and two-person hypergraph representation. Our proposed ME-Former outperforms state-of-the-art methods on two different skeleton-based action recognition benchmarks.

2. Related Work

2.1. Skeleton-Based Action Recognition

With respect to methods for modeling human skeleton data, skeleton-based action recognition methods can be simply categorized into convolutional neural networks (CNNs), recurrent neural networks (RNNs), GCNs, and Transformers. CNNs process the skeleton data as a pseudo-image [21,22,23,24], while RNNs flattens the skeleton data into a sequence [25,26,27,28]. Nevertheless, the above processing methods lose the semantic information about the skeleton data and the higher-order relations of the joints.
GCN-based approaches. GCNs modeled the skeleton data as a graph [29,30], in which the nodes are joints and the edges are bones. It turned out that GCNs were better at modeling the spatial features of skeleton data. Yan et al. [31] proposed ST-GCN which first extracts and aggregates the spatio-temporal information about the skeleton using the topological diagram, whereas ST-GCN used a handcrafted skeleton graph, which cannot characterize the relevance of joints that are not naturally connected, limiting the ability of GCNs to extract and aggregate features. To improve the capacity of GCNs, Shi et al. [13] proposed 2s-AGCN which included the second-order information (the lengths and directions of bones) of the skeleton data in the range of feature extraction, and used two streams to process the information about different modalities and integrate it. The model constructed an adaptive topology map, which can be updated automatically with the backpropagation algorithm of the neural network to characterize the node connection strength better. None of the above models considered that each channel can express richer potential action features until CTR-GCN [32] modeled channel-wise topologies by dividing the channels into groups and only sharing the skeleton topology between groups. CTR-GCN learned a shared topology and channel-specific correlations simultaneously to enhance the feature extraction and aggregation ability of GCNs without significantly increasing the parameters. Lee et al. [33] proposed HD-GCN which decomposes the human skeleton graph into six hierarchical subsets, highlighting the distance relationship between joints. More importantly, it allowed nodes, that are not naturally connected, to be connected in a sub-graph, which is better for transmitting information between the joints. One disadvantage of GCNs was their inability to model long-distance spatio-temporal dependence due to the local nature of graph convolution.
Transformer-based approaches. Transformer was a novel deep-learning model, which gradually appeared in the field of action recognition because of its strong capacity for parallel processing sequence data and modeling long-term dependency. Transformers focused primarily on modeling spatio-temporal correlations and embedding semantic information specific to the skeleton joints. Wang et al. [34] proposed a Transformer-based network (IIP-Transformer) that models the intra-part and inter-part dependencies on the spatial dimension, enabling a more detailed representation of features. Plizzari et al. [35] proposed a novel Spatial–Temporal Transformer network (ST-TR) in which the spatial self-attention module and temporal self-attention module are used to capture the correlation between different nodes in a frame and the dynamic relationship between the same node in the whole frames. Zhang et al. [36] proposed a Spatial-Temporal Specialized Transformer (STST), which is composed of the spatial transformer block and directional temporal transformer block. The spatial transformer block was designed to model the dependency between joints and semantic representation in each frame separately, and the directional temporal transformer block was designed to model the long-term temporal dependency with a direction-aware strategy. Ying et al. [19] proposed a Graphformer to incorporate the joint connection information into attention computation via a distance-based relative positional embedding method. The natural human topology graph contained only two joints on each edge and this pairwise structure limited the representation of higher-order kinematic dependencies. Zhou et al. [20] further proposed a Hypergraph Transformer. The hypergraph contained hyperedges that are capable of connecting two or more vertices. By embedding hypergraph representations into attention calculations, the Transformer could learn high-order relations of skeleton joints.

2.2. Human Interaction Recognition

Human interaction recognition is a sub-task of action recognition. The methods of interaction recognition need to learn both individual information and interactive information between individuals. Ji et al. [37] proposed an interactive body part model. The model split the human body into five body parts and combined the limbs of two participants into eight interacting body part pairs. For each pair, they calculated the joint feature in three frames, called poselets. Yang et al. [38] introduced a pairwise graph that connects the joints of two people in pairs in a predefined way. Both Li et al. [39] and Zhu et al. [40] explicitly modeled the interactive relationship of the two persons into an additional inter-body matrix, and took the intra-body matrix and inter-body matrix in parallel into the graph convolution network to obtain the results. Perez et al. [41] proposed a two-stream interaction relation network called LSTM-IRN which models the internal joint relationship of the same subject and the inter-relationship of different subjects. Furthermore, they defined structured pairwise operations to extract and learn valuable extra information (distance and motion) from each pair of joints. Pang et al. [42] proposed a novel Interaction Graph Transformer (IGFormer) network to achieve skeleton-based interaction recognition by modeling interacting body parts as graphs. More specifically, IGFormer constructed interaction graphs based on semantic and distance correlations between the body parts of interactive persons and aggregated information about interacting body parts by the learned graph. Gao et al. [43] first treated the topology of two persons as a whole and used attention-driven modules to capture interactive correlations dynamically. Li et al. [44] designed several graph labeling strategies to accelerate the convergence of the model. In addition, they integrated the data pre-processing methods and developed a symmetry processing method to reduce the high intra-class variation caused by the order of people.

3. Preliminaries

In this section, we recap the definition of attention mechanism and hypergraph representation.

3.1. Attention Mechanism

The full name of attention is scaled product attention. The function of scaled dot-product attention is to map a query and a set of key-value pairs to an output, where the query (Q), key (K), value (V), and output are all vectors. The attention values are generated by applying the softmax function to the dot product of Q and K and dividing each element by d k . The attention value is multiplied by V to get a weighted output:
Attention ( Q , K , V ) = SoftMax ( Q K T d k ) V .
It is worth noting that we added a multi-head attention mechanism in practice to improve the performance of the Transformer. This divides the channel dimension into groups and applies attention to each group in parallel to efficiently learn distinctive features. The multi-head attention is calculated as follows:
MultiHead ( Q , K , V ) = Concat ( h e a d 1 , h e a d 2 , , h e a d h ) W O ,
h e a d i = Attention ( Q W i Q , K W i K , V W i V ) , i [ 1 , h ] ,
where parameter matrices W O R C × d k , W i Q R h × C h e a d × d q , W i K R h × C h e a d × d k , and W i V R h × C h e a d × d v . h is the number of heads and C h e a d = C / h is the number of channels per head. In this work, we employ h = 8 .

3.2. Hypergraph Representation

In a standard graph, only two vertices can be connected to each edge. The hypergraph is not restricted and the hyperedge can connect more than two vertices. A hypergraph is defined as G = ( V , E , W ) which consists of a vertex set V , a hyperedge set E , and a symmetric matrix W which contains hyperedge weights. The hypergraph G can be denoted by a | V | × | E | incidence matrix H, with entries defined as follows:
h ( v , e ) = 1 i f v e 0 i f v e .
The degree of a vertex v V is defined as d ( v ) = e E w ( e ) h ( v , e ) and the degree of a hyperedge e E is defined as d ( e ) = v V h ( v , e ) . We use diagonal matrices D v and D e , respectively, to represent the degree matrix of vertices and hyperedges, where the elements on the diagonal are the degrees of each vertex or hyperedge.

4. The Proposed Method

In this section, we introduce the whole structure of our multi-modal enhancement transformer network (ME-Former) which is depicted in Figure 1. The ME-Former network consists of two stages: the feature extraction and fusion stage, and the feature refinement stage.
At the feature extraction and fusion stage, the ME-former comprises five parallel data streams. They are four skeletal modalities and the global context. The four skeletal modalities are joint coordinates (J), bone vector (B), joint motion (JM), and bone motion (BM). Each specific modal stream passes through the multi-modal enhancement module (ME) and multi-scale temporal convolution block (MS-TC) in one layer, for a total of four layers. The ME module has two functions: (1) Enhance and complement specific skeletal modalities with the global context. (2) Model long-distance spatial dependencies on each skeletal modality itself. We adopt the MS-TC block from [45] to model the temporal dependence of every specific modality. The global context, which is generated by the context progressive fusion block (CPF), aggregates information about the four skeletal modalities in each layer. The global context is used to enhance each specific modality so that the skeleton features from different modalities are gradually fused together. After four layers of feature extraction and fusion, the four kinds of skeleton features enhanced by the feature extraction and fusion stage are concatenated together to be fed into the feature refinement stage.
At the feature refinement stage, the ME-Former treats the fused features as a unified skeletal modality. The feature refinement stage has six layers; each layer uses TH-SA blocks and MS-TC blocks to model spatial and temporal skeleton features. Finally, the action prediction scores are obtained by feeding skeleton features into the global average pooling layer and the fully connected layer.

4.1. Modality Enhancement Module

The ME module is placed in each layer of the feature extraction and fusion stage. It can not only extract the spatial features of the specific modality but also utilize the global context to gradually enhance the skeleton features of each specific modality. The technical details of the ME module are shown in Figure 2. The inputs to the ME module are the global context information Y G [ i ] and the four skeleton modalities Y * [ i ] , where { J , B , JM , BM } , which act as the source and target modality, respectively. The source modality is used to enhance the target modality; this enhancement process can be denoted as source→target. The final output Y * of the ME module can be expressed as follows:
Y * [ i + 1 ] = ME G [ i ] ( Y G [ i ] , Y * [ i ] ) ,
where { J , B , JM , BM } and i represents the i-th layer of the ME-Former. The ME module does not change the feature shape of skeletal modalities; the size of both the input and output are the same, Y * R N × C × T × J , where N represents batch size, C represents channel number of features, T represents sequence length of the skeleton frames, and J represents the number of human body joints.
Specifically, the ME module consists of a two-person hypergraph self-attention block (TH-SA), which is responsible for modeling the spatial dependence of specific modal skeleton features, and a multi-head cross-modal attention block (MH-CA), which is responsible for enhancing the skeleton features of a specific modality with global context information. The intermediate outputs of the two blocks are computed as follows:
Y * [ i ] = TH-SA ( LN ( Y * [ i ] ) ,
Y G [ i ] = MultiHead ( LN ( Y * [ i ] ) , LN ( Y * [ i ] ) , LN ( Y G [ i ] ) ) ,
where LN represents the layer normalization operation. For a clear description, we first introduce the enhancement process of the specific modality. The detailed implementation of the TH-SA block is left at the end of this section. Then, the intermediate output of the two blocks is processed via the following dynamic filter:
W * [ i ] = Sigmoid ( Conv2d ( Y * [ i ] ) + Conv2d ( Y G [ i ] ) ) ,
Y * [ i ] = W * [ i ] Y G [ i ] + ( 1 W * [ i ] ) Y * [ i ] .
The weight of each branch can be dynamically determined by learnable parameters of the Conv2d function. This operation can filter out features that are not important in the uni-modal enhancement process. Finally, we use a simple convolution layer with residual connection to obtain the output of ME module Y * [ i + 1 ] .

4.2. Context Progressive Fusion Block

The context progressive fusion block (CPF) aggregates the skeleton features of each specific modality in the current layer and transforms them into a global context for cross-modal information interacting with the ME modules of the next layer. This is more efficient than making different skeletal modalities exchange information in pairs. The technical details of the CPF block are shown in Figure 3. The input to the CPF block is the global context Y G [ i ] of the current layer and four skeleton modalities Y * [ i ] , where { J , B , JM , BM } .
They are first converted to the features z J , z B , z JM , z BM , and z G with a range [−1, 1] by the Conv2d and Tanh functions. The magnitude of the feature values of different skeletal modalities also defines their importance. Next, we apply the softmax function to these features to get the weight for each skeletal modality. We then multiply the features of each modality with the corresponding weights and add them together to obtain global context Y G [ i + 1 ] . The calculation process is formulated as follows:
z * = Tanh ( Conv2d ( Y * [ i ] ) ) ,
w * = exp ( z * ) exp ( z J ) + exp ( z B ) + exp ( z JM ) + exp ( z BM ) + exp ( z G ) ,
Y G [ i + 1 ] = Y J [ i ] w J + Y B [ i ] w B + Y J M [ i ] w J M + Y B M [ i ] w B M + Y G [ i ] w G ,
where { J , B , JM , BM } , ⊙ means Hadamard Product, and exp means Exponential Function. This process is a variant of attention, which fuses important features from different skeletal modalities and the global context of the current layer into the global context of the next layer. Finally, we feed global context Y G [ i + 1 ] into a convolution layer with residual connections to obtain the output of the CPF block.

4.3. Two-Person Hypergraph Self-Attention Block

TH-SA blocks exist in both the two stages of the ME-Former network. Its function is to model the spatial dependencies of skeleton features. To better model spatial features, TH-SA embeds two kinds of skeleton structure information based on vanilla SA, namely two-person skeleton topology and two-person hypergraph representation. They indicate pairwise relations and higher-order relations between skeleton joints, respectively. The pipeline of the TH-SA block is shown in Figure 4.
Conventional separate single-person skeleton topologies cannot capture interaction information between two persons, so we design a simple yet effective two-person skeleton topology. The difference between two separate single-person skeleton topologies and our designed two-person skeleton topologies is that we consider the skeleton topology of two persons as a whole and then connect the limbs and heads of the two persons. Skeleton features that belong to two persons can interact through these additional connections by the skeleton topology embedding method. Their visualization is shown in Figure 5a,b.
The attention value A consists of three components, which are vanilla SA, skeleton topology embedding, and two-person hypergraph (2p-Hypergraph) embedding. Given the input Q = K = V = Y * [ i ] , where { J , B , JM , BM } , the attention value is calculated as follows:
A = Q K T + Q R k T + Q E h y p e r T .
Skeleton topology embedding and 2p-hypergraph embedding are realized by multiplying structural information R k and 2p-hypergraph features E h y p e r with skeleton features, respectively. The detailed procedure for obtaining R k and E h y p e r is in Section 4.3.1 and Section 4.3.2. Both embedding methods are designed to make the transformer aware of the specific spatial semantics of each human joint. By matrix multiplication of the attention value A with V, the output of TH-SA can be obtained as Y * [ i ] .
Y * [ i ] = s o f t m a x ( A d k ) V .

4.3.1. Encoding Structural Information about Two-Person Skeleton Topology

The joints of the human body are naturally connected by bones, forming a bio-mechanical model. The movement of each joint in an action has a chain effect on the other joints. Therefore, it is necessary to embed the structural information about the human body into each position of skeleton features.
We refer to the existing relative positional embedding (RPE) method and propose a K-hop RPE method suitable for skeleton data. We first obtain the set P R J × J using the Shortest Path Distance (SPD) algorithm [19], where p i j indicates the distance or number of hops between joint i and joint j. Then, we define a learnable parameter W k R k × C to represent the weights corresponding to different distances (max hop is k). Finally, each element in set P is assigned the corresponding weight in W k to obtain the structural information about skeleton topology R k R J × J × C .
R k = i = 0 J j = 0 J W k [ p i , j ] .

4.3.2. Deriving Two-Person Hypergraph Feature

The definition of hypergraph has been explained in Section 3.2. We set the degree of each vertex d ( v ) = 1 for efficient realization. In this particular case, the human joints can be divided into | E | discrete subsets and the incidence matrix H is equivalent to a partition matrix. Each row is a one-hot vector denoting the group to which each joint belongs. So how to formulate a partition strategy to obtain hypergraph representation is crucial.
We demonstrate two common skeleton topology partition strategies based on body prior knowledge in Figure 5c,d. The body parts partition strategy divides the skeletal topology into four limbs and a trunk, while the upper and lower partition strategy divides the skeletal topology into upper and lower halves. Although the above partitioning strategies can reflect the relations between multiple joints to a certain extent, these manual methods are not the optimal solution and have poor generalization. Unlike these, in the TH-SA block, we use a learnable partition matrix that can automatically transform a two-person skeleton topology into a two-person hypergraph representation.
Although the binary incidence matrix H R | V | × | E | can be regarded as a partition matrix, its values are discrete. We parameterize it to make values continuous by applying a softmax function along the column axis:
H ˜ = h ˜ v e = exp ( h v e ) e = 1 | E | exp ( h v e ) : v = 1 | V | , e = 1 . . . | E | ,
Continuous partition matrix H ˜ can be optimized jointly with other Transformer parameters. At the end of the optimization, we re-convert the values of the partition matrix from continuous to discrete by applying the argmax function along each row of H ˜ :
H ¯ = argmax ( H ˜ ) ,
The visualization of the learned partition matrix H ¯ is shown in Figure 5e.
Then, we utilize the H ¯ to get a two-person hypergraph representation and embed it into the skeleton features. Given the partition matrix H ¯ R | V | × | E | and skeleton joint features X R | V | × C , the representation of hyperedge set E R | E | × C can be calculated by the following formula:
E = D e 1 H ¯ T X W e ,
where the inverse degree matrix D e of hyperedges plays the role of normalization and W e is a learnable parameter matrix. Finally, we assign the weight of each hyperedge to the corresponding subsidiary joints to get the two-person hypergraph feature E h y p e r R | V | × C :
E h y p e r = H D e 1 H ¯ T X W e .

5. Experiment

5.1. Datasets

NTU-RGB+D 60 (NTU-60) [46] is a large-scale dataset for RGB+D human action recognition with more than 56 thousand video samples and 4 million frames, collected from 40 distinct subjects. It contains 60 different action classes including daily, mutual, and health-related actions. There are two standard evaluation protocols for this dataset including the cross-subject (X-Sub) protocol in which training data comes from 20 subjects and testing data comes from the other 20 subjects, and the cross-view (X-View) protocol where two cameras are used for training and the third one is used for testing. NTU-60 dataset contains 11 two-person interaction classes involving ‘punch/slap’, ‘pat on the back’, ‘giving object’, ‘walking towards’, ‘kicking’, ‘point finger’, ‘touch pocket’, ‘walking apart’, ‘pushing’, ‘hugging’, and ‘handshaking’; we denote this part of the dataset as NTU-60* in the following experiments.
NTU-RGB+D120 (NTU-120) [47] is a large 3D joint dataset of human movements. This dataset expands the NTU-RGB+D dataset by adding 57,367 skeletal sequences and 60 additional action categories. This dataset was captured with three cameras, with a total of 113,945 samples, by 106 volunteers. The dataset was also partitioned between the training and testing sets using two different partitions. (1) Cross-subject (X-Sub): the training data is from 53 subjects and the test set is from the remaining subjects. (2) Cross-setup (X-Setup): half of the setups are used for training and the remaining ones are used for testing. This dataset contains 15 additional human interaction classes including ‘hit with object’, ‘wield knife’, ‘knock over’, ‘grab stuff’, ‘shoot with gun’, ‘step on foot’, ‘high-five’, ‘cheers and drink’, ‘carry object’, ‘take a photo’, ‘follow’, ‘whisper’, ‘exchange things’, ‘support somebody’, and ‘rock–paper–scissors’, resulting a total of 26 interaction classes; we denote this part of the dataset as NTU-120* in the following experiments.

5.2. Implementation Details

We set the batch size to 16 and resize the frame number of a single sample to 64; the other data pre-processing follows [44,48]. Our model has ten layers and the number of channels in each layer is 192. The attention mechanism uses eight heads. ME-Former is trained by Stochastic Gradient Descent (SGD) with momentum set to 0.9, and weight decay set to 0.0005. The total number of training epochs is 140 with the first 5 epochs to warm up [49]. Our base learning rate (LR) is 0.1, the LR decay rate is 0.1, and the decay step is [110, 120]. We set standard cross-entropy as a loss function. All experiments are conducted on the Pytorch deep-learning framework with two NVIDIA GeForce RTX 3090 GPUs and 64G RAM.

5.3. Comparison with State-of-the-Art Methods

We compare the recognition accuracy of our proposed ME-Former with other state-of-the-art methods on the NTU-60* and NTU-120* datasets. We categorize the methods into three types and rank them by publication time and accuracy. The results are shown in Table 1.
The first category includes RNN-based methods for human interaction recognition. Although RNN-based methods can model the interactive information about two persons to a certain extent, it is difficult to capture the skeleton’s pairwise connection information and semantic information with them. Under the X-Sub and X-View protocols of the NTU-60*, the recognition accuracy of our proposed method ME-Former is 4.87 % and 4.10 % higher than the latest LSTM-IRN model, respectively.
The second category includes GCN-based methods for human interaction recognition. GCN-based methods mainly capture interactive information by modeling additional topologies and are generally better than the former methods. Their common disadvantage is poor generalization and requires a lot of extra computational overhead. At the same time, the skeleton topology cannot properly represent the connection information about the various skeletal modalities that are gradually merging. Under the X-Sub protocols of the NTU-60*, the recognition accuracy of our model is 0.88 % higher than the EfficientGCN-B1 and 0.03 % higher than the 2s-AIGCN, respectively. Under the X-Sub protocols of the NTU-120*, the recognition accuracy of our model is 0.20 % higher than the EfficientGCN-B1 and 0.13 % higher than the 2s-AIGCN, respectively.
The third category includes Transformer-based methods for human interaction recognition. Transformer can model long-distance dependency, which has advantages over GCNs that can only extract local features. Transformer-based methods are lightweight and easier for multi-modal fusion. After embedding the spatial structural information about the skeleton, Transformer-based methods gradually became mainstream. Under the X-Sub and X-View protocols of the NTU-120*, the recognition accuracy of our model is 5.44 % and 4.83 % higher than the IGFormer, and 0.56 % and 0.20 % higher than the STSA-Net, respectively.
To further demonstrate the performance of our approach, we present the normalized confusion matrix of the EfficientGCN-B1 model (baseline) and our proposed ME-Former model on the NTU-60* (X-Sub) benchmark in Figure 6. The confusion matrix contains eleven action classes, and we can observe that our model outperforms the EfficientGCN model on almost all action classes except the actions ‘hugging’ and ‘pat on the back’. The most significant improvements in accuracy are ‘pocket touches’ and ‘kicking’, which improved 2.18 % and 2.9 % compared to baseline, respectively. The most difficult actions to recognize in the NTU-60* (X-Sub) dataset are ‘punch/slap’ and ‘giving object’, and we believe that better recognition of two-person interactions requires more detailed modeling of hand gestures. Benefiting our proposed methods, the ME-Former model has outperformed state-of-the-art RNN-based, GCN-based, and Transformer-based methods in human interaction recognition.

5.4. Ablation Studies

In this section, we conduct extensive ablation studies on the NTU-60* dataset using the cross-subject (X-Sub) protocol to validate the effectiveness of the proposed multi-modal enforcement transformer network (ME-Former). To prove the validity of our proposed two-person skeleton topology and two-person hypergraph partition strategy, we further conduct detailed ablation experiments on the components in the TH-SA block. The visualizations of two skeleton topologies and three partition strategies are shown in Figure 5.
We set the EfficientGCN [16] as our baseline, which consists of two stages: the feature extraction and fusion stage, and the feature refinement stage. EfficientGCN uses spatial graph convolution (SGC) blocks to model the spatial features of the skeleton data in both stages. Since there are four skeletal modalities (four input branches), four independent SGC blocks are required in the feature extraction and fusion stage. It is worth noting that both our ME-Former and the baseline use the same multi-scale temporal convolution block (MS-TC) for modeling temporal features of the skeleton data, so we omit it for clarity in the ablation studies.

5.4.1. Effectiveness of Multi-Modal Enforcement Transformer

ME-Former adopts a similar framework to the baseline model; the difference is that ME-Former uses the TH-SA blocks to model the spatial features of the skeleton data in both parts. ME-Former adds two additional units to the input branch, enhancing each single modality by interchanging information with the other skeletal modalities. They are the multi-modal enforcement module (ME) and the context progressive fusion block (CPF). The ME module comprises a multi-head cross-modal attention block (MH-CA) and a two-person hypergraph self-attention block (TH-SA). We will gradually replace and add our methods to the baseline model to demonstrate the effectiveness of our proposed methods. The overall ablation results of the multi-modal enforcement transformer network (ME-Former) are shown in Table 2.
  • We first replace the SGC block in the baseline model with four TH-SA blocks to test the effect of the TH-SA block. We can observe that the recognition accuracy is improved by 0.49 % but the number of parameters in the model is also greatly increased. This is because the baseline is a lightweight model with a lot of fine-tuning in the model structure.
  • Next, we add MH-CA blocks to each input branch of the feature extraction and fusion stage, so that MH-CA and TH-SA are integrated into the ME modules. The recognition accuracy improves by 0.28 % , which strongly demonstrates the significance of cross-modal information transfer between different input branches. Features of the other skeletal modalities can indeed enhance the skeleton features of each single modality.
  • Finally, we add the CPF block to the feature extraction and fusion stage, and the result shows that the accuracy improves by 0.11 % . This proves that using the global context can better aggregate the features among the various skeleton modalities and facilitate the enhancement of each single modality.
The results of all the ablation experiments prove that our proposed multi-modal enhancement transformer network is advanced and its recognition accuracy is 0.90 % higher than the baseline model.

5.4.2. Effectiveness of Two-Person Hypergraph Self-Attention Block

To eliminate distractions, we test the effects of each component of the TH-SA block when both the two stages of ME-Former only have the TH-SA block. As shown in Table 3, the results prove the effects of the two extra components in the TH-SA block. They are skeleton topology embedding and two-person hypergraph embedding.
  • We first compare the effect between single-person skeleton topology embedding and two-person skeleton topology embedding, showing that the performance of the latter is 0.21 % higher than the former. This result proves that additional connections in the two-person skeleton topology facilitate learning interactive information between two persons. Adding two-person skeleton topology embedding improves accuracy by 1.18 % compared to vanilla self-attention (SA).
  • Then we add two-person hypergraph embedding to the vanilla SA and the results show that they improve the accuracy by 0.70 % . The effect of two-person hypergraph embedding is slightly better than that of skeleton topology embedding, indicating that the higher-order relationship between joints can better reflect the difference between different actions than the pairwise relationship.
The full TH-SA is 1.61 % higher than the vanilla SA while the number of parameters does not increase significantly. This proves that it is necessary to embed skeleton-specific connection information into the transformer.
In addition, we verify the influence of different partitioning strategies for skeleton topology in Table 4 and the results demonstrate the superiority of the two-person hypergraph. Our proposed two-person hypergraph partition strategy is 0.45 % and 0.34 % higher than the upper and lower body partition strategy, and body parts partition strategy, respectively. The two-person hypergraph partition strategy that can be learned automatically not only has higher recognition performance but also has stronger generalization.

6. Conclusions

In this work, we propose a multi-modal enhancement transformer network (ME-Former) for skeleton-based human interaction recognition. The ME-Former framework consists of two main components: the multi-modal enhancement module (ME) and the context progressive fusion block (CPF). The ME module is composed of a multi-head cross-modal attention block (MH-CA) and a two-person hypergraph self-attention block (TH-SA). These blocks have distinct roles in enhancing specific skeletal modalities and capturing long spatial dependencies between joints using a specific modality, respectively. On the other hand, the CPF block allows for information transfer between multiple skeletal modalities. The CPF block aggregates the skeleton features of different modalities and transforms them into a global context for multi-modal enhancement. With our proposed methods, ME-Former not only can efficiently fuse information from different skeletal modalities but also better recognize two-person interactions. Extensive experiment results show that our model has advanced performance on the NTU-60 and NTU-120 datasets. We believe that information transfer between different modalities can be more efficient and recommend further research into the feature fusion process. We expect that our proposed multi-modal enhancement transformer network will be applied in bionic robots and human-robot interaction interfaces.

Author Contributions

Conceptualization, Q.H. and H.L.; methodology, Q.H.; software, Q.H.; validation, Q.H. and H.L; writing—original draft preparation, Q.H.; writing—review and editing, H.L.; supervision, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The datasets used in this paper are available online.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhao, S.; Zhao, G.; He, Y.; Diao, Z.; He, Z.; Cui, Y.; Jiang, L.; Shen, Y.; Cheng, C. Biomimetic Adaptive Pure Pursuit Control for Robot Path Tracking Inspired by Natural Motion Constraints. Biomimetics 2024, 9, 41. [Google Scholar] [CrossRef]
  2. Kwon, J.Y.; Ju, D.Y. Living Lab-Based Service Interaction Design for a Companion Robot for Seniors in South Korea. Biomimetics 2023, 8, 609. [Google Scholar] [CrossRef]
  3. Song, F.; Li, P. YOLOv5-MS: Real-time multi-surveillance pedestrian target detection model for smart cities. Biomimetics 2023, 8, 480. [Google Scholar] [CrossRef]
  4. Liu, M.; Meng, F.; Liang, Y. Generalized Pose Decoupled Network for Unsupervised 3d Skeleton Sequence-based Action Representation Learning. Cyborg Bionic Syst. 2022, 2022, 0002. [Google Scholar] [CrossRef]
  5. Zhang, Y.; Xu, X.; Zhao, Y.; Wen, Y.; Tang, Z.; Liu, M. Facial Prior Guided Micro-Expression Generation. IEEE Trans. Image Process. 2024, 33, 525–540. [Google Scholar] [CrossRef]
  6. Huang, Z.; Du, C.; Wang, C.; Sun, Q.; Xu, Y.; Shao, L.; Yu, B.; Ma, G.; Kong, X. Bionic Design and Optimization on the Flow Channel of a Legged Robot Joint Hydraulic Drive Unit Based on Additive Manufacturing. Biomimetics 2023, 9, 13. [Google Scholar] [CrossRef]
  7. Wang, X.; Gao, Y.; Ma, X.; Li, W.; Yang, W. A Bionic Venus Flytrap Soft Microrobot Driven by Multiphysics for Intelligent Transportation. Biomimetics 2023, 8, 429. [Google Scholar] [CrossRef]
  8. Zhang, Q.; Liu, Q.; Duan, J.; Qin, J. Research on Teleoperated Virtual Reality Human–Robot Five-Dimensional Collaboration System. Biomimetics 2023, 8, 605. [Google Scholar] [CrossRef] [PubMed]
  9. Bultmann, S.; Memmesheimer, R.; Behnke, S. External Camera-based Mobile Robot Pose Estimation for Collaborative Perception with Smart Edge Sensors. arXiv 2023, arXiv:2303.03797. [Google Scholar]
  10. Chun, S.; Park, S.; Chang, J.Y. Learnable human mesh triangulation for 3d human pose and shape estimation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 2850–2859. [Google Scholar]
  11. Chun, S.; Park, S.; Chang, J.Y. Representation learning of vertex heatmaps for 3D human mesh reconstruction from multi-view images. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023; pp. 670–674. [Google Scholar]
  12. Xue, S.; Gao, S.; Tan, M.; He, Z.; He, L. How does color constancy affect target recognition and instance segmentation? In Proceedings of the 29th ACM International Conference on Multimedia, Virtual, 20–24 October 2021; pp. 5537–5545. [Google Scholar]
  13. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12026–12035. [Google Scholar]
  14. Mehmood, F.; Zhao, H.; Chen, E.; Guo, X.; Albinali, A.A.; Razzaq, A. Extended Multi-Stream Adaptive Graph Convolutional Networks (EMS-AAGCN) for Skeleton-Based Human Action Recognition. Res. Sq. 2022; preprint. [Google Scholar] [CrossRef]
  15. Trivedi, N.; Sarvadevabhatla, R.K. Psumnet: Unified modality part streams are all you need for efficient pose-based action recognition. In Computer Vision—ECCV 2022 Workshops, Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 211–227. [Google Scholar]
  16. Song, Y.F.; Zhang, Z.; Shan, C.; Wang, L. Constructing stronger and faster baselines for skeleton-based action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1474–1488. [Google Scholar] [CrossRef]
  17. Wilms, M.; Eickhoff, S.B.; Hömke, L.; Rottschy, C.; Kujovic, M.; Amunts, K.; Fink, G.R. Comparison of functional and cytoarchitectonic maps of human visual areas V1, V2, V3d, V3v, and V4 (v). Neuroimage 2010, 49, 1171–1179. [Google Scholar] [CrossRef] [PubMed]
  18. Milošević, N.T.; Ristanović, D. Fractality of dendritic arborization of spinal cord neurons. Neurosci. Lett. 2006, 396, 172–176. [Google Scholar] [CrossRef]
  19. Ying, C.; Cai, T.; Luo, S.; Zheng, S.; Ke, G.; He, D.; Shen, Y.; Liu, T.Y. Do transformers really perform badly for graph representation? Adv. Neural Inf. Process. Syst. 2021, 34, 28877–28888. [Google Scholar]
  20. Zhou, Y.; Li, C.; Cheng, Z.Q.; Geng, Y.; Xie, X.; Keuper, M. Hypergraph transformer for skeleton-based action recognition. arXiv 2022, arXiv:2211.09590. [Google Scholar]
  21. Li, B.; Dai, Y.; Cheng, X.; Chen, H.; Lin, Y.; He, M. Skeleton based action recognition using translation-scale invariant image mapping and multi-scale deep CNN. In Proceedings of the IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China, 10–14 July 2017; pp. 601–604. [Google Scholar]
  22. Soo Kim, T.; Reiter, A. Interpretable 3d human action analysis with temporal convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 20–28. [Google Scholar]
  23. Li, C.; Zhong, Q.; Xie, D.; Pu, S. Co-occurrence feature learning from skeleton data for action recognition and detection with hierarchical aggregation. arXiv 2018, arXiv:1804.06055. [Google Scholar]
  24. Tas, Y.; Koniusz, P. Cnn-based action recognition and supervised domain adaptation on 3d body skeletons via kernel feature maps. arXiv 2018, arXiv:1806.09078. [Google Scholar]
  25. Liu, J.; Shahroudy, A.; Xu, D.; Wang, G. Spatio-temporal lstm with trust gates for 3d human action recognition. In Computer Vision—ECCV 2016, Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 816–833. [Google Scholar]
  26. Du, Y.; Wang, W.; Wang, L. Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1110–1118. [Google Scholar]
  27. Song, S.; Lan, C.; Xing, J.; Zeng, W.; Liu, J. An end-to-end spatio-temporal attention model for human action recognition from skeleton data. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  28. Zhang, P.; Lan, C.; Xing, J.; Zeng, W.; Xue, J.; Zheng, N. View adaptive recurrent neural networks for high performance human action recognition from skeleton data. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2117–2126. [Google Scholar]
  29. Liu, J.; Wang, X.; Wang, C.; Gao, Y.; Liu, M. Temporal Decoupling Graph Convolutional Network for Skeleton-based Gesture Recognition. IEEE Trans. Multimed. 2023, 26, 811–823. [Google Scholar] [CrossRef]
  30. Wang, X.; Zhang, W.; Wang, C.; Gao, Y.; Liu, M. Dynamic Dense Graph Convolutional Network for Skeleton-based Human Motion Prediction. IEEE Trans. Image Process. 2024, 33, 1–15. [Google Scholar] [CrossRef]
  31. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  32. Chen, Y.; Zhang, Z.; Yuan, C.; Li, B.; Deng, Y.; Hu, W. Channel-wise topology refinement graph convolution for skeleton-based action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 13359–13368. [Google Scholar]
  33. Lee, J.; Lee, M.; Lee, D.; Lee, S. Hierarchically Decomposed Graph Convolutional Networks for Skeleton-Based Action Recognition. arXiv 2022, arXiv:2208.10741. [Google Scholar]
  34. Wang, Q.; Peng, J.; Shi, S.; Liu, T.; He, J.; Weng, R. Iip-transformer: Intra-inter-part transformer for skeleton-based action recognition. arXiv 2021, arXiv:2110.13385. [Google Scholar]
  35. Plizzari, C.; Cannici, M.; Matteucci, M. Spatial temporal transformer network for skeleton-based action recognition. In Proceedings of the Pattern Recognition. ICPR International Workshops and Challenges, Virtual, 10–15 January 2021; Springer: Cham, Switzerland, 2021; pp. 694–701. [Google Scholar]
  36. Zhang, Y.; Wu, B.; Li, W.; Duan, L.; Gan, C. STST: Spatial-temporal specialized transformer for skeleton-based action recognition. In Proceedings of the the 29th ACM International Conference on Multimedia, Virtual, 20–24 October 2021; pp. 3229–3237. [Google Scholar]
  37. Ji, Y.; Ye, G.; Cheng, H. Interactive body part contrast mining for human interaction recognition. In Proceedings of the IEEE International Conference on Multimedia and Expo Workshops (ICMEW), Chengdu, China, 14–18 July 2014; pp. 1–6. [Google Scholar]
  38. Yang, C.L.; Setyoko, A.; Tampubolon, H.; Hua, K.L. Pairwise adjacency matrix on spatial temporal graph convolution network for skeleton-based two-person interaction recognition. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 2166–2170. [Google Scholar]
  39. Li, J.; Xie, X.; Cao, Y.; Pan, Q.; Zhao, Z.; Shi, G. Knowledge embedded gcn for skeleton-based two-person interaction recognition. Neurocomputing 2021, 444, 338–348. [Google Scholar] [CrossRef]
  40. Zhu, L.; Wan, B.; Li, C.; Tian, G.; Hou, Y.; Yuan, K. Dyadic relational graph convolutional networks for skeleton-based human interaction recognition. Pattern Recognit. 2021, 115, 107920. [Google Scholar] [CrossRef]
  41. Perez, M.; Liu, J.; Kot, A.C. Interaction relational network for mutual action recognition. IEEE Trans. Multimed. 2021, 24, 366–376. [Google Scholar] [CrossRef]
  42. Pang, Y.; Ke, Q.; Rahmani, H.; Bailey, J.; Liu, J. IGFormer: Interaction Graph Transformer for Skeleton-based Human Interaction Recognition. In Computer Vision—ECCV 2022, Proceedings of the 17th European Conference, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 605–622. [Google Scholar]
  43. Gao, F.; Xia, H.; Tang, Z. Attention Interactive Graph Convolutional Network for Skeleton-Based Human Interaction Recognition. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; pp. 1–6. [Google Scholar]
  44. Li, Z.; Li, Y.; Tang, L.; Zhang, T.; Su, J. Two-person Graph Convolutional Network for Skeleton-based Human Interaction Recognition. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 3333–3342. [Google Scholar] [CrossRef]
  45. Liu, Z.; Zhang, H.; Chen, Z.; Wang, Z.; Ouyang, W. Disentangling and unifying graph convolutions for skeleton-based action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 143–152. [Google Scholar]
  46. Shahroudy, A.; Liu, J.; Ng, T.T.; Wang, G. Ntu rgb+ d: A large scale dataset for 3d human activity analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1010–1019. [Google Scholar]
  47. Liu, J.; Shahroudy, A.; Perez, M.; Wang, G.; Duan, L.Y.; Kot, A.C. Ntu rgb+ d 120: A large-scale benchmark for 3d human activity understanding. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2684–2701. [Google Scholar] [CrossRef]
  48. Zhang, P.; Lan, C.; Zeng, W.; Xing, J.; Xue, J.; Zheng, N. Semantics-guided neural networks for efficient skeleton-based human action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1112–1121. [Google Scholar]
  49. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  50. Liu, J.; Shahroudy, A.; Xu, D.; Kot, A.C.; Wang, G. Skeleton-based action recognition using spatio-temporal LSTM network with trust gates. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 3007–3021. [Google Scholar] [CrossRef]
  51. Liu, J.; Wang, G.; Hu, P.; Duan, L.Y.; Kot, A.C. Global context-aware attention lstm networks for 3d action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1647–1656. [Google Scholar]
  52. Liu, J.; Wang, G.; Duan, L.Y.; Abdiyeva, K.; Kot, A.C. Skeleton-based human action recognition with global context-aware attention LSTM networks. IEEE Trans. Image Process. 2017, 27, 1586–1599. [Google Scholar] [CrossRef]
  53. Liu, J.; Shahroudy, A.; Wang, G.; Duan, L.Y.; Kot, A.C. Skeleton-based online action prediction using scale selection network. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1453–1467. [Google Scholar] [CrossRef]
  54. Nguyen, X.S. Geomnet: A neural network based on riemannian geometries of spd matrix space and cholesky space for 3d skeleton-based interaction recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 13379–13389. [Google Scholar]
  55. Li, M.; Chen, S.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Actional-structural graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3595–3603. [Google Scholar]
  56. Song, Y.F.; Zhang, Z.; Shan, C.; Wang, L. Stronger, faster and more explainable: A graph convolutional baseline for skeleton-based action recognition. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1625–1633. [Google Scholar]
  57. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Decoupled spatial-temporal attention network for skeleton-based action-gesture recognition. In Computer Vision—ACCV 2020, Proceedings of the 15th Asian Conference on Computer Vision, Kyoto, Japan, 30 November–4 December 2020; Springer: Cham, Switzerland, 2020. [Google Scholar]
  58. Qiu, H.; Hou, B.; Ren, B.; Zhang, X. Spatio-temporal segments attention for skeleton-based action recognition. Neurocomputing 2023, 518, 30–38. [Google Scholar] [CrossRef]
  59. Wen, Y.; Tang, Z.; Pang, Y.; Ding, B.; Liu, M. Interactive spatiotemporal token attention network for skeleton-based general interactive action recognition. arXiv 2023, arXiv:2307.07469. [Google Scholar]
Figure 1. The overall architecture of our proposed multi-modal enhancement transformer network (ME-Former). The ME-Former network consists of two stages: the feature extraction and fusion stage, and the feature refinement stage. Each modality in the feature extraction and fusion stage has an independent information flow to learn distinctive spatio-temporal features. At the same time, each layer in the feature extraction and fusion stage also has a cross-modal information flow. The CPF block converts the features of all skeletal modalities in the previous layer into a global context. Then, it transfers cross-modal information into four ME modules in the current layer to enhance four single modalities. At the end of the feature extraction and fusion stage, ME-Former fuses the four skeletal modalities into a unified modality and models its spatial and temporal features. Finally, the skeleton features are sent into the fully connected layer (FC) to obtain the predicted scores.
Figure 1. The overall architecture of our proposed multi-modal enhancement transformer network (ME-Former). The ME-Former network consists of two stages: the feature extraction and fusion stage, and the feature refinement stage. Each modality in the feature extraction and fusion stage has an independent information flow to learn distinctive spatio-temporal features. At the same time, each layer in the feature extraction and fusion stage also has a cross-modal information flow. The CPF block converts the features of all skeletal modalities in the previous layer into a global context. Then, it transfers cross-modal information into four ME modules in the current layer to enhance four single modalities. At the end of the feature extraction and fusion stage, ME-Former fuses the four skeletal modalities into a unified modality and models its spatial and temporal features. Finally, the skeleton features are sent into the fully connected layer (FC) to obtain the predicted scores.
Biomimetics 09 00123 g001
Figure 2. Illustration of the ME module. The ME module consists of a multi-head cross-modal attention(MH-CA) block and a two-person hypergraph self-attention (TH-SA) block. The skeleton features are gradually enhanced from the source modality to the target modality.
Figure 2. Illustration of the ME module. The ME module consists of a multi-head cross-modal attention(MH-CA) block and a two-person hypergraph self-attention (TH-SA) block. The skeleton features are gradually enhanced from the source modality to the target modality.
Biomimetics 09 00123 g002
Figure 3. Illustration of the CPF block. This converts the features of four skeleton modalities and the previous global context into a higher-level global context for enhancing four single skeleton modalities in the next layer.
Figure 3. Illustration of the CPF block. This converts the features of four skeleton modalities and the previous global context into a higher-level global context for enhancing four single skeleton modalities in the next layer.
Biomimetics 09 00123 g003
Figure 4. Illustration of the two-person hypergraph self-attention block (TH-SA). In addition to vanilla SA, this adds skeleton topology embedding and 2p-hypergraph embedding, allowing the transformer to better recognize the spatial structure and semantic information specific to each human joint.
Figure 4. Illustration of the two-person hypergraph self-attention block (TH-SA). In addition to vanilla SA, this adds skeleton topology embedding and 2p-hypergraph embedding, allowing the transformer to better recognize the spatial structure and semantic information specific to each human joint.
Biomimetics 09 00123 g004
Figure 5. A brief visualization of different skeleton topologies and partition strategies. (a,b) are two kinds of skeleton topologies that contain different pairwise relations of skeleton joints. (ce) are partition strategies that depict the higher-order relations of joints in a two-person skeleton topology. Different colors represent different higher-order relations. Joints with the same higher-order relation are set into a group and the entire two-person skeleton topology is divided into five groups. Multiple joints with the same high-order relation are labeled with the same color.
Figure 5. A brief visualization of different skeleton topologies and partition strategies. (a,b) are two kinds of skeleton topologies that contain different pairwise relations of skeleton joints. (ce) are partition strategies that depict the higher-order relations of joints in a two-person skeleton topology. Different colors represent different higher-order relations. Joints with the same higher-order relation are set into a group and the entire two-person skeleton topology is divided into five groups. Multiple joints with the same high-order relation are labeled with the same color.
Biomimetics 09 00123 g005
Figure 6. Confusion matrix diagram for two models testing on the NTU-60* (X-Sub) benchmark.
Figure 6. Confusion matrix diagram for two models testing on the NTU-60* (X-Sub) benchmark.
Biomimetics 09 00123 g006
Table 1. Comparisons of interaction recognition performance on NTU-60* and NTU-120* datasets. The recognition accuracy of our proposed ME-Former are listed in the last row.
Table 1. Comparisons of interaction recognition performance on NTU-60* and NTU-120* datasets. The recognition accuracy of our proposed ME-Former are listed in the last row.
TypeMethodConf./Jour.NTU-60* (%)NTU-120* (%)
X-SubX-ViewX-SubX-Set
RNNST-LSTM [50]TPAMI 201783.0087.3063.0066.60
GCA-LSTM [51]CVPR 201785.9089.0070.6073.70
2s-GCA-LSTM [52]TIP 201787.2089.9073.0073.30
FSNET [53]TPAMI 201974.0180.5061.2269.70
GeomNet [54]ICCV 202193.6296.3286.4987.58
LSTM-IRN [41]TMM 202290.5093.5077.7079.60
GCNST-GCN [31]AAAI 201889.3193.7280.6980.27
AS-GCN [55]CVPR 201989.3093.0082.9083.70
2s-AGCN [13]CVPR 201993.3696.6787.8389.21
ST-GCN-PAM [13]ICIP 2020--83.2888.36
Pa-ResGCN-B19 [56]ACM MM 202094.3497.5589.6489.94
CTR-GCN [32]ICCV 202191.6094.3083.2084.40
DR-GCN [40]PR 202193.6894.0985.3684.49
K-GCN [39]NC 202193.7096.80--
2S-DRAGCN [40]PR 202194.6897.1990.5690.41
AIGCN [43]ICME 202293.8997.2287.8087.96
EfficientGCN-B1 [16]TPAMI 202294.4997.2390.6490.21
2s-AIGCN [43]ICME 202295.3498.0090.7190.65
TransformerDSTA-NET [57]ACCV 2020--88.9290.10
IGFormer [42]ECCV 202293.6496.5085.4086.50
STSA-Net [58]NC 2023--90.2891.13
ISTA-Net [59]IROS 2023--90.6091.87
Our ME-Former-95.3797.6090.8491.33
Table 2. Ablation study on multi-modal enforcement transformer network.
Table 2. Ablation study on multi-modal enforcement transformer network.
Model DesignParams. (M)NTU-60*X-Sub (%)
Feature Extraction and Fusion StageFeature Refinement Stage
4 × SGC (Baseline)SGC1.294.49
4 × TH-SATH-SA5.694.98
4 × METH-SA7.895.26
4 × (ME + CPF)TH-SA9.295.37
Table 3. Ablation study on components of TH-SA block.
Table 3. Ablation study on components of TH-SA block.
Model DesignParams. (M)NTU-60*X-Sub (%)
Vanilla SA4.493.37
Vanilla SA + Single-person Skeleton Topology Embedding4.894.34
Vanilla SA + Two-person Skeleton Topology Embedding4.894.55
Vanilla SA + Two-person Hypergraph Embedding5.194.67
Entire TH-SA5.694.98
Table 4. Ablation study on partition strategies of TH-SA block.
Table 4. Ablation study on partition strategies of TH-SA block.
Partition StrategyNTU-60*X-Sub (%)
Upper and Lower Body94.53
Body Parts94.64
Two-person Hypergraph94.98
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, Q.; Liu, H. Multi-Modal Enhancement Transformer Network for Skeleton-Based Human Interaction Recognition. Biomimetics 2024, 9, 123. https://doi.org/10.3390/biomimetics9030123

AMA Style

Hu Q, Liu H. Multi-Modal Enhancement Transformer Network for Skeleton-Based Human Interaction Recognition. Biomimetics. 2024; 9(3):123. https://doi.org/10.3390/biomimetics9030123

Chicago/Turabian Style

Hu, Qianshuo, and Haijun Liu. 2024. "Multi-Modal Enhancement Transformer Network for Skeleton-Based Human Interaction Recognition" Biomimetics 9, no. 3: 123. https://doi.org/10.3390/biomimetics9030123

Article Metrics

Back to TopTop