Next Article in Journal
A Dockerized Approach to Dynamic Endpoint Management for RESTful Application Programming Interfaces in Internet of Things Ecosystems
Previous Article in Journal
Borehole Radar Experiment in a 7500 m Deep Well
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

GT-SRR: A Structured Method for Social Relation Recognition with GGNN-Based Transformer

1
College of Science, Qingdao University of Technology, Qingdao 266000, China
2
College of Information and Control Engineering, Qingdao University of Technology, Qingdao 266000, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 2992; https://doi.org/10.3390/s25102992
Submission received: 24 March 2025 / Revised: 25 April 2025 / Accepted: 1 May 2025 / Published: 9 May 2025
(This article belongs to the Section Sensor Networks)

Abstract

:
Social relationship recognition (SRR) holds significant value in fields such as behavior analysis and intelligent social systems. However, existing methods primarily focus on modeling individual visual traits, interaction patterns, and scene-level contextual cues, often failing to capture the complex dependencies among these features and the hierarchical structure of social groups, which are crucial for effective reasoning. In order to overcome these restrictions, this essay suggests a SRR model that integrates Gated Graph Neural Network (GGNN) and Transformer. The task for SRR in this model is image-based. Specifically, the purpose of a novel and robust hybrid feature extraction module is to capture individual characteristics, relative positional information, and group-level cues, which are used to construct relation nodes and group nodes. A modified GGNN is then employed to model the logical dependencies between features. Nevertheless, GGNN alone lacks the capacity to dynamically adjust feature importance, which may result in ambiguous relationship representations. The Transformer’s multi-head self-attention (MSA) mechanism is integrated to improve feature interaction modeling, allowing the model to capture global context and higher-order dependencies effectively. By fusing pairwise features, graph-structured features, and group-level information. Experimental results on public datasets such as PISC demonstrate that the proposed approach outperforms comparison models including Dual-Glance, GRM, GRRN, Graph-BERT, and SRT in terms of accuracy and mean average precision (mAP), validating its effectiveness in multi-feature representation learning and global reasoning.

1. Introduction

Social relationships and interactive behaviors are fundamental to human social structures, significantly influencing societal development. SRR leverages visual data analysis to infer interpersonal relationships by examining individual attributes and interactions within images. This approach has been extensively used in fields including innovative city management [1], group emotion detection [2], behavioral analytics, and scene interpretation.
Recent advancements in deep learning and computer vision have propelled SRR research [3,4,5,6,7,8,9,10,11]. However, the complexity and diversity of social relationships, coupled with varying visual contexts and intricate semantic interrelations among features, continue to pose challenges to the accuracy and robustness of SRR systems. Existing studies have primarily focused on extracting and modeling individual, interactional, and contextual features. Techniques based on Convolutional Neural Networks (CNNs) [12] have been employed for individual feature extraction, analyzing facial expressions, attire, and postures to infer social relationships [9,13,14]. However, these methods often overlook interpersonal interactions, limiting their effectiveness in dyadic scenarios. To address this, researchers have incorporated interactional features, such as relative positioning [15,16] and proximity, recognizing their significance in reflecting social bonds. In social scenarios, individuals with intimate or familiar relationships are typically positioned in close physical proximity and tend to orient themselves toward each other, while strangers usually maintain greater interpersonal distance. However, interaction modeling approaches that rely solely on spatial or geometric features often fail to perform well in complex multi-person environments, primarily due to their limited capacity to incorporate rich contextual information.
Additionally, integrating contextual features has become prevalent in SRR research [13,14], providing environmental insights into interpersonal dynamics. For example, domestic settings often suggest intimate relationships, while professional settings imply professional associations. However, existing approaches frequently treat individual, interactional, and contextual features in isolation, lacking comprehensive modeling of their interdependencies, which hinders performance in diverse relational contexts. Moreover, research on modeling the interplay between group characteristics and dyadic relationships remains relatively underexplored. Group-level features offer insights into broader social behaviors, enriching the understanding of dyadic relationships by providing essential contextual information. For example, group activities can reveal interaction patterns and role dynamics, which are crucial for understanding relationships in complex scenarios. Ignoring group semantics leads to a fragmented understanding of dyadic relationships. Moreover, conventional SRR methods rely on single-pass reasoning, failing to capture higher-order correlations among features. GNN-based techniques [9,17,18,19] construct relational graphs to model node interactions, effectively capturing local feature relationships. However, the iterative nature of GNNs presents challenges in modeling global contexts, limiting the representation of complex semantic associations.
To address these challenges, this paper introduces a novel SRR model that effectively combines GGNN and Transformer architectures. The main contributions are as follows:
  • Hybrid Feature Extraction: The model constructs relational and group nodes by leveraging multi-layer networks to extract individual features, interactional positional data, and group information. An enhanced GGNN captures the logical relationships among these nodes, while an attention mechanism dynamically weights scene features, effectively modeling complex inter-feature interactions.
  • Integrated Feature Fusion Strategy: The model integrates pairwise features, relational graph data, and group characteristics. Group features supplement pairwise recognition by offering comprehensive contextual insights. A feature weighting mechanism dynamically adjusts each feature’s impact type, enhancing the model’s expressiveness and robustness, particularly in complex relational scenarios.
  • Advanced Relationship Modeling with GGNN and Transformer: Building upon initial modeling, the combined GGNN and Transformer framework leverages the MSA feature of the Transformer to enhance global context modeling and conduct higher-order correlation analysis. While the GGNN effectively captures local feature relationships, the Transformer enables deep inter-feature interactions through global reasoning, ultimately improving the representation of social relationships and boosting classification performance.
The remainder of the paper is organized as follows: Section 2 reviews relevant work on Transformer-based relational modeling and SRR. Section 3 details the proposed GT-SRR framework, including hybrid feature extraction, hierarchical graph reasoning, and Transformer-based global fusion. Section 4 presents the experimental setup, implementation details, and comparative evaluations on benchmark datasets. Finally, Section 5 wraps up the paper.

2. Related Work

2.1. SRR

SRR is a pivotal research area in computer vision, focusing on inferring interpersonal interactions from visual data. This involves extracting individual features and modeling interactions, spatial configurations, and social contexts to elucidate complex social networks and interpersonal relationships.
Early approaches predominantly utilized handcrafted features, emphasizing attributes such as skin and hair color [20], as well as other appearance-related characteristics [21], primarily for identifying basic kinship relations. The advent of the People in Social Context (PISC) dataset [16] facilitated more sophisticated modeling by providing extensive data support. This led to the adoption of advanced neural structures that can extract facial, bodily, and contextual scene features to recognize intricate social relationships. For instance, Li et al. [16] employed CNNs to extract body features from segmented individual and joint regions, concurrently capturing visual context. Zhang et al. [12] introduced a multi-granularity inference framework incorporating Graph Convolutional Networks (GCNs) to model facial keypoints, significantly enhancing recognition accuracy for complex relationships. Additionally, Goel et al. [22] proposed an end-to-end neural network model capable of automatically generating social relationship graphs from input images, facilitating the inference of diverse relationship types.
The proliferation of GNNs has shifted research towards graph-based methodologies for SRR. GNNs adeptly model complex relational structures by capturing dependencies and interactions among various relationship types through graph nodes and edges. Wang et al. [23] developed a model integrating graph reasoning to enhance recognition by merging relational and environmental data. Guo et al. [24] suggested a method of social relation reasoning based on triangular constraints, introducing the Triangular Reasoning Graph Attention Network (TRGAT), which leverages attention and contrastive learning to capture high-order context and effectively improve reasoning accuracy and consistency. Li et al. [25] introduced a GNN-based social recommendation algorithm, enhancing system efficiency by learning interaction data within social graphs. Qing et al. [14] combined GGNNs and GCNs to fuse global and local information, effectively capturing dynamic relationships and local features between nodes.
Despite advancements, many methods struggle to fully integrate multiple feature modalities and model complex relational contexts, particularly in group interactions and higher-order semantic relationships. Addressing these challenges, this essay suggests a novel model combining GNNs and Transformer architectures for SRR. The model aims to enhance recognition accuracy and robustness in complex social relationship scenarios by effectively fusing individual, interactional, and group features.

2.2. Transformer-Based Relationship Modeling

Transformer-based models have significantly advanced, moving from theoretical foundations to practical applications across various fields. Vaswani et al. [26] presented the mechanism of self-attention, providing a framework for capturing long-range dependencies and overcoming limitations associated with traditional Recurrent Neural Networks (RNNs). Building upon this, Gao et al. [27] proposed an approach to connection modeling that utilizes adaptive token division, which effectively combines the advantages of dual-stream and single-stream tracking. It reduces confusion between target and background by utilizing selective interaction, attention masking, and Gumbel-Softmax for efficient training, achieving state-of-the-art performance on various benchmarks.
However, standard Transformer models face challenges in representing structured relationships, such as overlapping or nested relationships. To address this, Soares et al. [28] proposed the Virtual Token Contrastive Learning framework, incorporating learnable [REL] tokens to explicitly define relationship boundaries. Incorporating graph-structured data, Zhang et al. [29] introduced Graph-BERT, which encodes graph features, such as node degree centrality, into Transformer position embeddings, achieving an AUROC of 0.912 in protein interaction prediction tasks. Diao et al. [30] introduced a relational attention mechanism, enhancing Transformer models by integrating prior relationships between entities as attention biases. Chen et al. [31] proposed a structure-aware Transformer, augmenting the perceptual and modeling capabilities of Transformer architectures. These developments underscore the dynamic and evolving landscape of Transformer-based relationship modeling, highlighting their efficacy and adaptability across various complex relationship modeling tasks.

3. Methods

In this section, we introduce the GT-SRR framework. We first describe the model architecture, followed by discussions on node feature extraction, graph inference using GGNN, and relationship classification through the Transformer.

3.1. GT-SRR Framework Overview

The GT-SRR model, similar to prior SRR models [10,13,14], focuses on pairwise relationship analysis. As shown in Figure 1, it consists of three key modules: (i) Hierarchical graph construction via hybrid feature extraction; (ii) Graph-guided relational inference; (iii) Classification via Transformer-based cross-modal fusion. Given an input image with N individuals, the model generates M = C ( N , 2 ) binary RNs (RN-i, i = 1 , , M ) and a single GN, where the group and global scene nodes are treated as one entity. Here, C ( N , 2 ) denotes the combination function, defined as C ( N , 2 ) = N ( N 1 ) 2 , which calculates the number of unique unordered pairs formed from N individuals.
Relation nodes are formed by combining appearance, spatial positions, and interaction features, while the group node encodes collective semantic and dynamic group characteristics. These nodes are connected via learnable weighted edges, forming a fully connected social relationship graph. A message-passing mechanism enables multi-hop reasoning between nodes: local interaction information flows from relation nodes to the group node, while global context from the group node enhances relation node representations.
The model integrates binary relationship features, graph-optimized structural features, and group features, which are processed through an enhanced MSA Transformer. This module adjusts the contribution of each feature modality, optimizing cross-level representations and effectively identifying social relationships. Higher-order feature representations are aggregated in latent space to capture key patterns. Finally, a fully connected classification layer processes the fused high-order features, yielding the social relationship probability distribution and completing recognition.

3.2. Node Feature Extraction

To capture social relationship features from images, we propose a node generation module that integrates various features to construct relationship nodes for further inference. The module architecture is shown in Figure 2.
As described in Section 3.1, for a picture that has M individuals, M = C ( N , 2 ) relationship nodes are generated, while only one group node is created, irrespective of the number of individuals. For example, with three individuals, three relationship nodes are formed.The node generation process for two individuals is as follows: the features of these individuals are extracted through separate channels.
As outlined in previous studies [8,13,14], the spatial distance between individuals is utilized to capture their positional and relational proximity. A completely interconnected layer is used to extract the location information, which is then encoded into the node representation. In the equation:
pos = [ x min P 1 , y min P 1 , x max P 1 , y max P 1 , area P 1 ] , [ x min P 2 , y min P 2 , x max P 2 , y max P 2 , area P 2 ] .
In this equation, x min P 1 , y min P 1 , x max P 1 , y max P 1 represent the boundary coordinates of the bounding box of P 1 , and area P 1 indicates the area of P 1 .
Similarly, x min P 2 , y min P 2 , x max P 2 , y max P 2 , area P 2 represent the corresponding coordinates and area for P 2 . These features are normalized to the range of [−1, 1], using P 1 ’s position as an example:
area P 1 = ( y min P 1 y max P 1 ) × ( x max P 1 x min P 1 )
area norm = 2 × area P 1 area max 1 , area norm [ 1 , 1 ]
where area max represents the maximum boundary area in the dataset. Equation (1) uses this normalization for all elements. At the same time, we make reference to the attraction formula from SIF theory [32], where the interaction force between P i and P j is proportional to the attraction between them and inversely related to the square of the distance. The formula is given as:
F att = λ F i T F j 1 r 2
Here, F att denotes the attraction force, and F i and F j are the individual features of persons P i and P j , respectively. λ is a constant set to 0.5. r represents the Euclidean distance between P i and P j , which is defined as:
r = F i F j 2
The attraction and positional information are fused via a fully connected (FC) layer and encoded as interaction features, with an output dimension of 512.
To extract individual and joint features, the ImageNet dataset is used to pre-train the ResNet-101 and ResNet-50 models. These models can capture spatial coordinates and relational semantics from visual data. The individual and joint regions are resized to 244 × 244, yielding a 4096-dimensional feature vector. This vector is further processed through the ResNet backbone to enhance relational feature representation. To improve recognition performance, feature maps from the final ResNet layer are fine-tuned for the SRR task. Additionally, group-level characteristics are extracted using a ResNet-50 model trained regarding the Places365-Standard dataset. After resizing the input image to 448 × 448 pixels, a scene feature vector with 4096 dimensions is produced. This vector is subsequently fused with individual and global image features to form a comprehensive feature representation for social relationship inference.
All ResNet models are fine-tuned to adapt to the specific characteristics of the SRR dataset. After feature extraction, additional individual features are integrated via another ResNet-101 model. Finally, a ResNet-50 network aggregates these features, producing a 512-dimensional group-level feature representation for relational reasoning.

3.3. Graph Inference Based on GGNN

To implement the GGNN-based graph inference, we first construct the relational graph. Since the features of individual nodes are not always equally significant, we assign different weights to the relationship and group nodes based on their relevance. In this work, the social relationship reasoning task is formulated as a graph-based inference problem, where the constructed graph G = ( V , E ) is an undirected graph. The node set V consists of M = C ( N , 2 ) relation nodes, each representing a pairwise relationship among N detected individuals, along with a single group node that encodes holistic scene-level semantics. Thus, the total number of nodes is M + 1 . Edges E are defined only between each relation node and the group node, resulting in a star-shaped topology where every relation node is connected to the group node via a single undirected edge. No edges exist between relation nodes. The total number of edges is therefore M. This graph does not incorporate explicit edge features; instead, the dependencies between nodes are captured through dynamically learned attention weights during the message-passing process.
First, we perform a layer-wise projection to obtain the initial state h R N i ( 0 ) for each relationship node and h G N ( 0 ) for the group node, as shown below:
h R N i ( 0 ) = ReLU W R X R N i + b r
h G N ( 0 ) = ReLU W G X G N + b G
where X R N i represents the feature vector for relationship node i, and X G N represents the feature vector for the group node. The weight matrices W R , W G and bias terms b r , b G are learned during training. The initial feature dimensions for both the relation nodes and the group node are set to 512. Specifically, each relation node feature h R N i ( 0 ) R 512 is obtained by projecting the concatenated visual, positional, and interaction features of a pair of individuals into a 512-dimensional space. Similarly, the group node feature h G N ( 0 ) R 512 is derived by projecting the group feature into the same 512-dimensional space. This consistent dimensionality facilitates unified representation and efficient integration in subsequent GGNN and Transformer-based modules.
Next, we aim to aggregate information from different relationship nodes to enrich the information captured by each node, which is essential for identifying relationships. A learnable weight α is assigned, and the features from the relationship nodes and the group node are fused as follows:
e i = LeakyReLU α T h R N i ( t ) , h G N ( t )
where e i denotes the attention score, and LeakyReLU is applied to ensure that negative values do not saturate the output. The attention mechanism normalizes these values using softmax:
a i = exp ( e i ) k = 1 M exp ( e k ) , i = 1 , 2 , , M
where a i represents the attention weight for each relationship node, reflecting the degree of attention each node gives to the group node. A higher value of a i indicates a stronger relationship with the group node.
In the graph inference phase, we incorporate message passing from the relationship nodes to the group node. This process propagates information through the graph, enhancing the semantic representation of each node based on its neighbors. After this step, the hidden states of the group node h G N and the relationship nodes h R N i are updated. The update rule for the group node is:
m G N ( t + 1 ) = i = 1 M a i W p h R N i ( t )
where W p is the learned weight matrix for propagating the messages, and the feature vector is updated iteratively. After message propagation, we use the Gated Recurrent Unit (GRU) to further update the group and relationship nodes. The GRU helps integrate information from multiple hops of message passing to refine the node features continuously. The update rules are as follows:
h G N ( t + 1 ) = GRU h G N ( t ) , m G N ( t + 1 )
h R N i ( t + 1 ) = GRU h R N i ( t ) , m R N i ( t + 1 )
where h G N ( t ) and h R N i ( t ) represent the hidden states at the t-th iteration, and m G N ( t + 1 ) and m R N i ( t + 1 ) represent the messages passed between the nodes.
This iterative process continues until the node states converge or reach a predefined number of iterations. The graph output dimension is adjusted to 512 to facilitate subsequent fusion, as it is not directly used for classification.

3.4. Social Relationship Modeling via Transformer

Although graph inference provides a comprehensive structural foundation, we further enhance the model by incorporating Transformer-based fusion inspired by multi-head attention (MHA) in modern natural language processing (NLP) architectures. Specifically, the Transformer component is designed to integrate diverse features for relational modeling.
As shown in Figure 3, features in the model are fully utilized through interaction across visual and structural domains. The input to the feature fusion module is formulated as:
S i n p u t = [ X L , X g r a p h , X p a i r , X G N ]
where X L , X g r a p h , X p a i r , and X G N denote the learnable embeddings, graph-structural features, dyadic relational features, and group-level features respectively, each in R 512 . The input sequence maintains a fixed length and dimensionality of 512, which remains unaffected by the number of individuals or the size of the graph. As a result, the Transformer module can process inputs without requiring structural adjustments for different graph sizes. This design ensures consistent cross-modal feature fusion and significantly enhances the model’s ability to capture contextual dependencies. The entire transformation process is defined as:
S l = MSA ( LN ( S k 1 ) ) + S k 1 , k = 1 , 2 , , K
where K denotes the number of Transformer layers. Multi-head attention mechanisms allow dynamic weighting of feature contributions, enabling effective fusion of global embeddings and inter-node dependencies. The final output of the Transformer encodes the social relationship representations.
In the GGNN message passing process, each node updates its features by aggregating the weighted features of its neighbouring nodes. As described in Equations (10)–(12), the weights α i are determined by the similarity between nodes, reflecting their dependencies. However, this weighting scheme is primarily based on local neighbour information, making it difficult to capture more complex relationships between nodes and global dependencies. To address this limitation, we integrate Transformer’s MSA mechanism into the GGNN framework. To facilitate the integration, we use a projection layer to map the GGNN node features into a feature space compatible with MSA. The attention weights α i computed by MSA are then introduced into the GGNN message passing process, dynamically adjusting the influence of each relational node on the group node. This allows the model to capture complex dependencies between nodes, overcoming the limitations of static graph structures. In particular, the dynamic attention weights α i computed by MSA optimise the flow of information in each message-passing step. As a result, information propagation is no longer limited to local neighbour relationships, but can also capture dependencies from distant nodes, improving node feature updates. Furthermore, MSA’s multi-head mechanism enhances this process by allowing each node to focus on relationships at multiple scales, providing richer contextual information for feature updates and improving the model’s ability to capture both global and local relationships.
In this context, the GN plays a critical role. It not only aggregates information, but also influences the feature updates of the relational nodes. Specifically, during each iteration, the group node adjusts the influence of each relational node on itself through the dynamic attention weights α i computed by MSA. This dynamic weighting allows the group node to flexibly control the contribution of each relational node to the information propagation process based on its importance. As a result, the influence of the group node on the relational nodes is refined, allowing the model to capture more complex global dependencies in the message-passing process.
We further represent the M pairwise relationships from an image as:
S i n p u t i n t e r = [ r 1 , r 2 , , r M ] , r i R 512 , i ( 1 , 2 , , M )
Analogous to the previous formulation, Equation (13) constructs the Transformer-enhanced representation of each social relationship. The MSA mechanism captures contextual similarities across all pairwise relations in an image, thereby generating informative and semantically meaningful relational embeddings—an essential component in SRR tasks. For example, an image might illustrate three types of relationships: two pairs labelled as “friends” and one pair labelled as “no relation”. In our Intra-TRM MSA module, we refine input features by aggregating similarities between the current node and other relationship representations. This process improves feature integration, enabling us to capture underlying affinities more effectively.
Notably, relationships with high internal similarity are assigned greater attention weights, while semantically dissimilar relations maintain lower affinities. This mechanism highlights important patterns while reducing confusion from visual similarities between different elements. As shown in Figure 4, this attention-guided fusion mitigates semantic ambiguity and enhances relationship discrimination.

3.5. Classification and Optimization

In the Transformer-based relation generation module, the final relation representation r final is used to compute the classification probability distribution over social relation categories. The predicted probability for each class p i is obtained through the softmax function as:
S R i = { p 1 , p 2 , , p m } = softmax ( F C ( r final ) )
Here, p i (for i = 1 , 2 , , m ) represents the probability score for the i-th relation class, and m is the total number of relation categories. The relation types are defined according to task-specific classification schemes (e.g., PISC-C and PISC-F). S R denotes the final probability distribution of social relation classification for each instance.
We introduce a similarity-aware optimization strategy based on maximum likelihood estimation to improve classification performance further. Specifically, a Gaussian kernel–based similarity loss is employed to improve the discriminative power of relation representations. The process includes the following two aspects:
  • Computing similarity between samples, including inter-class and intra-class similarity.
  • Enhancing classification by aligning the predicted distribution with class-specific target distributions.
The final classification loss function is defined as:
loss = log 1 + y y e z y z y + δ
where z y and z y represent the outputs from the final fully connected layer for the ground-truth class y and other classes y , respectively. The adaptive similarity margin δ is calculated as:
δ = n u m max n u m y · cosine_similarity ( y , y )
In this formula, n u m max denotes the maximum number of samples across all classes in the training set, and num y denotes the number of samples belonging to class y, while cosine_similarity ( y , y ) represents the cosine similarity between classes y and y , which reflects the semantic proximity between different relation classes and contributes to more adaptive optimization in classification.

4. Experiments and Results

This section introduces the datasets and implementation specifics, baseline comparisons, and ablation studies to demonstrate the efficacy of the suggested model.

4.1. Dataset

We evaluate our method on the People in Social Context (PISC) dataset, a widely used benchmark for SRR. The dataset contains 22,670 images and 76,568 annotated relationship samples. Following prior works, we employ a two-level hierarchical classification approach and take mAP as the primary evaluation metric. As shown in Table 1, the dataset is divided into training, validation, and test sets to ensure reproducibility and fair comparison. The coarse-level classification (PISC-C) categorizes samples into three major categories: no relation, non-intimate, and intimate. The fine-level classification (PISC-F) further refines the relationship categories into six classes: no relationship, professional, commercial, couple, friend, and family.

4.2. Implementation Details

Considering the class imbalance issue in the PISC-F dataset (as illustrated in Table 2), we applied data augmentation strategies to mitigate overfitting and improve model generalization. Specifically, we utilized fundamental image transformations (e.g., flipping, rotation, cropping), random occlusion, and Gaussian noise injection with a standard deviation of 0.02.
For visual feature extraction, we used pre-trained ResNet-101 and ResNet-50 as backbone networks, and incorporated a location encoder to capture bounding box coordinates and region area information. The multi-modal features—including visual, spatial, and interaction cues—were concatenated and linearly projected before passing through the graph inference module and a 4-layer Transformer encoder for relational classification.
The PyTorch 2.1 framework was used for all experiments. The model training used the Adam optimizer, starting with a 0.0001 learning rate and the OneCycleLR learning rate scheduler (peak learning rate set to 0.001). A batch size of 64 was used, and the training process spanned 100 epochs. All training and evaluation procedures were performed on an NVIDIA GeForce RTX 3060 GPU.

4.3. Comparative Experiments

To assess the performance of the proposed GT-SRR model, we conduct comparisons against five approaches:
Dual-Glance: The model consists of two stages that simulate human visual perception. In the first stage, it focuses on the appearance and spatial cues of the target pair to make a coarse prediction. The second stage improves these predictions by examining contextual features using attention mechanisms and region proposals, such as sual perception. The first stage attends to the appearance and spatial cues of the target pair for coarse prediction. The second stage refines predictions by exploring contextual features via attention mechanisms and region proposals.
GRM: A high-fidelity 3D reconstruction framework that recovers multi-view visual information from a single image using a pre-trained transformer backbone. It achieves accurate reconstruction within 0.1 s.
GRRN: A novel trainable GNN that constructs social relation graphs and models both pairwise relationships and contextual attributes. It employs GRUs for iterative message passing between nodes.
Graph-BERT: A Transformer-based model that incorporates BERT-style pretraining for graph-structured relational reasoning. It encodes pairwise and contextual features using attention mechanisms and achieves competitive performance without explicit graph construction.
SRT: A simple relation transformer designed for social understanding. It directly models relationship semantics through stacked Transformer layers, without relying on intermediate graph modules, but shows limitations in fine-grained categories due to lack of explicit structure modeling.
To ensure a fair comparison of performance under consistent training conditions, all baseline models were trained using the same settings as the proposed GT-SRR. Specifically, all models were implemented using the PyTorch framework and optimized with the Adam optimizer, with an initial learning rate set to 0.0001. The OneCycleLR scheduler was employed, with a peak learning rate of 0.001. A batch size of 64 was used, and training was conducted for 100 epochs. The cross-entropy loss function was adopted for all models. All experiments were carried out on an NVIDIA GeForce RTX 3060 GPU. This unified setup eliminates discrepancies due to training configurations and enables a more objective evaluation of each model’s effectiveness in social relationship recognition.
Table 3 gives the quantitative comparison of the performance of all methods on the PISC-C and PISC-F datasets. The proposed GT-SRR model achieves the highest overall mAP of 64.48% on PISC-C, demonstrating superior relational representation capabilities. Notably, GT-SRR achieves the highest accuracy of 80.10% in the Intimate category, outperforming the strongest baseline (GRRN) by approximately 1%. The advantage is even more pronounced in the Non-intimate category, where GT-SRR improves performance by over 5% compared to other models. On the more challenging PISC-F dataset, GT-SRR again achieves the best overall mAP of 64.06%, indicating its strong generalization ability across diverse relationship categories. It consistently outperforms all baselines in categories such as Friend, Family, Couple, Professional, and Commercial. For instance, in the Commercial category, GT-SRR reaches 63.06% accuracy, significantly surpassing Dual-Glance (25.71%), GRM (46.33%), and GRRN (52.63%). To further validate the effectiveness of GT-SRR, we include two representative comparison models, Graph-BERT and SRT, which adopt graph-based and rule-driven reasoning mechanisms, respectively. Graph-BERT achieves a competitive mAP of 71.30% on the PISC-C dataset, showing balanced performance across coarse-grained categories. However, its performance drops to 59.79% on PISC-F, with particularly low accuracy in the Commercial category (10.73%). In contrast, SRT demonstrates weaker overall performance, with mAP scores of 52.35% on PISC-C and 45.28% on PISC-F, and performs poorly in fine-grained categories such as Couple and Commercial. These results further underscore the robustness and superiority of GT-SRR in handling both coarse- and fine-grained social relationship recognition tasks.
These results clearly demonstrate the effectiveness of GT-SRR’s dual-stream feature fusion and hierarchical relational reasoning. Compared to Dual-Glance’s coarse ROI-based interaction modeling, GRM’s rigid geometric encoding, and GRRN’s limited GRU-based local reasoning, the improvements are evident. Graph-BERT’s BERT-enhanced graph inference with MLP fusion, and SRT’s rule-based interaction without an explicit graph structure, GT-SRR benefits from hybrid interaction modeling, hierarchical graph structure, and cross-modal attention fusion. As further detailed in Table 4, GT-SRR captures both local and global context using a hierarchical graph of relation and group nodes, and leverages GGNN-Transformer reasoning for comprehensive social relationship understanding.
To further assess performance across relation categories, we report F1-scores, which jointly consider precision and recall. The F1-score is computed as:
F 1 = 2 × Precision × Recall Precision + Recall
F1 scores provide deeper insights into model performance on different relationship types. Figure 5 and Figure 6 illustrate F1-scores across PISC-C and PISC-F, showing GT-SRR’s superiority in the categories of Intimate, Non-intimate, Friend, Family, Professional, and Commercial.
To explore class-level prediction errors, we further visualize the confusion matrix in Figure 7. GT-SRR achieves the highest recognition rate of 81% for the Professional class, but it suffers from some misclassification between Couple–Commercial and Friends–Family due to overlapping semantics. Commercial relations are often confused with Friends due to shared visual-textual contexts. There is a classification error between the Couple and Commercial categories. The underlying cause of this error may lie in the visual similarity between the two categories, as well as a potential data distribution bias, which makes it challenging for the model to accurately differentiate between the two. These observations indicate that nuanced context modeling is essential for clarifying ambiguity in SRR. Future research should integrate more advanced semantic embeddings to enhance disambiguation abilities.
To better understand the differences in prediction results across models and the underlying causes of misclassification, we present representative visual samples of each relationship category in the PISC-F dataset and their corresponding recognition results, as illustrated in Figure 8. Visual analysis shows that the proposed GT-SRR model successfully identifies complex categories such as Family and Couple, whereas the GRM and Dual-Glance models demonstrate significant misclassification issues, including incorrect predictions of Familyrelationships as Friend, and Couple relationships as Friend. Additionally, the GRRN model tends to misclassify Commercial relationships as Professional. We further observe that Graph-BERT, although capturing Friend and Family relations correctly, misclassifies No Relation as Professional, indicating challenges in distinguishing contextual independence. Similarly, SRT incorrectly identifies Couple as Commercial and Friend as Family, showing limitations in fine-grained relation understanding. This analysis highlights the GT-SRR model’s superior capability in capturing subtle category boundaries, demonstrating its enhanced sensitivity and accuracy in SRR tasks, and further demonstrating the effectiveness of the proposed approach.

4.4. Ablation Study

This seeks to confirm the efficacy of the suggested multi-modal fusion strategy and graph reasoning mechanism. The findings are displayed in Table 5. The results show that using only the GGNN module yields an mAP of 55.54%, but it performs poorly in the non-intimate category (30.11%). The standalone Transformer module performs slightly better, with an mAP of 54.07%. When combining GGNN and Transformer without incorporating group features, the performance improves substantially to an mAP of 74.84%, indicating the benefit of structural and attention-based fusion. However, the whole model—integrating GGNN, Transformer, and group-level features—achieves the best overall performance, reaching an mAP of 80.52%, with a significant gain in the intimate category (73.69%). These results confirm the complementary strengths of the two modules and highlight the critical role of group-level semantics in enabling more effective modeling of spatial and contextual relationships.
To further validate the role of GGNN in the GT-SRR framework, we replaced the GGNN with a Graph Convolutional Network (GCN) and a Graph Attention Network (GAT), while keeping all other settings consistent. The results of this comparison are presented in the Table 6.
We evaluate the model under different iteration settings to investigate the impact of the number of GGNN iterations. As shown in Table 7 and Figure 9, the best performance is achieved when T = 3, with an mAP of 80.52%. Fewer iterations result in insufficient feature aggregation, while excessive iterations may lead to overfitting.
To further investigate the impact of different feature fusion strategies, we conducted ablation experiments on the PISC-C dataset, comparing four configurations: simple concatenation, attention-based fusion, gated fusion, and our proposed multi-head self-attention (MSA) fusion. The results are presented in Table 8. As shown, the gated fusion method achieves relatively high accuracy in the Intimate category (80.62%), but its performance in the Non-intimate and No-relation categories is less competitive. Attention-based fusion performs more consistently across all classes and outperforms simple concatenation. Most notably, our proposed MSA-based fusion demonstrates superior performance across all relationship categories. In particular, it significantly boosts accuracy in the Non-intimate category to 73.69%, while achieving the highest overall mAP of 80.52%. This confirms that the MSA module effectively captures global context and fine-grained dependencies, offering a more robust and expressive representation for social relationship recognition than traditional fusion methods.

5. Conclusions

In this research, we propose GT-SRR, a context-aware framework for social relation recognition that integrates GGNN and Transformer modules. The model effectively captures individual attributes, interaction cues, and group-level context through a hybrid feature extraction mechanism. By constructing relation and group nodes and modeling their structural dependencies with GGNN, the framework enables local interaction reasoning. Additionally, a Transformer encoder performs global context modeling and captures higher-order semantic associations. Incorporating group semantics provides complementary contextual information that enhances the robustness of pairwise relation inference. Experimental results on the PISC dataset demonstrate that GT-SRR achieves superior performance compared to existing methods, particularly in complex relational scenarios. This work advances multi-feature fusion and hierarchical relation modeling in social relation recognition tasks.

Author Contributions

Conceptualization, D.H. and S.G.; methodology, D.H.; software, D.H.; validation, D.H., M.X. and R.C.; formal analysis, D.H. and M.X.; investigation, M.X.; resources, R.C.; data curation, X.K.; writing—preparation of the first draft, D.H.; writing—editing and review, D.H.; visualization, D.H.; supervision, S.G.; project administration, S.G.; funding acquisition, S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Shandong Province Natural Science Foundation [grant number ZR2020QF028].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The People in Social Context (PISC) dataset is available at https://zenodo.org/records/1059155 (accessed on 19 January 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sou, K.; Shiokawa, H.; Yoh, K.; Doi, K. Street design for hedonistic sustainability through AI and human co-operative evaluation. Sustainability 2021, 13, 9066. [Google Scholar] [CrossRef]
  2. Quiroz, M.; Patiño, R.; Diaz-Amado, J.; Cardinale, Y. Group emotion detection based on social robot perception. Sensors 2022, 22, 3749. [Google Scholar] [CrossRef] [PubMed]
  3. Zheng, S.; Chen, S.; Jin, Q. Vrdformer: End-to-end video visual relation detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 18836–18846. [Google Scholar]
  4. Fu, J.; Yang, F.; Dang, Y.; Liu, X.; Yin, J. Learning constrained dynamic correlations in spatiotemporal graphs for motion prediction. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 14273–14287. [Google Scholar] [CrossRef] [PubMed]
  5. Huang, C.; Jiang, F.; Han, Z.; Huang, X.; Wang, S.; Zhu, Y.; Jiang, Y.; Hu, B. Modeling Fine-Grained Relations in Dynamic Space-Time Graphs for Video-Based Facial Expression Recognition. IEEE Trans. Affect. Comput. 2025, 1–17. [Google Scholar] [CrossRef]
  6. Gao, K.; Chen, L.; Zhang, H.; Xiao, J.; Sun, Q. Compositional prompt tuning with motion cues for open-vocabulary video relation detection. arXiv 2023, arXiv:2302.00268. [Google Scholar]
  7. Wang, M.; Du, X.; Shu, X.; Wang, X.; Tang, J. Deep supervised feature selection for social relationship recognition. Pattern Recognit. Lett. 2020, 138, 410–416. [Google Scholar] [CrossRef]
  8. Wang, Y.; Qing, L.; Wang, Z.; Cheng, Y.; Peng, Y. Multi-level transformer-based social relation recognition. Sensors 2022, 22, 5749. [Google Scholar] [CrossRef] [PubMed]
  9. Li, W.; Duan, Y.; Lu, J.; Feng, J.; Zhou, J. Graph-based social relation reasoning. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 18–34. [Google Scholar]
  10. Tang, W.; Qing, L.; Li, L.; Wang, Y.; Zhu, C. Progressive graph reasoning-based social relation recognition. IEEE Trans. Multimed. 2023, 26, 6012–6024. [Google Scholar] [CrossRef]
  11. Yu, X.; Yi, H.; Tang, Q.; Huang, K.; Hu, W.; Zhang, S.; Wang, X. Graph-based social relation inference with multi-level conditional attention. Neural Netw. 2024, 173, 106216. [Google Scholar] [CrossRef] [PubMed]
  12. Zhang, J.; Kalantidis, Y.; Rohrbach, M.; Paluri, M.; Elgammal, A.; Elhoseiny, M. Large-scale visual relationship understanding. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 9185–9194. [Google Scholar]
  13. Li, L.; Qing, L.; Wang, Y.; Su, J.; Cheng, Y.; Peng, Y. HF-SRGR: A new hybrid feature-driven social relation graph reasoning model. Vis. Comput. 2022, 38, 3979–3992. [Google Scholar] [CrossRef]
  14. Qing, L.; Li, L.; Wang, Y.; Cheng, Y.; Peng, Y. SRR-LGR: Local–global information-reasoned social relation recognition for human-oriented observation. Remote Sens. 2021, 13, 2038. [Google Scholar] [CrossRef]
  15. Ramanathan, V.; Yao, B.; Fei-Fei, L. Social role discovery in human events. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2475–2482. [Google Scholar]
  16. Li, J.; Wong, Y.; Zhao, Q.; Kankanhalli, M.S. Dual-glance model for deciphering social relationships. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2650–2659. [Google Scholar]
  17. Guo, X.; Polania, L.; Zhu, B.; Boncelet, C.; Barner, K. Graph neural networks for image understanding based on multiple cues: Group emotion recognition and event recognition as use cases. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Seattle, WA, USA, 13–19 June 2020; pp. 2921–2930. [Google Scholar]
  18. Zhang, M. Graph neural networks: Link prediction. In Graph Neural Networks: Foundations, Frontiers, and Applications; Springer: Singapore, 2022; pp. 195–223. [Google Scholar]
  19. Gao, J.; Qing, L.; Li, L.; Cheng, Y.; Peng, Y. Multi-scale features based interpersonal relation recognition using higher-order graph neural network. Neurocomputing 2021, 456, 243–252. [Google Scholar] [CrossRef]
  20. Fang, R.; Tang, K.D.; Snavely, N.; Chen, T. Towards computational models of kinship verification. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 1577–1580. [Google Scholar]
  21. Yan, H.; Lu, J.; Zhou, X. Prototype-based discriminative feature learning for kinship verification. IEEE Trans. Cybern. 2014, 45, 2535–2545. [Google Scholar] [CrossRef] [PubMed]
  22. Goel, A.; Ma, K.T.; Tan, C. An end-to-end network for generating social relationship graphs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11186–11195. [Google Scholar]
  23. Wang, Z.; Chen, T.; Ren, J.; Yu, W.; Cheng, H.; Lin, L. Deep reasoning with knowledge graph for social relationship understanding. arXiv 2018, arXiv:1807.00504. [Google Scholar]
  24. Guo, Y.; Yin, F.; Feng, W.; Yan, X.; Xue, T.; Mei, S.; Liu, C.L. Social relation reasoning based on triangular constraints. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 737–745. [Google Scholar]
  25. LYU, Y.x.; HAO, S.; QIAO, G.t.; XING, Y. A Social Recommendation Algorithm Based on Graph Neural Network. J. Northeast. Univ. Nat. Sci. 2024, 45, 10. [Google Scholar]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  27. Gao, S.; Zhou, C.; Zhang, J. Generalized Relation Modeling for Transformer Tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 18686–18695. [Google Scholar]
  28. Soares, L.B.; FitzGerald, N.; Ling, J.; Kwiatkowski, T. Matching the blanks: Distributional similarity for relation learning. arXiv 2019, arXiv:1906.03158. [Google Scholar]
  29. Zhang, J.; Zhang, H.; Xia, C.; Sun, L. Graph-bert: Only attention is needed for learning graph representations. arXiv 2020, arXiv:2001.05140. [Google Scholar]
  30. Diao, C.; Loynd, R. Relational attention: Generalizing transformers for graph-structured tasks. arXiv 2022, arXiv:2210.05062. [Google Scholar]
  31. Chen, D.; O’Bray, L.; Borgwardt, K. Structure-aware transformer for graph representation learning. In Proceedings of the International Conference on Machine Learning, PMLR, Vancouver, BC, Canada, 18–22 June 2022; pp. 3469–3489. [Google Scholar]
  32. Zhou, C.; Han, M.; Liang, Q.; Hu, Y.F.; Kuai, S.G. A social interaction field model accurately identifies static and dynamic social groupings. Nat. Hum. Behav. 2019, 3, 847–855. [Google Scholar] [CrossRef] [PubMed]
Figure 1. GT-SRR Framework. The GT-SRR framework comprises four stages: (1) node generation extracts relation nodes (RNs) and a group node (GN); specifically, the Relation Node Generator (RNG) encodes individual or relational features, while the Group Node Generator (GNG) captures holistic group-level representations; (2) a hierarchical graph captures local and global relational contexts; (3) multi-source features are fused for enhanced representations; (4) a Transformer-based reasoning module outputs final predictions. This unified design effectively models fine-grained interactions and global social semantics.
Figure 1. GT-SRR Framework. The GT-SRR framework comprises four stages: (1) node generation extracts relation nodes (RNs) and a group node (GN); specifically, the Relation Node Generator (RNG) encodes individual or relational features, while the Group Node Generator (GNG) captures holistic group-level representations; (2) a hierarchical graph captures local and global relational contexts; (3) multi-source features are fused for enhanced representations; (4) a Transformer-based reasoning module outputs final predictions. This unified design effectively models fine-grained interactions and global social semantics.
Sensors 25 02992 g001
Figure 2. Node Generation.The node generation module extracts relation nodes (RNs) based on individual, pairwise, and positional features, and generates a group node (GN) from the global scene context to represent group-level semantics.
Figure 2. Node Generation.The node generation module extracts relation nodes (RNs) based on individual, pairwise, and positional features, and generates a group node (GN) from the global scene context to represent group-level semantics.
Sensors 25 02992 g002
Figure 3. Initial Relation Modeling.This layer integrates multi-modal features and performs initial SRR via self-attention.
Figure 3. Initial Relation Modeling.This layer integrates multi-modal features and performs initial SRR via self-attention.
Sensors 25 02992 g003
Figure 4. Contextual Refinement.This layer refines relational features by modeling inter-node dependencies and enhancing contextual semantics.
Figure 4. Contextual Refinement.This layer refines relational features by modeling inter-node dependencies and enhancing contextual semantics.
Sensors 25 02992 g004
Figure 5. F1 score comparison of different models on the PISC-F dataset.
Figure 5. F1 score comparison of different models on the PISC-F dataset.
Sensors 25 02992 g005
Figure 6. F1 score comparison of different models on the PISC-C dataset.
Figure 6. F1 score comparison of different models on the PISC-C dataset.
Sensors 25 02992 g006
Figure 7. GT-SRR Confusion Matrix.GT-SRR shows high accuracy in most classes, with occasional confusion among similar relation types such as Friends, Family, and Couple.
Figure 7. GT-SRR Confusion Matrix.GT-SRR shows high accuracy in most classes, with occasional confusion among similar relation types such as Friends, Family, and Couple.
Sensors 25 02992 g007
Figure 8. Examples of confused identification by different models. This figure presents visual examples comparing predicted relation categories across different models. Each column shows one ground-truth relation type, along with corresponding model predictions and correctness.
Figure 8. Examples of confused identification by different models. This figure presents visual examples comparing predicted relation categories across different models. Each column shows one ground-truth relation type, along with corresponding model predictions and correctness.
Sensors 25 02992 g008
Figure 9. Impact of GGNN iterations on mAP in the PISC-C dataset.
Figure 9. Impact of GGNN iterations on mAP in the PISC-C dataset.
Sensors 25 02992 g009
Table 1. Task split statistics for different SRR levels.
Table 1. Task split statistics for different SRR levels.
TrainingValidationTesting
PISC-C13,142/49,0174000/14,5364000/15,497
PISC-F23,207/72,612500/15051250/3961
Note: 16,826/55,400 indicates that 16,826 training images correspond to 55,400 social relation samples in PISC-F.
Table 2. PISC-F Fine-level Sample Distribution.
Table 2. PISC-F Fine-level Sample Distribution.
Rel.FriFamCouProComNo
Samples12,6867818155220,842652311,979
Note: Fri: Friend; Fam: Family; Cou: Couple; Pro: Professional; Com: Commercial; No: No relation.
Table 3. Performance Comparison on PISC-C and PISC-F Datasets.
Table 3. Performance Comparison on PISC-C and PISC-F Datasets.
ModelPISC-CPISC-F
Int.Non.No.mAPFri.Fam.Cou.Pro.Com.No.mAP
Dual-Glance78.2565.0463.9976.4056.5139.4463.9275.0625.7159.7557.15
GRM76.1268.2868.1374.8438.1460.4161.1867.7246.3354.9756.45
GRRN79.0864.2569.3677.3058.6757.6177.2564.1048.5952.6363.17
Graph-BERT55.9567.9070.2471.3041.3962.0442.1974.5910.7360.7259.79
SRT58.6213.5879.9752.3529.3730.586.6462.008.0374.9545.28
GT-SRR (Ours)80.1073.6964.4880.5260.0861.6048.6381.1218.0863.0664.06
Note: We report per-class accuracy (%) and mean Average Precision (mAP, %) for PISC-C and PISC-F datasets. Int. = Intimate, Non. = Non-intimate, No. = No-relation. Best results are highlighted in bold.
Table 4. Structural Comparison of GT-SRR and Baseline Models.
Table 4. Structural Comparison of GT-SRR and Baseline Models.
ComponentDual-GlanceGRMGRRNGraph-BERTSRTGT-SRR (Ours)
Individual Feature Extraction
Interaction Modeling
(Position + Behavior)
ROI-based Glance Module3D Geometric EncodingGRU-based Edge ModelingGraph-Based InteractionSimple Distance RuleHybrid Interaction
Scene Context Integration
Group Semantics Modeling
Feature Fusion StrategyConcat-
enation
Geometric-Spatial FusionContextual Feature FusionMLP-based FusionDirect ConcatenationCross-modal Attention Fusion
Graph Structure
(RNs + GN)
Static GraphBERT-Enhanced GraphHierarchical Graph
Relational Reasoning MethodGRU-based Local ReasoningGraph-BERT LayersRule-based PredictionGGNN
+
Transformer
Note: ✓ indicates the model includes the corresponding component. The proposed GT-SRR introduces hierarchical graph reasoning and integrates both local and global contextual modeling, achieving a more comprehensive social relationship representation compared to existing models.
Table 5. Performance comparison of individual modules and their combination on PISC-C.
Table 5. Performance comparison of individual modules and their combination on PISC-C.
Int.Non.No.mAP
GGNN70.0230.1161.5255.54
Transformer68.9324.1864.1654.07
GGNN+Transformer (without group feature)83.0952.4162.0374.84
GGNN+Transformer+group feature (ours)80.1073.6964.4880.52
Table 6. Comparison of different graph reasoning modules on PISC-C.
Table 6. Comparison of different graph reasoning modules on PISC-C.
Int.Non.No.mAP
GCN66.6267.9070.0873.52
GAT64.7570.2668.7074.17
GGNN80.1073.6964.4880.52
Table 7. Comparison of mAP performance with varying GGNN iterations on PISC-C.
Table 7. Comparison of mAP performance with varying GGNN iterations on PISC-C.
Iterations (T)12345
mAP (%)75.4277.8680.5278.5378.10
Table 8. Comparison of Different Feature Fusion Strategies on PISC-C.
Table 8. Comparison of Different Feature Fusion Strategies on PISC-C.
Fusion MethodInt.Non.No.mAP
Simple Concatenation72.1245.3361.4970.35
Attention-based Fusion78.3752.8468.0073.24
Gated Fusion80.6256.2056.8473.97
MSA (ours)80.1073.6964.4880.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, D.; Xia, M.; Chang, R.; Kong, X.; Guo, S. GT-SRR: A Structured Method for Social Relation Recognition with GGNN-Based Transformer. Sensors 2025, 25, 2992. https://doi.org/10.3390/s25102992

AMA Style

Huang D, Xia M, Chang R, Kong X, Guo S. GT-SRR: A Structured Method for Social Relation Recognition with GGNN-Based Transformer. Sensors. 2025; 25(10):2992. https://doi.org/10.3390/s25102992

Chicago/Turabian Style

Huang, Dejiao, Menglei Xia, Ruyi Chang, Xiaohan Kong, and Shuai Guo. 2025. "GT-SRR: A Structured Method for Social Relation Recognition with GGNN-Based Transformer" Sensors 25, no. 10: 2992. https://doi.org/10.3390/s25102992

APA Style

Huang, D., Xia, M., Chang, R., Kong, X., & Guo, S. (2025). GT-SRR: A Structured Method for Social Relation Recognition with GGNN-Based Transformer. Sensors, 25(10), 2992. https://doi.org/10.3390/s25102992

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop