Next Article in Journal
Effect of Grain Size Distribution on Frictional Wear and Corrosion Properties of (FeCoNi)86Al7Ti7 High-Entropy Alloys
Previous Article in Journal
Dynamic Balance: A Thermodynamic Principle for the Emergence of the Golden Ratio in Open Non-Equilibrium Steady States
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multimodal Semantic-Enhanced Attention Network for Fake News Detection

National Key Laboratory of Information Systems Engineering, National University of Defense Technology, No. 109 Deya Street, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(7), 746; https://doi.org/10.3390/e27070746
Submission received: 28 April 2025 / Revised: 25 June 2025 / Accepted: 9 July 2025 / Published: 12 July 2025
(This article belongs to the Section Multidisciplinary Applications)

Abstract

The proliferation of social media platforms has triggered an unprecedented increase in multimodal fake news, creating pressing challenges for content authenticity verification. Current fake news detection systems predominantly rely on isolated unimodal analysis (text or image), failing to exploit critical cross-modal correlations or leverage latent social context cues. To bridge this gap, we introduce the SCCN (Semantic-enhanced Cross-modal Co-attention Network), a novel framework that synergistically combines multimodal features with refined social graph signals. Our approach innovatively combines text, image, and social relation features through a hierarchical fusion framework. First, we extract modality-specific features and enhance semantics by identifying entities in both text and visual data. Second, an improved co-attention mechanism selectively integrates social relations while removing irrelevant connections to reduce noise and exploring latent informative links. Finally, the model is optimized via cross-entropy loss with entropy minimization. Experimental results for benchmark datasets (PHEME and Weibo) show that SCCN consistently outperforms existing approaches, achieving relative accuracy enhancements of 1.7% and 1.6% over the best-performing baseline methods in each dataset.

1. Introduction

The exponential expansion of social media platforms has generated an abundance of digital content while simultaneously diversifying information access channels. News has also spread rapidly in multimodal format with the universal benefit of comprehensive content [1], with multimodal news being more eye-catching and persuasive compared to text-only news [2]. However, this simultaneously facilitates the growth of a considerable quantity of fake news, defined as misinformation that is intentionally fabricated or distorted to mislead or deceive news consumers [3]. From an information-theoretic standpoint, the increasing volume and diversity of accessible information directly increase the uncertainty, complexity, and entropy inherent in fake news detection tasks. Consequently, automated fake news detection systems have been developed to identify misinformation on social media platforms, aiming to mitigate its harmful societal effects while ensuring that users receive authentic information.
The essence of the fake news detection problem is regarded as a binary classification problem, where the traditional content-based methods generally determine the authenticity of fake news by analyzing its content through different modalities. Initially, automated fake news detection approaches generally used the textual content to complete semantic feature modeling in order to make judgments [4]. Due to multimedia technology decreasing the accuracy of unimodal detection approaches, researchers then attempted to extract multimodal features for fake news detection [5,6]. Specifically, multimodal learning models [7] have been developed to comprehensively capture both textual and visual features through variational encoders [8] and self-attention mechanisms [9], reducing the cross-modal fusion gap [10]. Moreover, because the background features of the news entity may impact detection accuracy, knowledge graphs have been introduced as additional information to support detection frameworks [11].
The above-mentioned methods simply take into consideration the content without mining the deeper intrinsic structural relations [12]; thus, structure-aware methods have been designed to solve such issues. Specifically, traditional neural network approaches such as RNN and CNN have been utilized to explore the potential structural features. In detail, social relationships are collected for application in the field of fake news detection [13]; these relationships generally consist of the comments following news, the forwarding users, and their additional comments. Moreover, GAT and GCN have proved to be effective at reflecting the structural features of social relations and accurately obtaining representations of node features [14].
Although current approaches have achieved satisfying performance, there remain two main shortcomings. First, simply extracting the feature vectors from text and images fails to capture the entire semantic information, potentially leading to semantic bias in the fusion process of cross-modal features. For example, in the second news story in Figure 1, hardly any clues can be found to prove that it is fake only through the text and image. However, negative opinions arguing that the news may be fake can be read in the comments of different users. Second, the noise in the social relation graphs may lead to bias during information aggregation of nodes by traditional GNNs, resulting in decreased model performance. For example, it is evident that there is a common user (user 2) commenting on all news stories in Figure 1, and their comments on both the first and second news stories take the same position. Obviously, this means there is a high probability that the last news story is fake, even though its text and image provide no evidence to prove that it is. Similarly, there may be some insignificant comments below each news story, such as the first comment on the first news story, the third comment on the second news story, etc. Overall, these comments are irrelevant to the news and could influence the performance of models; thus, they should be removed.
All of the above discussions are centered on the multimodal fake news detection task. Multimodal fake news detection is essentially a simple binary classification task performed to identify whether posts are true or false. Specifically, it aims to map posts to the most likely labels by learning their rich feature inputs, including unimodal features, multimodal features, and social relations. Compared with previous binary classification problems, its uniqueness lies in the use of features from different modalities to improve detection efficiency. For example, text, images, and social relations are all modalities, and the interactive fusion generated among them constitutes multimodal feature fusion, which is a non-negligible factor affecting the classification effect.
In this study, we propose a novel Semantic-enhanced Cross-modal Co-attention Network (SCCN) for multimodal fake news detection. It not only utilizes entity extraction to semantically enhance news but also enables the effective integration of text, images, and social relations. To better capture the structural features, we first model social relations as a graph and make some improvements to GAT. On one hand, we detect latent connections among nodes to enhance the social relation graph while simultaneously reducing noise interference. On the other, we introduce a sign strategy to capture neighbor-to-node correlations and amplify the positive–negative feedback to optimize the node representation. Unlike social relations, we encode text and images directly in addition to enhancing them with entity features. Next, we map their feature vectors to the same semantic space by performing cross-modal alignment operations on different modalities. Finally, self-supervised loss and co-attention mechanisms are introduced to reduce noise and improve the fused feature representation for cross-modal fusion. By learning the social relation graph representation and cross-modal fusion, we can enhance the capability of the model for fake news detection.
The main contributions in this study are summarized as follows:
  • We propose a novel Semantic-enhanced Cross-modal Co-attention Network for multimodal fake news detection, which focuses on utilizing the entity features for semantic enhancement and exploring the social relation graph.
  • We narrow the gap between different modalities by combining the entities and semantics of different modalities. In addition, we introduce self-supervised learning along with co-attention mechanisms for denoising, ultimately enhancing news semantics.
  • We introduce social relations as the structural features and design an improved GAT to process the social relation graph, thereby improving the representations from learning social relation graphs through enhanced neighborhood aggregation.
The rest of this paper is organized as follows: Section 2 reviews the existing approaches for fake news detection and analyzes their shortcomings; Section 3 introduces our proposed SCCN approach; Section 4 provides the experimental details, including the research questions, datasets, baselines, and experimental setup; Section 5 analyzes the results of experiments and discusses the findings of our experiments from different perspectives; finally, Section 6 concludes our current study and outlines prospects for future studies.

2. Related Studies

Existing studies on multimodal fake news detection can be categorized into two types: those with semantic-based features and those with structural-based features.
Semantic-based features are expressed not only in text but also in images [15]. This is due to evidence showing that multimodal approaches outperform unimodal approaches in fake news detection [16]. In detail, Wang et al. [2] proposed EANN, an end-to-end model capable of capturing features that are robust to event variations. However, its reliance on annotated data and computationally intensive training processes could limit scalability and training stability. Another end-to-end network is MVAE [1], which utilizes a bimodal variational framework to learn a probabilistic latent variable model. However, its dependence on paired multimodal data and complex architecture could limit its applicability in scenarios with incomplete or single-modality content. In addition, QSAN [17] employs quantum-inspired text analysis and a unique attention system to accurately identify rumors while making its decisions understandable through clear explanations and by highlighting relevant user comments. Due to its complex design and quantum-based computations, this model requires more computing power than traditional detection approaches. Furthermore, Zhou et al. [15] evaluated both the internal relationships within each modality and the cross-modal similarities between text and images for fake news detection. Chen et al. [18] solved the misclassification problem caused by inconsistency across modalities with cross-modal ambiguity learning. In [18], an attention-guided module was proposed to realize more accurate modal alignment and increase the interpretability of modal relationships [19].
However, the aforementioned approaches ignore the rich structural features in news and the plentiful factual information in knowledge graphs. To better explore the structural features in news, researchers have explored ways of representing news content through graph-based structures [20]. For example, EBGCN [21] boosts detection accuracy by dynamically adjusting unreliable edges in rumor propagation graphs; however, its Bayesian framework requires more training data to stabilize probabilistic inferences compared to deterministic models. The GLAN model [13] effectively merges local semantic relationships with global structural patterns through attention mechanisms, achieving superior rumor detection performance. However, its effectiveness can be compromised by noisy data in early propagation phases, and requires precise hyperparameter tuning to maintain optimal performance. Based on Yuan et al. [13], Qian et al. [22] integrated the modalities to capture important information in posts and further utilized a learning principal of knowledge-aware multimodal graphs to adaptively learn features from news stories. MPFN [5] improves fake news identification through the layered integration of text and image data, but requires greater processing power due to its advanced architecture and multi-stage blending components. Finally, MFAN [10] advances rumor detection through its novel integration of multimodal features and attention mechanisms; yet, its effectiveness depends on comprehensive data coverage across all modalities, and requires substantial computational resources for optimal operation.

3. Approach

In this section, we detail the proposed SCCN, the framework of which is presented in Figure 2. As demonstrated in Figure 2, SCCN comprises four fundamental components. First, data preparation (Section 3.1) includes the preparation of images, text, and relations along with the problem definition. Second, information extraction and feature encoding (Section 3.2) extracts and encodes the entities from image and text, respectively. In addition, the feature vectors of visual, textual, and social relations are obtained by the corresponding encoders. Third, information enhancement and cross-modal fusion (Section 3.3) enhance the text and image features through entity extraction; meanwhile, an improved GAT is utilized to refine the characteristics of social relations. In addition, we utilize a co-attention mechanism to realize the integration of images, text, and social relations. Finally, optimization and classification (Section 3.4) is mainly used to optimize the cross-modal fusion features and realize fake news detection through classifiers.
Below, we denote the data by mathematical symbols for convenience of description. Specifically, let D = X , Y be a set of news on social media containing multimodal features and their labels, which can be denoted as X = x 1 , x 2 , , x n and Y = y 1 , y 2 , , y n , respectively. For each news x j = t j , v j , u j , c j X , which is a quadruple, t j , v j , u j , and  c j represent the text, images, users, and comments, respectively. Moreover, c j = c j 1 , c j 2 , , c j i denotes the set of comments on news x j , where each comment c j i corresponds to a user u j i . To better understand the mathematical symbols mentioned above, we take Figure 1 as an example to illustrate the specific meaning of each symbol corresponding to Figure 1. First, the three posts and their labels can be represented as X = x 1 , x 2 , x 3 and Y = y 1 , y 2 , y 3 , respectively. Then, we focus on the first post ( x 1 ), where t 1 is the text “Syrian girl selling chewing gum in the street of Jordan”, v 1 is the image in this post, u 1 contains two users (user 1 and user 2), and  c 1 includes three comments posted by two users. To highlight the connections between news, users, and comments, we build a heterogeneous graph G = V , A , E , called a social relation graph, where V represents the set of nodes, including the above-mentioned three types of nodes. Moreover, A R | V | | V | is an adjacency matrix which reflects the connections between nodes, i.e., the value is 1 if there is a link between nodes and 0 if there is none, with E as the set of edges between nodes.
In this paper, we define fake news detection as a binary classification task. The value of label y is 0 or 1, where y = 1 represents fake news and y = 0 legitimate news. The objective of our study is to predict the label of a given news item D through our proposed SCCN model.

3.1. Data Preparation

Real-world news stories posted on social media contain rich content and structural features such as images, text, and social relations [23]. Specifically, each news story contains various entities and covers different abundant modalities. The special relationships between different modalities, including images, text, and social relations, provide us with ideas for fake news detection.
In the data preparation section, we extract the data that our proposal requires from the datasets, such as images, text, users, and comments. For the users and their comments on the news, we visualize the relationship between them in a graph, which can be called the social relation graph. To infer hidden links and remove noisy links, it is necessary to process the social relation graph. According to the homogeneity of graph neural networks, we assume that similar nodes are more likely to be associated with each other than dissimilar nodes [24]. Therefore, it is possible to infer the links between nodes with high similarity and remove noisy links between nodes with low similarity by calculating the similarity of features between different nodes.
Specifically, for the embeddings of nodes in G , the vectors of text in news and the vectors of sentences in comments are used as the initial embeddings of news nodes and comment nodes, respectively. Then, we adopt the mean value of all comment embeddings published by one user as this user’s embedding vector. To facilitate the subsequent computation, we construct the node embedding matrix B R | V | d , where each row represents the embedding of one node and d represents the dimension size of the embeddings.
After that, we use the cosine similarity to calculate the similarity coefficient α i j between nodes n i and n j , as follows:
α i j = b i · b j b i b j ,
where b i and b j denote the embeddings of nodes n i and n j , respectively.
Based on the findings of a previous study [10], we argue that if there is currently no link between nodes, the potential link may exist if the similarity coefficient is above 0.5. Moreover, if there is already a link between nodes, this link can be regarded as noise if the similarity coefficient is below 0.2. This can be formulated as follows:
δ i j = 1 , i f α i j < 0.2 0 , i f 0.2 α i j 0.5 1 , o t h e r w i s e ,
where δ i j is a transition variable that measures the similarity between nodes n i and n j .
Next, we improve the original adjacency matrix A based on δ i j , i.e., removing the noisy links and adding the potential links, which can be formulated as follows:
a i j = 0 , i f δ i j = 1 o r i f δ i j = 0 a n d a i j = 0 1 , o t h e r w i s e
where a i j and a i j are the elements in the initial and improved adjacency matrices A and A , respectively.
Up to this point, we have enhanced the performance of the model and reduced its training time by removing distracting noise such as malicious comments and bot users from the social relation graph. In this way, we obtain the improved social relation graph G = V , A , E and input the image, text, and graph G to the next step.

3.2. Information Extraction and Feature Encoding

In the information extraction and feature encoding module, we adopt a synchronous processing mode for processing the input data to obtain the required feature vectors, i.e., learning the text and image representations and the social relation graph representation.

3.2.1. Text and Image Representations

First, we extract the corresponding entities from the image and text of each news story using the existing entity linking tool TAGME (https://sobigdata.d4science.org/group/tagme/ (accessed on 27 March 2023)), which can precisely identify the entities and provide their confidence coefficient. Then, we retain the entities with confidence coefficients larger than 0.2 as the final entity extraction results. In addition, a pretrained TransE model [25] attaches them to the Freebase [26] knowledge graph to obtain the background knowledge features e I and e T as the entity embeddings.
Subsequently, we extract the feature vectors from the text and image using the respective encoders. Specifically, we utilize the pretrained BERT [27] and ResNet-50 [28] as the encoders for text and image, respectively. However, the length of the text generally varies for different news stories. Therefore, to facilitate subsequent operations, we set the text of different news to be of the same length, i.e., L, through padding or truncating, which can be denoted as follows:
T i R L d = t 1 i , t 2 i , , t L i
where d is the dimension of the word embeddings and  T i is the text embedding of the i-th news item, which consists of L word embeddings.
Next, this embedding sequence is fed into Bi-LSTM [29] to obtain the text feature vector a T , as follows:
a T = W T ( B i L S T M ( T i ) ) + b T
where W T is the learnable weight matrix and  b T is the bias vector.
For the image, we extract the output of the second-last layer of ResNet-50 [28] and subsequently feed it through a fully connected layer to generate a feature vector a I with the same dimensions as the text feature a T . This can be formulated as
a I = s i g m o i d W I R I ,
where R I is the output of the second-last layer of ResNet-50 and  W I is the weight matrix of the fully connected layer.

3.2.2. Social Relation Graph Representation

Unlike the above two types of representations, the social relation graph contains abundant structural features. Inspired by Velickovic et al. [30], Graph Attention (GAT) networks can capture the graph structure features for graph processing. However, traditional GAT suffers from poor interpretability and inconsistent performance across datasets when the processed graphs are complicated [24]. Thus, we propose an improved GAT to capture the correlations of the neighbor nodes for better graph feature representation, the pipeline of which is shown in Figure 3.
  • Attention Weight
Specifically, for node n i and its set of neighbor nodes N i = n i 1 , n i 2 , , n i k in graph G , we first calculate the attention weights E i = ε i 1 , ε i 2 , , ε i k , where the element ε i k E denotes the attention weight between n i and n i k . In detail, it can be seen in Figure 3 that we integrate two common attention mechanisms of the traditional GAT, i.e., the single-layer neural network and the dot product [30]. This can be formulated as
e i k = a ˜ W b i W b k · s i g m o i d [ W b i ] · W b k ε i k = L e a k y R e L U e i k ,
where ‖ is the concatenation operation, a ˜ is a parameter in the single-layer neural network, W is a learnable weight matrix, and  b i and b k are the embeddings of node n i and its neighbor node n i k N i , respectively.
  • Attention Coefficient
Next, to facilitate the following calculation, we employ the softmax function to normalize the obtained attention weight set E i . In addition, we note that some of the attention weights become very small after normalization, which means that these neighbor nodes have very little influence on node n i . Such situations generally occur when the attention weights are negative. In fact, the attention weights objectively reflect the potential influence of the neighbor nodes on node n i , which include both positive and negative influences, and these influences’ reflection in the values of the attention weights is correspondingly positive or negative. However, the direct application of the softmax function ignores the negative influence, which we also need to focus on.
For instance, given a specific node n p , we obtain its attention weights as E p = 0.7 , 0.2 , 0.1 , 0.2 , 0.8 ; the subsequently obtained attention coefficients after normalization are Φ p = 0.36 , 0.22 , 0.20 , 0.14 , 0.08 . It can be observed that the neighbor node with attention weight ε p 5 = 0.8 has a normalized attention coefficient ϕ p 5 = 0.08 , which indicates that its contribution to the output is almost negligible. This large negative influence implies that the embeddings of these two nodes are in opposite directions, which may be beneficial for fake news detection. This situation may be an instance of “cheating” or “camouflage” [14]. In certain cases, there may be paid-for real comments on fake news stories or malicious comments added to discredit real news [14].
As a result, inspired by [17], we introduce a sign mechanism to correctly handle the positive and negative relations between nodes. In detail, we obtain E ˜ i after taking the opposite number of attention weights E i for node n i , i.e.,  E ˜ i = E i = ε ˜ i 1 , ε ˜ i 2 , , ε ˜ i k . Afterwards, normalization is implemented with the softmax function for the two attention weights E i and E ˜ i to obtain the corresponding attention coefficients. This can be formulated as follows:
ϕ i j = s o f t m a x j ε i j = e x p L e a k y R e L U ε i j Σ k | N i | e x p L e a k y R e L U ε i k ϕ i j = s o f t m a x j ε ˜ i j = e x p L e a k y R e L U ε ˜ i j Σ k | N i | e x p L e a k y R e L U ε ˜ i k
where ϕ i j Φ i and ϕ ˜ i j Φ i are the two attention coefficients.
  • Graph Feature
To completely capture the interactions between nodes, we obtain the weighted aggregation of the neighbor nodes of n i with Φ i and Φ i , respectively. Subsequently, the above vectors are concatenated and fed into a fully connected layer to obtain the final vector representation of node n i . Note that we utilize multihead attention in order to adapt to the complex graph structure in this process. This allows the model to fully take into account the correlation and importance between different nodes, resulting in improved expressive ability. The multihead attention can be formulated as follows:
b i K = k = 1 K σ W k Φ i k B i Φ i k B i
where K is the number of heads, σ is the activation function, W k is the weight matrix of the fully connected layer of the k-th head, and  B i is the embedding matrix of the neighbor nodes.
To this point, the final embeddings of all nodes are calculated by Equation (9) and denoted as the node embedding matrix B . Finally, the multihead attention mechanism [31] is employed to acquire the features of the social relation graph as follows:
G = k = 1 K σ B
where K is the number of heads and the i-th column of G denotes the graph feature of the i-th news item.

3.3. Information Enhancement and Cross-Modal Fusion

In this section, we perform information enhancement and cross-modal fusion using the feature vectors obtained from Section 3.2 for text, image, and the social relation graph. For the feature vectors of the text and images, we use their entity embeddings to implement self-information enhancement. Specifically, taking an image as an example, we concatenate a I and e I before feeding them into a multilayer perceptron to obtain the information-enhanced feature vector Z I . Moreover, the information-enhanced feature vector Z T of the text can be calculated in the same way. For the social relation graph, we acquire the feature vector Z R with the same dimension as Z I by feeding G into a multilayer perceptron [31]. This can be formulated as
Z I = σ W I a I e I + b I Z T = σ W T a T e T + b T Z R = σ W R G + b R ,
where σ is the activation function, W I , W T , and  W R are the learnable weights matrix, and b I , b T , and  b R are the bias vectors.
It cannot be ignored that performing the cross-modal fusion operations inevitably causes intrinsic loss of information between modalities, which should be intrinsic to the representation of different modalities in the original news story. This leads to the features extracted from disparate modalities potentially exhibiting substantial semantic gaps. To address this issue, a novel cross-modal alignment with self-supervised loss is introduced to refine the feature representations. For example, we map the produced feature vectors Z I and Z T to the same semantic space as, follows:
Z I ^ = W I ^ Z I Z T ^ = W T ^ Z T
where W I ^ and W T ^ are the learnable parameters. After this, we adopt the MSE loss to narrow the distance between Z I ^ and Z T ^ :
L a l i g n T I = i = 1 n ( Z I ^ Z T ^ ) 2 n
where n is the total number of news stories.
Similarly, we can map Z T and Z R along with Z I and Z R to the same semantic space and calculate their MSE loss as L a l i g n T R and L a l i g n I R , respectively. Then, we add the above three losses to obtain the final loss of the cross-modal alignment, as follows:
L a l i g n = L a l i g n T I + L a l i g n T R + L a l i g n I R .
Up to this point, we have obtained the cross-modal aligned graph feature vectors of images, text, and social relations as Z I ˜ , Z T ˜ , and  Z R ˜ , respectively. In addition, considering that there are three feature vectors from different modalities, we need to integrate their embeddings in order to improve the credibility before detection. Inspired by Lu et al. [32], we adopt a cross-modal fusion approach with a co-attention mechanism.
Specifically, we adopt the co-attention mechanism between every two modalities to obtain the cross-modal features containing the key information between the two corresponding modalities. Taking the fusion of text and images as an example, we use Q T = Z T ˜ W T Q , K I = Z I ˜ W I K , and  V I = Z I ˜ W I V to calculate the query matrix, key matrix, and value matrix, respectively. In this case, W T Q , W I K , and  W I V R d d K are the linear transformation matrices, where K is the number of heads and d is the dimension of the feature vectors.
Then, we can produce the cross-modal feature f T I between the text and images, calculated as follows:
f T I = k = 1 K s o f t m a x Q T K I d V I W T I
where W T I represents the linear transformation matrix.
However, it is evident that f T I is actually the visual feature enhanced by a textual feature, which does not fully reflect the characteristics and relations between the two modalities. Therefore, we exchange the representation of text and image in Equation (15) to obtain another cross-modal feature f I T , which is the textual feature enhanced by the visual feature. For the other two groups of modalities, we obtain the mutually reinforced cross-modal features in the same way. For convenience of distinction, we denote these as f T R , f R T , f I R , and  f R I , respectively. Finally, we concatenate them as the final multimodal fusion feature, as follows:
Z = c o n c a t f T I , f I T , f T R , f R T , f I R , f R I .

3.4. Optimization and Classification

In the optimization and classification module, we feed the final multimodal fusion feature Z into a fully connected layer to predict the labels of news stories, as follows:
y ^ = s o f t m a x M L P Z ,
where y ^ denotes the predicted scores.
In our model, the entropy can be used to measure the dispersion and uncertainty of the experimental results. In the proposed framework, high entropy reflects significant divergence between detection results and ground-truth labels, whereas low entropy indicates stable convergence. Information theory offers quantitative frameworks for three critical aspects of probabilistic systems: entropy for uncertainty assessment, KL divergence for distributional differences, and mutual information for dependency measurement. Inspired by the theoretical understanding of entropy in information theory, we adopt the cross-entropy loss function as the loss function for this binary classification problem. The optimization strategy of SCCN explicitly targets entropy minimization through cross-modal feature learning, as follows:
L c l s = y l o g y ^ 1 y l o g 1 y ^
where y indicates the ground truth of the fake news detection label.
Considering that cross-modal alignment is conducted to narrow the semantic gaps between modalities, the generated loss may contribute less to the final classification. Thus, we introduce two parameters, λ a and λ b , to respectively regulate L c l s and L a l i g n in the total loss function. We then combine L c l s and L a l i g n to compute the final loss function of SCCN, as follows:
L t o t a l = λ a L c l s + λ b L a l i g n .
However, we find that the text of some news items does not strictly follow the grammar rules in practice, which may reduce the efficiency of our model to some extent. To resolve this issue, we add an adversarial perturbation mechanism, i.e., the Projected Gradient Descent (PGD) [33], to enhance the robustness of our model when extracting the text embeddings. The effectiveness of PGD has been proven by Madry et al. [33]. In detail, we compute the gradients of the text features in training and add perturbations to the text features. We then recalculate its gradient and repeat the above process T times. Finally, all the adversarial perturbations are accumulated into the original gradient and the parameters are updated.
More details of the PGD are provided in Algorithm 1. Specifically, in this algorithm, s i g n ( g ) is a function that returns the sign of the gradient g, indicating the direction in which the input should be perturbed in order to maximize the loss. Moreover, c l i p and p r o j e c t are functions which respectively ensure that the perturbations are within acceptable bounds and that the perturbed inputs are projected back into the valid data space.
Algorithm 1 Procedure of PGD
Require: 
The original dataset D = x i , y i i = 1 N . The initialized model parameter θ . Set the epoch number T, the step size β and the perturbation limit δ .
Ensure: 
The model parameter θ .
1:
Initialize θ randomly.
2:
for each epoch t = 1 to T do
3:
    for each batch x b a t c h , y b a t c h in D do
4:
          Calculate the gradient of the loss function:
5:
           g = x J θ , x b a t c h , y b a t c h .
6:
          Compute the perturbation in the direction of the gradient:
7:
           p = β s i g n g
8:
          Apply the perturbation to the data, ensuring it does not exceed the limit δ :
9:
           x a d v = x b a t c h + p     x a d v = c l i p x a d v , x b a t c h δ , x b a t c h + δ
10:
        Project the perturbed data back into the valid data space:
11:
         x a d v = p r o j e c t x a d v , 0 , 1
12:
        Update the model parameter θ :
13:
         g θ = θ J θ , x a d v , y b a t c h     θ = θ β g θ
14:
    end for
15:
end for
16:
return the updated θ .

4. Experiments

In this section, we first introduce the datasets in the social media field used to evaluate performance of SCCN in Section 4.1. Next, the discussed baselines are listed in Section 4.2. Finally, we describe the implementation details of our experiments in Section 4.3.

4.1. Datasets

The performance of the proposed SCCN and the baselines are evaluated using two widely used datasets: the PHEME [34] and Weibo [35].
The PHEME Dataset (https://figshare.com/articles/PHEME_dataset_of_rumours_and_non-rumours/4010619 (accessed on 24 October 2016)) is a collection of posts on Twitter about multiple breaking news stories and their related information. Similarly, the Weibo Dataset (https://drive.usercontent.google.com/download?id=14VQ7EWPiFeGzxp3XC2DeEHi-BEisDINn&export=download (accessed on 22 May 2021)) is a Chinese dataset that collects a large quantity of posts from the most widely used social media platforms within China. Both of the above datasets contain rich information derived from news stories, including text, images, users, and comments. In our model, we mainly focus on text, images, and social relation networks to detect fake news. Thus, we perform data cleaning on the raw datasets. We delete those news stories that have only unimodal information such as text or images, as our model is focused on multimodal conditions. Moreover, we make sure that each news story contains at least one piece of information that we need in our model, including text, images, and social relations. The processed statistics relating to the PHEME and Weibo datasets are shown in Table 1.

4.2. Baselines

We validate the effectiveness of our proposed SCCN by comparing it with the following competitive baseline models:
  • EANN [2]: EANN utilizes a cross-modal feature extractor and a fake posts detector to support fake news detection, which can derive event-invariant features that make it easier to detect newly emerging events.
  • MVAE [1]: MVAE uses a bimodal variational auto-encoder to model images and text in order to achieve classification.
  • QSAN [17]: QSAN incorporates quantum-driven text encoding along with a signing mechanism within its framework, which can utilize conflicting information to provide clues for detection. In addition, this method is interpretable.
  • SAFE [15]: SAFE is a fake news detection approach that emphasizes the similarity between textual and visual content more than other methods.
  • EBGCN [21]: EBGCN identifies unreliable relationships existing in rumors and enables the detection of fake news by training an edge consistency framework.
  • GLAN [13]: GLAN is a global–local network that captures structural information for fake news detection by jointly coding global and local information.
  • MPFN [5]: MPFN is able to recognize the level of information represented in the different modalities and use this to build a strong hybrid modality.
  • KMAGCN [22]: KMAGCN is an adaptive graph convolutional network that converts posts into graphs to capture discontinuous semantic relations.
  • MFAN [10]: MFAN introduces the element of comments in posts while considering the complement and alignment between different modalities for better integration.

4.3. Implementation Details

Following previous studies Zheng et al. [10], we divide the PHEME and the Weibo datasets into a training set, validation set, and test set in a ratio of 7:1:2, respectively. We then sequentially initialize the word embeddings in our model with vectors of dimension 300, i.e., d = 300 . Those words that are not in the pretrained word vectors are initialized from a uniform distribution. The number of heads K is set to 8 for all the multi-head attention mechanisms involved in this paper. The values of λ a and λ b in Equation (19) are 2.15 and 1.55, according to previous experience. In our experiments, the learning rate during data training is set to 0.002 and the max length L = 50 . The convolutional kernel sizes are configured to (3, 4, 5), with 100 kernels allocated for each size. The final result of the model is determined by taking the average of the results obtained from five successive runs. Finally, we measure the performance of our model using the common indicators of accuracy, precision, recall, and F1 score.

5. Results and Discussion

In Section 5.1, we evaluate the overall performance of our proposed SCCN and the baselines. Following that, we conduct an ablation study to investigate the impact of each component within our proposal in Section 5.2. We then perform a quantitative analysis to visually evaluate the performance of the SCCN and its variants in Section 5.3 based on an ablation study, as well as comparing it to several baselines for further verification of its validity. Next, we perform a convergence analysis to verify the overfitting risk of the proposed SCCN in Section 5.4. In Section 5.5, we analyze the influence of hyperparameters on our model from different angles. Finally, to further understand our SCCN model, we conduct case studies in Section 5.6.

5.1. Overall Performance

First of all, we evaluate the fake news detection performance of the SCCN as well as the baselines in Table 2.
For the discussed baselines, we find that the performance of the models that consider social relations is much better than that of the other methods. Note that the baselines listed here are all multimodal fake news detection methods. In detail, the KMAGCN model achieves up to 3.1% and 1.1% performance improvement over the MPFN model on the PHEME and Weibo datasets, respectively. This illustrates the importance of capturing structural information from social relations for fake news detection.
Next, we zoom in on the comparisons of our SCCN against the baselines. Generally, it is evident that SCCN exhibits superior performance in comparison to all the baselines on both the PHEME and Weibo datasets. For instance, SCCN achieves respective performance improvements of up to 1.7% and 1.6% over the best baseline method on the PHEME and Weibo datasets. We consider that the advantage of SCCN is due to its ability to combine the entities and semantics of different modalities along with its optimized representation of social relation graphs.
Because the PHEME dataset is sourced from Twitter, most of the posts are related to specific things and the correlations between them are minor; as a result, this dataset is more likely to result in overfitting. Nevertheless, SCCN performs better on both datasets. This advantage can be explained by the fact that SCCN leverages a co-attention mechanism to enhance text and image representations with entity features. Thus, it can refine feature vectors, which is more conducive to subsequent fusion. In addition, we introduce social relation graphs consisting of news, users, and their comments in order to extract the potential structure features from news items, allowing SCCN to further complete the feature representation.
To further explore the practical deployability of SCCN, we use the three metrics of time, inference latency, and memory. Table 3 shows the time, inference latency, and memory of SCCN on the PHEME and Weibo datasets, respectively. Based on these results, all three capability metrics for our model are within a reasonable range of intervals, showing good potential for practical deployment.

5.2. Ablation Study

Next, we perform an ablation study to gain deeper insights into each component of our proposed SCCN and the different modal embeddings contained within it.

5.2.1. Effect of Modules

To assess the impacts of key components of SCCN on detection performance, we construct the following model variants: (1) SCCN w/o DPS: Removes the processing of all raw data in the data preparation section and replaces it with the direct utilization of raw data as input. (2) SCCN w/o SRG: Eliminates the social relations component, considering only the text and image modalities for final fake news detection. (3) SCCN w/o IEM: Replaces the output of the information enhancement modules with the original embedding vectors for the subsequent cross-modal fusion operation. (4) SCCN w/o CMF: Replaces the output of the cross-modal fusion module with the sum of the feature vectors for text, images, and social relations.
Table 4 shows the results of the ablation experiments. We use the accuracy and F1 score indicators, as these most clearly reflect the performance of the different models. Overall, the different SCCN variants all underperform the original model in terms of accuracy and F1 score, suggesting that each module in our model makes a significant contribution to the final performance. Specifically, we summarize the following points:
  • Comparing the four variants of SCCN, we find that the SCCN w/o DPS, SCCN w/o SRG, and SCCN w/o IEM variants all show obvious performance drops, indicating that introducing data processing, social relations, and information enhancement strengthens the performance of our model. From an information-theoretic perspective, greater performance degradation indicates higher uncertainty and entropy in the variants, demonstrating enhanced effectiveness of our model architecture when incorporating the proposed modules. Moreover, the results of the ablation study indicate that combining text, images, and social relations can facilitate cross-modal feature fusion in a way that is crucial for fake news detection.
  • The performance decrease for the SCCN w/o CMF variant proves that the cross-modal fusion module implemented with the co-attention mechanism helps to improve the performance of our model. Furthermore, all variants show similar performances on both the PHEME and Weibo datasets, while the complete SCCN model performs better on Weibo than PHEME dataset, demonstrating that these modules play a greater role in the Weibo dataset.

5.2.2. Effects of Different Types of Modal Embeddings

To more deeply investigate the effects of different modalities and fusion strategies on detection efficacy, the next experiment removes specific modalities from the model. In Figure 4, -Text indicates the removal of the text modality, -Image indicates the removal of the image modality, and -Relations indicates the removal of the social relations modality.
As shown in Figure 4, the full SCCN model outperforms all other variants with different types of modal embeddings. Moreover, in comparing the three variants it can be seen that the detection effect is greatly reduced by removing the text modality, which means that the textual content is more important for fake news detection. We consider the cause of this phenomenon to be the fact that the textual description usually reflects the essential features of a post more accurately, while visual features and social relations may be contaminated by less critical or even irrelevant information [36]. Similarly, -Image outperforms -Relations, which is due to the social relation graph containing comment nodes. Moreover, comment nodes contain information about users’ comments, which is also represented as textual content. Thus, the text modality assumes a more prominent role in detection.

5.3. Quantitative Analysis

The ablation experiments allow us to obtain a basic view of the respective impacts of the different modules on the whole model. To more intuitively show the contribution of each module in our model, we visualize the SCCN and its variants on the PHEME dataset. In detail, we adopt heat maps to realize data visualization for the PHEME dataset. Considering that most of the posts in the PHEME dataset are related to some specific events, we randomly select twenty news items from the dataset, including ten pieces of real news and ten pieces of fake news, then input them into the SCCN model and its variants to obtain Figure 5. Meanwhile, we use t-SNE visualization to perform an in-depth analysis of the features learned by SCCN and several baselines, with the results shown in Figure 6.
Figure 5 compares the fake news detection capabilities of our model and its variants using the PHEME dataset. We observe that SCCN provides clear boundaries between real and fake news while showing similarities between the classes. However, the rest of the methods tend to blur the boundaries between real news and fake news, meaning that their performance is poorer and that they are less likely to recognize fake news. The visualization results here are similar to those of the ablation experiments in Section 5.2, with both showing that the performance of the SCCN w/o SRG and SCCN w/o IEM variants is inferior to that of SCCN w/o DPS and SCCN w/o CMF. This suggests that our visualization experiments are consistent with the performance results of the ablation experiments, providing a certain degree of credibility. Social relations are not considered by the SCCN w/o SRG variant, which only integrates text and image features are; this may result in some critical information hidden in the comments being missed. Moreover, the lack of structure information can lead to difficulty in excavating the correlations between news stories to some extent, resulting in reduced model performance. Compared to the SCCN, the SCCN w/o IEM variant not only fails to utilize entity features for the information enhancement of text and image modalities but also struggles to process the information in the social relation graphs without the improved GAT. This can lead to the omission of important information when processing modal information; in other words, it can amplify the noise generated by non-essential information, which can seriously affect model performance. In general, each module of our proposed model plays a favorable role in fake news detection in both the ablation experiments and the quantitative analysis.
Unlike the above format, we next use the t-SNE visualization method to visualize the test set data of the Weibo dataset, as shown in Figure 6. We choose the three most representative baselines for comparison with our model, of which only MPFN does not utilize social relations. Specifically, the boundary between real news and fake news is clear and obviously noticeable with SCCN, indicating that its classification effect is the best among the different models. Similarly, the boundaries obtained with MPFN seem rather indistinct compared to the first three models. It is hypothesized that this phenomenon may be attributable to underutilization of certain social relationships within the dataset, such as users and comments. From an information-theoretic perspective, when feature representations from distinct categories appear in close proximity within the visualization space, this suggests reduced inter-class informational divergence and increased mutual information between categories, both of which represent undesirable characteristics for effective classification. Additionally, the visualization results show general consistency with those in Table 2, demonstrating the rigor of our experiments.

5.4. Convergence Analysis

To assess the contributions of L c l s and L a l i g n during training and verify the convergence of our SCCN model, we generate average loss curves for SCCN in Figure 7. It is evident that the total loss results on both the PHEME and Weibo datasets show a decreasing trend before eventually stabilizing. This is an effective indication that our model has good stability and robustness. In addition, we provide the values of L c l s and L a l i g n at each epoch, finding that the initial value of L c l s is higher and the overall downward trend is more obvious than that of L a l i g n . We believe that the reason for this phenomenon is that there are too many modules in the model. This causes large classification loss in the whole process, although it still plays an important role in training.

5.5. Parameter Analysis

Our proposed SCCN method adopts fixed lengths for the text and convolution kernel sizes; thus, selecting these hyperparameters may have a significant effect on its performance. In the next experiments, we examine the parameter sensitivity in order to analyze the influence of different hyperparameters in our model. According to the principle of control variables, we minimize the influence of other irrelevant parameters. We then demonstrate the feasibility of the parameter settings by performing parameter sensitivity experiments on the text length and the size of the convolution kernel.

5.5.1. Length of Text

Based on the objective fact of text length in the datasets, we set the length of text L = 20 , 30 , 40 , 50 , 60 and record the accuracy and F1 score. Table 5 demonstrates the performance of the model with different text lengths. It can be seen that the text length has an effect on our model to some extent. In detail, the performance of SCCN decreases if the text length is too short or too long. A possible reason for this trend is that when the text length is too short, it cannot provide sufficient information for SCCN to accurately distinguish fake news. Moreover, although longer text lengths can provide enough information, they introduce noise which may affect the performance of the model.

5.5.2. Size of the Convolution Kernel

Figure 8 shows the accuracy comparison of different convolution kernel sizes on the PHEME and Weibo, where the x-axis indicates the kernel size. Experiments on both datasets indicate that when the kernel size is set to a single fixed value, the performance increases as the kernel size increases, peaking around sizes of 3 and 4. In addition, we conduct comparative experiments utilizing various convolutional kernel configurations. The results demonstrate that a hybrid architecture incorporating kernel sizes of 3, 4, and 5 yields optimal detection accuracy; consequently this configuration is implemented in the full SCCN model. The effectiveness of combining multiple convolution kernel sizes can be understood based on the fact that different kernel sizes can help the model to capture distinctive information from multiple perspectives, which is conducive to promoting the exploration of potential key information.

5.6. Case Studies

To further illustrate the importance of social relations and entity extraction in the proposed SCCN, we provide some examples of fake news identified by our model to demonstrate its effectiveness.
Figure 9 shows an example of fake news detected by our proposed SCCN model on the PHEME test set. In Figure 9a,b, the attached image and text examples appear normal without any anomalies, and there is no significant difference in the content that they intend to convey. This provides reasonable grounds to consider them real news. However, our model identifies irregularities in their social relations, leading to their classification as fake news. Specifically, the comment section under the news post in Figure 9a exhibits a pattern of highly repetitive user interactions, with multiple comments containing identical or near-identical minimal responses (e.g., “Seems these cops think too much and know too little.”). Similarly, Figure 9b exhibits an analogous pattern, indicating that our model can effectively identify evidence of fake news from social relationships and flag news items accordingly.
Although social relations have proven instrumental in fake news detection, the intrinsic characteristics of textual and visual entities remain indispensable and should not be neglected. In Figure 9c,d, “Ottawa”, “Sydney”, and “Lindt Cafe” are key pieces of information that enable readers to recognize whether the news is true or not. Nevertheless, not all readers are familiar with the meaning of these entities. Moreover, it is not possible to determine via simple observation whether entities in the text and images convey similar messages to readers. Our model can extract entities from both images and text while performing comprehensive analysis of textual and visual features, thereby contributing to more effective fake news detection.

6. Conclusions

This study introduces an innovative Semantic-enhanced Cross-modal Co-attention Network (SCCN) designed for enhanced fake news detection. The proposed SCCN introduces social relations with structure information and employs a co-attention mechanism to achieve cross-modal fusion with text and images. We not only successfully enhance text and images via entity extraction but also utilize an improved GAT to extract the structural features of social relations. Moreover, we deduce hidden links in the social relation graphs and remove the possible noisy links. We utilize a perturbation mechanism to enhance the robustness of the SCCN. Furthermore, we adopt cross-entropy as our optimization objective and conduct comprehensive information-theoretic analyses of both the model architecture and experimental outcomes. The results of our experiments demonstrate that the performance of the proposed SCCN is better than that of various baselines. The source code for SCCN can be found at https://github.com/asufdahu/SCCN (accessed on 20 June 2025).
In subsequent research, we intend to introduce external knowledge to improve the representation capability of the model, which can enhance the confidence coefficient of the entities and feature vectors. In addition, we will incorporate Mutual Information Maximization (MIM) into our framework to enhance model performance by optimizing the mutual information between input and output representations.

Author Contributions

Conceptualization, W.C.; methodology, W.C.; validation, W.C.; writing—original draft, W.C.; investigation, Y.D.; writing—reviewing and editing, Y.D.; supervision, X.Z.; resources, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Acknowledgments

The anonymous reviewers are acknowledged for their constructive comments. The authors would like to express gratitude for the support provided by COSTA: Complex System Optimization Team of the College of System Engineering at NUDT.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have influenced or appeared to influence the work reported in this study.

References

  1. Khattar, D.; Goud, J.S.; Gupta, M.; Varma, V. MVAE: Multimodal Variational Autoencoder for Fake News Detection. In Proceedings of the WWW’19: The Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2915–2921. [Google Scholar] [CrossRef]
  2. Wang, Y.; Ma, F.; Jin, Z.; Yuan, Y.; Xun, G.; Jha, K.; Su, L.; Gao, J. EANN: Event Adversarial Neural Networks for Multi-Modal Fake News Detection. In Proceedings of the KDD’18: The 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018; pp. 849–857. [Google Scholar] [CrossRef]
  3. Islam, M.S.; Sarkar, T.; Khan, S.H.; Kamal, A.H.M.; Hasan, S.M.; Kabir, A.; Yeasmin, D.; Islam, M.A.; Chowdhury, K.I.A.; Anwar, K.S.; et al. COVID-19–related infodemic and its impact on public health: A global social media analysis. Am. J. Trop. Med. Hyg. 2020, 103, 1621. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, Y.; Wu, Y.B. Early Detection of Fake News on Social Media Through Propagation Path Classification with Recurrent and Convolutional Networks. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 354–361. [Google Scholar] [CrossRef]
  5. Jing, J.; Wu, H.; Sun, J.; Fang, X.; Zhang, H. Multimodal fake news detection via progressive fusion networks. Inf. Process. Manag. 2023, 60, 103120. [Google Scholar] [CrossRef]
  6. Yu, C.; Ma, Y.; An, L.; Li, G. BCMF: A bidirectional cross-modal fusion model for fake news detection. Inf. Process. Manag. 2022, 59, 103063. [Google Scholar] [CrossRef]
  7. Tseng, Y.; Yang, H.; Wang, W.; Peng, W. KAHAN: Knowledge-Aware Hierarchical Attention Network for Fake News detection on Social Media. In Proceedings of the WWW’22: The ACM Web Conference 2022 Virtual Event (Companion Volume), Lyon, France, 25–29 April 2022; pp. 868–875. [Google Scholar] [CrossRef]
  8. Wu, Y.; Zhan, P.; Zhang, Y.; Wang, L.; Xu, Z. Multimodal Fusion with Co-Attention Networks for Fake News Detection. In Proceedings of the ACL/IJCNLP (Findings); Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 2560–2569. [Google Scholar] [CrossRef]
  9. Liu, P.; Qian, W.; Xu, D.; Ren, B.; Cao, J. Multi-Modal Fake News Detection via Bridging the Gap between Modals. Entropy 2023, 25, 614. [Google Scholar] [CrossRef] [PubMed]
  10. Zheng, J.; Zhang, X.; Guo, S.; Wang, Q.; Zang, W.; Zhang, Y. MFAN: Multi-modal Feature-enhanced Attention Networks for Rumor Detection. In Proceedings of the IJCAI, Vienna, Austria, 23–29 July 2022; pp. 2413–2419. [Google Scholar] [CrossRef]
  11. Hu, L.; Yang, T.; Zhang, L.; Zhong, W.; Tang, D.; Shi, C.; Duan, N.; Zhou, M. Compare to The Knowledge: Graph Neural Fake News Detection with External Knowledge. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 754–763. [Google Scholar] [CrossRef]
  12. Li, P.; Sun, X.; Yu, H.; Tian, Y.; Yao, F.; Xu, G. Entity-Oriented Multi-Modal Alignment and Fusion Network for Fake News Detection. IEEE Trans. Multim. 2022, 24, 3455–3468. [Google Scholar] [CrossRef]
  13. Yuan, C.; Ma, Q.; Zhou, W.; Han, J.; Hu, S. Jointly Embedding the Local and Global Relations of Heterogeneous Graph for Rumor Detection. In Proceedings of the ICDM, Beijing, China, 8–11 November 2019; pp. 796–805. [Google Scholar] [CrossRef]
  14. Yang, X.; Lyu, Y.; Tian, T.; Liu, Y.; Liu, Y.; Zhang, X. Rumor Detection on Social Media with Graph Structured Adversarial Learning. In Proceedings of the IJCAI, Yokohama, Japan, 7–15 January 2021; pp. 1417–1423. [Google Scholar] [CrossRef]
  15. Zhou, X.; Wu, J.; Zafarani, R. SAFE: Similarity-Aware Multi-modal Fake News Detection. In Advances in Knowledge Discovery and Data Mining, PAKDD 2020; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12085, pp. 354–367. [Google Scholar] [CrossRef]
  16. Wei, Z.; Pan, H.; Qiao, L.; Niu, X.; Dong, P.; Li, D. Cross-Modal Knowledge Distillation in Multi-Modal Fake News Detection. In Proceedings of the ICASSP, Singapore, 23–27 May 2022; pp. 4733–4737. [Google Scholar] [CrossRef]
  17. Tian, T.; Liu, Y.; Yang, X.; Lyu, Y.; Zhang, X.; Fang, B. QSAN: A Quantum-probability based Signed Attention Network for Explainable False Information Detection. In Proceedings of the CIKM, Virtual, 19–23 October 2020; pp. 1445–1454. [Google Scholar] [CrossRef]
  18. Chen, Y.; Li, D.; Zhang, P.; Sui, J.; Lv, Q.; Lu, T.; Shang, L. Cross-modal Ambiguity Learning for Multimodal Fake News Detection. In Proceedings of the WWW, Virtual, 25–29 April 2022; pp. 2897–2905. [Google Scholar] [CrossRef]
  19. Wang, L.; Zhang, C.; Xu, H.; Xu, Y.; Xu, X.; Wang, S. Cross-modal Contrastive Learning for Multimodal Fake News Detection. In Proceedings of the ACM Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 5696–5704. [Google Scholar] [CrossRef]
  20. Shu, K.; Cui, L.; Wang, S.; Lee, D.; Liu, H. dEFEND: Explainable Fake News Detection. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, 4–8 August 2019; pp. 395–405. [Google Scholar] [CrossRef]
  21. Wei, L.; Hu, D.; Zhou, W.; Yue, Z.; Hu, S. Towards Propagation Uncertainty: Edge-enhanced Bayesian Graph Convolutional Networks for Rumor Detection. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 3845–3854. [Google Scholar] [CrossRef]
  22. Qian, S.; Hu, J.; Fang, Q.; Xu, C. Knowledge-aware Multi-modal Adaptive Graph Convolutional Networks for Fake News Detection. ACM Trans. Multim. Comput. Commun. Appl. 2021, 17, 98:1–98:23. [Google Scholar] [CrossRef]
  23. Ma, X.; Wu, J.; Xue, S.; Yang, J.; Zhou, C.; Sheng, Q.Z.; Xiong, H.; Akoglu, L. A Comprehensive Survey on Graph Anomaly Detection With Deep Learning. IEEE Trans. Knowl. Data Eng. 2023, 35, 12012–12038. [Google Scholar] [CrossRef]
  24. Kim, D.; Oh, A. How to Find Your Friendly Neighborhood: Graph Attention Design with Self-Supervision. arXiv 2021, arXiv:2204.04879. [Google Scholar] [CrossRef]
  25. Huang, Y.; Gao, M.; Wang, J.; Yin, J.; Shu, K.; Fan, Q.; Wen, J. Meta-prompt based learning for low-resource false information detection. Inf. Process. Manag. 2023, 60, 103279. [Google Scholar] [CrossRef]
  26. Wang, J.; Zheng, J.Z.J.; Yao, S.; Wang, R.; Du, H. TLFND: A Multimodal Fusion Model Based on Three-Level Feature Matching Distance for Fake News Detection. Entropy 2023, 25, 1533. [Google Scholar] [CrossRef] [PubMed]
  27. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers); Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the CVPR, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  29. Chen, M.; Lai, Y.; Lian, J. Using Deep Learning Models to Detect Fake News about COVID-19. ACM Trans. Internet Technol. 2023, 23, 25:1–25:23. [Google Scholar] [CrossRef]
  30. Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar] [CrossRef]
  31. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2023, arXiv:1706.03762. [Google Scholar] [CrossRef]
  32. Lu, J.; Batra, D.; Parikh, D.; Lee, S. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. arXiv 2019, arXiv:1908.02265. [Google Scholar] [CrossRef]
  33. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. arXiv 2018, arXiv:1706.06083. [Google Scholar] [CrossRef]
  34. Zubiaga, A.; Liakata, M.; Procter, R. Exploiting Context for Rumour Detection in Social Media. In Social Informatics. SocInfo 2017. Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2017; Volume 10539, pp. 109–123. [Google Scholar] [CrossRef]
  35. Song, C.; Yang, C.; Chen, H.; Tu, C.; Liu, Z.; Sun, M. CED: Credible Early Detection of Social Media Rumors. IEEE Trans. Knowl. Data Eng. 2021, 33, 3035–3047. [Google Scholar] [CrossRef]
  36. Liu, K.; Xue, F.; Guo, D.; Sun, P.; Qian, S.; Hong, R. Multimodal Graph Contrastive Learning for Multimedia-Based Recommendation. IEEE Trans. Multim. 2023, 25, 9343–9355. [Google Scholar] [CrossRef]
Figure 1. Three pieces of news in the PHEME dataset, including text, images, users, and comments; the top two news stories are in the training set and are labeled as fake, while the last one is in the test set.
Figure 1. Three pieces of news in the PHEME dataset, including text, images, users, and comments; the top two news stories are in the training set and are labeled as fake, while the last one is in the test set.
Entropy 27 00746 g001
Figure 2. The framework of our proposed SCCN. Specifically, we first extract the features of images, text, and social relations from news stories. We then obtain their feature vectors through feature extraction, where the text and images are the content features and the social relations are the structural features. Meanwhile, we extract entities from images and text and encode them to enhance the content features before performing modal alignment with a self-supervised loss to facilitate subsequent feature fusion. In addition, we obtain the enhanced features between modalities through a co-attention mechanism. Finally, we integrate all the enhanced features of different modalities to obtain a cross-modal fusion feature for fake news detection.
Figure 2. The framework of our proposed SCCN. Specifically, we first extract the features of images, text, and social relations from news stories. We then obtain their feature vectors through feature extraction, where the text and images are the content features and the social relations are the structural features. Meanwhile, we extract entities from images and text and encode them to enhance the content features before performing modal alignment with a self-supervised loss to facilitate subsequent feature fusion. In addition, we obtain the enhanced features between modalities through a co-attention mechanism. Finally, we integrate all the enhanced features of different modalities to obtain a cross-modal fusion feature for fake news detection.
Entropy 27 00746 g002
Figure 3. Pipeline of the improved GAT, where we combine the green and red box parts to compose a new mixed form. The red circles represent attention that is not normalized, while the blue diamond is the transition node.
Figure 3. Pipeline of the improved GAT, where we combine the green and red box parts to compose a new mixed form. The red circles represent attention that is not normalized, while the blue diamond is the transition node.
Entropy 27 00746 g003
Figure 4. Comparison of accuracy and F1 score when removing different types of modal embeddings.
Figure 4. Comparison of accuracy and F1 score when removing different types of modal embeddings.
Entropy 27 00746 g004
Figure 5. Visualization of the results of SCCN and its variants on the PHEME dataset. The values in the boxes indicate the correlation with the news, which is represented by the correlation of the news prediction labels.
Figure 5. Visualization of the results of SCCN and its variants on the PHEME dataset. The values in the boxes indicate the correlation with the news, which is represented by the correlation of the news prediction labels.
Entropy 27 00746 g005
Figure 6. Visualization results of SCCN and several baselines on the Weibo dataset. Dots of the same color belong to the same label.
Figure 6. Visualization results of SCCN and several baselines on the Weibo dataset. Dots of the same color belong to the same label.
Entropy 27 00746 g006
Figure 7. Average training loss on the PHEME and Weibo datasets.
Figure 7. Average training loss on the PHEME and Weibo datasets.
Entropy 27 00746 g007
Figure 8. Accuracy of different convolutional kernel sizes on the PHEME and Weibo databases. Here, “single” indicates a single fixed convolutional kernel size.
Figure 8. Accuracy of different convolutional kernel sizes on the PHEME and Weibo databases. Here, “single” indicates a single fixed convolutional kernel size.
Entropy 27 00746 g008
Figure 9. Images and text of example fake news stories detected by SCCN: (a) “This is what #Ferguson PD did to #HenryDavis in 2009 when they mistook him for a man with an outstanding warrant”; (b) “According to reports, one gunman is dead in #Ottawa Parliament shooting”; (c) “Ottawa shooting: Soldier dies of injuries, gunman shot dead”; (d) “DRAMATIC VIDEO: #Sydney siege hostages run from #Lindt Cafe, more remain trapped”.
Figure 9. Images and text of example fake news stories detected by SCCN: (a) “This is what #Ferguson PD did to #HenryDavis in 2009 when they mistook him for a man with an outstanding warrant”; (b) “According to reports, one gunman is dead in #Ottawa Parliament shooting”; (c) “Ottawa shooting: Soldier dies of injuries, gunman shot dead”; (d) “DRAMATIC VIDEO: #Sydney siege hostages run from #Lindt Cafe, more remain trapped”.
Entropy 27 00746 g009
Table 1. Statistics of the datasets used in our experiments after data cleaning.
Table 1. Statistics of the datasets used in our experiments after data cleaning.
StatisticsNon RumorsRumorsImagesUsersComments
PHEME142859020188947388
Weibo87759014679854534
Table 2. Performance comparison between SCCN and other methods. The global best and second-best results are highlighted in bold and underlined, respectively. SCCN-F and SCCN-R represent the results of the fake and real classes in prediction performance. To validate the stability of our method, we conducted experiments under five different randomized seeds, then confirmed the statistical significance of the pairwise differences in the SCCN relative to the strongest baseline using a t-test p-value < 0.01 (1 × 10 −2).
Table 2. Performance comparison between SCCN and other methods. The global best and second-best results are highlighted in bold and underlined, respectively. SCCN-F and SCCN-R represent the results of the fake and real classes in prediction performance. To validate the stability of our method, we conducted experiments under five different randomized seeds, then confirmed the statistical significance of the pairwise differences in the SCCN relative to the strongest baseline using a t-test p-value < 0.01 (1 × 10 −2).
PHEME [34]Weibo [35]
Method Accuracy Precision Recall F1 Score Accuracy Precision Recall F1 Score
EANN [2]0.7710.7140.7000.7040.8090.8020.7970.799
MVAE [1]0.7760.7350.7220.7280.7170.7050.7020.703
QSAN [17]0.7510.6990.6580.6690.7100.7100.6750.676
SAFE [15]0.8150.7990.7950.7970.8490.8500.8490.849
EBGCN [21]0.8300.8130.7930.7980.8310.8540.8180.814
GLAN [13]0.8330.8120.7710.7850.8240.8240.8090.813
MPFN [5]0.8330.8270.8210.8100.8380.8650.8780.882
KMAGCN [22]0.8640.8510.8550.8530.8490.8470.8500.849
MFAN [10]0.8870.8700.8560.8610.8890.8890.8810.883
SCCN-F (ours)0.9040.9180.9490.9330.9050.8830.9720.925
SCCN-R (ours)0.9040.8650.7960.8290.9050.9490.8020.869
SCCN (ours)0.9040.9020.9040.9030.9050.9090.9050.903
p-value5.44 × 10 −53.32 × 10 −54.04 × 10 −82.43 × 10 −76.49 × 10 −33.10 × 10 −41.80 × 10 −56.65 × 10 −3
Table 3. Results showing the practical deployability of SCCN.
Table 3. Results showing the practical deployability of SCCN.
DatasetTime (s/epoch)Inference Latency (s/epoch)Memory (GB)
PHEME [34]30.240.8681.99
Weibo [35]29.080.8692.45
Table 4. Performance comparison between the full SCCN model and different variants.
Table 4. Performance comparison between the full SCCN model and different variants.
Method-w/o DPS-w/o SRG-w/o IEM-w/o CMFSCCN
PHEME [34]Accuracy0.8790.8800.8750.8860.904
Precision0.8770.8810.8780.8840.902
Recall0.8770.8800.8750.8850.904
F1 Score0.8780.8810.8760.8820.903
Weibo [35]Accuracy0.8750.8710.8780.8880.905
Precision0.8740.8710.8780.8980.909
Recall0.8750.8710.8770.8880.905
F1 Score0.8740.8700.8780.8850.903
Table 5. Accuracy and F1 score results with different text lengths on the two datasets. SCCN-20 means that the text length is set to 20.
Table 5. Accuracy and F1 score results with different text lengths on the two datasets. SCCN-20 means that the text length is set to 20.
DatasetMetricSCCN-20SCCN-30SCCN-40SCCN-50SCCN-60
PHEME [34]Accuracy0.7430.7630.8650.9040.875
F1 Score0.7330.7440.8670.9030.873
Weibo [35]Accuracy0.6070.8000.8740.9050.817
F1 Score0.4580.8020.8730.9030.802
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, W.; Dang, Y.; Zhang, X. A Multimodal Semantic-Enhanced Attention Network for Fake News Detection. Entropy 2025, 27, 746. https://doi.org/10.3390/e27070746

AMA Style

Chen W, Dang Y, Zhang X. A Multimodal Semantic-Enhanced Attention Network for Fake News Detection. Entropy. 2025; 27(7):746. https://doi.org/10.3390/e27070746

Chicago/Turabian Style

Chen, Weijie, Yuzhuo Dang, and Xin Zhang. 2025. "A Multimodal Semantic-Enhanced Attention Network for Fake News Detection" Entropy 27, no. 7: 746. https://doi.org/10.3390/e27070746

APA Style

Chen, W., Dang, Y., & Zhang, X. (2025). A Multimodal Semantic-Enhanced Attention Network for Fake News Detection. Entropy, 27(7), 746. https://doi.org/10.3390/e27070746

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop