You are currently on the new version of our website. Access the old version .
MathematicsMathematics
  • Article
  • Open Access

21 January 2026

A Multimodal Ensemble-Based Framework for Detecting Fake News Using Visual and Textual Features

,
,
,
,
,
,
and
1
School of Computer Science and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China
2
School of Information Science and Engineering, Hunan University, Changsha 410082, China
3
Institute of Information and Computational Technologies, Almaty 050010, Kazakhstan
4
Department of Computer Science, Swansea University, Swansea SA1 8EN, UK
This article belongs to the Special Issue Research on Machine Learning, Data Mining, Natural Language Processes, and Optimization Methods

Abstract

Detecting fake news is essential in natural language processing to verify news authenticity and prevent misinformation-driven social, political, and economic disruptions targeting specific groups. A major challenge in multimodal fake news detection is effectively integrating textual and visual modalities, as semantic gaps and contextual variations between images and text complicate alignment, interpretation, and the detection of subtle or blatant inconsistencies. To enhance accuracy in fake news detection, this article introduces an ensemble-based framework that integrates textual and visual data using ViLBERT’s two-stream architecture, incorporates VADER sentiment analysis to detect emotional language, and uses Image–Text Contextual Similarity to identify mismatches between visual and textual elements. These features are processed through the Bi-GRU classifier, Transformer-XL, DistilBERT, and XLNet, combined via a stacked ensemble method with soft voting, culminating in a T5 metaclassifier that predicts the outcome for robustness. Results on the Fakeddit and Weibo benchmarking datasets show that our method outperforms state-of-the-art models, achieving up to 96% and 94% accuracy in fake news detection, respectively. This study highlights the necessity for advanced multimodal fake news detection systems to address the increasing complexity of misinformation and offers a promising solution.

1. Introduction

The dissemination of misinformation through fake news to mislead or deceive the public has historical roots and is increasingly exploited in modern times via internet-based channels. Fake news has been extensively used in various contexts and types of content, including satire, conspiracy, news manipulation, and clickbait [1]. Consequently, this term has become increasingly prevalent and is strongly associated with events that pose significant social concerns, such as misinformation about COVID-19 vaccines, which likely led to more illness and deaths by discouraging vaccinations [2], and its potential influence on public opinion during the 2016 US presidential election [3].
Using social media for news updates has both advantages and disadvantages, as it allows for the free expression of personal views but also facilitates the spreading of false news [4]. Modern devices have amplified the use of social media and enabled users to create and share information, posts, and news more quickly. The wide adoption of the internet has impacted information quality [5] and enabled false claims to spread about technical and complex subjects that are usually harder to verify [6]. Fake news is often perceived as credible and spreads rapidly on social media, making disinformation a major concern. Thus, creating an automated system for effective fake news detection is crucial.
Multimodal fake news detection, which adopts various data types or modalities like text and images, integrates information from these diverse sources to provide a more comprehensive understanding and improve the accuracy of identifying and classifying fake news [7,8]. Detecting false content with such systems is difficult due to the complex relationship between language patterns and the underlying data. Language is deeply contextual, with meaning often depending on subtle nuances, such as tone, syntax, and cultural references. When combined with the massive scale of data—spanning various platforms, formats, and contexts—this complexity makes it hard for detection systems to distinguish between true and false information accurately [4]. Researchers have proposed solutions based on deep neural networks, such as Convolutional Neural Networks (CNNs) [9], Recurrent Neural Networks (RNNs) [10], and linguistic features such as sentiment analysis. This method uses vector embeddings, including stylometric and domain name analysis, and news source history [11] to differentiate real from fake news through text and visuals [12]. Most sentiment analysis models generate context-free or static embeddings and do not consider the varying meanings of words that may have different contexts. Thus, we focus on ensemble approaches that blend linguistic and deep learning models, effectively addressing these challenges [13].
Figure 1 illustrates multimodal data manipulation of visual and textual information. The first row compares real news with an image of Robert Joel Halderman during his trial for extortion against David Letterman, against fake news showing Barack Obama falsely accused of illegal activities. The second row features an image of Bill Gates and factual reporting on his 2018 discussion about population control through health initiatives. The fake news distorts this into a conspiracy theory, falsely suggesting that Gates is using deceptive tactics to reduce the population, further exaggerating this narrative by adding an ’Illuminati’ symbol. These examples underscore how real images and text can be manipulated to spread disinformation, lending false credibility to misrepresented stories.
Figure 1. Examples of multimodal data manipulation contrasting real and fake news to mislead the audience [14].
Developing a model that cohesively, efficiently, and scalably combines the strengths of natural language processing and image analysis technologies is primarily challenged by the contextual variations between images and text, which complicate the alignment and interpretation of multimodal content [15], and by the complexity and substantial computational resources required for feature extraction from text and images [16]. Furthermore, semantic gaps between textual and visual content often complicate detection, especially when information that is true at one point becomes false later, or vice versa. For example, health recommendations or policies may change as new information becomes available. Models need to account for this temporal sensitivity. Another significant issue in multimodal fake news detection is the struggle of previous models to align visual elements, such as images, with textual claims due to subtle or blatant inconsistencies between these modalities.
In this light, there is a need for an advanced detection model that can effectively analyse and integrate multiple data types to identify fake news. Such a model must not only process textual and visual content with high accuracy but also adapt to the evolving tactics used in misinformation. This study aims to address these issues. The rest of this article is structured as follows. Section 2 reviews the state of the art in fake news detection. Section 3 motivates our approach and outlines its unique contribution to the field. Section 4 describes the dataset and methodology used in this study. Section 5 describes the proposed ensemble model. Section 6 provides a description of the experimental setup. Section 7 reports and comments on the results. Section 8 discusses the results and identifies the limits of the proposed model. Section 9 concludes this work.

2. Related Work

The transition from classical machine learning approaches to advanced deep learning (DL) techniques has marked an indicative evolution in the fight against fake news [17]. Initial efforts to detect fake news used conventional machine learning algorithms, including Logistic Regression, Support Vector Machine (SVM), and Naive Bayes, focussing on the extraction of features and the classification of textual content [18]. These algorithms generally emphasise generic textual features, analysing the text through linguistic dimensions such as lexicon, syntax, discourse, and semantics. In contrast, latent textual features generate embeddings from text data from news articles, at the word, sentence, or document level. These embeddings are transformed into latent vectors and then used as inputs for classification tools such as SVMs.
Despite the success of these techniques, they often fail to handle the complexity of modern news content, which can span multiple domains. To address this, advances in Natural Language Processing (NLP) and computational intelligence techniques have led to more sophisticated hybrid systems capable of dealing with this problem [19]. In [7], a heuristic algorithm for optimisation is used to tune the hyperparameters of the used NLP model for multi-label news classification, demonstrating superior performance to traditional CNNs and validated across four public news datasets. Furthermore, they proposed the HyproBert model in [20], which adds extra fake news detection capabilities by integrating CNN, DistilBERT, BiGRU, and CapsNet layers. Similarly, Recurrent Neural Networks (RNNs), which can effectively capture temporal information [10], and CNNs are used together to extrapolate semantic and syntactic aspects of the text in [21]. Models such as BerConvoNet and CAME-BiLSTM showed significant improvements in multilabel classification but are difficult to tune [5,22].
Challenges emerge when addressing multimodal information, as existing models for multi-label classification have proven inadequate in simultaneously processing textual and visual data as required. Due to the proliferation of fake news in multimedia formats, such as images, audio, and video, there has recently been a growing demand for developing models for multimodal fake news detection. In [23], an interesting text- and image-based multimodal model is proposed as an alternative to sequence-independent classification. Further innovations include the use of the Attention-Residual Network for long-range data dependencies [24], the introduction of feature fusion systems based on multimodal factorised bilinear pooling [25], and additional event classification as a supplementary task [26]. The latter method has proven to improve generalisability by identifying event-invariant features in multiple media formats [27]. A different approach in [28] integrates text-based and image-based features using pre-trained models like BERT [29] for text and XLNet for images. It is worth mentioning the Feature Gradient Method with Feature Regularisation Adversarial Training (FGM-FRAT), which employs adversarial training to enhance the model’s robustness against unseen data [30].
The approach involving a visual and textual shared feature space, as detailed in [31], paves the way for an innovative similarity-based methodology within this domain. However, it currently encounters challenges in effectively capturing multimodal inconsistencies attributable to semantic gaps. Meanwhile, the Fake-News-Revealer (FNR) method introduced in [32] suggests using Vision Transformer and BERT to extract image and text features separately and then determining their similarities through loss functions. These methods show innovation and reflect the current efforts made by researchers to improve multimodal fake news detection by linking media features, but they struggle to capture complex cross-modal correlations.
Ensemble learning has emerged as a promising approach to enhance the detection of multimodal fake news by leveraging the strengths of multiple models and integrating diverse sources of information [28]. These methods combine predictions from multiple classifiers trained with different modalities (such as text, images, and videos), improving both accuracy and robustness [33]. The ensemble approach is grounded in key principles such as combining weak learners to create a stronger learner. This is based on the bias–variance tradeoff, where an ensemble reduces both bias and variance compared to individual models, leading to more accurate and robust predictions. A notable example is the SEMI-FND framework, which combines NasNet Mobile for image analysis with an ensemble of BERT and ELECTRA for text analysis, offering a minimal-parameter solution for fake news detection [34,35]. The model benefits from the ensemble principle of stacking, where multiple base models are trained separately, and their predictions are then combined through a meta-learner. The effectiveness of stacking lies in its ability to combine diverse model strengths, improving generalisability and predictive performance. Similarly, the Ensemble Learning-based Framework (ELD-FN) uses V-BERT to generate embeddings from both text and images, followed by a deep learning ensemble model for training and evaluation [13]. This framework integrates text and image features into a unified space, emphasising feature fusion, which enhances the model’s ability to detect fake news by combining complementary information from both modalities.
The A BERT-Based Multimodal Framework for Enhanced Fake News Detection Using Text and Image Data Fusion integrates text from images using Optical Character Recognition (OCR), demonstrating the potential of multimodal approaches to boost detection accuracy [36]. It employs the feature fusion principle, merging multiple data sources (text and image) into a shared space, which helps the model capture richer, more complementary information. In the context of improving machine learning models for lithofacies identification, How to Improve Machine Learning Models for Lithofacies Identification by Practical and Novel Ensemble Strategy and Principles introduces a strategy to make machine learning models more accessible for non-experts while improving performance [37]. The paper employs ensemble pruning, optimising the ensemble size by removing less useful models to improve computational efficiency without sacrificing performance.
The Truth Be Told: A Multimodal Ensemble Approach for Enhanced Fake News Detection uses a stacked ensemble model that integrates textual and visual data, improving efficiency in distinguishing fake news [38]. By applying a boosting strategy, the model focuses on correcting misclassifications in weak learners, which increases overall predictive accuracy. Finally, the EnsembleNet model combines GANBERT and BiLSTM for fake news detection, enhancing accuracy across multiple datasets [39]. This model follows the bagging principle, where multiple base models are trained on different subsets of the data and their predictions are aggregated. Bagging reduces variance and prevents overfitting, leading to more robust performance, particularly in noisy or imbalanced datasets.
These innovations reflect the theoretical grounding of ensemble learning principles such as stacking, boosting, bagging, and feature fusion, which contribute to the strength and robustness of multimodal fake news detection models. By combining predictions from multiple models, these techniques enhance accuracy, generalisability, and resilience, making ensemble learning a powerful tool in this domain. Detecting fake news on social platforms is made even more challenging by the vast amount of data and the time-sensitive nature of the task. The model [40], which processes image and textual features via a Quantum Convolutional Neural Network, and the ensemble approach combining pre-trained BERT with DL models (Bi-LSTM and/or Bi-GRU architectures on GloVe and FastText embeddings) for multi-aspect hate speech detection [41] are promising research directions to address these issues. Also worth mentioning are the Hierarchical Multi-modal Contextual Attention Network [42,43] and the Multi-modal Co-Attention Network [44], which extract spatial and frequency features from images and text, and the BERT-based multimodal models which encode and enhance interactions between text and images using ensemble learning [45,46].
In summary, recent studies highlight three key biases in multimodal fake news detection: image enhancement of text, text–image discrepancies signalling fake news, and improved detection through their integration. These findings underscore the value of combining text and visual cues for accuracy. Challenges include integrating NLP with image analysis, handling contextual variability, computational demands, and bridging semantic gaps. Table 1 summarises this review.
Table 1. Advances on multimodal fake news detection.

3. Rationale and Contributions of the Proposed Approach

We propose an Ensemble Learning-based Multimodal Fake News Detection (EMFND) model that leverages stacked ensemble methods to improve fake news detection by integrating text and visual data. This approach combines multiple DL and transformer-based models, followed by soft voting, drawing on insights from our literature review to leverage the strengths of proven methods. EMFND uses sophisticated models like Bi-GRU [47], Transformer-XL [48], DistilBERT [49], and XLNet [50] to process features and ensure accurate predictions despite contextual variations between text and images. Designed for scalability, accuracy, and robustness across media formats, it reduces bias and overfitting. Parameter sharing is used to minimise training time, memory use, and model complexity.
We summarise the contributions of this study below.
  • The proposed EMFND framework leverages ViLBERT’s two-stream architecture with cross-modality transformers [51] to process text and visual data simultaneously, capturing complex relationships and bridging the semantic gap for improved multimodal fake news detection.
  • The model addresses the challenge of visual–text inconsistencies in multimodal fake news detection by using transformers to align cross-modal information and detect misinformation.
  • We combine XLNet and Transformer-XL to capture long-term dependencies, improving the detection of evolving fake news with incremental modifications.
  • We address the limitations of relying solely on text or image with our method, using Bi-GRU to process extracted features and manage sequential dependencies, and DistilBERT to capture bidirectional text context. Our stacked ensemble method with soft voting and a meta-classifier combines predictions from baseline models for the final result.

4. Materials and Methods

A graphical outline of the entire methodology is displayed in Figure 2.
Figure 2. Research methodology work-flow chart.
Details on the dataset, problems, and preprocessing phases are reported in the remainder of this section.

4.1. Problem Formulation

The task of detecting fake news consists of analysing a dataset containing a set of m multimedia news posts, represented as P = ( p 1 , p 2 , , p m ) . Each post p k ( K = 1 , 2 , , m ) consists of a textual component T k , a corresponding set of images I k , and a label Y k indicating whether the news is real ( Y k = 1 ) or fake ( Y k = 0 ). Textual features f t ( T k ) transform text into a vector of characteristics in space R d t . Image features f i ( I k ) map the images to a feature space R d i . These features are integrated using a fusion function F, where F k represents the unified feature vector for the k-th post: F k = ϕ ( f t ( T k ) , f i ( I k ) ) . A classifier C uses this fused feature vector F k to assess the authenticity of the news post with a predicted label y ^ k . The label classifies the post as real ( y ^ k = 1 ) or fake ( y ^ k = 0 ): y ^ k = C ( F k ) This prediction takes advantage of the combined strengths of text and image analysis to accurately determine the integrity of each news post in the dataset.

4.2. Datasets

We conduct experiments on established benchmark datasets, namely Fakeddit and Weibo. Fakeddit [14] is an extensive multimodal fake news dataset featuring text and images (Figure 3). It comprises 1,063,106 samples across various fake news categories. The primary textual features include post titles, captions, or content, while the visual features comprise images associated with the posts.
Figure 3. Fakeddit dataset overview [14].
The Weibo dataset [52,53] is a collection of 49,713 posts and 25,513 images from nine different domains, providing a comprehensive resource for detecting fake news. It includes rich data comprising texts, images, and social context information that are integral for multimodal analysis.
For both datasets, the focus in this study is on multimodal posts—those that contain both text and image—which are most commonly used in multimodal fake news detection research. In alignment with standard practices in multimodal fake news detection, the primary features exploited for model training are text (typically the cleaned post title or content) and visual content (the images). This approach ensures that the models are trained using the most informative and commonly utilised features in the field.
Statistics for these datasets are shown in Table 2. To perform experiments, we split them into three subsets: 70% of the data is randomly selected for training, 15% for testing, and 15% for validation.
Table 2. Datasets summary.
To tackle class imbalance, we use oversampling for Fakeddit and SMOTE [54] for Weibo datasets, plus Class Weight Adjustment [55] to reduce bias in model training. Figure 4 graphically displays the effect of these data balancing techniques.
Figure 4. Balancing the datasets.
Note that both the Fakeddit (Figure 4a) and Weibo (Figure 4b) datasets are plagued by a significant class imbalance, introducing undesired biases into the models, but with opposite minority classes. In the case of Fakeddit, oversampling is applied to augment true news. This technique is preferred for this dataset over SMOTE because of its large size. Oversampling prevents overfitting, enhances minority class learning, and maintains multimodal data diversity without adding excessive noise. For Weibo, which has a smaller sample size, this approach was unfeasible due to the necessity of generating synthetic data, which motivates the use of SMOTE to increase samples in the fake news class.
Results are satisfactory, with nearly the same amount of samples in the two classes after the balancing process for both datasets. This guarantees a good degree of generalisation and a bias-free training phase.

4.3. Data Preprocessing

Every textual component T k of a news post p k undergoes tokenisation, which splits the text into n individual tokens for analysis, i.e.,  T k = { t k 1 , t k 2 , , t k n } . Subsequently, all characters are normalised to lowercase, stop words and punctuation marks are removed, and words are lemmatised, i.e.,  T k = normalise ( T k ) = lemma ( lower ( T k { stop _ words } { punctuation } ) ) . After normalisation, the text is vectorised to create numerical representations as v T k = vectorise ( T k ) . For the visual component I k , all images from each news post p k are uniformly resized to a 224 × 224 format using bilinear interpolation and their pixel values are normalised via Min–Max.

5. The Proposed Model

Algorithm 1 describes our model. Each step is detailed in the remainder of this section.
Algorithm 1: EMFND Pseudocode
1:
Input: News post p k , textual and visual Features extracted using ViLBERT.   ▹ Section 5.1
2:
Output: Predicted label y ^ (0 for fake news, 1 for real news).
3:
for each classifier C i  {Bi-GRU, Transformer-XL, DistilBERT, XLNet} do     ▹ Section 5.5
4:
     Process input p k using C i .
5:
     Obtain probability distribution: P C i ( y ) = P C i ( y = 0 ) , P C i ( y = 1 )
6:
end for 
7:
Compute the averaged probability distribution: P final ( y ) = 1 4 C i P C i ( y )     ▹ Section 5.9
8:
Concatenate outputs of all classifiers: P concat = P Bi - GRU , P Transformer - XL , P DistilBERT , P XLNet     ▹ Section 5.10
9:
Metaclassifier: P meta ( y ) = T 5 ( P concat )                     ▹ Section 5.11
10:
Choose final class label y ^ with the highest probability: y ^ = arg max P meta ( y )
11:
Return  y ^ as the predicted label.                      ▹ Section 5.12

5.1. Vision and Language BERT (ViLBERT)

We use ViLBERT [56] to extract and combine text and visual features into a unified representation. This model extends BERT, having two streams that allow for text and image processing. It also employs cross-modal co-attention to learn inter-modal relationships between them.

5.1.1. Textual Stream

The input text E T k is tokenised and fed to BERT’s encoder to generate contextual embeddings for each token t k i . The latter passes through 12 BERT layers with multi-head self-attention (Equation (1)) to capture word relationships, followed by a feed-forward neural network, layer normalisation, and residual connections.
Attention ( Q , K , V ) = softmax Q K T d t V , Q = query , K = key , V = values d t = dimensionality of the token
The output is a set of context-aware hidden representations of the text, as shown in Equation (2). These hidden states capture the deeper semantics of the text, such as sentiment, intent, and misleading cues, which are essential for detecting misinformation in news posts.
H T k = { h t k 1 , h t k 2 , , h t k n } , h t k i R d t

5.1.2. Sentiment Analysis

Valence Aware Dictionary for sEntiment Reasoning (VADER) [57], a lexicon- and rule-based sentiment analysis tool, computes sentiment scores (positive, negative, neutral, and compound) for the entire news article and captures emotional tone. It is selected for several key reasons: (1) it demonstrates strong performance in processing informal language, slang, emojis, punctuation-based emphasis, and negation—features that are prevalent in social media text and that contemporary transformer architectures may insufficiently weight without explicit guidance; (2) it is lightweight, interpretable, and requires no additional training, thereby ensuring computational efficiency and facilitating seamless integration into large-scale transformer-based pipelines; (3) in contrast to purely contextual embeddings derived from BERT-like models, VADER yields explicit sentiment polarity scores that directly capture emotional exaggeration and polarisation, both of which constitute strong signals for the presence of misinformation; and (4) prior studies [19,58] have demonstrated its effectiveness in fake news detection tasks, particularly when it is used in conjunction with deep learning–based feature representations.
The sentiment scores are concatenated with BERT’s embeddings to create an enriched representation as shown in Equations (3) and (4).
s T k = S pos ( T k ) , S neg ( T k ) , S neu ( T k ) , S comp ( T k )
H T k final = H T k s T k R d t + 4
This enriched representation H T k final incorporates both contextual text features and sentiment scores, allowing the model to better capture the emotional manipulations that are common in fake news. This sentiment-enhanced text representation is then used in the cross-modal co-attention mechanism with visual features, improving the overall ability of the model to detect fake news.

5.1.3. Visual Stream

CNN models like ResNet are utilised on images I k for extracting region-based visual features E I k . The embeddings are then passed through transformer layers to model relationships between different regions and result in a set of hidden representations for the image, as shown in the second case of Equation (5).
E I k = { e i 1 , e i 2 , , e i m } , e i j R d i H I k = { h i 1 , h i 2 , , h i m } , h i j R d i

5.2. Cross-Modal Co-Attention

ViLBERT’s core innovation is its cross-modal co-attention mechanism, which enables interaction between textual and visual streams during processing, enhancing understanding of their relationship. Incorporating sentiment analysis from VADER further enriches this mechanism by offering emotional insights into the text. By combining sentiment scores with ViLBERT’s contextual embeddings, the model effectively detects emotionally charged content commonly present in misleading or fake news. Text and image representations are aligned through a generalised attention mechanism, which is formulated as Attention X ( Q X , K X , V X ) = softmax Q X K X T d c V X , where X { T I , I T } indicates the direction of attention (text relating to the image or image relating to the text). For text relating to image ( T I ), Q X = W T H T k final , K X = W I H I k , and V X = W I H I k . For image relating to text ( I T ), Q X = W I H I k , K X = W T H T k final , and V X = W T H T k final . The resulting co-attended hidden states for the text and image are formalised in Equation (6).
H T k = Attention T I ( H T k final , H I k ) H I k = Attention I T ( H I k , H T k final )
These updated hidden states, H T k and H I k , represent integrated information from text and image, allowing the model to capture inconsistencies between these modalities. For example, if an article describes a medical breakthrough, but the accompanying image is irrelevant or misleading, the co-attention mechanism can detect this discrepancy as H T k H I k .

5.3. Image–Text Contextual Similarity

Another component in assessing the alignment between visual and textual content is Image–Text Contextual Similarity. Due to the use of misaligned or irrelevant imagery to manipulate or deceive, this similarity measure becomes particularly useful for detecting fake news. Mathematically, this is simply computed as the cosine similarity between the final co-attended text and image embeddings, as formulated in Equation (7).
S sim ( H T k , H I k ) = H T k · H I k H T k H I k ,
where H T k and H I k are the co-attended text and image embeddings, respectively. The cosine similarity score S sim ranges from −1 to 1, where S sim = 1 indicates perfect alignment between text and image, while S sim = 1 indicates complete dissimilarity, and S sim = 0 suggests no discernible relationship. This contextual similarity score helps the model detect contrasting narratives between the image and text, such as when the text in a news article is factual, but the image used is misleading or unrelated.
The notion of image–text similarity is extended by introducing three different types of similarities, namely textual similarity ( S i m text ( T k , I k ) ), which represents the relationship between textual information extracted from both images and news content; semantic similarity ( S i m sem ( T k , I k ) ), which measures the semantic alignment between text and image; and contextual similarity ( S i m cont ( T k , I k ) ), which evaluates how well the image and text align within the broader context of the article. These similarities are calculated based on information extracted directly from the original image and text or related knowledge using BERT embeddings.

Post-Training Similarity

To improve fake news detection, a novel image–text similarity measure called post-training similarity ( S i m post ( T k , I k ) ) is introduced. After training ViLBERT, a state-of-the-art multimodal fake news detection classifier, the hidden representations of both text and image are extracted. The cosine similarity, denoted as Sim post ( T k , I k ) , measures how these representations align after the model has learned to differentiate between fake and real news. Specifically, given a news article N = ( T k , I k ) , image features are extracted using Faster R-CNN with ResNet-101 as the backbone, while text features are processed using ViLBERT’s linguistic tokens. The final image and text representations are captured by the vectors corresponding to the IMG and CLS tokens, respectively. The cosine similarity between these representations is defined through Equation (8).
S i m post ( T k , I k ) = h IMG · h CLS h IMG h CLS
This similarity captures the relationships between image and text as processed during model training, offering valuable insight into how image–text similarity evolves within the fake news detection model.

5.4. Unified Representation and Classification

The classicifaction output y ^ k = C ( H T k , H I k ) { 0 , 1 } ( 0 = fake, 1 = real) is obtained by processing final co-attended embeddings H T k and H I k with a classifier C {DistilBERT, Transformer-XL, XLNet}. We implement a stacked ensemble approach using the available classifiers to then make the final decision with a T5 meta-classifier with a soft voting mechanism to improve robustness. Formally, y ^ k = C ( H T k , H I k ) . The ensemble classifier prediction is modelled via Equation (9), where y ^ k { 0 , 1 } indicates whether the news post is classified as fake (0) or real (1).
The key advantages of this unified representation include effective bridging of semantic gaps through cross-modal co-attention, enabling precise detection of text–image inconsistencies common in fake news; enhanced robustness and accuracy by integrating diverse transformer strengths (e.g., bidirectional context from DistilBERT and long-range dependencies from XLNet); and improved generalisation across datasets, as unified embeddings capture complementary multimodal cues more comprehensively than single-modality approaches.
y ^ k = C ( H T k , H I k ) ,

5.5. Bi-GRU Implementation

The Bi-GRU (Bidirectional Gated Recurrent Unit) network [47] is utlised to capture both forward and backward dependencies within a sequence of tokens. The architecture is shown in Figure 5, providing a high-level overview of the model’s structure and the flow of data through the layers. This makes it highly effective for understanding the full temporal context of a text. The textual embeddings H T k generated from ViLBERT are passed into the Bi-GRU network. These embeddings represent the tokens in the sequence and provide local and global context from the news post, with n being the number of tokens and d the dimensionality of each embedding. Implemented in three bidirectional layers, the Bi-GRU model captures short-term dependencies in the first layer while the second and third layers refine these representations by capturing longer-term dependencies and integrating context from both directions. This means that two GRUs operate in opposite directions, i.e., one forward and one backward.
Figure 5. High-level view of the Bi-GRU architecture.
The forward GRU processes the sequence from the first token to the last, while the backward GRU processes it in reverse, from the last token to the first. For each token t i , the forward GRU produces a hidden state h t i forward , and the backward GRU produces a hidden state h t i backward . The hidden state updates for each direction are defined with Equations (10) and (11).
h t i forward = GRU forward ( h t i 1 forward , h t i )
h t i backward = GRU backward ( h t i + 1 backward , h t i )
At each time step i, the forward and backward hidden states are concatenated to form the final hidden state h t i for that time step as h t i = [ h t i forward , h t i backward ] , where h t i R 2 d and d is the size of the hidden state for each direction. To enhance the network’s focus on important tokens, an attention mechanism α t i is applied on top of the Bi-GRU, assigning a weight to each hidden state h t i to represent the importance of the corresponding token in the final prediction. The attention score for each token is computed through Equation (12),
α t i = exp ( v T tanh ( W h h t i ) ) j exp ( v T tanh ( W h h t j ) ) ,
where v and W h are learnable parameters, and α t i represents the attention weight for token t i . The final context vector is then computed as a weighted sum of the hidden states (Equation (13)).
c = i α t i h t i
Once all hidden states h t i have been generated, they are aggregated to form a summary of the entire sequence using the Final Time Step method. This means that hidden state of the final token h t n is used as the representation for the entire sequence. By doing h final = h t n , we capture the refined context of the entire sequence, combining information from both directions.
The final hidden state h final is fed to a connected dense layer to generate the output probabilities as indicated in Equation (14),
P Bi - GRU ( y ) = softmax ( W · h final + b ) .
Here, W is the weight matrix of the dense layer, b is the bias vector, and y represents the output class (0 for fake news, 1 for real news).
Then, a dense layer applies a linear transformation followed by a softmax activation function to produce the probability distribution over the class labels (real or fake news). This ensures that the output is a valid probability distribution over the classes, summing to 1 (Equation (15)).
P Bi - GRU ( y = 1 ) = exp ( W 1 · h final + b 1 ) exp ( W 0 · h final + b 0 ) + exp ( W 1 · h final + b 1 )
With reference to Equation (15), P Bi - GRU ( y = 1 ) denotes the probability of the news being real, while P Bi - GRU ( y = 0 ) denotes the probability of it being fake. The final classification decision is made by selecting the class with the highest probability. This is easily done via Equation (16).
y ^ = arg max P Bi - GRU ( y ) ,
where y ^ represents the predicted class, with 0 indicating fake news and 1 indicating real news.

5.6. Transformer-XL

The Transformer-XL [48,59] is used to capture long-range dependencies in sequential data by introducing a segment-level recurrence mechanism. This allows the model to maintain a memory of past segments, which is especially useful for long-form text, such as news articles, where information might be distributed across multiple segments. The textual embeddings H T k , generated from ViLBERT, are passed into Transformer-XL. The sequence of embeddings is split into segments, where each segment contains l tokens, and the entire sequence may span multiple segments. H T k = { h t 1 , h t 2 , , h t n } where n is the number of tokens in the sequence, and each segment has l tokens. Transformer-XL processes each segment, transferring the hidden state from prior segments to capture long-range dependencies. The hidden state for each token in the current segment is updated with Equation (17),
h t i current = Transformer - XL ( h t i previous , h t i ) ,
where h t i previous is the hidden state from the previous segment.
A relative positional encoder is used to capture the relative distances between tokens t i and t j , and their attention score is computed. Once all segments have been processed, the hidden state of the last token in the sequence h t n is taken as the final representation of the entire sequence as h final = h t n . The final hidden state h final is passed through a dense layer followed by a softmax function to produce the output probabilities for the class labels. This is done through Equation (18),
P Transformer - XL ( y ) = softmax ( W · h final + b ) ,
where W and b are learnable parameters. The final class is determined by selecting the label with the highest probability with y ^ = arg max P Transformer - XL ( y ) .

5.7. DistilBERT

DistilBERT [49,60] is a smaller and faster variant of BERT, retaining 97% of BERT’s performance. It uses 6 transformer layers instead of 12, making it ideal for real-time predictions. Textual embeddings H T k , generated from ViLBERT, are passed into DistilBERT and processed through its transformer layers, where a multi-head self-attention mechanism is applied, allowing each token to attend to every other token in the sequence. After the attention score computation, the representation of each token is passed through a feed-forward network to refine its embedding. The hidden state of the [CLS] token, h [ CLS ] , is taken as the representation for the entire sequence as h final = h [ CLS ] . The final hidden state is passed through a dense layer followed by a softmax function to generate the class probabilities as P DistilBERT ( y ) = softmax ( W · h final + b ) . The final class is determined by selecting the label with the highest probability using Equation (19):
y ^ = arg max P DistilBERT ( y )

5.8. XLNet

XLNet [50,61] is a permutation-based transformer model that improves on BERT by allowing the model to capture bidirectional context without masking. It does this by training on all possible permutations of the input sequence. The textual embeddings H T k , generated from ViLBERT, are passed into XLNet. Instead of applying self-attention to the original sequence, XLNet processes multiple permutations of the sequence and allows it to capture relationships in all possible orderings. The attention mechanism is applied, and each token’s embedding is passed through a feed-forward network. The hidden state of the final token h t n is used as the representation of the entire sequence as h final = h t n . The final hidden state is passed through a dense layer and softmax function to compute the output probabilities as P XLNet ( y ) = softmax ( W · h final + b ) . The final class is selected by taking the label with the highest probability using Equation (20):
y ^ = arg max P XLNet ( y )

5.9. Ensemble Approach

In this approach, four base classifiers—Bi-GRU, Transformer-XL, DistilBERT, and XLNet—are combined using soft voting and a stacking ensemble method, with the T5 meta-classifier making the final prediction. This ensemble method maximises the strengths of each model, improving the overall accuracy and robustness of fake news detection. Each base classifier C i , where C i {Bi-GRU, Transformer-XL, DistilBERT, XLNet}, outputs a probability distribution over the class labels y { 0 , 1 } , with y = 0 representing fake news and y = 1 representing real news. The output probabilities for each classifier can be represented as
P C i = P C i ( y = 0 ) , P C i ( y = 1 ) , C i { Bi - GRU , Transformer - XL , DistilBERT , XLNet }
These vectors P C i represent the probabilities that each classifier C i assigns to the news post being either fake or real. The outputs from these classifiers provide valuable insights into the classification task by focusing on different aspects of the input data, such as short-term dependencies with Bi-GRU, long-range dependencies with Transformer-XL, and contextual representations with DistilBERT and XLNet.

5.10. Soft Voting Mechanism

After obtaining the probability distributions from each base classifier, a soft voting mechanism is applied to aggregate their outputs. The combined probability distribution for class y is calculated by averaging the individual outputs of the four classifiers via Equation (22), where P final ( y ) represents the averaged probability for class y (fake or real).
P final ( y ) = 1 4 P Bi - GRU ( y ) + P Transformer - XL ( y ) + P DistilBERT ( y ) + P XLNet ( y )
Soft voting averages predicted class probabilities rather than using hard majority votes, thereby utilising richer confidence information from each base classifier. This approach offers several key advantages: (1) it balances the complementary strengths of diverse models (e.g., Bi-GRU’s sequential processing, XLNet’s permutation-based dependencies), reducing individual biases and variance; (2) it mitigates the impact of over-confident but erroneous predictions from any single classifier; and (3) it consistently yields higher accuracy and robustness in ensemble settings, particularly on noisy multimodal data such as social media fake news, as supported by recent ensemble-based detection studies.
The resulting P final ( y ) is then concatenated with individual base probabilities and fed to the T5 meta-classifier for the ultimate decision.

5.11. Meta-Classifier T5

Alongside soft voting, we employ a stacking ensemble approach with a T5 meta-classifier, where the T5 model [62] is trained to combine the outputs of the base classifiers and make the final prediction by learning how these outputs relate to each other. The concatenated probability outputs of all base classifiers form the input to the T5 model as indicated in Equation (23).
P concat = P Bi - GRU ( y = 0 ) , P Bi - GRU ( y = 1 ) , P Transformer - XL ( y = 0 ) , P Transformer - XL ( y = 1 ) , P DistilBERT ( y = 1 ) , P XLNet ( y = 0 ) , P XLNet ( y = 1 )
This 8-dimensional vector contains the probability scores for both class labels from each of the four base classifiers. Because of its transformer-based architecture, the T5 model processes the concatenated vector P concat to learn and optimise the final prediction based on the relationships between the outputs of the base classifiers. Formally, the T5 meta-classifier generates the final probability distribution over the class labels. To make the final classification decision y ^ , the class with the highest probability from the T5 meta-classifier’s output is selected (Equation (24)). The final prediction y ^ , 0 (fake) or 1 (real), is the optimal decision from the ensemble of classifiers.
y ^ = arg max P meta ( y )

5.12. Loss Function

The overall loss function L used for training the ensemble model minimises the classification error between the true labels y k and the predicted labels y ^ k generated from the ensemble’s output. The total loss accounts for both text and image modalities. For each news post p k , the total loss is calculated as Equation (25).
L ( y k , y ^ k ) = α · L text ( y k , y ^ k text ) + β · L image ( y k , y ^ k image )
where α and β are weighting factors that balance the contributions of the text and image modalities. L text and L image represent the cross-entropy loss for the text and image components, respectively. The cross-entropy loss L for each modality is defined as Equation (26).
L ( y , y ^ ) = i = 1 C y i log ( y ^ i )
where C is the number of classes (e.g., fake or real news), and y i and y ^ i represent the true and predicted probabilities for each class. To further balance the contributions of the text and image features, we introduce a trade-off parameter μ , where μ [ 0 , 1 ] . The final loss function for training the stacked ensemble with the T5 meta-classifier is optimised by minimising the weighted sum of the losses from the text and image predictions. The updated loss function is shown in Equation (27).
L ensemble ( y k , y ^ k ) = i = 1 C μ · y k i log ( y ^ k i text ) + ( 1 μ ) · y k i log ( y ^ k i image )
Here, μ controls the contribution of the text-based predictions y ^ k i text , and ( 1 μ ) controls the contribution of the image-based predictions y ^ k i image . C represents the number of classes, and y k i is the true label. This final ensemble loss ensures both text and image features contribute effectively to improving fake news detection accuracy during training.

6. Experimental Phase

The ensemble learning methodology for detecting fake news, using a soft voting strategy combined with a T5 meta-classifier, is outlined in Algorithm 1, with detailed steps explained in the following subsections. The approach begins by extracting both textual and visual features from a news post using ViLBERT. Four classifiers—Bi-GRU, Transformer-XL, DistilBERT, and XLNet—are then used to process the input independently, each producing a probability distribution for the news being fake or real. These distributions are averaged through soft voting to produce a final probability. The outputs of all classifiers are concatenated and fed into a T5 meta-classifier, which refines the prediction by generating a final probability distribution. The class with the highest probability is selected as the predicted label, providing a robust approach to fake news detection by combining the strengths of multiple classifiers with a meta-classifier.

6.1. Hyperparameter Tuning

The hyperparameters of the EMFND model are optimised through a combination of grid search and manual tuning. It is important to note that the concept of memory length applies only to models like Transformer-XL, which utilise segment-level recurrence to manage long-range dependencies. In contrast, models such as Bi-GRU, DistilBERT, and XLNet do not require a memory length parameter, as they rely on mechanisms such as gated units and bidirectional attention to handle sequential information.
Both Bi-GRU and DistilBERT implement dropout rates to prevent overfitting during training. Transformer-XL omits standard dropout to better capture long-term dependencies. Meanwhile, XLNet employs a dropout rate of 0.1 for effective regularisation. All models are trained for 40 epochs, which is sufficient to achieve convergence without overfitting, thus balancing performance with training time. The AdamW optimiser is selected for all models because it effectively decouples weight decay from gradient optimisation, improving generalisation and reducing overfitting.
The hyperparameters listed in Table 3 are designed to optimise the ensemble’s performance by striking a balance between complexity, efficiency, and generalisation.
Table 3. Hyperparameters for the classifiers used in the EMFND Model.

6.2. Evaluation Metrics

Established metrics are used to assess the performance of the EMFND mode, such as accuracy, precision, recall, F1-score, and AUC-ROC. For details, see [63].

7. Results and Analysis

We report confusion matrices for the classification results and compare the models in the ensemble. The performance of EMFND for each dataset is summarised in Table 4.
Table 4. Performance of the EMFND framework on Fakeddit and Weibo datasets.
The model achieved an accuracy of 96% for detecting fake news and 97% for true news on the Fakeddit dataset. Precision values are 96% for fake news on Fakeddit and 93% on Weibo. These values demonstrate the model’s strength in identifying fake news. Recall scores are 95% for fake news and 96% for true news on Fakeddit. On Weibo, recall is 94% for both fake news and true news. F1-scores are 97% for fake news and 95% for true news on Fakeddit, and 95% for fake news and 94% for true news on Weibo. These metrics highlight the model’s ability to balance precision and recall. AUC-ROC values of 0.94 on Fakeddit and 0.91 on Weibo demonstrate the model’s discriminative power in distinguishing between fake and true news. These results confirm the model’s effectiveness across both datasets.

7.1. Performance Analysis

The model performs slightly better on the Fakeddit dataset due to its larger size and more diverse set of examples compared to Weibo. The F1-score and AUC-ROC metrics further confirm the reliability of the model in detecting fake news and distinguishing it from real news.
Table 5 presents the classifier-wise performance of the proposed ensemble framework on the Fakeddit and Weibo datasets. The results clearly show how each classifier performs in detecting fake and real news, highlighting the strengths and limitations of individual models compared to the ensemble approach.
Table 5. Classifier-wise performance on Fakeddit and Weibo datasets (%).
Bi-GRU performs reasonably well but falls short compared to the other models, especially on the Weibo dataset, where it exhibits lower accuracy and precision due to its focus on short-term patterns. In contrast, Transformer-XL captures long-range dependencies, thereby achieving up to 94% accuracy for fake news on Fakeddit and clearly outperforming Bi-GRU. DistilBERT achieves 96% accuracy for true news on Fakeddit, making it a reliable choice across both datasets. XLNet also performs well, maintaining precision and recall above 92% on both datasets. Note that EMFND (i.e., the output of the T5 meta-classifier) significantly outperforms the individual models, especially on Fakeddit, with 97% accuracy for true news and 96% for fake news, as shown in Figure 6. This demonstrates that better overall performance is obtained by integrating the strengths of all models.
Figure 6. Classifier-wise model performance analysis.
To improve our model, we incorporate logical constraints to guide the learning process of the latent variables z. By varying the trade-off parameter μ in the loss function, we investigate how different levels of logical supervision impact the quality of z and the overall performance of the model.
As shown in Figure 7, low μ (0.1) results in low hard and soft logic accuracies ( Acc h and Acc s ), indicating poor detection of deceptive patterns. Increasing μ improves performance, reaching a peak at μ = 0.5 , after which performance stabilises or slightly declines, suggesting the need for balanced logical supervision. Notably, overall accuracy remains stable across different μ values, demonstrating the robustness of our approach.
Figure 7. EMFND evaluation at varying loss function weights.

7.2. Impact of Sentiment Analysis

The sentiment scores are classified into two categories: real and fake. Table 6 presents the aggregated emotion scores, highlighting the differences in performance when sentiment analysis is enabled or disabled in the textual features for fake news detection.
Table 6. Influence of sentiment analysis on Fakeddit and Weibo Datasets (%).
In the Fakeddit dataset, the model achieves higher accuracy (96.83% instead of 94.12%) and a higher F1-score (97.11% instead of 95.17%) when sentiment analysis is enabled, as visualised in Figure 8. Similarly, for Weibo, it achieved an accuracy of 94.83% instead of 92.49% and an F1-score of 95.89% instead of 90.11%, as illustrated in Figure 9.
Figure 8. Impact of sentiment analysis on EMFND evaluation with Fakeddit dataset.
Figure 9. Impact of sentiment analysis on EMFND evaluation with Weibo dataset.
True news typically has neutral sentiment, while false news uses extreme sentiment to evoke extreme emotions. Enabling sentiment analysis helps the model distinguish between true and false news by detecting these emotional cues, improving accuracy and F1-scores, especially in political or sensitive topics.
Figure 10 and Figure 11 show that semantic similarity ( s i m sem ( T k , I k ) ) is consistently higher than textual similarity ( s i m text ( T k , I k ) ) across real and fake news in both the Fakeddit and Weibo datasets. This can be attributed to BERT, whose embeddings capture richer contextual information and lead to a higher semantic alignment between text and image. Interestingly, both s i m text ( T k , I k ) and s i m sem ( T k , I k ) tend to be higher for fake news compared to real news. This is likely because fake news often uses emotionally or semantically aligned language and imagery to mislead readers. In the Fakeddit dataset, e.g., many fake news items show high semantic similarity scores in the range [ 0.25 , 0.40 ] . This reflects the model’s sensitivity to semantic manipulation in fake news. No clear pattern for post-training similarity ( s i m post ( T k , I k ) ) is observed across datasets, which could suggest that the learnt representations of fake and real news vary depending on how the modalities interact during training. However, the cumulative distribution functions (Figure 10 and Figure 11) confirm that image–text similarity tends to be higher for fake news across most cases.
Figure 10. Statistical evaluation of the Fakeddit dataset shows significant differences between real and fake news in terms of (a) textual, (b) semantic, and (c) post similarities, (df) both real and fake news probability ratio with Mann–Whitney test (p-values are 3.62 × 10−7, 1.69 × 10−9, and 2.61 × 10−8), respectively.
Figure 11. Statistical evaluation of the Weibo dataset shows significant differences between real and fake news based on (a) textual, (b) semantic, and (c) post similarities, (df) both real and fake news probability ratio with Mann–Whitney test (p-values of 1.70 × 10−10, 1.57 × 10−9, and 1.81 × 10−11), respectively.
The evolution of image–text similarity from pre-training (using s i m text ( T k , I k ) and s i m sem ( T k , I k ) ) to post-training ( s i m post ( T k , I k ) ) further demonstrates the model’s ability to distinguish between real and fake news more effectively after training. Initially, real and fake news have similar scores, but after training, fake news shows greater divergence, especially in Fakeddit, where higher image–text similarity improves detection performance.
Figure 12a–c illustrates several fake news examples from the Fakeddit and Weibo datasets. It highlights retrieved image contexts, deceptive patterns, and predicted authenticity labels. In Figure 12a, the text mentions a CNY 20 insurance purchase alongside references to chemotherapy and liver cancer. The model retrieves image contexts related to chemotherapy and treatment, identifies ‘Image Manipulation’, and predicts the label as fake due to inconsistencies between the text and image. In Figure 12b, the text describes the Boston Explosion involving an 8-year-old child, while the retrieved images relate to fitness, power t-shirts, and a marathon. The model predicts Fake, detecting inconsistency between the tragic event and the unrelated image contexts. In Figure 12c, the text discusses a 10th-grade math exercise in a U.S. textbook, with retrieved images aligning with a mathematics textbook and high school context. The model predicts Real, as the image and text are consistent and authentic.
Figure 12. Fake (a,b) and Real (c) news examples from Fakeddit and Weibo dataset with EMFND prediction.

7.3. Ablation Study

Ablation studies are vital for assessing component contributions in complex models and revealing the most crucial features influencing model performance. We do this on both the Fakeddit and Weibo datasets and report results in Table 7. Fake News Detection Without Social Context ( FND S ) is performed where the social context feature is removed. The model relies solely on knowledge extraction from text and image data for fake news detection. Fake News Detection Without Image ( FND I ) is excluded in this variant. The binary classifier processes knowledge derived from the combination of shared text and social context features across all classifiers, without the inclusion of image data. Fake News Detection Without Image and Social Context ( FND S + I ) is performed as image and social context features are removed in this model. The model relies only on the knowledge extracted from the text as input for training. Fake News Detection Without Knowledge Extraction ( FND K ) is executed without the knowledge extraction (KE) component. The other components operate without additional external knowledge integration. Fake News Detection Without Social Context + Knowledge Extraction ( FND S + K ) removes the knowledge extraction and social context features. It leaves only image and text data as input for the model’s training. In Fake News Detection Without Image + Knowledge Extraction ( FND I + K ), the model is trained without the image feature and the knowledge extraction module, utilising only text and social context data. Fake News Detection Without Image, Social Context, and Knowledge Extraction ( FND S + I + K ) is the final variation. The model is trained without image, social context, or knowledge extraction. It only uses the text data for training, and the averaged outputs are passed into a binary classifier for fake news detection.
Table 7. Ablation study performance analysis for Fakeddit and Weibo datasets.
Removing individual components like social context ( FND S ) or images ( FND I ) causes a noticeable performance drop. The absence of both social context and images ( FND S + I ) results in an even greater decrease in accuracy, precision, and recall. It highlights their importance in multimodal fake news detection. The sharp decline in models like FND S + I + K shows that combining textual, visual, and contextual features is essential for effective fake news detection. This reinforces the conclusion that the ensemble model, by using all features, provides the most accurate performance on both datasets.
Removing individual components such as social context ( FND S ) or images ( FND I ) causes a performance drop. This is expected, as both social context and image features provide supplementary information that is essential for accurate fake news detection. Social context provides metadata or external cues that help in interpreting the news content, such as user engagement or trends, which are not always captured by text or images alone. Likewise, image data plays a vital role in identifying manipulative or misleading visuals that often accompany fake news. The absence of both social context and image features ( FND S + I ) results in an even greater decrease in accuracy, precision, and recall. This highlights the importance of both visual and contextual features in multimodal fake news detection. The sharp decline in performance when removing social context, image, and knowledge extraction ( FND S + I + K ) demonstrates that combining textual, visual, and contextual features is essential for effective fake news detection. Without these multiple sources of information, the model is unable to capture the full context and nuances, leading to a significant performance drop.
This ablation study reinforces the conclusion that the EMFND model, which uses all features (text, image, social context, and knowledge extraction), provides the most accurate performance across both datasets. It confirms that multimodal learning, which combines different sources of information, is the key to achieving the best results in fake news detection.

7.4. Baseline Models

The baseline models selected for comparison are carefully chosen to cover a range of approaches, from simple traditional machine learning techniques to advanced deep learning architectures. Each baseline model offers insights into how different methods handle multimodal fake news detection by focusing on text, images, or both. These choices provide a well-rounded comparison and are consistent with the multimodal ensemble approach used in the EMFND model. To ensure a fair comparison, hyperparameter tuning was conducted for all baseline models, following the same methodology as for the EMFND model. Grid search was used to optimise key hyperparameters for each model, such as learning rate, batch size, dropout rate, and number of layers. This optimisation process ensured that each baseline model was trained under optimal conditions for a fair evaluation of performance. Hyperparameters were selected based on performance in a validation set, and we ensured that the models were trained for a sufficient number of epochs to achieve convergence without overfitting.
  • BERT + ResNet (Late Fusion) [64] combines BERT for text analysis and ResNet [65] for image analysis. Text and image data are processed separately and then fused at a later stage for final classification.
  • VGG-16 + LSTM [66] uses VGG-16 for image feature extraction and LSTM for text analysis. The extracted features are concatenated and passed through a fully connected layer for classification.
  • CLIP (Contrastive Language-Image Pretraining) [67] is a multimodal model by OpenAI that processes text and images by projecting them into a shared latent space, trained via contrastive learning to match images with their text descriptions.
  • Logistic Regression (LR) + Handcrafted Features (HF) [68] serves as a baseline, using TF-IDF (Term Frequency-Inverse Document Frequency) for text and colour histograms and pixel-based features for images.
  • Random Forest [69] is an ensemble algorithm that builds multiple decision trees and averages their predictions. In this baseline, text features from BERT embeddings and image features from ResNet are concatenated and classified using Random Forest.
The comparative analysis using the selected baseline models provides a comprehensive understanding of how different techniques handle multimodal fake news detection on Fakeddit and Weibo datasets (Table 8: Each baseline model contributes differently in terms of evaluation metrics.
Table 8. Comparative analysis with baseline models on Weibo and Fakeddit datasets (%).
The comparative analysis shows that advanced deep learning models like BERT + ResNet (Late Fusion) and CLIP outperform traditional machine learning methods such as Logistic Regression and Random Forest across both datasets. BERT + ResNet achieves an accuracy of 0.89 for fake news and 0.90 for true news on Fakeddit, while on Weibo, it reaches 0.90 and 0.91, respectively. CLIP demonstrates even higher performance, achieving 0.93 accuracy for fake news and 0.92 for true news on Fakeddit, and 0.91 for both on Weibo, due to its contrastive learning technique that aligns text and image features effectively. VGG-16 + LSTM, with 0.85 accuracy for fake news and 0.84 for true news on Fakeddit, performs moderately well but struggles with more complex data relationships compared to deeper models. Logistic Regression with Handcrafted Features and Random Forest show the limitations of traditional techniques, with Logistic Regression reaching only 0.75 accuracy for fake news and 0.74 for true news on Fakeddit, while Random Forest fares slightly better with 0.82 for fake news and 0.83 for true news. Ultimately, our proposed EMFND model outperforms all baselines, achieving the highest accuracy and F1-scores across both Fakeddit and Weibo, confirming the effectiveness of our multimodal ensemble approach in fake news detection.

7.5. Comparitive Analysis

To validate the effectiveness of the proposed model, it is compared against other state-of-the-art multimodal fake news detection models using the same dataset in Table 9 and Table 10. The key models used for comparison are as follows:
  • CLIP [70] introduced a multimodal framework for fake news detection by integrating text and visual data. It employs NLP for text preprocessing, the DeepL translator for language consistency, and LSTM networks for analysing text sequences. For image analysis, it uses the CLIP model, and the combined features are classified as real or fake in the decision-making layer.
  • FakeNED [13] presented an ensemble learning-based method for detecting multimodal fake news. It utilised Visual BERT (V-BERT) to generate embeddings for text and Faster-RCNN for images. These embeddings are input into a deep-learning ensemble model for training and testing.
  • Ref. [32] extracted textual and visual features to improve detection accuracy. Textual features are obtained using pre-trained BERT, GRU, and CNN models, while image features are extracted with ResNet-CBAM. These features are fused, and dimensionality is reduced by using an auto-encoder. The features are classified using an FLN classifier to detect fake news.
  • CLIP-GCN [71] proposed a Clip-GCN multimodal fake news detection model that uses the Clip pre-training model for joint semantic feature extraction from image–text data. The model employed adversarial training to extract inter-domain invariant features and graph convolutional networks (GCNs) to utilise intra-domain knowledge.
  • FND-Clip [27] detects fake news by using CLIP to extract and combine deep text and image features. It weights these features based on similarity and uses a modality-wise attention module to improve feature aggregation for accurate detection.
  • Event-Radar [72] performed event-level multimodal analysis and credibility-based multi-view fusion for detecting fake news effectively.
  • MACCN [73] improved fake news detection by fusing textual and visual features through distinct encoders and an Adaptive Co-attention Fusion Network. It strengthened correlations between the modalities for better representation.
  • NSLM [74] Neuro-Symbolic Latent Model detected news accuracy and deceptive patterns using two-valued latent variables learned through variational inference and symbolic logic rules.
  • TMGWO [75] uses the TMGWO genetic algorithm to extract optimal features from text, metadata, and author embedding data. Fusion methods combine these multimodal features, and a two-layer MLP is used for fake news detection.
  • Multimodal CNN [76] performed a fine-grained classification of fake news using unimodal and multimodal approaches. The best results were achieved using a multimodal approach based on a Convolutional Neural Network (CNN) that combines text and image data.
  • HiPo [77] is a multimodal method that combines features from graphical and textual content. It assesses the truthfulness of a social media post by constructing its historical context from previous similar, off-label posts, enabling online detection without the need for a context or knowledge database.
  • SDSA [78] refers to the Semantic Distillation and Structural Alignment (SDSA) network. It includes a semantic distillation module that retains task-relevant information and removes redundancies from modality-specific features. A triple similarity alignment module is proposed to maintain structural information by aligning intra-modal, inter-modal, and fused feature similarities.
  • Fakefind [79] is a hybrid model that combines CNN and RNN to integrate multimodal features for efficient rumour detection. It uses a CNN-based knowledge extractor to extract stance from post–reply pairs and incorporates stance representations for fake news detection.
Table 9. A comparative analysis of the proposed EMFND model with existing studies on the fakeddit dataset.
Table 9. A comparative analysis of the proposed EMFND model with existing studies on the fakeddit dataset.
ReferenceModelFake NewsReal News
AccuracyF1Prec.RecallAccuracyF1Prec.Recall
[76]CNN0.870.870.860.870.860.850.850.86
[74]HiPo0.860.870.850.860.850.860.870.85
[79]Fakefind0.840.840.830.840.830.820.810.83
[70]LSTM CLIP0.930.930.920.930.920.920.910.92
[13]V-BERT CNN LSTM0.880.910.890.900.890.900.890.89
[73]MACCN0.900.890.880.890.910.900.900.91
[74]NSLM0.920.910.910.920.930.920.910.93
[75]TMGWO0.900.850.840.850.880.870.870.88
[78]SDSA0.940.940.930.940.930.930.920.93
ProposedEMFND0.960.970.970.960.970.950.960.96
Table 10. Comparison of models by accuracy, F1, precision, and recall on Weibo dataset.
Table 10. Comparison of models by accuracy, F1, precision, and recall on Weibo dataset.
ReferenceModelFake NewsReal News
AccuracyF1Prec.RecallAccuracyF1Prec.Recall
[27]FND-CLIP0.900.900.890.900.890.890.880.89
[74]HiPo0.880.910.870.890.870.890.880.87
[79]Fakefind0.930.880.870.880.920.870.860.88
[32]LSTM CLIP0.880.870.860.870.870.860.850.87
[71]CLIP GCN0.860.870.850.860.850.860.860.85
[72]Event-Radar0.910.910.900.910.900.890.880.89
[73]MACCN0.920.900.910.900.910.890.880.90
[74]NSLM0.880.890.870.880.870.880.860.88
[78]SDSA0.910.920.900.920.900.910.910.90
ProposedEMFND0.940.950.940.950.930.940.920.94
The comparative analysis for the Fakeddit dataset shows that the proposed EMFND model outperforms all others in both fake and real news detection. With 0.96 accuracy for fake news and 0.97 for real news, and F1-scores of 0.97 and 0.95, it demonstrates superior performance by leveraging multimodal features. SDSA and LSTM-CLIP perform well, but their results are close but not as consistent. EMFND’s combination of text, image, and social context offers a clear advantage. HiPo, MACCN, and NSLM also perform respectably, but their F1-scores are lower. This reinforces the effectiveness of the ensemble approach in EMFND. It confirms the importance of multimodal feature integration and ensemble learning for accurate fake news detection.
For the Weibo dataset, The EMFND model shows the best performance, with 0.94 accuracy and an F1-score of 0.95 for fake news, and 0.93 accuracy and 0.94 F1 for real news. This highlights the efficiency of the model, especially in text-heavy datasets. Although models like FND-CLIP, Event-Radar, and MACCN perform well, they fall short of EMFND’s consistency across all metrics. The strong results of SDSA, Event-Radar, and HiPo show that multimodal approaches work well for Weibo. However, integration of EMFND features into an ensemble framework proves to be the most effective for multimodal and text-centric data.

8. Discussion

The experimental results show that the proposed EMFND framework significantly outperforms individual baseline models and recent state-of-the-art multimodal approaches. This advantage mainly stems from its stacked ensemble strategy, which captures complementary textual and visual features and integrates them via soft voting and a T5 meta-classifier. The use of transformer-based models (DistilBERT, XLNet, Transformer-XL) provides strong contextual understanding of complex linguistic patterns, while ViLBERT’s two-stream cross-modality attention mechanism enables precise alignment between images and text. The inclusion of VADER sentiment analysis and Image–Text Contextual Similarity further enhances the model’s sensitivity to emotional manipulation and cross-modal inconsistencies—common hallmarks of sophisticated fake news.
One of the key observations is that the ensemble model performs better in cases where fake news exhibits multimodal inconsistencies, such as when the accompanying text and image do not align (e.g., a misleading headline paired with an unrelated image). For instance, in scenarios where the text discusses a serious event like a terrorist attack, but the retrieved image represents an unrelated context, such as fitness or fashion, the ensemble model can effectively identify these inconsistencies and classify the news as fake. This suggests that the combination of models allows for a more nuanced understanding of multimodal content, which is critical for detecting image-heavy fake news. Moreover, the stacked ensemble approach leverages the complementary strengths of individual models. Transformer-based models excel at handling nuanced linguistic features, such as sarcasm, misinformation cues, and sentiment shifts, which are common in text-heavy fake news. ViLBERT’s image–text alignment capabilities play a crucial role in identifying inconsistencies in image-heavy fake news where visual deception is prevalent. By combining the outputs of these specialised models, the ensemble approach achieves more accurate results across both text-heavy and image-heavy fake news cases.
The findings also highlight broader implications for misinformation countermeasures. By achieving 96% accuracy on the large-scale Fakeddit dataset and 94% on the culturally distinct Weibo dataset, the proposed approach demonstrates strong generalisation across different languages, platforms, and misinformation styles. This suggests that ensemble-based multimodal architectures can mitigate dataset-specific biases more effectively than single-model solutions. Furthermore, the framework’s ability to detect evolving fake news tactics—such as incremental modifications to existing true stories—offers promise for combating emerging threats, including AI-generated deepfakes and synthetic media, where visual authenticity is increasingly difficult to assess manually.
From a societal perspective, improving automated fake news detection at this level can support platform moderation, fact-checking organisations, and public awareness initiatives, potentially reducing the spread of harmful misinformation during critical events (e.g., elections, public health crises). However, deployment must be accompanied by careful consideration of ethical issues, including potential biases in training data and the risk of over-reliance on automated systems for content moderation.
Despite their differences—Fakeddit’s informal, satirical content and Weibo’s culturally specific, domain-specific news—the EMFND framework generalises well and performs strongly on both datasets. These datasets contain potential biases that may limit generalisation to real-world misinformation. Fakeddit shows lexical biases, such as over-reliance on entities (e.g., proper nouns) and a Reddit-specific satire style, which can push the model toward surface cues instead of deeper context, especially for text-heavy fake news. Weibo has time-related, domain-specific (daily-life news), and cultural biases, affecting performance on other platforms or languages. Moreover, platforms differ in linguistic structures, posting styles, user interactions, language formality, and content types (image- vs. text-heavy), further constraining transferability. These platform-specific biases may limit the model’s ability to generalise across domains. To address this, we used Image–Text Contextual Similarity analysis to detect misalignments between text and images in news posts, such as misleading headlines paired with unrelated images. This is key to improving robustness to text–image discrepancies. Although oversampling and SMOTE addressed class imbalance, further work is needed to strengthen robustness against evolving misinformation tactics (e.g., AI-generated content and deepfakes) and to enhance generalisation across platforms, languages, and posting styles in real-world settings.
Although the EMFND framework integrates several large pre-trained models, this design choice is justified by the clear performance advantages over baselines, as the ensemble leverages complementary architectural strengths to address multimodal challenges that individual models cannot adequately handle. Each component adds distinct, complementary capabilities: ViLBERT aligns modalities, Bi-GRU captures bidirectional sequences, Transformer-XL models long-range dependencies, DistilBERT encodes context efficiently, XLNet improves robustness via permutation training, and T5 fuses ensemble predictions. Computational cost is reduced through lightweight variants and parameter sharing, enabling training and inference on standard GPUs. Future work includes quantisation and knowledge distillation to further speed up inference and improve real-time scalability. Ongoing work on quantisation and knowledge distillation will further reduce resource requirements, enhancing real-time applicability and deployment on edge devices while preserving the detection accuracy that justifies the current design.

Limitations

Although the ensemble model is successful, it has limitations. The ensemble approach requires balanced textual and visual information. Predominantly text-based articles or those lacking meaningful images hinder the model’s multimodal capabilities. This results in lower accuracy for detecting text-only or image-scarce fake news. Additionally, fake news using advanced image manipulation closely matching the text might elude detection, as the model depends on clear image–text inconsistencies.

9. Conclusions

In this study, we proposed the Ensemble-based Multimodal Fake News Detection framework, which integrates ViLBERT’s cross-modal architecture with VADER sentiment analysis and Image–Text Contextual Similarity features. By leveraging a stacked ensemble of Bi-GRU, Transformer-XL, DistilBERT, and XLNet classifiers—combined via soft voting and a T5 meta-classifier—the model effectively captures textual nuances, visual cues, and cross-modal inconsistencies. Experimental results show that the proposed model outperforms the current state of the art on the Fakeddit and Weibo multimodal datasets, showing robust performance across domains with strong generalisation across diverse domains and languages. Ablation studies further validate the contributions of individual components, confirming the framework’s efficiency and robustness in detecting sophisticated misinformation. Its efficient design, confirmed by an ablation study, still allows for enhancements, particularly in handling complex fake news cases with video or audio. We plan to incorporate multimedia sources to improve accuracy and integrate domain-specific features to address challenges and broaden its platform applications in the future. Moreover, we will work towards minimising the computational complexity of the proposed framework, which is currently reliant on large pre-trained models (ViLBERT and XLNet), necessitating considerable memory and processing resources. This poses scalability challenges, particularly in the context of real-time applications or when handling extensive datasets. This limitation will be addressed in future work through the implementation of quantisation and knowledge distillation. Our objective is to reduce the size and memory demands of the model without significantly affecting its performance.

Author Contributions

Conceptualization, H.Z., H.E. and F.C.; Methodology, M.A., H.Z., A.J. and M.S.; Software, M.A. and O.M.; Investigation, O.M.; Resources, O.M.; Data curation, M.S.; Writing—original draft, M.A., A.J., Z.T. and F.C.; Writing—review and editing, O.M., H.E. and F.C.; Visualization, A.J., Z.T. and F.C.; Supervision, H.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no funding.

Data Availability Statement

The data presented in this study are openly available in Fakedit at https://doi.org/10.48550/arXiv.1911.03854, reference number [14], as well as in Weibo at https://doi.org/10.21227/hjmr-kr22, reference [47].

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this article.

References

  1. Biradar, S.; Saumya, S.; Chauhan, A. Combating the infodemic: COVID-19 induced fake news recognition in social media networks. Complex Intell. Syst. 2023, 9, 2879–2891. [Google Scholar] [CrossRef] [PubMed]
  2. Bilal, M.; Almazroi, A.A. Effectiveness of fine-tuned BERT model in classification of helpful and unhelpful online customer reviews. Electron. Commer. Res. 2023, 23, 2737–2757. [Google Scholar] [CrossRef]
  3. Angizeh, L.B.; Keyvanpour, M.R. Detecting Fake News using Advanced Language Models: BERT and RoBERTa. In Proceedings of the 2024 10th International Conference on Web Research (ICWR), Tehran, Iran, 24–25 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 46–52. [Google Scholar]
  4. Mende, M.; Ubal, V.O.; Cozac, M.; Vallen, B.; Berry, C. Fighting Infodemics: Labels as Antidotes to Mis- and Disinformation?! J. Public Policy Mark. 2024, 43, 31–52. [Google Scholar] [CrossRef]
  5. Choudhary, A.; Arora, A. Assessment of bidirectional transformer encoder model and attention based bidirectional LSTM language models for fake news detection. J. Retail. Consum. Serv. 2024, 76, 103545. [Google Scholar] [CrossRef]
  6. Asudani, D.S.; Nagwani, N.K.; Singh, P. Impact of word embedding models on text analytics in deep learning environment: A review. Artif. Intell. Rev. 2023, 56, 10345–10425. [Google Scholar] [CrossRef]
  7. Nadeem, M.I.; Ahmed, K.; Li, D.; Zheng, Z.; Naheed, H.; Muaad, A.Y.; Alqarafi, A.; Abdel Hameed, H. SHO-CNN: A metaheuristic optimization of a convolutional neural network for multi-label news classification. Electronics 2022, 12, 113. [Google Scholar] [CrossRef]
  8. Gao, X.; Wang, X.; Chen, Z.; Zhou, W.; Hoi, S.C. Knowledge Enhanced Vision and Language Model for Multi-Modal Fake News Detection. IEEE Trans. Multimed. 2024, 26, 8312–8322. [Google Scholar] [CrossRef]
  9. Islam, S.; Ab Ghani, N.; Ahmed, M. A review on recent advances in Deep learning for Sentiment Analysis: Performances, Challenges and Limitations. Compusoft 2020, 9, 3775–3783. [Google Scholar]
  10. Xie, B.; Li, Q. Detecting fake news by RNN-based gatekeeping behavior model on social networks. Expert Syst. Appl. 2023, 231, 120716. [Google Scholar] [CrossRef]
  11. Truică, C.O.; Apostol, E.S.; Karras, P. DANES: Deep neural network ensemble architecture for social and textual context-aware fake news detection. Knowl.-Based Syst. 2024, 294, 111715. [Google Scholar] [CrossRef]
  12. Liu, Z.; He, X.; Liu, L.; Liu, T.; Zhai, X. Context matters: A strategy to pre-train language model for science education. In Proceedings of the International Conference on Artificial Intelligence in Education, Tokyo, Japan, 3–7 July 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 666–674. [Google Scholar]
  13. Luqman, M.; Faheem, M.; Ramay, W.Y.; Saeed, M.K.; Ahmad, M.B. Utilizing ensemble learning for detecting multi-modal fake news. IEEE Access 2024, 12, 15037–15049. [Google Scholar] [CrossRef]
  14. Nakamura, K.; Levy, S.; Wang, W.Y. r/Fakeddit: A new multimodal benchmark dataset for fine-grained fake news detection. arXiv 2019, arXiv:1911.03854. [Google Scholar]
  15. Sheng, Y.; Zhang, G.; Zhang, Y.; Luo, M.; Pang, Y.; Wang, Q. A multimodal data sensing and feature learning-based self-adaptive hybrid approach for machining quality prediction. Adv. Eng. Inform. 2024, 59, 102324. [Google Scholar] [CrossRef]
  16. Li, F.; Rosli, M.M.; Wang, Y. A Review of Image and Text Feature Extraction Methods in Fake News Detection Tasks. Ing. Syst. Inf. 2024, 29, 409–420. [Google Scholar] [CrossRef]
  17. Jouhar, J.; Pratap, A.; Tijo, N.; Mony, M. Fake News Detection using Python and Machine Learning. Procedia Comput. Sci. 2024, 233, 763–771. [Google Scholar] [CrossRef]
  18. Abdullah, M.; Hongying, Z.; Javed, A.; Mamyrbayev, O.; Caraffini, F.; Eshkiki, H. A joint learning framework for fake news detection. Displays 2025, 90, 103154. [Google Scholar] [CrossRef]
  19. Abraham, T.M.; Wen, T.; Wu, T.; Chen, Y.w. Leveraging data analytics for detection and impact evaluation of fake news and deepfakes in social networks. Humanit. Soc. Sci. Commun. 2025, 12, 1040. [Google Scholar] [CrossRef]
  20. Nadeem, M.I.; Mohsan, S.A.H.; Ahmed, K.; Li, D.; Zheng, Z.; Shafiq, M.; Karim, F.K.; Mostafa, S.M. Hyprobert: A fake news detection model based on deep hypercontext. Symmetry 2023, 15, 296. [Google Scholar] [CrossRef]
  21. Wang, W.; Yang, R.; Guo, C.; Qin, H. CNN-based hybrid optimization for anomaly detection of rudder system. IEEE Access 2021, 9, 121845–121858. [Google Scholar] [CrossRef]
  22. Choudhary, M.; Chouhan, S.S.; Pilli, E.S.; Vipparthi, S.K. BerConvoNet: A deep learning framework for fake news classification. Appl. Soft Comput. 2021, 110, 107614. [Google Scholar] [CrossRef]
  23. Uppada, S.K.; Patel, P. An image and text-based multimodal model for detecting fake news in OSN’s. J. Intell. Inf. Syst. 2023, 61, 367–393. [Google Scholar] [CrossRef]
  24. Fanni, S.C.; Febi, M.; Aghakhanyan, G.; Neri, E. Natural language processing. In Introduction to Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2023; pp. 87–99. [Google Scholar]
  25. Nadeem, M.; Abbas, P.; Zhang, W.; Rafique, S.; Iqbal, S. Enhancing Fake News Detection with a Hybrid NLP-Machine Learning Framework. ICCK Trans. Intell. Syst. 2024, 1, 203–214. [Google Scholar] [CrossRef]
  26. Raj, C.; Meel, P. ConvNet frameworks for multi-modal fake news detection. Appl. Intell. 2021, 51, 8132–8148. [Google Scholar] [CrossRef]
  27. Zhou, Y.; Yang, Y.; Ying, Q.; Qian, Z.; Zhang, X. Multimodal Fake News Detection via CLIP-Guided Learning. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 2825–2830. [Google Scholar] [CrossRef]
  28. Hamed, S.K.; Ab Aziz, M.J.; Yaakub, M.R. A review of fake news detection approaches: A critical analysis of relevant studies and highlighting key challenges associated with the dataset, feature representation, and data fusion. Heliyon 2023, 9, e20382. [Google Scholar] [CrossRef] [PubMed]
  29. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  30. Mahadevan, S., sr.; Ahmad, S. BERT based Blended approach for Fake News Detection. J. Big Data Artif. Intell. 2024, 2, 7–15. [Google Scholar]
  31. Hu, X.; Yu, W.; Wu, Y.; Chen, Y. Multi-modal recommendation algorithm fusing visual and textual features. PLoS ONE 2023, 18, e0287927. [Google Scholar] [CrossRef]
  32. Almarashy, A.H.J.; Feizi-Derakhshi, M.R.; Salehpour, P. Elevating Fake News Detection Through Deep neural networks, Encoding Fused Multi-Modal Features. IEEE Access 2024, 12, 82146–82155. [Google Scholar] [CrossRef]
  33. Azri, A.; Favre, C.; Harbi, N.; Darmont, J.; Noûs, C. Rumor classification through a multimodal fusion framework and ensemble learning. Inf. Syst. Front. 2023, 25, 1795–1810. [Google Scholar] [CrossRef]
  34. Ghorbanpour, F.; Ramezani, M.; Fazli, M.A.; Rabiee, H.R. FNR: A similarity and transformer-based approach to detect multi-modal fake news in social media. Soc. Netw. Anal. Min. 2023, 13, 56. [Google Scholar] [CrossRef]
  35. Singh, P.; Srivastava, R.; Rana, K.; Kumar, V. SEMI-FND: Stacked ensemble based multimodal inferencing framework for faster fake news detection. Expert Syst. Appl. 2023, 215, 119302. [Google Scholar] [CrossRef]
  36. Al-alshaqi, M.; Rawat, D.B.; Liu, C. A BERT-Based Multimodal Framework for Enhanced Fake News Detection Using Text and Image Data Fusion. Computers 2025, 14, 237. [Google Scholar] [CrossRef]
  37. Dong, S.Q.; Sun, Y.M.; Xu, T.; Zeng, L.B.; Du, X.Y.; Yang, X.; Liang, Y. How to improve machine learning models for lithofacies identification by practical and novel ensemble strategy and principles. Pet. Sci. 2023, 20, 733–752. [Google Scholar] [CrossRef]
  38. Mohawesh, R.; Obaidat, I.; AlQarni, A.A.; Aljubailan, A.A.; Al-Shannaq, M.A.; Salameh, H.B.; Al-Yousef, A.; Saifan, A.A.; Alkhushayni, S.M.; Maqsood, S. Truth be told: A multimodal ensemble approach for enhanced fake news detection in textual and visual media. J. Big Data 2025, 12, 197. [Google Scholar] [CrossRef]
  39. Brinda, B.; Rajan, C.; Geetha, K. EnsembleNet: A novel deep ensemble learning model with GANBERT and BiLSTM for automated fake news detection. Knowl. Inf. Syst. 2025, 67, 11015–11040. [Google Scholar] [CrossRef]
  40. Qu, Z.; Meng, Y.; Muhammad, G.; Tiwari, P. QMFND: A quantum multimodal fusion-based fake news detection model for social media. Inf. Fusion 2024, 104, 102172. [Google Scholar] [CrossRef]
  41. Mazari, A.C.; Boudoukhani, N.; Djeffal, A. BERT-based ensemble learning for multi-aspect hate speech detection. Clust. Comput. 2024, 27, 325–339. [Google Scholar] [CrossRef]
  42. Cui, S.; Gong, L.; Li, T. Hmltnet: Multi-modal fake news detection via hierarchical multi-grained features fused with global latent topic. Neural Comput. Appl. 2025, 37, 5559–5575. [Google Scholar] [CrossRef]
  43. Wu, F.; Chen, S.; Gao, G.; Ji, Y.; Jing, X.Y. Balanced multi-modal learning with hierarchical fusion for fake news detection. Pattern Recognit. 2025, 164, 111485. [Google Scholar] [CrossRef]
  44. Wu, Y.; Zhan, P.; Zhang, Y.; Wang, L.; Xu, Z. Multimodal fusion with co-attention networks for fake news detection. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, 1–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 2560–2569. [Google Scholar]
  45. Zhang, W.; Gui, L.; He, Y. Supervised contrastive learning for multimodal unreliable news detection in COVID-19 pandemic. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual Event, 1–5 November 2021; ACM Digital Library: New York, NY, USA, 2021; pp. 3637–3641. [Google Scholar]
  46. Sciucca, L.D.; Mameli, M.; Balloni, E.; Rossi, L.; Frontoni, E.; Zingaretti, P.; Paolanti, M. FakeNED: A deep learning based-system for fake news detection from social media. In Proceedings of the International Conference on Image Analysis and Processing, Lecce, Italy, 23–27 May 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 303–313. [Google Scholar]
  47. Hu, Y.L.; Zhao, Q.S. Bi-GRU model based on pooling and attention for text classification. Int. J. Wirel. Mob. Comput. 2021, 21, 26–31. [Google Scholar] [CrossRef]
  48. Dai, Z. Transformer-xl: Attentive language models beyond a fixed-length context. arXiv 2019, arXiv:1901.02860. [Google Scholar]
  49. Wei, S.; Yu, D.; Lv, C. A Distilled BERT with Hidden State and Soft Label Learning for Sentiment Classification. J. Phys. Conf. Ser. 2020, 1693, 012076. [Google Scholar] [CrossRef]
  50. Wang, C.; Zhang, F. The performance of improved XLNet on text classification. In Proceedings of the Third International Conference on Artificial Intelligence and Electromechanical Automation (AIEA 2022), Changsha, China 8–10 April 2022; SPIE: Bellingham, WA, USA, 2022; Volume 12329, pp. 154–159. [Google Scholar]
  51. Lu, J.; Batra, D.; Parikh, D.; Lee, S. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. In Proceedings of the NIPS’19: 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; ACM Digital Library: New York, NY, USA, 2019; Volume 32. [Google Scholar]
  52. Nan, Q.; Cao, J.; Zhu, Y.; Wang, Y.; Li, J. MDFEND: Multi-domain fake news detection. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Virtual Event, 1–5 November 2021; ACM Digital Library: New York, NY, USA, 2021; pp. 3343–3347. [Google Scholar]
  53. Bing, W. Multimodal Fake News Dataset Weibo23. 2023. Available online: https://ieee-dataport.org/documents/multimodal-fake-news-dataset-weibo23 (accessed on 18 January 2026).
  54. Mujahid, M.; Kına, E.; Rustam, F.; Villar, M.G.; Alvarado, E.S.; De La Torre Diez, I.; Ashraf, I. Data oversampling and imbalanced datasets: An investigation of performance for machine learning and feature engineering. J. Big Data 2024, 11, 87. [Google Scholar] [CrossRef]
  55. Gupta, A.; Tatbul, N.; Marcus, R.; Zhou, S.; Lee, I.; Gottschlich, J. Class-weighted evaluation metrics for imbalanced data classification. In Proceedings of the International Conference on Learning Representations ICLR 2021, Vienna, Austria, 4 May 2021; OpenReview: Amherst, MA, USA, 2021. [Google Scholar]
  56. Liu, C.; Wu, X.; Yu, M.; Li, G.; Jiang, J.; Huang, W.; Lu, X. A two-stage model based on BERT for short fake news detection. In Proceedings of the Knowledge Science, Engineering and Management: 12th International Conference, KSEM 2019, Athens, Greece, 28–30 August 2019; Proceedings, Part II 12; Springer: Berlin/Heidelberg, Germany, 2019; pp. 172–183. [Google Scholar]
  57. Elbagir, S.; Yang, J. Twitter sentiment analysis using natural language toolkit and VADER sentiment. In Proceedings of the International Multiconference of Engineers and Computer Scientists, Hong Kong, 13–15 March 2019; International Association of Engineers: Hong Kong, China, 2019; Volume 122. [Google Scholar]
  58. Samaras, L.; García-Barriocanal, E.; Sicilia, M.A. Sentiment analysis of COVID-19 cases in Greece using Twitter data. Expert Syst. Appl. 2023, 230, 120577. [Google Scholar] [CrossRef]
  59. Kimiyoung. Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context. 2024. Available online: https://github.com/kimiyoung/transformer-xl (accessed on 23 November 2024).
  60. Face, H. DistilBERT: A Smaller Version of BERT. 2024. Available online: https://huggingface.co/distilbert/distilbert-base-uncased (accessed on 23 November 2024).
  61. Face, H. XLNet: Generalized Autoregressive Pretraining for Language Understanding. 2024. Available online: https://huggingface.co/xlnet/xlnet-base-cased (accessed on 23 November 2024).
  62. Adewumi, T.; Sabry, S.S.; Abid, N.; Liwicki, F.; Liwicki, M. T5 for Hate Speech, Augmented Data, and Ensemble. Sci 2023, 5, 37. [Google Scholar] [CrossRef]
  63. Rainio, O.; Teuho, J.; Klén, R. Evaluation metrics and statistical tests for machine learning. Sci. Rep. 2024, 14, 6086. [Google Scholar] [CrossRef]
  64. Ding, N.; Tian, S.W.; Yu, L. A multimodal fusion method for sarcasm detection based on late fusion. Multimed. Tools Appl. 2022, 81, 8597–8616. [Google Scholar] [CrossRef]
  65. Koonce, B.; Koonce, B. ResNet 50. In Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization; Apress: Berkeley, CA, USA, 2021; pp. 63–72. [Google Scholar]
  66. Arjun, K.; Kumar, K.S. A combined approach of VGG 16 and LSTM transfer learning technique for skin melanoma classification. Int. J. Health Sci. 2022, 6, 13504–13516. [Google Scholar] [CrossRef]
  67. Hafner, M.; Katsantoni, M.; Köster, T.; Marks, J.; Mukherjee, J.; Staiger, D.; Ule, J.; Zavolan, M. CLIP and complementary methods. Nat. Rev. Methods Primers 2021, 1, 20. [Google Scholar] [CrossRef]
  68. Geng, Y.; Li, Q.; Yang, G.; Qiu, W. Logistic regression. In Practical Machine Learning Illustrated with KNIME; Springer: Berlin/Heidelberg, Germany, 2024; pp. 99–132. [Google Scholar]
  69. Thair Ali, N.; Falih Hassan, K.; Najim Abdullah, M.; Salam Al-Hchimy, Z. The Application of Random Forest to the Classification of Fake News. BIO Web Conf. 2024, 97, 00049. [Google Scholar] [CrossRef]
  70. Kumari, S.; Singh, M.P. A Deep Learning Multimodal Framework for Fake News Detection. Eng. Technol. Appl. Sci. Res. 2024, 14, 16527–16533. [Google Scholar] [CrossRef]
  71. Zhou, Y.; Pang, A.; Yu, G. Clip-GCN: An adaptive detection model for multimodal emergent fake news domains. Complex Intell. Syst. 2024, 10, 5153–5170. [Google Scholar] [CrossRef]
  72. Ma, Z.; Luo, M.; Guo, H.; Zeng, Z.; Hao, Y.; Zhao, X. Event-Radar: Event-driven Multi-View Learning for Multimodal Fake News Detection. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Stroudsburg, PA, USA, 2024; pp. 5809–5821. [Google Scholar]
  73. Yi, Z.; Lu, S.; Tang, X.; Wu, J.; Zhu, J. MACCN: Multi-Modal Adaptive Co-Attention Fusion Contrastive Learning Networks for Fake News Detection. In Proceedings of the ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 6045–6049. [Google Scholar]
  74. Dong, Y.; He, D.; Wang, X.; Jin, Y.; Ge, M.; Yang, C.; Jin, D. Unveiling Implicit Deceptive Patterns in Multi-Modal Fake News via Neuro-Symbolic Reasoning. Proc. AAAI Conf. Artif. Intell. 2024, 38, 8354–8362. [Google Scholar] [CrossRef]
  75. Uppada, S.K.; Ashwin, B.; Sivaselvan, B. A novel evolutionary approach-based multimodal model to detect fake news in OSNs using text and metadata. J. Supercomput. 2024, 80, 1522–1553. [Google Scholar] [CrossRef]
  76. Segura-Bedmar, I.; Alonso-Bartolome, S. Multimodal Fake News Detection. Information 2022, 13, 284. [Google Scholar] [CrossRef]
  77. Xiao, T.; Guo, S.; Huang, J.; Spolaor, R.; Cheng, X. HiPo: Detecting Fake News via Historical and Multi-Modal Analyses of Social Media Posts. In Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, Birmingham, UK, 21–25 October 2023; ACM Digital Library: New York, NY, USA, 2023; pp. 2805–2815. [Google Scholar]
  78. Liu, S.; Yue, X.; Wu, F.; Sun, J.; Feng, Y.; Ji, Y. Semantic Distillation and Structural Alignment Network for Fake News Detection. In Proceedings of the ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 6620–6624. [Google Scholar]
  79. Sengan, S.; Vairavasundaram, S.; Ravi, L.; AlHamad, A.Q.M.; Alkhazaleh, H.A.; Alharbi, M. Fake news detection using stance extracted multimodal fusion-based hybrid neural network. IEEE Trans. Comput. Soc. Syst. 2023, 11, 5146–5157. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.