Next Article in Journal
Multi-Objective Parametric Optimization Design for Mirrors Combined with Non-Dominated Sorting Genetic Algorithm
Next Article in Special Issue
Readability Metrics for Machine Translation in Dutch: Google vs. Azure & IBM
Previous Article in Journal
An Effective Motion-Tracking Scheme for Machine-Learning Applications in Noisy Videos
Previous Article in Special Issue
Predicting Location of Tweets Using Machine Learning Approaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Does Context Matter? Effective Deep Learning Approaches to Curb Fake News Dissemination on Social Media

1
School of Information and Physical Sciences, College of Engineering Science and Environment, The University of Newcastle, Callaghan, Newcastle, NSW 2308, Australia
2
Department of Computer Science, King Khalid University, Abha 62521, Saudi Arabia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(5), 3345; https://doi.org/10.3390/app13053345
Submission received: 4 February 2023 / Revised: 26 February 2023 / Accepted: 4 March 2023 / Published: 6 March 2023
(This article belongs to the Special Issue Text Mining, Machine Learning, and Natural Language Processing)

Abstract

:
The prevalence of fake news on social media has led to major sociopolitical issues. Thus, the need for automated fake news detection is more important than ever. In this work, we investigated the interplay between news content and users’ posting behavior clues in detecting fake news by using state-of-the-art deep learning approaches, such as the convolutional neural network (CNN), which involves a series of filters of different sizes and shapes (combining the original sentence matrix to create further low-dimensional matrices), and the bidirectional gated recurrent unit (BiGRU), which is a type of bidirectional recurrent neural network with only the input and forget gates, coupled with a self-attention mechanism. The proposed architectures introduced a novel approach to learning rich, semantical, and contextual representations of a given news text using natural language understanding of transfer learning coupled with context-based features. Experiments were conducted on the FakeNewsNet dataset. The experimental results show that incorporating information about users’ posting behaviors (when available) improves the performance compared to models that rely solely on textual news data.

1. Introduction

Online social networks (OSNs) have gained importance due to their easy accessibility. They are tools used for exchanging information and influencing public opinion, rather than just a means of communicating between individuals [1]. In recent years, fake news has spread more widely due to the ease with which it can be created and distributed online. This type of news is not actual news, it is fake news that is made real for a specific purpose [2].
Twitter is a common social media platform that people use to express their opinions and share their ideas with others. User-generated content (UGC) on OSNs, especially Twitter, is gaining more importance in the research community owing to its value in discovering patterns that significantly benefit various applications. According to the fake news velocity study, tweets that include falsified information reach people six times faster than tweets that include trustworthy information [3]. This indicates how terribly fake news disseminates and how it can have adverse social effects. The concern lies in the quick reactions, such as retweets, likes, and shares of a tweet (fake news story) received on Twitter without pre-thinking, aggravating the problem even more. According to [4], false news (particularly political news) on Twitter is usually retweeted by more users and spreads extremely rapidly. As Wang et al. [5] pointed out, the proliferation of fake news on social media begins with user-posting behaviors.
Much of the current work on fake news detection focuses on textual news content (see Section 2 for more details), ignoring user behavior clues that could have the potential to advance fake news detection performance. Moreover, most of these studies are based on extensions of the bag-of-words representation or other context-independent embeddings, which cannot capture deep semantically contextualized information. The ability to capture and understand semantics and contextualized information about the input text is crucial to identifying fake news.
Given the benefits of user behavior clues in identifying fake news, we argue that the interplay between social user engagements (via feature engineering) and feature learning derived from state-of-the-art (SOTA) deep context-aware representation methods (such as BERT) has the potential to boost fake news classification performance. Given the power of user-behavior cues and the characteristics of SOTA pre-trained language models (PLMs) in generating deep-semantical contextual patterns of a given input text (i.e., news article), their complementary effects need to be investigated.
As such, in this study, we investigated the impacts of user posting behaviors on fake news detection using the FakeNewsNet dataset and explored two research questions: (1) Does modeling both news content and user-posted behavioral cues improve fake news detection performance? and (2) How effective are user-posted behavioral cues in improving the detection performance of the proposed deep learning frameworks?
The key contributions of this paper are as follows:
  • An examination of the performance of several deep learning algorithms and the current state-of-the-art word embeddings, such as BERT, on a benchmark dataset of fake news.
  • For detecting fake news, new hybrid CNN-RNN architectures using attention modules were developed.
  • Extensive experiments on two real-world fake news datasets demonstrate the effectiveness of the proposed frameworks for detecting fake news.
The rest of this paper is organized as follows. Section 2 summarises the literature related to the detection of fake news. Section 3 provides an overall background of the models we used. We describe the problem statement in Section 4. The proposed deep-learning approaches are described in Section 5. Experiments on the performance of the predictive models are presented in Section 6. The experimental results are discussed in Section 7. Finally, Section 8 concludes the paper.

2. Related Works

Current research in fake news detection can be generally classified into content- and context-based fake news detection. According to [6], the former takes into account text-based features, such as general features, which are catalysts for describing content style from four linguistic levels: lexicon, syntax, discourse, and semantics. Some statistical techniques are applied to the lexicon level, such as the bag-of-words (BoW) model [7]. At the syntax level, part of speech (POS) (e.g., nouns and verbs) frequency is assessed using POS taggers [7]. A rhetorical structure theory and rhetorical parsing tool can be used at the discourse level to capture the resulting frequency of rhetorical relations among sentences as features [7]. On the semantic level, these frequencies correspond to lexicons or expressions assigned to each psycholinguistic category [8] (e.g., LIWC [9]). A second type of text-based feature is called a latent textual feature, which is used for news text embedding. These features can be conducted either at the word level [10], sentence-level [11], or document-level [11], which result in dense vector representations that can then be processed further. Some of the studies on detecting fake news are mainly based on the content; in [12], to detect fake news, the authors applied CNN and BiLSTM to embed textual and speaker metadata. In [13], the authors presented twenty-six linguistic-based textual features to detect fake news. In [9], the authors employed linguistic features (e.g., punctuation, readability, syntax, and psycholinguistic features) to distinguish between true and fraudulent news items, while the authors of [14] developed an enhanced set of linguistic features for distinguishing fake from true news.
In [15], a Chinese WeChat clickbait dataset was created, and a multi-featured method of detecting WeChat clickbait was proposed, using semantic, syntactic, and auxiliary information. MFWCD-BERT and MFWCD-BiLSTM are models based on the MFWCD framework with varying parameter scales, which use Bidirectional Encoder Representation from Transformers (BERT) and lightweight Bidirectional Long Short-Term Memory (Bi-LSTM) networks with attention mechanisms to encode headline semantics, respectively. The researchers also presented an improved Graph Attention Network (GAT) with attention mechanisms to capture valuable structures in titles based on local syntactic structures. MFWCD performs better than compared baseline methods in clickbait detection, proving its effectiveness and interpretability. For other related research conducted for clickbait detection, see [16,17].
It is inherently difficult to detect fake news because fake news is usually written intentionally to mislead the reader. Among the potential clues to the detection is the context-based features (aka auxiliary contextual information), such as user interactions on online social networks. Researchers combined content- and context-based features for characterizing fake news. For example, Mouratidis et al. [18] combined network account features with linguistic features. Shu et al. [19] developed the social article fusion (SAF) model, which combines social engagement features with linguistic aspects using the FakeNewsNet dataset. Authors in [20] presented a deep learning model based on the hierarchical attention network using news content and user comments on Twitter for detecting fake news, and their framework achieved state-of-the-art results in the FakeNewsNet dataset.
In the context of Twitter, post-based features refer to those retrieved from source tweets. Several studies used contextual information from tweets, such as temporal patterns embedded in a series of replies (i.e., comments) posted by users on social media and other elements representing their interactions and engagements. Several studies have used similar temporal patterns to identify useful patterns for fake news detection. For example, Ma et al. [21] proposed a technique called SVM-TS that employs time series of aggregated news attributes to detect fake news. Linguistic-based features can also be extracted from each tweet. Furthermore, features extracted from the topic using topic modeling approaches, such as Latent Dirichlet Allocation (LDA) [21]. Credibility features can also be extracted for posts in order to assess the reliability degree [22]. Using user-based clues could help differentiate between fake and real news. This is due to the fact that individuals who are more likely to spread incorrect information have distinct characteristics and attributes than those who are not [23]. These findings prompted researchers to look into user-based features for detecting fake news. For example, in [23], the authors examined user profiles and extracted features to distinguish between fake and real news. Wang [12] uses user profile features as input to a hybrid CNN model for fake news detection. In [24], the authors concluded that the idea of incorporating a speaker profile could noticeably enhance performance. Most importantly, social propagation features, such as followers and retweet counts, can be used as useful insights for determining how fake news can spread in social media [19].
In recent research, many useful methods for detecting fake news have relied on sequential neural networks for encoding news content and social context information, in which text sequences were analyzed unidirectionally [25]. With this bidirectional training approach, long-distance and semantic dependencies can be captured in the given sentences.
A deep learning approach (FakeBERT) is proposed in [25] that uses BERT (Bidirectional Encoder Representations from Transformers) and the parallel blocks of a deep Convolutional Neural Network (CNN) with varying kernel sizes and filters combined with BERT. They found that this combination is useful in dealing with ambiguity, which is the most difficult challenge in natural language understanding. However, the authors overlooked the benefit of modeling user behavior clues, which could improve classification performance. Using the LIAR dataset, a machine learning framework for automatic fake news detection using BERT b a s e was presented by Alghamdi et al. [26]. In order to extract local features, the input text was encoded using BERT b a s e , and then the output sequence was fed into a CNN network. Then, a CNN followed by a BiLSTM was used to encode the user profile features. The final output of the two components is fed into a classification layer. The authors concluded that user behavior features are important for fake news detection. A comparative study using different machine learning and deep learning models was conducted [27]; the authors showed the power of BERT and its variations in improving detection performance across different datasets.
In a study conducted by Natali Ruchansky et al. [28], the researchers analyzed social media data and developed a hybrid deep learning model that showed an accuracy of 0.892 on Twitter data and 0.953 on Weibo data. The researchers concluded that capturing the temporal behavior of articles and learning source characteristics about users’ behavior was important for detecting fake news. Exploiting useful clues from user behavior on social media is of great importance to detect fake news, and as such, in this work, we investigate the impact of user posting behavior for fake news detection using the FakeNewsNet dataset. In addition, we assess the effectiveness of a different set of user posting behavior features. A majority of existing studies used features extracted from a given news article to model fake news detection. This approach is reasonable since news text is most commonly provided in isolation in public datasets; however, content-based features are often insufficient since malicious entities are frequently manipulating content, following success at detection, to mimic trustworthy content. A natural question arises: Can associated user posting behavior information on social media platforms (e.g., Twitter) improve the performance of fake news detection?
To this end, noting the power of user behavioral features, this study focused on leveraging joint learning of news content and the user posting behavioral clues on Twitter for fake news detection. We hypothesize that news content and the user posting behavior attributes contain complementary information that needs to be encoded and captured simultaneously for fake news detection. Furthermore, the majority of existing studies modeled text input using representations extracted from pre-trained context-free embedding models, such as GloVe, to train fake news detection models. A key limitation of this process is that it assigns fixed vector values to each word in a given sequence regardless of its context. Progress in the realm of natural language processing has centered on deep contextualized models that leverage the deeper long-term semantical contextual relationships between words and sentences in a given corpus. In this work, we leveraged the power of context-aware embedding models, such as BERT, to encode the news text. Generally speaking, the textual news content is initially encoded using BERT to extract deep contextual representations, followed by a deep learning architecture. Next, user features are processed by stacking BiGRU layers. These two components are concatenated to combine the features extracted from news articles and their associated user behavior clues into a single feature vector. This combined feature vector is then fed into a classification layer with a single neuron, activated using a sigmoid function.

3. Background: Models

3.1. Word Embedding

An embedding is typically learned from a large corpus of text. In NLP, there are two main alternatives for obtaining a distributed representational vector that captures the semantics of the input text.
  • A word-based representation.
  • A context-based representation.

3.1.1. Word-Based Representations (Non-Contextualized Embeddings)

Context-independent methods produce the same embedding vector for each word in the vocabulary, irrespective of its context. Examples of the most common context-independent pre-trained embeddings include:
  • A 300-dimensional vector was generated by Word2Vec. using a large corpus of news articles with 300 million tokens.
  • Pre-trained GloVe used 27 billion tokens in a huge corpus of tweets, resulting in a 200-dimensional vector.

3.1.2. Context-Based Representations (Contextualized Embeddings)

By capturing only short-range co-occurrence context, non-contextual embedding models are unable to capture deeper contextual relationships. In contrast, context-aware embedding models can compute the embedding for a word based on its context. An example of the most common deep contextualized pre-trained embedding models is BERT. In Devlin et al. [29], the bidirectional encoder representation from transformer (BERT) is introduced as the first deeply bidirectional, unsupervised language representation, which functions as a transformer encoder with multiple layers (performs self-attention in both directions) that conditions both left and right contexts simultaneously. Therefore, BERT generates embeddings that are context-aware.
Furthermore, to remove the unidirectionality constraint, BERT performs pre-training using an unsupervised prediction task, which includes a masked language model (MLM) in charge of understanding context and then making predictions (of words). Thus, the model can produce a vector representation that can capture the general information of the input text. These semantic representations of each word in the input text can be improved using an attention mechanism when different words in a context show different effects in boosting semantical representation. As a core component of transformer architecture, the attention mechanism’s underlying role is to assign less or more weights to different parts of text towards the output (i.e., differentiate the contribution of different parts of the input on the output). Attention can be considered as a function that maps queries and follows key-value and output vector pairs; the scaled dot-product attention formula can be seen in Equation (1).
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d k ) V
Query, key, and value are marked by the letters Q, K, and V, respectively. d k represents the dimension of the key vector k and query vector q. A Softmax activation function is used by attention to normalize the inputs to a value between 0 and 1. Although BERT uses a transformer’s encoder, it employs a multi-head attention mechanism, as shown in Equation (2), where each distinct head and its corresponding weight matrices are identified by the subscript i.
M u l t i H e a d ( Q , K , V ) = C o n c a t ( h e a d 1 , , h e a d h ) W O
where each h e a d i is calculated using the formula below, and W O represents a weight matrix that was trained along with the model:
h e a d i = A t t e n t i o n ( Q W i Q , K W i K , V W i V )
Even though BERT contains millions of parameters (i.e., BERT b a s e contains 110 million parameters and BERT l a r g e has 340 million parameters) [29], it is relatively inexpensive to apply BERT to downstream tasks if the parameters are fine-tuned using a pre-trained model. In this work, we use BERT b a s e .

3.2. CNN

Convolutional neural networks (CNNs) have demonstrated effectiveness in several NLP tasks in which one-dimensional convolutional neural networks (Conv1D) are used. As a result, Conv1D iterates over training data using a filter with a fixed window size (each filter cell is initially set to a weight) and uses the sliding window to iterate over the training data. At each step, the inputs (word vectors) are multiplied by these filter weights to generate a feature map (filter output array) that encodes informative features derived from input training data. A CNN is well known for its ability to automatically extract local features and capture them more accurately. As such, CNNs are generally better at extracting features from input data, thereby reducing dimensionality and increasing robustness [30].

3.3. BiGRU

A variant of the gated recurrent unit (GRU) consists of only two gates: an update gate that acts as a catalyst for combining the forget and input gates and controls the information flowing to the current state, and a reset gate that determines when to ignore the previous hidden state [31]. The update and reset gates are computed similarly to LSTM as follows [31]:
r t = δ ( W r h t 1 + U r x t + b r ) ,
z t = δ ( W z h t 1 + U z x t + b z ) ,
h t = ( 1 z t ) h t 1 + z t h t ˜ ,
h t ˜ = t a n h ( W h t ˜ ( h t 1 r t ) + U h t ˜ x t ) .
In the formulas above, δ ( . ) signifies the logistic sigmoid function, W and U are gate weight matrices, and h t and b are the hidden state and bias vectors, respectively. Basic RNNs simply consider the previous context and lack the ability to capture future context. As a result of the breakthrough design, bi-directional long short term memory (BiLSTM) and bi-directional gated recurrent unit (BiGRU) are suitable options for accounting for the future and preceding contexts. To do this, the forward and backward hidden layers are integrated, controlling the temporal information flow in both directions and leading to better learning.

3.4. Attention Mechanism

Recently, the attention paradigm, inspired by human biological systems, has gained traction in natural language processing (NLP) fields. The NLP community is preparing for the paradigm shift by designing models that can assign weights to various parts of a given input text, capturing more relevant information for further processing. The attention model aims to mimic humans’ biological systems, where, given a piece of text, humans can selectively identify what is most vital and relevant in a given context while ignoring irrelevant information. As a consequence, certain parts of the input can be considered adaptively by the model [32]. Furthermore, the attention mechanism enables the model to learn to attend to the relevant parts of the given input (e.g., hidden states of a BiGRU network) to generate a single salient vector representation. In light of this, the model may identify the words that are being concentrated on for each input text by “attending to” particular areas of the input when processing the data. Typically, weighting different words in a sentence are done by combining all hidden states, generating a single vector representation, as described below.
α t = exp ( v T · h t ˜ ) t exp ( v · h t ˜ )
S A w = t α t h t
where v is a trainable parameter [33] and the hidden states are computed in Equations (6) and (7).

4. Problem Definition

We approach the challenge of fake news detection as a binary classification, in which a model is built to predict whether a piece of news is fake or real based on the provided attributes. Formally, assume A = { a 1 , a 2 , , a l } is a set of news articles (where l is the length of the input) and U = { u 1 , u 2 , , u n } represents user posting behavior information towards such news (e.g., the number of retweets such a news article received on Twitter). Given news articles and the associated user-posted behavioral cues, the purpose of fake news identification is to determine whether the input text represents fake or legitimate content.
That is, f : f ( A , U ) →{fake, real} ∋,
F ( A ) = f a k e , if A is fake content , r e a l , otherwise .

5. The Proposed Approaches

5.1. BERT b a s e -CNN-BiGRU-ATT

In this subsection, we will introduce the first deep learning architecture proposed in this paper for fake news detection. Specifically, as illustrated in Figure 1, we first perform BERT b a s e on input text—suppose x i j R , represents the d-dimensional word embedding of the j-th word in the i-th sentence, to obtain deep contextualized representations since BERT b a s e plays a bidirectional attention mechanism on input text, ensuring better coverage of global semantic understanding. Then, a CNN layer is applied to capture local features as follows. Given a sentence, assume X 1 : i : m = x 11 x 12 x 1 m , where m denotes the length of a given sentence i and ⊕ refers to a concatenation operation. A feature map F is generated by performing a convolutional operation with a filter T R h × d on X i : i + h 1 using the equation f i = σ ( T · X i : i + h 1 + b ) where σ represents a non-linear activation function and b denotes a bias term. The final feature map F is obtained by applying the previous step to all sliding windows of h words in a sentence. This F is fed into a max-pooling layer and to a global average pooling layer for capturing the most relevant and salient information, generating F m = [ F m 1 , F m 2 , , F m k ] and F g = [ F g 1 , F g 2 , , F g k ] , respectively, with k represents the number of filters used. The resulting feature map is max-pooled and global averaged-pooled. In addition, BiGRU is applied to the output of the CNN layer. The BiGRU layer consists of a forward G R U and a backward G R U to process the information from both left-to-right and right-to-left directions.
h i j = G R U ( F , h i ( j 1 ) ) , j { 1 , , m }
h i j = G R U ( F , h i ( j 1 ) ) , j { m , , 1 }
We denote the combination of the hidden state obtained from both the forward and backward GRUs as h i j = [ h i j h i j ] , which, in turn, captures full salient information about a given input text. In order to better capture salient and relevant features from the final output of BiGRU networks, the attention mechanism (see Equations (13)–(15)) is applied to the resultant output of BiGRU and the models, to assign different weights to different parts of the given input.
u i j = t a n h ( W w h i j + b w )
α i j = exp ( u i j u w ) j exp ( u i j u w )
where u i j is a hidden representation of h i j , u w is a trainable parameter that is randomly initialized during the course of training, and the result of the product ( u i j u w ) is used to calculate the importance score of word j, which is then normalized using the softmax function, generating the importance weight α i j for each word.
Finally, the weighted sum of the concatenated word representations with the hidden states h i j form the news vector n i .
n i = t α i j h i j
Ultimately, we concatenate the final output as follows: C = F m F g n i , where ⊕ denotes the concatenation operation. We encode user posting behavior features using BiGRU. Assume U is the final representation of users encoded features, a concatenation layer is used to extract the final informative features of the two components (news text and user posting behavior features) as follows: O = U C , where O is the final output representation of the given news article with its associated user-posted behavioral cues. Finally, the output is passed to a classification layer with one unit, activated using the sigmoid function.

5.2. BERT b a s e -BiGRU-CNN-ATT

Here, we describe the second proposed model, which we call BERT b a s e -BiGRU-CNN-ATT. The architecture can be seen in Figure 2. Given a sentence S = { w 1 , w 2 , , w N } , similar to the CNN-BiGRU-ATT model, BERT b a s e is used to assign each word w i deep contextualized representation X = { x 1 , x 2 , , x N } , which is then processed by a BiGRU, which consists of a forward G R U and a backward G R U , where combining the resultant annotations forms the final representation h t . The CNN architecture was used to encode the final representation of BiGRU. The CNN architecture we used here was adapted from the research of [34].
Each convolutional layer is connected to max-pooling and global average pooling layers for capturing rich–local semantical information. Following the multi-layer CNN is an attention layer that helps reflect the correlation between features by assigning different importance scores to different parts of a given input. After obtaining the output of the attention layer, a concatenation layer is used to concatenate the outputs of the max-pooling and global average pooling layers with the attention layer’s output. Next, the user posting behavior features are encoded using BiGRU layers stacked on top of each other and the resulting output is concatenated with the output of the above-mentioned concatenation layer; the final output is passed to a classification layer with a single unit activated with a sigmoid function.

5.3. BERT b a s e -CNN-BiGRU

Recently, the authors in [35] showed the effectiveness of joint learning of different deep learning models, including BERT, CNN, and BiGRU networks. Inspired by this work, we investigate the effectiveness of exploiting the corresponding representation generated by the (CLS) special token in the BERT b a s e model, where this token returns a vector that carries the meaning of a full sentence. The resulting representations of BERT b a s e are used as input to a CNN layer, followed by max-pooling and global average pooling layers in the first module. A BiGRU layer is also performed on the output of BERT b a s e , and a concatenation layer is used to concatenate the outputs of both the max- and global average-pooling layers with the pooled output of BERT b a s e and the resulting representation of BiGRU. In the second module, user posting behavioral features are encoded in the same way as described with the previous model; the final output is concatenated with the output of the first module and the output is passed into a classification layer. The architecture can be seen in Figure 3.

6. Methodology

6.1. Experimental Setup

This subsection presents the experiments to assess the effectiveness of the proposed architectures. Python was used to implement the experiments, and the TensorFlow library was used to build the models. We used the pre-trained BERT b a s e available on TensorFlowHub (https://www.tensorflow.org/hub, accessed on 2 January 2023). We used a macOS computer with an Intel Core i5 processor running at 2.3 GHz and 8 GB of memory. The source code is available (https://github.com/Jal-ghamdi/Context_Aware_Fake_News_Detection, accessed on 2 January 2023.)

6.2. Evaluation Metrics

In order to evaluate the performance of the models, we used the following four evaluation criteria: accuracy, precision, recall, and F1 score (calculated as in the equations below):
  • Accuracy (A): is a measure of a classifier’s ability to correctly identify information as false or true. Equation (16) can be used to calculate the accuracy:
    A c c u r a c y = T P + T N T P + T N + F P + F N
  • Precision (P): is a metric that measures the exactness of the classifier, with a low value indicating a high amount of false positives. The precision is determined using Equation (17) and indicates the number of positive predictions divided by the total number of positive class values predicted.
    P r e c i s i o n = T P T P + F P
  • Recall (R): is calculated by dividing the total number of true positives by the total number of true positives and false negatives, as shown in Equation (18); it is thought to be a measure of a classifier’s completeness (for example, a low recall value suggests a high number of false negatives).
    R e c a l l = T P T P + F N
  • F 1 s c o r e (F1): Equation (19) is used to calculate the F 1 s c o r e as the weighted harmonic mean of the classifier’s precision and recall measures.
    F 1 = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l = 2 T P 2 T P + F P + F N
    where T P , T N , F P , and F N represent true positive, true negative, false positive, and false negative, respectively.

6.3. Datasets

We empirically investigated our research questions using the FakeNewsNet dataset. FakeNewsNet (https://github.com/KaiDMML/FakeNewsNet, accessed on 15 January 2023) is a comprehensive dataset collected from two fact-checking platforms: PolitiFact and GossipCop. The news content is labeled (e.g., articles) and the context information is provided (e.g., user responses to the news item on social networks, such as Twitter). Table 1 shows the statistics of FakeNewsNet dataset.

6.4. Compared Fake News Detection Methods

In this study, we compared our proposed methods with the state-of-the-art algorithms to assess the predictive power and viability of the proposed features for fake news detection.
  • SAF [19]: A model that combines social engagement features with linguistic aspects using the FakeNewsNet dataset.
  • BiLSTM-BERT [36]: The natural language inference approach (i.e., inferring the veracity of the news item) uses BiLSTM and BERT embeddings in the PolitiFact dataset.
  • LNN-KG [37]: A model trained in the PolitiFact dataset using both textual patterns and embeddings of concepts in the input text.
  • Logistic regression (N-Gram) [38]: N-gram-based logistic regression model for fake news detection.
  • dEFEND [20]: A framework that comprises an encoder for news content, an encoder for user comments, and a model of co-attention for sentence–comment interaction to detect fake news.

7. Results

Fake News Detection Performance

User behavior clues are important for detecting fake news [19]. Fake news detection has previously been achieved using word embeddings and bag-of-words (see Section 2 for more details). However, neither of these representations can capture contextual information. As opposed to that, our system uses highly sophisticated representations based on bidirectional encoder representations from transformers (BERT) [29]. By training a transformer bidirectionally, BERT can learn contextual relationships among words. In addition, BERT reads the entire word sequence at once, in contrast to earlier systems that either looked at a text sequence from left to right or combined left-to-right and right-to-left training. This trait enables the model to understand a word’s context based on all of its surroundings. The tables below present the results of the proposed models and the official baseline models in the two datasets. The best performance scores are highlighted in bold.
Modeling user posting behavior information in addition to news text was effective for detecting fake news. The test results in the PolitiFact dataset shown in Table 2 demonstrate that the highest-performing model is BERT-CNN-BiGRU-ATT, which outperformed the other official baselines. The BERT b a s e -BiGRU-CNN-ATT model obtained 87.05% accuracy, 86.90% for precision, 91.25% for recall, and 89.02% for the F1 score. The BERT b a s e -CNN-BiGRU model obtained a performance of 93.59% for precision, 91.37% for accuracy, 91.25% for recall, and 92.41% for the F1 score. Summaries of the prediction results of each model in the PolitiFact dataset are shown in the forms of confusion matrices in Figure 4, Figure 5 and Figure 6. Regarding PolitiFact, all models showed similar confusion matrix results, with BERT b a s e -CNN-BiGRU-ATT yielding the best results. Figure 7 shows the overall scores across all of the models. The comparison of results of the GossipCop dataset (using the combination of user-behavior cues with news text) is described in Table 3 and shows that the BERT b a s e -BiGRU-CNN-ATT model outperformed the existing baselines and other counterpart models. The test results show that the BERT b a s e -BiGRU-CNN-ATT model provides the best performance with accuracy and an F1 score of 90.34% with 93.79% for precision. The BERT b a s e -CNN-BiGRU-ATT model obtains a performance of 88.49% for accuracy, 91.07% for precision, 94.14% for recall, and 92.58% for the F1 score. The last proposed BERT b a s e -CNN-BiGRU model yields an accuracy of 90.10%, precision of 90.73%, recall of 96.91%, and an F1 score of 93.72%. Summaries of the prediction results of the models in the GossipCop dataset are shown in the forms of confusion matrices in Figure 8, Figure 9, Figure 10 and Figure 11, showing the overall scores across all of the models.
We were also eager to further assess the effectiveness of the proposed architectures. Therefore, we conducted a case study where we tested the model’s performance by ignoring user posting behavior features and comparing the model’s performance. We observed that when neglecting user posting behavior information, the performance deteriorated in the two datasets. The experimental results are tabulated in Table 4 and Table 5 for the PolitiFact and GossipCop datasets, respectively. The best-performing model is shown to be BERT b a s e -CNN-BiGRU in the two datasets. The BERT b a s e -CNN-BiGRU-ATT model obtained a performance of 87.77% for accuracy, 86.21% for precision, 93.75% for recall, and 89.82% for the F1 score in the PolitiFact dataset, lower than the results obtained by existing studies [36,37], while the model achieved an accuracy of 85.84%, precision of 88.49%, recall of 93.16%, and 90.98% for the F1 score in the GossipCop dataset. The BERT b a s e -BiGRU-CNN-ATT model yielded an accuracy of 88.49%, and precision, recall, and F1 score of 86.44%, which is on par with [36,37] in the Politifact dataset and higher than the official baselines in the GossipCop dataset. Summaries of the prediction results of the models in the PolitiFact and GossipCop datasets are shown in the forms of confusion matrices in Figure 12, Figure 13 and Figure 14 and Figure 15, Figure 16 and Figure 17, respectively. Figure 18 and Figure 19 show the overall scores across all of the models in both datasets, respectively.
Table 2. Comparison (%) of the results obtained in the PolitiFact dataset using news content and user posting behavior features.
Table 2. Comparison (%) of the results obtained in the PolitiFact dataset using news content and user posting behavior features.
ModelAccuracy (%)Precision (%)Recall (%)F1 (%)
SAF [19]0.6910.6380.7890.706
BiLSTM-BERT [36]0.885NANANA
LNN-KG [37]0.8800.90110.8800.8892
Logistic Regression (N-Gram) [38]0.800.790.780.78
dEFEND [20]0.9040.9020.9560.928
BERT b a s e -CNN-BiGRU-ATT0.92810.93750.93750.9375
BERT b a s e -BiGRU-CNN-ATT0.87050.86900.91250.8902
BERT b a s e -CNN-BiGRU0.91370.93590.91250.9241
Table 3. Comparison (%) of the results obtained in the GossipCop dataset using news content and user posting behavior features.
Table 3. Comparison (%) of the results obtained in the GossipCop dataset using news content and user posting behavior features.
ModelAccuracy (%)Precision (%)Recall (%)F1 (%)
SAF [19]0.7960.8200.7530.785
Logistic Regression (N-Gram) [38]0.820.750.790.77
dEFEND [20]0.8080.7290.7820.755
BERT b a s e -CNN-BiGRU-ATT0.88490.91070.94140.9258
BERT b a s e -BiGRU-CNN-ATT0.90340.91970.95680.9379
BERT b a s e -CNN-BiGRU0.90100.90730.96910.9372
Table 4. Comparison (%) of the results obtained in the PolitiFact dataset w/o user posting behavior features.
Table 4. Comparison (%) of the results obtained in the PolitiFact dataset w/o user posting behavior features.
ModelAccuracy (%)Precision (%)Recall (%)F1 (%)
SAF [19]0.6910.6380.7890.706
BiLSTM-BERT [36]0.885NANANA
LNN-KG [37]0.8800.90110.8800.8892
Logistic Regression (N-Gram) [38]0.800.790.780.78
dEFEND [20]0.9040.9020.9560.928
BERT b a s e -CNN-BiGRU-ATT0.87770.86210.93750.8982
BERT b a s e -BiGRU-CNN-ATT0.88490.86440.86440.8644
BERT b a s e -CNN-BiGRU0.89930.93420.88750.9103
Table 5. Comparison (%) of the results obtained in the GossipCop dataset w/o user posting behavior features.
Table 5. Comparison (%) of the results obtained in the GossipCop dataset w/o user posting behavior features.
ModelAccuracy (%)Precision (%)Recall (%)F1 (%)
SAF [19]0.7960.8200.7530.785
Logistic Regression (N-Gram) [38]0.820.750.790.77
dEFEND [20]0.8080.7290.7820.755
BERT b a s e -CNN-BiGRU-ATT0.85840.88490.93610.9098
BERT b a s e -BiGRU-CNN-ATT0.85010.89020.91650.9031
BERT b a s e -CNN-BiGRU0.86190.86190.97510.9150
Based on the experimental results, we made the following observations.
  • To answer the first question raised in the introduction, based on the experiments, we observed that the interplay between the news text and the user posting behavioral features improves the detection performance, achieving state-of-the-art results. This seems to confirm our hypothesis—stated in Section 2—that user posting behavior attributes and news content contain complementary information that must be encoded and captured simultaneously in order to detect fake news.
  • To answer the second question raised in the introduction, the experimental results demonstrate that by excluding context-based features, such as user-posted behavioral cues, the performance deteriorated in the two datasets, showing the power of harnessing such clues in the modeling process.
Mining user interactions with news articles on social media can aid in the identification of false information [39]. As a result, adding pertinent information from user–news interactions would provide additional knowledge and improve detection performance. The experimental findings showed that, to a certain extent, user features could be used to efficiently quantify detection performance. Experimental findings of two real-world datasets show how the proposed models can improve detection performance significantly. Additionally, we found that simplified versions of our models that utilized specific components only exhibited degraded performance, although still superior to baseline models, but were outperformed by the full proposed models that integrate the semantically rich information of news with user posting behavior features. The proposed model structures, on the other hand, are relatively complex. We anticipate that by using DistilBert, which has fewer parameters than BERT b a s e , our model will be trained efficiently in a large volume of real-world data. It has been demonstrated that combining content and context features provides discriminatory power for distinguishing fake from real content. This study emphasizes the existence of a strong and consistent correlation between news content and user behavior information when semantic contextualized relations derived from content and context features are combined for improved detection quality.

8. Conclusions

This study leverages context and content-based features for fake news detection. More specifically, tests were conducted to determine the impacts of considering the interplay between news content and the user posting behavior features on the performance of the detection models. The test results show that the combination of such features positively improves the detection performance. Our study exploits the power of the context-aware BERT b a s e model for modeling the input text of news content. In addition, several deep learning models, such as CNN and bidirectional GRU (with attention module) were evaluated in this study using two well-known real-world fake news datasets. Based on the test results, BERT b a s e -CNN-BiGRU-ATT outperformed other models and achieved state-of-the-art results in the PolitiFact dataset, while BERT b a s e -BiGRU-CNN-ATT provided the best performance score in the GossipCop dataset.
The following will be used in the future to improve the proposed architectures:
  • Other context-based features, e.g., user comments, replies, spatiotemporal features, and propagation features, can improve detection performance.
  • Experiments can be conducted using other context-aware models, such as RoBERTa.
  • Data augmentation techniques can be applied to balance the GossipCop dataset to reduce the potential bias and improve the detection performance.

Author Contributions

Conceptualization, algorithm development, experiments, and formal analysis, J.A.; writing, J.A.; writing, review, editing, analysis, interpretation of the experimental data, and supervision, Y.L. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The FakeNewsNet data are available at https://github.com/KaiDMML/FakeNewsNet, accessed on 15 January 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kwak, H.; Lee, C.; Park, H.; Moon, S. What is Twitter, a Social Network or a News Media? In Proceedings of the 19th International Conference on World Wide Web, WWW’10, Raleigh, NC, USA, 26–30 April 2010; Association for Computing Machinery: New York, NY, USA, 2010; pp. 591–600. [Google Scholar] [CrossRef] [Green Version]
  2. Park, K.; Rim, H. Social media hoaxes, political ideology, and the role of issue confidence. Telemat. Inform. 2019, 36, 1–11. [Google Scholar] [CrossRef]
  3. Langin, K. Fake news spreads faster than true news on Twitter—Thanks to people, not bots. Sci. Mag. 2018. [Google Scholar] [CrossRef]
  4. Vosoughi, S.; Roy, D.; Aral, S. The spread of true and false news online. Science 2018, 359, 1146–1151. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, S.; Pang, M.S.; Pavlou, P.A. Cure or Poison? Identity Verification and the Posting of Fake News on Social Media. J. Manag. Inf. Syst. 2021, 38, 1011–1038. [Google Scholar] [CrossRef]
  6. Conroy, N.K.; Rubin, V.L.; Chen, Y. Automatic deception detection: Methods for finding fake news. Proc. Assoc. Inf. Sci. Technol. 2015, 52, 1–4. [Google Scholar] [CrossRef] [Green Version]
  7. Zhou, X.; Jain, A.; Phoha, V.V.; Zafarani, R. Fake News Early Detection: An Interdisciplinary Study. arXiv 2019, arXiv:1904.11679. [Google Scholar]
  8. Zhou, X.; Zafarani, R. A survey of fake news: Fundamental theories, detection methods, and opportunities. ACM Comput. Surv. (CSUR) 2020, 53, 1–40. [Google Scholar] [CrossRef]
  9. Pérez-Rosas, V.; Kleinberg, B.; Lefevre, A.; Mihalcea, R. Automatic detection of fake news. arXiv 2017, arXiv:1708.07104. [Google Scholar]
  10. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  11. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the International Conference on Machine Learning, PMLR, Beijing, China, 22–24 June 2014; pp. 1188–1196. [Google Scholar]
  12. Wang, W.Y. “Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection. arXiv 2017, arXiv:1705.00648. [Google Scholar]
  13. Hakak, S.; Alazab, M.; Khan, S.; Gadekallu, T.R.; Maddikunta, P.K.R.; Khan, W.Z. An ensemble machine learning approach through effective feature extraction to classify fake news. Future Gener. Comput. Syst. 2021, 117, 47–58. [Google Scholar] [CrossRef]
  14. Gravanis, G.; Vakali, A.; Diamantaras, K.; Karadais, P. Behind the cues: A benchmarking study for fake news detection. Expert Syst. Appl. 2019, 128, 201–213. [Google Scholar] [CrossRef]
  15. Liu, T.; Yu, K.; Wang, L.; Zhang, X.; Zhou, H.; Wu, X. Clickbait detection on WeChat: A deep model integrating semantic and syntactic information. Knowl.-Based Syst. 2022, 245, 108605. [Google Scholar] [CrossRef]
  16. Razaque, A.; Alotaibi, B.; Alotaibi, M.; Amsaad, F.; Manasov, A.; Hariri, S.; Yergaliyeva, B.B.; Alotaibi, A. Blockchain-Enabled Deep Recurrent Neural Network Model for Clickbait Detection. IEEE Access 2022, 10, 3144–3163. [Google Scholar] [CrossRef]
  17. Razaque, A.; Alotaibi, B.; Alotaibi, M.; Hussain, S.; Alotaibi, A.; Jotsov, V. Clickbait Detection Using Deep Recurrent Neural Network. Appl. Sci. 2022, 12, 504. [Google Scholar] [CrossRef]
  18. Mouratidis, D.; Nikiforos, M.N.; Kermanidis, K.L. Deep learning for fake news detection in a pairwise textual input schema. Computation 2021, 9, 20. [Google Scholar] [CrossRef]
  19. Shu, K.; Mahudeswaran, D.; Wang, S.; Lee, D.; Liu, H. FakeNewsNet: A Data Repository with News Content, Social Context and Spatialtemporal Information for Studying Fake News on Social Media. arXiv 2019, arXiv:1809.01286. [Google Scholar] [CrossRef]
  20. Shu, K.; Cui, L.; Wang, S.; Lee, D.; Liu, H. DEFEND: Explainable Fake News Detection. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD’19, Anchorage, AK, USA, 4–8 August 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 395–405. [Google Scholar]
  21. Ma, J.; Gao, W.; Wei, Z.; Lu, Y.; Wong, K.F. Detect rumors using time series of social context information on microblogging websites. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 19–23 October 2015; pp. 1751–1754. [Google Scholar]
  22. Castillo, C.; Mendoza, M.; Poblete, B. Information credibility on twitter. In Proceedings of the 20th International Conference on World Wide Web, Hyderabad, India, 28 March–1 April 2011; pp. 675–684. [Google Scholar]
  23. Shu, K.; Zhou, X.; Wang, S.; Zafarani, R.; Liu, H. The role of user profiles for fake news detection. In Proceedings of the 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Vancouver, BC, Canada, 27–30 August 2019; pp. 436–439. [Google Scholar]
  24. Long, Y.; Lu, Q.; Xiang, R.; Li, M.; Huang, C.R. Fake News Detection Through Multi-Perspective Speaker Profiles. In Proceedings of the Eighth International Joint Conference on Natural Language Processing, Taipei, Taiwan, 27 November–1 December 2017; pp. 252–256. [Google Scholar]
  25. Kaliyar, R.K.; Goswami, A.; Narang, P. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimed. Tools Appl. 2021, 80, 11765–11788. [Google Scholar] [CrossRef] [PubMed]
  26. Alghamdi, J.; Lin, Y.; Luo, S. Modeling Fake News Detection Using BERT-CNN-BiLSTM Architecture. In Proceedings of the 2022 IEEE 5th International Conference on Multimedia Information Processing and Retrieval (MIPR), Virtual Event, 2–4 August 2022; pp. 354–357. [Google Scholar] [CrossRef]
  27. Alghamdi, J.; Lin, Y.; Luo, S. A Comparative Study of Machine Learning and Deep Learning Techniques for Fake News Detection. Information 2022, 13, 576. [Google Scholar] [CrossRef]
  28. Ruchansky, N.; Seo, S.; Liu, Y. Csi: A hybrid deep model for fake news detection. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 797–806. [Google Scholar]
  29. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  30. Liu, G.; Guo, J. Bidirectional LSTM with attention mechanism and convolutional layer for text classification. Neurocomputing 2019, 337, 325–338. [Google Scholar] [CrossRef]
  31. Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  32. Chaudhari, S.; Mithal, V.; Polatkan, G.; Ramanath, R. An attentive survey of attention models. ACM Trans. Intell. Syst. Technol. (TIST) 2021, 12, 1–32. [Google Scholar] [CrossRef]
  33. Sardelich, M.; Manandhar, S. Multimodal deep learning for short-term stock volatility prediction. arXiv 2018, arXiv:1812.10479. [Google Scholar]
  34. Kim, Y. Convolutional Neural Networks for Sentence Classification. arXiv 2014, arXiv:1408.5882. [Google Scholar]
  35. Liu, J.; Wang, X.; Tan, Y.; Huang, L.; Wang, Y. An Attention-Based Multi-Representational Fusion Method for Social-Media-Based Text Classification. Information 2022, 13, 171. [Google Scholar] [CrossRef]
  36. Sadeghi, F.; Jalaly Bidgoly, A.; Amirkhani, H. Fake News Detection on Social Media Using A Natural Language Inference Approach. Multimed. Tools Appl. 2022, 81, 33801–33821. [Google Scholar] [CrossRef]
  37. Koloski, B.; Stepišnik-Perdih, T.; Robnik-Šikonja, M.; Pollak, S.; Škrlj, B. Knowledge Graph informed Fake News Classification via Heterogeneous Representation Ensembles. arXiv 2021, arXiv:2110.10457. [Google Scholar] [CrossRef]
  38. Oriola, O. Exploring N-gram, word embedding and topic models for content-based fake news detection in FakeNewsNet evaluation. Int. J. Comput. Appl. 2021, 975, 8887. [Google Scholar] [CrossRef]
  39. Shu, K.; Wang, S.; Liu, H. Beyond news contents: The role of social context for fake news detection. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, Australia, 11–15 February 2019; pp. 312–320. [Google Scholar]
Figure 1. BERT b a s e -CNN-BiGRU-ATT architecture.
Figure 1. BERT b a s e -CNN-BiGRU-ATT architecture.
Applsci 13 03345 g001
Figure 2. BERT b a s e -BiGRU-CNN-ATT architecture.
Figure 2. BERT b a s e -BiGRU-CNN-ATT architecture.
Applsci 13 03345 g002
Figure 3. BERT b a s e -CNN-BiLSTM architecture.
Figure 3. BERT b a s e -CNN-BiLSTM architecture.
Applsci 13 03345 g003
Figure 4. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in PolitiFact.
Figure 4. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in PolitiFact.
Applsci 13 03345 g004
Figure 5. Confusion matrix of BERT b a s e -BiGRU-CNN-ATT in PolitiFact.
Figure 5. Confusion matrix of BERT b a s e -BiGRU-CNN-ATT in PolitiFact.
Applsci 13 03345 g005
Figure 6. Confusion Matrix of BERT b a s e -CNN-BiGRU in PolitiFact.
Figure 6. Confusion Matrix of BERT b a s e -CNN-BiGRU in PolitiFact.
Applsci 13 03345 g006
Figure 7. Performance comparisons (%) of all models in PolitiFact.
Figure 7. Performance comparisons (%) of all models in PolitiFact.
Applsci 13 03345 g007
Figure 8. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in GossipCop.
Figure 8. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in GossipCop.
Applsci 13 03345 g008
Figure 9. Confusion Matrix of BERT b a s e -BiGRU-CNN-ATT in GossipCop.
Figure 9. Confusion Matrix of BERT b a s e -BiGRU-CNN-ATT in GossipCop.
Applsci 13 03345 g009
Figure 10. Confusion matrix of BERT b a s e -CNN-BiGRU in GossipCop.
Figure 10. Confusion matrix of BERT b a s e -CNN-BiGRU in GossipCop.
Applsci 13 03345 g010
Figure 11. Performance comparisons (%) of all models in GossipCop.
Figure 11. Performance comparisons (%) of all models in GossipCop.
Applsci 13 03345 g011
Figure 12. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in PolitiFact w/o user posting behavior features.
Figure 12. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in PolitiFact w/o user posting behavior features.
Applsci 13 03345 g012
Figure 13. Confusion matrix of BERT b a s e -BiGRU-CNN-ATT in PolitiFact w/o user posting behavior features.
Figure 13. Confusion matrix of BERT b a s e -BiGRU-CNN-ATT in PolitiFact w/o user posting behavior features.
Applsci 13 03345 g013
Figure 14. Confusion matrix of BERT b a s e -CNN-BiGRU in PolitiFact w/o user posting behavior features.
Figure 14. Confusion matrix of BERT b a s e -CNN-BiGRU in PolitiFact w/o user posting behavior features.
Applsci 13 03345 g014
Figure 15. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in GossipCop w/o user posting behavior features.
Figure 15. Confusion matrix of BERT b a s e -CNN-BiGRU-ATT in GossipCop w/o user posting behavior features.
Applsci 13 03345 g015
Figure 16. Confusion matrix of BERT b a s e -BiGRU-CNN-ATT in GossipCop w/o user posting behavior features.
Figure 16. Confusion matrix of BERT b a s e -BiGRU-CNN-ATT in GossipCop w/o user posting behavior features.
Applsci 13 03345 g016
Figure 17. Confusion matrix of BERT b a s e -CNN-BiGRU in GossipCop w/o user posting behavior features.
Figure 17. Confusion matrix of BERT b a s e -CNN-BiGRU in GossipCop w/o user posting behavior features.
Applsci 13 03345 g017
Figure 18. Performance comparisons (%) of models in PolitiFact w/o user posting behavior features.
Figure 18. Performance comparisons (%) of models in PolitiFact w/o user posting behavior features.
Applsci 13 03345 g018
Figure 19. Performance comparisons (%) of models in GossipCop w/o user posting behavior features.
Figure 19. Performance comparisons (%) of models in GossipCop w/o user posting behavior features.
Applsci 13 03345 g019
Table 1. The statistics of FakeNewsNet dataset.
Table 1. The statistics of FakeNewsNet dataset.
DatasetPolitiFactGossipCop
# Candidate news69418,676
# True news35614,129
# Fake news3384547
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alghamdi, J.; Lin, Y.; Luo, S. Does Context Matter? Effective Deep Learning Approaches to Curb Fake News Dissemination on Social Media. Appl. Sci. 2023, 13, 3345. https://doi.org/10.3390/app13053345

AMA Style

Alghamdi J, Lin Y, Luo S. Does Context Matter? Effective Deep Learning Approaches to Curb Fake News Dissemination on Social Media. Applied Sciences. 2023; 13(5):3345. https://doi.org/10.3390/app13053345

Chicago/Turabian Style

Alghamdi, Jawaher, Yuqing Lin, and Suhuai Luo. 2023. "Does Context Matter? Effective Deep Learning Approaches to Curb Fake News Dissemination on Social Media" Applied Sciences 13, no. 5: 3345. https://doi.org/10.3390/app13053345

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop