Next Article in Journal
Remineralization Induced by Biomimetic Hydroxyapatite Toothpastes on Human Enamel
Next Article in Special Issue
YOLO-DRS: A Bioinspired Object Detection Algorithm for Remote Sensing Images Incorporating a Multi-Scale Efficient Lightweight Attention Mechanism
Previous Article in Journal
A New Concept of Sustainable Wind Turbine Blades: Bio-Inspired Design with Engineered Adhesives
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bio-Inspired Artificial Intelligence with Natural Language Processing Based on Deceptive Content Detection in Social Networking

1
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
2
Department of Information Systems, College of Computer Science, King Khalid University, P.O. Box 394, Abha 61421, Saudi Arabia
3
Department of Information science, College of Humanities and Social Sciences, King Saud University, P.O. Box 28095, Riyadh 11437, Saudi Arabia
4
Department of Information Technology, College of Computers and Information Technology, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia
5
Department of Electronics and Communication Engineering, New Horizon College of Engineering, Bengaluru 560103, India
6
Research Center, Future University in Egypt, New Cairo 11835, Egypt
*
Author to whom correspondence should be addressed.
Biomimetics 2023, 8(6), 449; https://doi.org/10.3390/biomimetics8060449
Submission received: 28 August 2023 / Revised: 7 September 2023 / Accepted: 16 September 2023 / Published: 23 September 2023
(This article belongs to the Special Issue Bioinspired Artificial Intelligence Applications)

Abstract

:
In recent research, fake news detection in social networking using Machine Learning (ML) and Deep Learning (DL) models has gained immense attention. The current research article presents the Bio-inspired Artificial Intelligence with Natural Language Processing Deceptive Content Detection (BAINLP-DCD) technique for social networking. The goal of the proposed BAINLP-DCD technique is to detect the presence of deceptive or fake content on social media. In order to accomplish this, the BAINLP-DCD algorithm applies data preprocessing to transform the input dataset into a meaningful format. For deceptive content detection, the BAINLP-DCD technique uses a Multi-Head Self-attention Bi-directional Long Short-Term Memory (MHS-BiLSTM) model. Finally, the African Vulture Optimization Algorithm (AVOA) is applied for the selection of optimum hyperparameters of the MHS-BiLSTM model. The proposed BAINLP-DCD algorithm was validated through simulation using two benchmark fake news datasets. The experimental outcomes portrayed the enhanced performance of the BAINLP-DCD technique, with maximum accuracy values of 92.19% and 92.56% on the BuzzFeed and PolitiFact datasets, respectively.

1. Introduction

In recent years, the popularity of social media has encouraged the spread of fake news like never before [1]. It is challenging to detect the fake news on social media as the false news is purposefully posted to mislead the public, leaving it challenging to distinguish the fake news from the original. So, in recent times, approaches to fake news recognition have captivated researchers [2]. In general, the framework for the detection of fake news related to social context encounters two difficulties in its modelling. The first issue is that the data in a social news context are heterogeneous and complex. Various kinds of entities are involved in social media content, for example, multiple types of connections, posts, replies, users, and re-posts. This creates following relations (user–user), responsive relations (post–post), and publishing relations (user–post) [3]. Though the heterogeneous features and connections of such an entity support the provision of evidence from diverse parties to confirm the news, a problem is their potential usage minus such evidence, with latter difficult to identify through modelling [4]. The second issue is the problem of distribution shift in modelling, where the training distribution varies with different test distributions.
Generally, the techniques for automatically identifying false data on social networking sites take advantage of social context data or news content [5]. Against this background, each dataset has its own strengths and weaknesses, and it is important not to work with incorrect data, which can create difficulties. Furthermore, despite the utility of social context data in enhancing the precision of approaches, various issues may cause substantial delays in fake news recognition [6]. These are both key considerations as the two most crucial factors in false data detection are early detection and accuracy [7]. Additionally, the highly adaptable nature of deceptive data plays a key role in obstructing the efficacy of the existing fake news recognition systems; here, a narrator can be useful to help ascertain the authenticity of news statements on a real-time basis. A further factor to consider is that it is hard to gain widespread data to train the systems for false news detection [8]. Moreover, it is a challenge to extract relevant attributes that can help in finding false news data in different fields [9]. Finally, it is difficult to find fake news about a new event as there are usually constrained data and knowledge concerning this event [10].
The current research article presents the Bio-inspired Artificial Intelligence with Natural Language Processing Deceptive Content Detection (BAINLP-DCD) technique in social networking. The goal of the BAINLP-DCD technique is to detect the presence of deceptive or fake content on social media. To accomplish this objective, the BAINLP-DCD algorithm uses a data preprocessing method to transform the input data into a meaningful format. For deceptive content detection, the BAINLP-DCD technique uses a Multi-Head Self-attention Bi-directional Long Short-Term Memory (MHS-BiLSTM) model. Finally, the African Vulture Optimization Algorithm (AVOA) is used for the selection of optimum hyperparameters for the MHS-BiLSTM model. The proposed BAINLP-DCD methodology was validated through simulation using two benchmark fake news datasets. The major contributions of the study are as follows.
  • Development of a novel BAINLP-DCD technique encompassing MHS-BiLSTM-based classification and AVOA-based hyperparameter tuning for deceptive content detection. To the best of the authors’ knowledge, the proposed BAINLP-DCD technique is a new contribution to the literature.
  • The parameter optimization of the MHS-BiLSTM model using the AVOA with cross-validation helps in boosting the predictive outcomes of the proposed model for unseen data.

2. Related Works

Nadeem et al. [11] introduced a hybrid technique named HyproBert for automated recognition of fake news. This technique utilizes DistilBERT for word embedding and tokenization. The embedding is fed as an input into the convolutional layer for extracting and highlighting the spatial features. Then, the output is presented to the Bidirectional Gated Recurrent Unit (BiGRU) for the extraction of the context features. Lastly, a dense layer is applied to integrate each feature for classification. In [12], the authors presented the Bidirectional Encoder Representation from Transformers (BERT)-related Deep Learning (DL) method by merging various parallel blocks of single-layer Deep Convolutional Neural Networks (DCNNs) and diverse kernel filters and sizes with BERT. Ahmad et al. [13] suggested new social-related and content-related attributes for the identification of rumors on social media networks. The study’s findings suggested that the presented features were highly useful in categorizing rumors compared with the existing baseline features. Likewise, the authors applied BiLSTM with a Recurrent Neural Network (Bi-LSTM-RNN) to text in order to detect rumors. This technique is a simple one, yet it has the potential to detect rumors. In the study, the experimentations on rumor detection were carried out using real-time datasets.
Kaliyar et al. [14] devised a deep CNN technique (FNDNet) for Fake News Detection (FND). Instead of depending on handcrafted features, the proposed method (FNDNet) was devised to discriminate between features automatically for the classification of false information using hidden layers, framed in the Deep Neural Network (DNN). The authors created a deep CNN to extract numerous features from all layers. Then, the study compared the performance of the presented method against various baseline approaches. In [15], a hyperparameter-tuned DL-related automatic FND (HDL-FND) method was proposed. This method accomplished results in terms of classification and detection of fake news. The proposed method encompassed a three-stage process involving BiLSTM-based classification, preprocessing, and feature extraction. Lee et al. [16] presented a DL structure for identifying fake news written in the Korean language. The earlier studies designed suitable FND methods for the English language, whereas the Korean language has two problems that cannot be overcome by the prevailing methods: first, it is hard for the DNN to function due to the feature scarcity for DL; second, due to morpheme ambiguity, semantic analysis is difficult.
Goldani et al. [17] modelled a CNN technique with margin loss and various embedding techniques for the detection of fake news. In this study, the static word embedding was compared against non-static embedding that offers the likelihood of gradually updating and up-training the word embeddings, during the training stage. The devised structure was assessed using two recent datasets in the domain, namely LIAR and ISOT. In [18], the authors carried out three experiments with Machine Learning (ML) methods, DL methods, and transformers. In all the experimentations, the authors relied upon the word embedding method for the extraction of contextual features. The experimental outcomes showed that the DL methods outpaced ML classifiers and BERT transformers in terms of accuracy. Additionally, the outcomes also displayed the same accuracy as those of the GRU and LSTM methods.
In [19], the performance of 16 meta-heuristic approaches in Artificial Neural Network (ANN) training was examined initially for the detection of non-linear systems. The study was designed to define the highly efficient meta-heuristic NN training systems. In [20], an NN-based method was presented for the detection of non-linear static methods. A modified algorithm named ABCES (ABC relies on Effective Scout Bee Stage) was established in this study for NN training. Via this approach, two essential alterations could be executed with ABCES. The authors of [21] developed a novel NN training approach named Hybrid Artificial Bee Colony algorithm based on the Effective Scout Bee Stage (HABCES). According to [22], the Feed-Forward NN (FFNN) training can be applied for Maximum Power Point Tracking (MPPT) by employing 13 Swarm-Intelligent (SI)-based optimizer systems.
Most of the studies conducted already on deceptive content detection on social media were based on off-the-shelf ML or DL models with default hyperparameters. So, a need still exists to explore the influence of hyperparameter tuning on the performance of the model, especially in the case of deceptive content detection. Since social media data come in a huge volume and often contain noisy and unstructured content, it is challenging to detect the deceptive content. Hyperparameter-tuning techniques can help in optimizing the model architectures and hyperparameters to handle this complexity in an effective manner. Since trial-and-error hyperparameter tuning is a tedious process, in this work, the AVOA is applied for hyperparameter tuning.

3. The Proposed Model

In the current study, we designed and developed the BAINLP-DCD technique for deceptive content detection in social media data. The major intention of the BAINLP-DCD technique is to detect the presence of deceptive or fake content in social media data. Figure 1 depicts the entire process of the proposed BAINLP-DCD approach. To accomplish its objective, the BAINLP-DCD technique follows a sequence of procedures such as data preprocessing, MHS-BiLSTM-based classification, and AVOA-based hyperparameter tuning. Initially, the input data are preprocessed to transform those into a useful format. Next, the detection process is conducted using the MHS-BiLSTM model. Finally, the AVOA is applied for the selection of the optimal hyperparameters for the MHS-BiLSTM model.

3.1. Data Preprocessing

In general, the input data are preprocessed in different ways to improve the quality of the data. The outcomes from the data-mining process depend on a lot of preprocessing [23]. This adapts the unreliable and lacking raw data into representative machine-readable data. To accomplish this task, NLP methods like character alteration to lowercase letters, stemming, tokenization, stop word elimination, and other such processes in the Keras library are employed. Words such as ‘the,’ ‘of,’ ‘there,’ etc., are called stop words and are the most frequently utilized words in day-to-day conversations. As such, a stop word may appear in a text several times, though it likely has a limited impact in terms of the whole contexts of phrases.

3.2. Detection Using the MHS-BiLSTM Model

For deceptive content detection, the BAINLP-DCD technique uses the MHS-BiLSTM model. LSTM-NN is a kind of RNN that presents a ‘gate’ model. It is capable of capturing the long-term semantic dependence and preventing the gradient-disappearing problems of the classical RNN, due to its long sequence [24]. Thus, the LSTM approach is applied for sentiment classification tasks. The computing method of the LSTM method is shown Equations (1)–(6).
i t = σ W i x t + U i h t 1 + V i c t 1    
C ~ t = t a n h W c x t + U c h t 1    
f t = σ W f x t + U f h t 1 + V f c t 1
c t = f t c t 1 + i t C ~ t
o t = σ W o x t + U o h t 1 + V o c t
h t = o t t a n h C
Here, the input at  t  time, the cell layer, the value of the forget gate, the value of the input gate, the value of the output gate, the layer of the candidate cell, and the outcome of the LSTM units are represented by    x t ,   c t ,   i t ,   f t ,   o t ,   C ~ t ,  and  h t ,  respectively. The sigmoid activation function is represented by  σ , while   indicates the dot product function among the weighted matrices and  { W * , U * , V * } * { i , f , c , 0 }  shows the parameter set from the LSTM unit. Hence, the BLSTM-NN technique is commonly utilized as the building block of the DL sentiment classification method to attain a better classification efficiency.
Word2vec translates the text data of the assessment information into a vector representation that is fed as an input into the BLSTM. The final result of the sample sentiment classification is attained by passing the BLSTM to the sigmoid layer.
Considering the sentence  S , a pretrained model and a standard tokenizer are used to attain  D  dimensional embedding for a single word in the sentence, where  S = e 1 , e 2 , ,   e N  and  S R N × D  from the input to the model [25]. It is necessary to recognize certain words to identify the sarcasm in sentence  S  that provide relevant clues like negative emotions and sarcastic connotations. The importance of such a cue word corresponds to various factors based on dissimilar contexts. In this work, Multi-Head Self-Attention (MHSA) is leveraged to detect the cue words from the input texts.
Here, the attention module is used to determine the design in the input that is critical for determining the presented task. The self-attention model helps in learning a task’s particular connection amongst various modules to generate the best series representation. In this self-attention model, three linear projections exist, where  K ,   Q ,   V R N × D ,  Key ( K ), Value ( V ), and Query ( Q ) of the given input order are created. The attention map is calculated according to the comparison among  K Q , and the outcomes of these modules.  A R N × D  denotes the scaled Dot-product between  V  and learned softmax attention  Q K T  as explained in Equation (7) as follows.
A = s o f t m a x Q K T D V  
Various copies of the self-attention model are utilized in parallel to MHSA. Every head captures the dissimilar connections among the individual keywords that support classification and identifies the words from the input text. In this work containing different heads  # H  in all the layers, a sequence of MHSA layers  # L  is used. Figure 2 demonstrates the framework of the attention BiLSTM layer.

3.3. Hyperparameter Tuning Using AVOA

In the current study, the AVOA is used for optimal fine-tuning of the hyperparameters for the MHS-BiLSTM model. AVOA is a new meta-heuristic swarm-based optimization approach inspired by the hunting style of the African vulture [26]. The African vulture is a kind of hunter that preys upon weak animals as its food. The AVOA is particularly inspired by its feeding and orienting behaviors. The algorithm consists of powerful operators while it also maintains a balance of exploration and efficiency in solving the continuous optimization problems.
In this method, there is an  N  count of the population of vultures and its values are adjusted to suit the problems that need to be resolved. The fitness of the vultures is measured after its arbitrary initialization. The best vulture is the vulture with the optimum solution, chosen to lead the first group; the second-best vulture is the vulture with next best solution, chosen to lead the rest of the groups. The remaining population is disseminated to make up both the groups, as in Equation (8). By applying the roulette wheel mechanism, the probability of choosing the group is calculated. The  α  and  β  parameters are predefined parameters so that the value lies in the range of [0, 1], where their sum is equivalent to  1 . When  α  is closer to  1  and  β  lies near  0 , the intensification increases in AVOA. If  β  is closer to 1 and  α  is closer to  0 , this increases the diversification. The following equation provides more details in this regard.
R i = B e s t V u l T u r e 1   i f   p i = α B e s t V u l T u r e 2   i f   p i = β
The starved vulture becomes aggressive; this stage defines the starvation rate of the vulture. A satiated vulture has abundant energy to travel a long distance, foraging for food; when it becomes starved, it becomes highly aggressive and finds food near other vultures. Equation (9) computes the satiation rate, which switches between the exploration and exploitation stages. Equation (10) ensures that the exploration stage reaches the exact estimate of the overall optimal solution and also that no early convergence takes place.
F = 2 × r a n d + 1 × z × 1 i i m a x + t
t = h × s i n γ π 2 × i i m a x + c o s π 2 × i i m a x 1
Here,  F  is the rate of starvation,  i  represents the existing iteration,  i m a x  shows the max iteration count, and  z  indicates a random integer within [ 1,1 ]; if  z  is negative, the vulture is starved, and if positive, the vulture is satisfied.  r a n d  denotes a random integer in [ 0,1 ] and  h  shows the random value within [ 2,2 ]. The last iteration of the AVOA performs the exploitation and exploration stages.
The transformation between the exploration and exploitation phases can be accomplished by shifting the probability of entering the exploration stage at the last stage. If the parameter  γ  reduces, then the probability of turning towards the exploration phase in the end stage reduces. At last, if  F  is less than 1, then the vulture is starved and finds prey in the following region, and accordingly, the AVOA enters the exploitation stage. If  F  exceeds one, then the vulture is satisfied and pursues food foraging in other regions. Therefore, the AVOA enters the exploration stage.
In the exploration stage, the vulture can find food and travel for a long distance. The vulture identifies a new area based on the satiation level, and a parameter  P 1  is used for the selection process, as shown in the Equations (11) and (12). Two strategies are used to ensure a wider exploration of the search range. The initial strategy considers the exploration of the search region, adjacent to a better vulture from the group. This technique enables the localized exploration from the neighborhood of the present optimum performance. On the other hand, the next approach allows exploration across the whole search range while following the specific upper and lower limitations. This method permits a wider exploration range without surpassing the limits.
P i + 1 = R i D i × F D i = X × R i P i P 1 r a n d P 1    
P i + 1 = R i F + r a n d 2 × u b l b × r a n d 3 + l b       P 1 < r a n d P 1
Here ,   P 1  denotes the random integer that lies in the range of [ 0,1 ],  R ( i )  indicates the  v u l t u r e i  leader, and  F  shows the rate of starvation, as set out in (9).  X  co-efficient is used for increasing the arbitrary movement and alteration with all the iterations.  u b  and  l b  indicate the upper and lower boundaries, respectively.  r a n d P 1 r a n d 2 , and  r a n d 3  are random integers within [0, 1].
When the rate of starvation is  | F | < l , then it enters the exploitation stage, which has two phases. If the values of the rate of starvation range between 1 and 0.5, then the AVOA technique enters the first exploitation stage. However, the vulture remains comparatively satisfied. At this point, a random number  r a n d P 2  is created in [0, 1] for choosing that specific approach to follow, and later, it is compared to a predetermined parameter  P 2 . The siege fight strategy is applied if  r a n d P 2  is equal to or greater than  P 2 . In this strategy, the vulture behaviors are simulated, i.e., when a strong vulture refuses to share food, the weaker vultures tend to tire this strong vulture out by surrounding and attacking it. The rotating flight is chosen if  r a n d P 2  is less than  P 2 ,  i.e., spiral motion of a vulture while finding food. Equations (13) and (14) represent the siege and rotation flights, respectively.
P i + 1 = D i × F + r a n d 4 d t d t = R i P i   P 2 r a n d P 2  
P i + 1 = R ( i ) ( S 1 + S 2 ) S 1 = R ( i ) × r a n d 5 × P i 2 π × c o s P i S 2 = R i × r a n d 6 × P i 2 π × s i n P i   P 2 < r a n d P 2
Here  D i  is evaluated in (10),  F  denotes the rate of starvation,  R ( i )  indicates the vulture’s leader,  P ( i )  stands for the present location of the vultures, and  r a n d 4 r a n d P 2  and  r a n d 6  denote the random numbers within [0, 1], which, in turn, increase the arbitrariness. The values of  s i n ( P ( i ) )  or  c o s P i  form an array vector of dimension  n × 1 , whereas  n  indicates the count of units generated.
If  | F | < 0.5 , the vulture is starved and aggressive in the second exploitation phase. Firstly, a random integer  r a n d P 3  within [0, 1] is compared to a predetermined parameter  P 3 . When  r a n d P 3  is equal to or greater than  P 3 , different kinds of vultures accumulate over the food. In other terms, an aggressive siege fight is selected as follows.
P i + 1 = A 1 + A 2 2 A 1 = B e s t V u l t u r e 1 i B e s t V u l t u r e 1 i × P i B e s t V u l t u r e 1 ( i ) × P ( i ) 2 × F   P 3 r a n d P 3
A 2 = B e s t V u l t u r e 2 ( i ) B e s t V u l t u r e 2 i × P i B e s t V u l t u r e 2 i × P ( i ) 2 × F P i + 1 = R i d t × F × L e v y d   P 3 < r a n d P 3
Here, the first and the second best vultures in the existing iteration are represented as  B e s t V u l t u r e 1 ( i )  and  B e s t V u l t u r e 2 i ,  respectively.  d  is the problem dimension.  P ( i )  denotes the existing location of the vultures.  R ( i ) ,   d ( t ) , and  F  were described earlier, while the levy flight increases the efficacy of the AVOA and is evaluated as follows.
L F x = 0.01 × u × σ 1 σ = Γ 1 + β × sin π β 2 Γ 1 + β 2 × β × 2 β 1 2 1 β    
In Equation (17),   β  is a set number that is equivalent to 1.5, whereas  u  and  v  are randomly generated values within [0, 1].
The AVOA system progresses a Fitness Function (FF) to realize the best classifier solution. It defines a positive integer to denote the best solution for candidate efficiencies. In this case, FF corresponds to the reduction in classifier errors, as expressed in Equation (18).
f i t n e s s x i = C l a s s i f i e r   E r r o r   R a t e x i = N o .   o f   m i s c l a s s i f i e d   i n s t a n c e s   T o t a l   n o .   o f   i n s t a n c e s   * 100

4. Experimental Validation

The proposed model was simulated using the Python 3.6.5 tool on a PC configured with i5-8600k, GeForce 1050Ti 4 GB, 16 GB RAM, 250 GB SSD, and 1 TB HDD. The parameter settings used for the investigation were as follows: learning rate: 0.01, dropout: 0.5, batch size: 5, epoch count: 50, and activation: ReLU.
The experimentation analysis for the BAINLP-DCD method was conducted using two datasets, Buzz Feed and PolitiFact. Table 1 shows the details of the BuzzFeed dataset [27].
Figure 3 illustrates the results accomplished by the BAINLP-DCD method on the BuzzFeed dataset. Figure 3a,b show the confusion matrices generated with the BAINLP-DCD technique at a 70:30 ratio of training phase (TRP)/testing phase (TSP). The figure signifies that the BAINLP-DCD approach identified and classified both the classes accurately. Also, Figure 3c exhibits the PR curve of the BAINLP-DCD algorithm. The figure denotes that the BAINLP-DCD method attained the optimum Precision-Recall (PR) performance on two class labels. Finally, Figure 3d demonstrates the ROC of the BAINLP-DCD method. The figure reveals that the BAINLP-DCD system produced effective results with higher ROC values under both the class labels.
The fake news detection results attained by the BAINLP-DCD technique on the BuzzFeed dataset are shown in Table 2 and Figure 4. The outcomes show that the BAINLP-DCD system properly categorized the real news class from the fake news. On the 70% TR set, the BAINLP-DCD technique gained average  a c c u y p r e c n r e c a l , and  F s c o r e  values of 91.99%, 95.81%, 91.99%, and 93.62%, respectively. In addition to this, on the 30% TS set, the BAINLP-DCD system achieved average  a c c u y p r e c n r e c a l , and  F s c o r e  values of 92.19%, 95.45%, 92.19%, and 93.38%, respectively.
Figure 5 demonstrates the training accuracy curves    T R _ a c c u y  and  V L _ a c c u y  accomplished by the BAINLP-DCD technique on the BuzzFeed dataset.  T L _ a c c u y  was determined by evaluating the BAINLP-DCD system on the TR dataset, whereas  V L _ a c c u y  was calculated by evaluating the performance on separate testing datasets. The experimental outcomes infer that the  T R _ a c c u y  and  V L _ a c c u y  values increase with an upsurge in the number of epochs. Therefore, the performance of the BAINLP-DCD method was enhanced on the TR and TS datasets with an increase in the number of epochs.
In Figure 6, the  T R _ l o s s  and  V R _ l o s s  analysis values achieved with the BAINLP-DCD method on the BuzzFeed dataset are presented. The  T R _ l o s s  is determined as the error between the predicted performance and original values of the TR data. The  V R _ l o s s  denotes a measure of performance of the BAINLP-DCD approach on separate validation data. The outcomes specify that the  T R _ l o s s  and  V R _ l o s s  values tend to lessen with a greater number of epochs. The outcomes establish the improved performance of the BAINLP-DCD system and its ability to accomplish accurate classification. The decreased  T R _ l o s s  and  V R _ l o s s  values show the better performance of the BAINLP-DCD approach in capturing the relationships and patterns.
In Table 3 and Figure 7, the comparative analysis outcomes of the BAINLP-DCD system on the BuzzFeed dataset are portrayed [1]. The experimental outcomes denote that the PBIC, CIMTDetect, and NBFND-PDA techniques accomplished poor performance. Simultaneously, the TFLI-FND, DF-IFND DNN, and EchoFakeD algorithm achieved closer outcomes. However, the BAINLP-DCD method achieves a superior performance with  a c c u y p r e c n r e c a l , and  F s c o r e  values of 92.19%, 95.45%, 92.19%, and 93.38%, respectively.
Table 4 provides a detailed description of the PolitiFact dataset.
Figure 8 portrays the classifier analysis outcomes of the BAINLP-DCD method on the PolitiFact dataset. Figure 8a,b show the confusion matrices generated by the BAINLP-DCD technique on a 70:30 TRP/TSP dataset. The outcomes signify that the BAINLP-DCD approach recognized both the classes precisely. Also, Figure 8c shows the PR curve of the BAINLP-DCD model. The figure denotes that the BAINLP-DCD system attained the maximum PR values under both the class labels. Finally, Figure 8d demonstrates the ROC of the BAINLP-DCD system. The figure depicts that the BAINLP-DCD method achieved efficient outcomes with higher ROC values on the two class labels.
The fake news detection outcomes of the BAINLP-DCD system on the PolitiFact dataset are shown in Table 5 and Figure 9. These values indicate that the BAINLP-DCD method appropriately categorized the real and fake news class labels. On the 70% TR set, the BAINLP-DCD approach achieved an increase in the average  a c c u y p r e c n r e c a l , and  F s c o r e  values, with 92.56%, 94.90%, 92.56%, and 93.59%, respectively. Additionally, on the 30% TS set, the BAINLP-DCD system gained average  a c c u y p r e c n r e c a l , and  F s c o r e  values of 88.89%, 95%, 88.89%, and 91.12%, respectively.
Figure 10 illustrates the training accuracy, i.e.,  T R _ a c c u y  and  V L _ a c c u y  values achieved by the BAINLP-DCD system on the PolitiFact dataset.  T L _ a c c u y  is determined by estimated the BAINLP-DCD algorithm on TR datasets, whereas  V L _ a c c u y  is calculated by evaluating the performance on a separate testing dataset. The investigational results show that the  T R _ a c c u y  and  V L _ a c c u y  values increase with an upsurge in the number of epochs. Thus, the performance of the BAINLP-DCD approach achieved enhanced accuracy values on both TR and TS datasets with an increase in the number of epochs.
In Figure 11, the  T R _ l o s s  and  V R _ l o s s  analysis outcomes achieved by the BAINLP-DCD system on the PolitiFact dataset are demonstrated.  T R _ l o s s  defines the error between the predicted performance and the original values on the TR data, while  V R _ l o s s  signifies a measure of performance for the proposed BAINLP-DCD system on a separate validation dataset. The experimental outcomes denote that the  T R _ l o s s  and  V R _ l o s s  values tend to reduce with an increase in the number of epochs. These outcomes depict the enriched performance of the BAINLP-DCD method and its ability to accomplish accurate classification. The minimized  T R _ l o s s  and  V R _ l o s s  values confirm the superior performance of the BAINLP-DCD approach in capturing the relationships and patterns.
In Table 6 and Figure 12, the comparison study outcomes of the BAINLP-DCD technique and other techniques using the PolitiFact dataset are given. The results show that the PBIC, CIMTDetect, and CITDetect models accomplished the worst performance. At the same time, the TFLI-FND, DF-IFND DNN, and EchoFakeD models attain closer results. However, the BAINLP-DCD technique reaches the maximum performance with  a c c u y p r e c n r e c a l , and  F s c o r e  values of 92.56%, 94.90%, 92.56%, and 93.59%, respectively.
Therefore, the proposed BAINLP-DCD system can be applied in automated fake news detection. The enhanced performance of the BAINLP-DCD technique is due to the inclusion of an AVOA-based hyperparameter-tuning process. Hyperparameters are settings that should be set prior to the training process as they are not learned during the training process. They exert a significant impact on the performance of the model, and selection of the optimal values can lead to better accuracy. By applying AVOA-based hyperparameter tuning, the proposed model achieves better results thanks to the selection of the optimal settings for the algorithm. These results confirm the improved performance of the proposed technique over other existing techniques.

5. Conclusions and Future Work

In the current study, we designed and developed the BAINLP-DCD technique for deceptive content detection in social media data. The major intention of the BAINLP-DCD technique is to detect the presence of deceptive or fake content in social media data. In order to accomplish this objective, the BAINLP-DCD technique follows a sequence of procedures like data preprocessing, MHS-BiLSTM-based classification, and AVOD-based hyperparameter tuning. For deceptive content detection, the BAINLP-DCD technique uses the MHS-BiLSTM model. Finally, the AVOA technique is applied for optimal hyperparameter selection of the MHS-BiLSTM network. The proposed BAINLP-DCD system was experimentally validated through simulation using two fake news datasets. The experimental outcomes confirmed the enhanced performance of the BAINLP-DCD method, with maximum accuracy values of 92.19% and 92.56% on the BuzzFeed and PolitiFact datasets, respectively.
In the future, the proposed model can be extended for sarcasm detection and classification processes. Furthermore, future works can extend the proposed model to multi-class classification, as well. Fine-grained classification techniques can also be explored to classify the deceptive content into different classes, like misinformation, disinformation, fake news, propaganda, etc., and thus provide more nuanced insights into the nature of deceptive content. In addition, the explainability and interpretability of the DL approaches can also be explored for deceptive content detection. Moreover, the computational complexity of the proposed model can be examined in the future.

Author Contributions

Conceptualization, A.A.A.; methodology, A.A.A., M.M., F.A.A., M.M.A. and A.K.; software, A.K.; validation, F.A.A., M.M.A. and A.K.; investigation, M.M.; data curation, F.A.A., A.S.; writing – original draft, A.A.A., M.M., F.A.A. and M.M.A.; writing – review & editing, M.M., M.M.A., A.S. and A.K.; visualization, A.K.; supervision, A.A.A.; project administration, M.M.A.; funding acquisition, A.A.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through large group Research Project under grant number (RGP2/ 235 /44). Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R191), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. Research Supporting Project number (RSPD2023R838), King Saud University, Riyadh, Saudi Arabia. This study is partially funded by the Future University in Egypt (FUE).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data sharing does not apply to this article as no datasets were generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through the contributions of all authors. All authors have approved the final version of the manuscript.

References

  1. Kaliyar, R.K.; Goswami, A.; Narang, P. EchoFakeD: Improving fake news detection in social media with an efficient deep neural network. Neural Comput. Appl. 2021, 33, 8597–8613. [Google Scholar] [CrossRef]
  2. Kong, S.H.; Tan, L.M.; Gan, K.H.; Samsudin, N.H. Fake news detection using deep learning. In Proceedings of the 2020 IEEE 10th Symposium on Computer Applications & Industrial Electronics (ISCAIE), Penang, Malaysia, 18–19 April 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 102–107. [Google Scholar]
  3. Sahoo, S.R.; Gupta, B.B. Multiple features based approach for automatic fake news detection on social networks using deep learning. Appl. Soft Comput. 2021, 100, 106983. [Google Scholar] [CrossRef]
  4. Hiramath, C.K.; Deshpande, G.C. Fake news detection using deep learning techniques. In Proceedings of the 2019 1st International Conference on Advances in Information Technology (ICAIT), Chikmagalur, India, 25–27 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 411–415. [Google Scholar]
  5. Aslam, N.; Ullah Khan, I.; Alotaibi, F.S.; Aldaej, L.A.; Aldubaikil, A.K. Fake detect: A deep learning ensemble model for fake news detection. Complexity 2021, 2021, 1–8. [Google Scholar] [CrossRef]
  6. Amine, B.M.; Drif, A.; Giordano, S. Merging deep learning model for fake news detection. In Proceedings of the 2019 International Conference on Advanced Electrical Engineering (ICAEE), Algiers, Algeria, 19–21 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–4. [Google Scholar]
  7. Ozbay, F.A.; Alatas, B. Fake news detection within online social media using supervised artificial intelligence algorithms. Phys. A Stat. Mech. Its Appl. 2020, 540, 123174. [Google Scholar] [CrossRef]
  8. Mahir, E.M.; Akhter, S.; Huq, M.R. Detecting fake news using machine learning and deep learning algorithms. In Proceedings of the 2019 7th International Conference on Smart Computing & Communications (ICSCC), Sarawak, Malaysia, 28–30 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  9. Lu, S.; Zhu, Z.; Gorriz, J.M.; Wang, S.H.; Zhang, Y.D. NAGNN: Classification of COVID-19 based on neighboring aware representation from deep graph neural network. Int. J. Intell. Syst. 2022, 37, 1572–1598. [Google Scholar] [CrossRef]
  10. Lu, S.; Wang, S.H.; Zhang, Y.D. Detection of abnormal brain in MRI via improved AlexNet and ELM optimized by chaotic bat algorithm. Neural Comput. Appl. 2021, 33, 10799–10811. [Google Scholar] [CrossRef]
  11. Nadeem, M.I.; Mohsan, S.A.H.; Ahmed, K.; Li, D.; Zheng, Z.; Shafiq, M.; Karim, F.K.; Mostafa, S.M. HyproBert: A Fake News Detection Model Based on Deep Hypercontext. Symmetry 2023, 15, 296. [Google Scholar] [CrossRef]
  12. Kaliyar, R.K.; Goswami, A.; Narang, P. FakeBERT: Fake news detection in social media with a BERT-based deep learning approach. Multimed. Tools Appl. 2021, 80, 11765–11788. [Google Scholar] [CrossRef]
  13. Ahmad, T.; Faisal, M.S.; Rizwan, A.; Alkanhel, R.; Khan, P.W.; Muthanna, A. Efficient Fake News Detection Mechanism Using Enhanced Deep Learning Model. Appl. Sci. 2022, 12, 1743. [Google Scholar] [CrossRef]
  14. Kaliyar, R.K.; Goswami, A.; Narang, P.; Sinha, S. FNDNet–A deep convolutional neural network for fake news detection. Cogn. Syst. Res. 2020, 61, 32–44. [Google Scholar] [CrossRef]
  15. Kanagavalli, N.; Priya, S.B.; Jeyakumar, D. Design of Hyperparameter Tuned Deep Learning based Automated Fake News Detection in Social Networking Data. In Proceedings of the 2022 6th International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 29–31 March 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 958–963. [Google Scholar]
  16. Lee, D.H.; Kim, Y.R.; Kim, H.J.; Park, S.M.; Yang, Y.J. Fake news detection using deep learning. J. Inf. Process. Syst. 2019, 15, 1119–1130. [Google Scholar]
  17. Goldani, M.H.; Safabakhsh, R.; Momtazi, S. Convolutional neural network with margin loss for fake news detection. Inf. Process. Manag. 2021, 58, 102418. [Google Scholar] [CrossRef]
  18. Amer, E.; Kwak, K.S.; El-Sappagh, S. Context-based fake news detection model relying on deep learning models. Electronics 2022, 11, 1255. [Google Scholar] [CrossRef]
  19. Kaya, E. A comprehensive comparison of the performance of metaheuristic algorithms in neural network training for nonlinear system identification. Mathematics 2022, 10, 1611. [Google Scholar] [CrossRef]
  20. Kaya, E.; Baştemur Kaya, C. A novel neural network training algorithm for the identification of nonlinear static systems: Artificial bee colony algorithm based on effective scout bee stage. Symmetry 2021, 13, 419. [Google Scholar] [CrossRef]
  21. Kaya, E. A new neural network training algorithm based on artificial bee colony algorithm for nonlinear system identification. Mathematics 2022, 10, 3487. [Google Scholar] [CrossRef]
  22. Kaya, E.; Baştemur Kaya, C.; Bendeş, E.; Atasever, S.; Öztürk, B.; Yazlık, B. Training of Feed-Forward Neural Networks by Using Optimization Algorithms Based on Swarm-Intelligent for Maximum Power Point Tracking. Biomimetics 2023, 8, 402. [Google Scholar] [CrossRef]
  23. Altheneyan, A.; Alhadlaq, A. Big data ML-based fake news detection using distributed learning. IEEE Access 2023, 11, 29447–29463. [Google Scholar] [CrossRef]
  24. Wang, C.; Yang, X.; Ding, L. Deep learning sentiment classification based on weak tagging information. IEEE Access 2021, 9, 66509–66518. [Google Scholar] [CrossRef]
  25. Akula, R.; Garibay, I. Interpretable multi-head self-attention architecture for sarcasm detection in social media. Entropy 2021, 23, 394. [Google Scholar] [CrossRef]
  26. Abuelrub, A.; Awwad, B. An improved binary African vultures optimization approach to solve the UC problem for power systems. Results Eng. 2023, 19, 101354. [Google Scholar] [CrossRef]
  27. Available online: https://www.kaggle.com/datasets/mdepak/fakenewsnet (accessed on 16 June 2023).
Figure 1. Overall procedure of the BAINLP-DCD method.
Figure 1. Overall procedure of the BAINLP-DCD method.
Biomimetics 08 00449 g001
Figure 2. Architecture of attention BiLSTM.
Figure 2. Architecture of attention BiLSTM.
Biomimetics 08 00449 g002
Figure 3. Classification results on BuzzFeed dataset.
Figure 3. Classification results on BuzzFeed dataset.
Biomimetics 08 00449 g003
Figure 4. Average values of the BAINLP-DCD technique on the BuzzFeed dataset.
Figure 4. Average values of the BAINLP-DCD technique on the BuzzFeed dataset.
Biomimetics 08 00449 g004
Figure 5. A c c u y  curve of the BAINLP-DCD technique on the BuzzFeed dataset.
Figure 5. A c c u y  curve of the BAINLP-DCD technique on the BuzzFeed dataset.
Biomimetics 08 00449 g005
Figure 6. Loss curve of the BAINLP-DCD technique on the BuzzFeed dataset.
Figure 6. Loss curve of the BAINLP-DCD technique on the BuzzFeed dataset.
Biomimetics 08 00449 g006
Figure 7. Comparative outcomes of the BAINLP-DCD algorithm on BuzzFeed dataset.
Figure 7. Comparative outcomes of the BAINLP-DCD algorithm on BuzzFeed dataset.
Biomimetics 08 00449 g007
Figure 8. Classification results on PolitiFact dataset.
Figure 8. Classification results on PolitiFact dataset.
Biomimetics 08 00449 g008
Figure 9. Average values of the BAINLP-DCD technique on PolitiFact dataset.
Figure 9. Average values of the BAINLP-DCD technique on PolitiFact dataset.
Biomimetics 08 00449 g009
Figure 10. A c c u y  curve of the BAINLP-DCD technique on the PolitiFact dataset.
Figure 10. A c c u y  curve of the BAINLP-DCD technique on the PolitiFact dataset.
Biomimetics 08 00449 g010
Figure 11. Loss curve of the BAINLP-DCD technique on the PolitiFact dataset.
Figure 11. Loss curve of the BAINLP-DCD technique on the PolitiFact dataset.
Biomimetics 08 00449 g011
Figure 12. Comparative outcomes of the BAINLP-DCD algorithm and other recent models on PolitiFact dataset.
Figure 12. Comparative outcomes of the BAINLP-DCD algorithm and other recent models on PolitiFact dataset.
Biomimetics 08 00449 g012
Table 1. Details of BuzzFeed dataset.
Table 1. Details of BuzzFeed dataset.
BuzzFeed Dataset
ClassNo. of Samples
Real News182
Fake News91
Total Samples273
Table 2. Fake news detection outcomes when using the BAINLP-DCD technique on the BuzzFeed dataset.
Table 2. Fake news detection outcomes when using the BAINLP-DCD technique on the BuzzFeed dataset.
Class   A c c u y   P r e c n   R e c a l   F s c o r e
Training Phase (70%)
Real News99.2493.5799.2496.32
Fake News84.7598.0484.7590.91
Average91.9995.8191.9993.62
Testing Phase (30%)
Real News100.0090.91100.0095.24
Fake News84.38100.0084.3891.53
Average92.1995.4592.1993.38
Table 3. Comparison analysis outcomes of the BAINLP-DCD method and other recent systems on the BuzzFeed dataset.
Table 3. Comparison analysis outcomes of the BAINLP-DCD method and other recent systems on the BuzzFeed dataset.
BuzzFeed Dataset
Methods   A c c u y   P r e c n   R e c a l   F s c o r e
PBIC on Twitter90.1273.5078.3075.60
CIMTDetect90.7472.9092.3081.30
TFLI-FND91.0885.2083.0083.50
NBFND-PDA90.8084.9085.2084.20
DF-IFND DNN91.0983.3386.9685.11
EchoFakeD91.3890.4787.3688.37
BAINLP-DCD92.1995.4592.1993.38
Table 4. Details of PolitiFact dataset.
Table 4. Details of PolitiFact dataset.
PolitiFact Dataset
ClassNo. of Samples
Real News240
Fake News120
Total Samples360
Table 5. Fake news detection outcomes of the BAINLP-DCD technique on the PolitiFact dataset.
Table 5. Fake news detection outcomes of the BAINLP-DCD technique on the PolitiFact dataset.
Class   A c c u y   P r e c n   R e c a l   F s c o r e
Training Phase (70%)
Real News98.2193.7598.2195.93
Fake News86.9096.0586.9091.25
Average92.5694.9092.5693.59
Testing Phase (30%)
Real News100.0090.00100.0094.74
Fake News77.78100.0077.7887.50
Average88.8995.0088.8991.12
Table 6. Comparison study outcomes of the BAINLP-DCD method and other recent methods on the PolitiFact dataset.
Table 6. Comparison study outcomes of the BAINLP-DCD method and other recent methods on the PolitiFact dataset.
PolitiFact Dataset
Methods   A c c u y   P r e c n   R e c a l   F s c o r e
PBIC on Twitter91.5377.7079.1078.30
CITDetect88.8567.9097.5079.10
CIMTDetect90.5080.3084.2081.80
TFLI-FND91.2687.2082.1084.30
DF-IFND DNN86.3982.1084.6084.04
EchoFakeD90.4386.3690.4888.37
BAINLP-DCD92.5694.9092.5693.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Albraikan, A.A.; Maray, M.; Alotaibi, F.A.; Alnfiai, M.M.; Kumar, A.; Sayed, A. Bio-Inspired Artificial Intelligence with Natural Language Processing Based on Deceptive Content Detection in Social Networking. Biomimetics 2023, 8, 449. https://doi.org/10.3390/biomimetics8060449

AMA Style

Albraikan AA, Maray M, Alotaibi FA, Alnfiai MM, Kumar A, Sayed A. Bio-Inspired Artificial Intelligence with Natural Language Processing Based on Deceptive Content Detection in Social Networking. Biomimetics. 2023; 8(6):449. https://doi.org/10.3390/biomimetics8060449

Chicago/Turabian Style

Albraikan, Amani Abdulrahman, Mohammed Maray, Faiz Abdullah Alotaibi, Mrim M. Alnfiai, Arun Kumar, and Ahmed Sayed. 2023. "Bio-Inspired Artificial Intelligence with Natural Language Processing Based on Deceptive Content Detection in Social Networking" Biomimetics 8, no. 6: 449. https://doi.org/10.3390/biomimetics8060449

Article Metrics

Back to TopTop