Next Article in Journal
Rational Sampling Numbers of Soil pH for Spatial Variation: A Case Study from Yellow River Delta in China
Next Article in Special Issue
hLSTM-Aging: A Hybrid LSTM Model for Software Aging Forecast
Previous Article in Journal
Potential for Energy Utilization of Air Compression Section Using an Open Absorption Refrigeration System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Model for a Secure Social Media Application

1
Computer Science Department, College of Computer Sciences & Information Technology, King Faisal University, Al-Ahsa 31982, Saudi Arabia
2
Mathematics Department, Faculty of Science, Zagazig University, Zagazig 44519, Egypt
3
College of Computing and Informatics, University of Sharjah, Sharjah 27272, United Arab Emirates
4
Faculty of Computers and Information, Mansoura University, Mansoura 35516, Egypt
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6375; https://doi.org/10.3390/app12136375
Submission received: 5 May 2022 / Revised: 19 June 2022 / Accepted: 20 June 2022 / Published: 23 June 2022
(This article belongs to the Special Issue Dependability and Security of IoT Network)

Abstract

:
The progress of data technology and wireless networks is generated by open online communication channels. Unfortunately, trolls are abusing the technology for executing cyberattacks and threats. An automated cybersecurity solution is vital for avoiding the threats and security issues from social media. This can be a requirement for tackling and considering cyberbullying in various aspects including prevention of such incidents and automated detection. This study introduces a novel Artificial Fish Swarm Algorithm with Weighted Extreme Learning Machine (AFSA-WELM) model for cybersecurity on social media. The proposed model is mostly intended to detect the existence of cyberbullying on social media. The proposed model starts by processing the dataset and making it ready for the next stages of the model. It then uses the TF-IDF vectorizer for word embedding. After that, it uses the WELM model for the identification and classification of cyberbullying. Finally, the optimal tunning parameters used in the WELM model are derived for the AFSA model. The experimental analysis has shown that the proposed model achieves maximum accuracy compared with existing algorithms. Moreover, our proposed model achieves maximum precision–recall performance with various datasets.

1. Introduction

Cyberbullying has been a progressive and always developing issue all through most recent years, and thus programmed recognition of it has been an engaging point for scientists. Cyberbullying or disdain discourse is the demonstration of posting scornful substance on social media, which is found to be harmful because of its tedious and oppressive nature [1]. Cyberbullying can influence individuals’ mental states and different parts of life. Particularly for youngsters, cyberbullying could even lead them to cause self-harm and self-destruction [2]. Scientists have likewise concentrated on the relationship between cyberbullying exploitation and the risk of self-destructive ideation [3,4]. It is a fundamental and provoking task to distinguish cyberbullying. Concerning its significance, although the mischief of cyberbullying should be limited as much as can be expected, it is more vital to foster a programmed model to foresee cyberbullying occasions, rather than checking physically or helping victims a while later.
The difficulties of cyberbullying discovery are as follows [5]. To begin with, according to the point of view of manual verification, the judgment of whether specific conduct is cyberbullying changes from one individual to another. For instance, when a sentence contains swear words, it may very well be considered cyberbullying activity, yet now and again, this is not accurate, such as when some teenagers speak with their companions on social networks. After that, where cyberbullying happens most often [6], social organization’s platforms generally have the idea of communicating with the public and unknown commenters. Thus, posts are more autonomous and context-free, and it is simple to cause doubt or misdirect others. Third, one significant issue with cyberbullying research is the absence of standard information [7]. Although the information utilized in numerous past investigations was obtained from similar social networks (e.g., Twitter), they were made by freely utilizing public API or by site scraping. Subsequently, individual data cannot be measured against each another and are not helpful for checking the technique’s comprehensiveness.
This study introduces a novel Artificial Fish Swarm Algorithm with Weighted Extreme Learning Machine (AFSA-WELM) model for cybersecurity on social media. The AFSA-WELM model presented here is mainly intended to detect the existence of cyberbullying on social media. The main contributions of this study are summarized as follows:
  • Introducing an efficient AFSA-WELM model to detect cyberbullying on social media using an artificial fish swarm algorithm and weighted extreme machine learning.
  • Using the TF-IDF model for feature extraction, using the WELM model for classification, and applying the AFSA model for parameter tuning.
  • The presented AFSA-WELM model has been validated through extensive experimental evaluations on various datasets. It achieved the maximum precision–recall performance for various datasets.
Finally, the presented AFSA-WELM has shown maximum accuracy compared with existing algorithms such as the SVM and Naïve Bayes algorithms.
The rest of this study is organized as follows. Section 2 presents the related work and discussion. Section 3 introduces the proposed model. Section 4 presents the detailed results and discussion. This is followed by the conclusions in Section 5.

2. Related Work

Regarding instant messages in social networks, the length of sentences is generally short and noisy. These messages can likewise be unstructured, e.g., messages could contain emoticons, emojis, or incorrect spellings, which befuddles models catching information from instant messages [8]. While a multitude of techniques have attempted to work on the exactness of the cyberbullying-related task somewhat, we accept that they have not completely used precise data on past text-based information, which we call the setting in the context of this study [9]. The main classification of setting is an inner setting that incorporates any non-text-based data that can be separated from the dataset itself, e.g., the client’s liked posts, adherents, or multimedia content connected to the social media post, e.g., pictures and recordings [10,11].
Bozyiğit et al. [12] described the significance of social networking characteristics in detecting cyberbullying. Initially, a balanced dataset consisting of 5000 labeled contents with numerous social networking characteristics was built. Next, the relationships between social media features and cyberbullying were analyzed by chi-square testing. It is obvious that a few characteristics (for example, the sender’s followers) are powerfully connected with online bullying events according to the tested outcomes. The researchers in [13] proposed a DNN method for detecting cyberbullying in three distinct types of social information, such as visual media, infographics, and textual content (embedded text and an image). The compact structure, CapsNet–ConvNet, comprises ConvNet to predict the visual bullying content and CapsNet-DNN with dynamic routing to predict text bullying content. The researchers in [14] presented an ML technique that (i) fine-tunes variants of BERT, a deep attention-based language technique that can detect patterns in a long and noisy body of text; (ii) extracts contextual data from various sources involving external knowledge sources, metadata, and images, and uses those characteristics for complementing the learning method; and (iii) effectively integrates textual and contextual characteristics through boosting and a wide-ranging framework. Fang et al. [15] presented the advantages of GRU cells and Bi-GRU for learning the fundamental relationships among words in these two directions. Rupesh et al. [16] used machine learning algorithms to efficiently detect hate speech on social networks. In addition, we presented a self-attention model which had the advantage of achieving great efficiency in cyberbullying classification tasks.
In this study, we conducted a comparative analysis of the presented the proposed model with some recent models, as shown in Table 1.
The table portrays a brief comparative study of the AFSA-WELM model and recent models [17].

3. The Proposed Model

This study introduces a novel model for cybersecurity on social media, as shown in Figure 1. The figure illustrates the overall process of the proposed model. The presented model is mainly intended to detect the existence of cyberbullying on social media. The proposed model is composed of five main stages:
  • Data input: The model accepts the training dataset.
  • Data pre-processing: The inputted data will be pre-processed to the formats that will be used in the next stages.
  • Feature extraction: The model uses a TF-IDF vectorizer for word embedding.
  • Classification and identification: The model uses WELM to efficiently identify and classify cyberbullying.
  • Parameter tunning: In this stage, the AFSA model is derived to perfectly obtain the optimal tuned parameters that are used in the proposed model.
Finally, in order to validate the introduced model, we introduced a comparative study that showed the efficiency of the introduced model over other methods in the related literature.

3.1. TF-IDF Vectorizer

In this work, the TF-IDF vectorizer was used for word embedding. TF-IDF is an extensively used feature extraction system for examining texts. The two substantial tasks are indexing and weighting; TF-IDF manages the weighting. It determines the weight of t in a document D. TF-IDF is derived from TF and IDF as follows [18]:
T F ( t ) = t D N D
I D F ( t ) = l o g d d t
Now t D ,   d , and d t are used to characterize the inclusive number of times of t appears in a document D , the total number of documents and the quantity of documents with term t .
The weight of each term according to the TF-IDF is estimated by:
W t , d = T F t , d ( t D d f , t )
where T F t , d and d f , t denote the frequency of t in document d and the sum of documents containing t , respectively.

3.2. WELM-Based Classification

Next, the WELM model was used for the identification and classification of cyberbullying. The WELM classification model was used to assign appropriate class labels. ELM was used for the classification of a balanced dataset, while WELM was used for the classification of an imbalanced dataset [19]. The training dataset has N distinct instances ,   ( x i ,   z i ) , i = 1 , 2 , , N . A single hidden layer NN with L hidden layer node was formulated by:
i = 1 L β i 1 ( w i x j + b i ) = z j , j = 1 , , N
whereas w i characterizes a single hidden layer’s input weight, l ( ) denotes the activation functions, β i indicates the output weights, and b i signifies the bias of the individual hidden layers.
S β = T
where S denotes the output matrix of a single hidden layer
S ( w 1 ,   ,   w L ,   b 1 ,   ,   b L ,   x 1 ,   ,   x N ) = ( l ( w 1 x 1 + b 1 l ( w L x 1 + b L ) l ( w 1 x N + b 1 ) l ( w L x N + b L ) ) N × L
Based on the Karush–Kuhn–Tucker concept, the Lagrangian factor was used to change the trained ELM as a dual problem [20]. The output weight β can be evaluated as follows:
β = S T ( 1 C + S S T ) 1 T
Here, C denotes the regularization coefficient. Therefore, the output function of the ELM classification is formulated by:
F ( x ) = s ( x ) S T ( 1 C + S S T ) 1 T = [ K ( x , x 1 ) K ( x , x N ) ] T ( 1 C + χ ) 1 T
where χ denotes the kernel matrix, which is calculated by:
χ = S S T = s ( x i ) s ( x j ) = K ( x i ,   x j )
It is obvious in (13) that the hidden layer feature maps s ( x ) denote the independence in classification results of ELM. The classification result is related to the kernel function, K ( x ,   y ) . K ( x ,   y ) denotes an inner product, and the number of hidden layers is no outcomes in the resulting output.
K ( x ,   y ) = exp   ( γ x y 2 )
Therefore, the KELM classification efficacy is distributed into two variables, namely the kernel function parameter, γ and the penalty parameter, C. The original goal of ELM is achieved in WELM by preserving the weight for distinct instances to handle imbalanced classification problems. It is calculated by the following equations:
F ( x ) = [ K ( x , x 1 ) K ( x , x N ) ] T ( 1 C + W χ ) 1 W T
W = d i a g   ( w i i ) , i = 1 , 2 ,   ,   N
where W denotes the weight matrix. WELM has two weightage models as follows:
w i i = 1 # ( z i )
w i i = { 0.618 # ( z i ) ,   i f   z i > z ¯ 1 # ( z i ) ,   o t h e r w i s e
where # ( z i ) denotes the number of instances assigned to class z i ,   i = 1 , , m ; m denotes the number of classes; and z ¯ indicates the average instance of all the classes. Figure 2 showcases the framework of ELM.

3.3. AFSA-Based Parameter Optimization

At this stage, the AFSA model is derived for optimal tuning of the parameters involved in the WELM model. The ASFA replicates the behavior of a prey fish swarm and it has been established for rapid searching. Exploration begins with arbitrary result maps for each fish. Fitness is assessed and the fish following techniques are initiated, and each fish follows the other fishes via the optimum outcome. Once the outcome is sub-optimum, a swarming method is initiated. Once the chosen outcome is not linked, the prey method is initiated. This process repeats until the required threshold or termination condition is encountered [21]. The AFSA has four functions that have been modeled on the nature of fish in a fish swarm. A primary function is free-moving action: in the environment, a fish travels in a swarm when it is not injured. The next function is prey action, since it uses smell, vision, and other available senses in the body. In AFSA, the area where an artificial fish senses a predator is modeled as a neighborhood via a visual-sized radius. The next function is following action when a fish identifies food; other swarm members follow it to reach the food. The last function is the swarm action, which replicates fishes as each fish continually attempts to be in the swarm and does not leave so it can be protected from the hunter.
Consider the condition vector of AFS is ( x 1 ,   x 2 , x n ) , where x 1 , x 2 , x n denotes the location of the fish. The food in this location of the fish is characterized by y = f ( x ) , in which y indicates the objective function value. The distance amongst artificial fish is d i j = X i X j ,   where   i and j represent arbitrary fish, Step indicates the maximum step size of artificial fish, and a denotes the degree of congestion factor.
Let X v be the visual location at a particular moment. X n e x t denotes the novel location. The development process is shown as follows:
X v = X i + V i s u a l × r a n d ( ) , i [ 0 ,   n ]
X n e x t = X + X v X X v X × s t e p × r a n d ( )
During the prey model, the behavior of fish is represented as follows:
p r e y   ( X i ) = { x i + s t e p x j x i x j x i i f     y j y i x i + ( 2 r a n d 1 )   .   s t e p e l s e
where rand denotes a random function within [ 0 , 1 ] .   In the swarming phase, the behavior of the swarm is represented as follows:
s w a r m ( X i ) = { X i + s t e p x j x i x j x i i f y c n f > δ y i p r e y ( X i ) e l s e
The next stage is represented in the following (Equation (19)):
f o l l o w   ( X i ) = { X i + s t e p x m a x x i x m a x x i i f y   max   n f > δ y i p r e y ( X i ) e l s e
where xmax is the last location that the fish can arrive and ymax is the maximum value of the objective function to guarantee that local and global exploration, and the exploration direction towards the optimum food source are carried out.

4. Results and Discussion

The presented AFSA-WELM model was simulated using a dataset collected from the Kaggle repository. The dataset holds 12,729 samples with 11,661 data classified as non-cyberbullying and 1068 data in the cyberbullying class.
Figure 3 illustrates the confusion matrices offered by the AFSA-WELM model for distinct sizes of training/testing datasets. With a training/testing data split of 90:10, the proposed model identified 1158 samples in the non-cyberbullying class and 104 samples in the cyberbullying class. In addition, with a training/testing data split of 80:20, the AFSA-WELM method identified 2323 samples in the non-cyberbullying class and 187 samples in the cyberbullying class. Meanwhile, with a training/testing data split of 70:30, the proposed model identified 3432 samples in the non-cyberbullying class and 215 samples in the cyberbullying class. Finally, with a training/testing data split of 60:40, the AFSA-WELM technique identified 4619 samples in the non-cyberbullying class and 347 samples in the cyberbullying class.
Table 2 reports the overall classifier results of our proposed model with different training/testing dataset sizes. With a training/testing dataset split of 90:10, the AFSA-WELM model resulted in an a c c u y , p r e c n , r e c a l , and F s c o r e of 99.14%, 95.22%, 99.53%, and 97.25%, respectively. Moreover, with a training/testing dataset split of 80:20, the proposed model resulted in an a c c u y , p r e c n , r e c a l , and F s c o r e of 98.59%, 96.26%, 94.25%, and 95.23%c correspondingly. In addition, with a training/testing dataset split of 70:30, the proposed model resulted in an a c c u y , p r e c n , r e c a l , and F s c o r e of 95.50%, 86.88%, 82.46%, and 84.49%, correspondingly. Lastly, with a training/testing dataset split of 60:40, the proposed model resulted in an a c c u y , p r e c n , r e c a l , and F s c o r e of 97.53%, 93.19%, 90.22%, and 91.64%, correspondingly.
A brief precision–recall examination of the AFSA-WELM model with different training/testing dataset sizes is portrayed in Figure 4. By observing the figure, it can be noticed that the AFSA-WELM model achieved maximum precision–recall performance for all datasets. The figure shows the recall and the precision values of the four different samples.
Figure 5 portrays the ROC investigation of the proposed model with different training/testing dataset sizes. The figure shows the false positive rate and the true positive rate. The figure has four different training/testing samples to verify the stability of the proposed model’s performance. The figure shows that the AFSA-WELM model produces adequate results with maximum ROC values for the different class labels.
Figure 6 reports a comparative a c c u y investigation of the AFSA-WELM model and existing models. The figure shows that the SVM (RBF algorithm) achieved a lower a c c u y of 87.28%. Next, the SVM (linear algorithm) and the Naïve Bayes algorithm achieved a slightly increased a c c u y of 95.17% and 95.59%, respectively. Along with that, the SVM (poly-algorithm) and SVM (sigmoid algorithm) models achieved a reasonable a c c u y of 98.55% and 98.29%, respectively. However, the AFSA-WELM model achieved the maximum a c c u y of 99.14%.
Figure 7 shows a comparative p r e c n investigation of our proposed model and existing models. The figure portrays that the SVM (RBF algorithm) obtained a lower p r e c n of 88.57%. Moreover, the SVM (linear algorithm) and the Naïve Bayes algorithm attained a slightly increased a c c u y of 93.75% and 95.24%, respectively. Next, the SVM (poly-algorithm) and SVM (sigmoid algorithm) models reached a reasonable p r e c n of 97.94% and 98.43%, correspondingly. Finally, the proposed method achieve a maximal a c c u y of 95.22%.
Figure 8 showcases a comparative r e c a l investigation of the proposed method and existing models. The figure shows that the SVM (RBF algorithm) achieved a lower r e c a l of 89.67%. Moreover, the SVM (linear algorithm) and the Naïve Bayes algorithm reached a somewhat increased r e c a l of 95.07% and 96.70%, correspondingly. At the same time, the SVM (poly-algorithm) and SVM (sigmoid algorithm) models reached a reasonable r e c a l of 98.795% and 98.43%, correspondingly. Finally, the proposed model achieved a maximum r e c a l of 99.53%.
After examining the results and discussion, it was verified that the proposed model resulted in maximum performance in terms of the detection and classification of cyberbullying.

5. Conclusions

This study presented a novel model for cybersecurity on social media. The proposed model developed here mostly detects the existence of cyberbullying on social media. The introduced model is composed of five main stages: input, pre-processing, feature extraction, classification and identification, and parameter tuning. Our model uses an artificial fish swarm algorithm and weighted extreme machine learning to effectively detect cyberbullying on social networks. It also uses the TF-IDF model to successfully carry out the feature extraction process with the dataset. Moreover, it uses WELM and AFSA for classification and parameter tuning, respectively. The experimental results show that the developed method achieved a maximum a c c u y and r e c a l in comparison with the SVM and Naïve Bayes algorithms. Future extensions of this study are as follows. Deep learning models can be introduced to improve the cyberbullying detection outcomes. Moreover, an extensive comparison needs to be conducted between the proposed model and some recent models in the related literature.

Author Contributions

Conceptualization, K.R. and M.E.; methodology, K.R. and M.E.; software, K.R. and M.E.; validation, K.R. and M.E.; formal analysis, K.R. and M.E.; investigation, K.R. and M.E.; resources, K.R. and M.E.; data curation, K.R. and M.E.; writing—original draft preparation, K.R. and M.E.; writing—review and editing, K.R. and M.E.; visualization, K.R.; supervision, K.R.; project administration, K.R.; funding acquisition, K.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported through the Annual Funding track by the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia [Project No. AN000108].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors extend their appreciation to the Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia, for funding this research work through the project number AN000108.

Conflicts of Interest

The authors declare that there is no conflict of interest.

References

  1. Kim, S.; Razi, A.; Stringhini, G.; Wisniewski, P.J.; De Choudhury, M. A Human-Centered Systematic Literature Review of Cyberbullying Detection Algorithms. Proc. ACM Hum.-Comput. Interact. 2021, 5, 1–34. [Google Scholar]
  2. Hall, D.L.; Silva, Y.N.; Wheeler, B.; Cheng, L.; Baumel, K. Harnessing the Power of Interdisciplinary Research with Psychology-Informed Cyberbullying Detection Models. Int. J. Bullying Prev. 2021, 4, 47–54. [Google Scholar] [CrossRef]
  3. Bharti, S.; Yadav, A.K.; Kumar, M.; Yadav, D. Cyberbullying detection from tweets using deep learning. Kybernetes 2021. [Google Scholar] [CrossRef]
  4. Perera, A.; Fernando, P. Accurate Cyberbullying Detection and Prevention on Social Media. Procedia Comput. Sci. 2021, 181, 605–611. [Google Scholar] [CrossRef]
  5. Arif, M. A Systematic Review of Machine Learning Algorithms in Cyberbullying Detection: Future Directions and Challenges. J. Inf. Secur. Cybercrimes Res. 2021, 4, 1–26. [Google Scholar] [CrossRef]
  6. Cheng, L.; Guo, R.; Silva, Y.N.; Hall, D.; Liu, H. Modeling Temporal Patterns of Cyberbullying Detection with Hierarchical Attention Networks. ACM/IMS Trans. Data Sci. 2021, 2, 1–23. [Google Scholar] [CrossRef]
  7. Kumar, A.; Sachdeva, N. A Bi-GRU with attention and CapsNet hybrid model for cyberbullying detection on social media. World Wide Web 2021, 1–14. [Google Scholar] [CrossRef]
  8. Alotaibi, M.; Alotaibi, B.; Razaque, A. A Multichannel Deep Learning Framework for Cyberbullying Detection on Social Media. Electronics 2021, 10, 2664. [Google Scholar] [CrossRef]
  9. Noviantho, S.I.; Ashianti, L. Cyberbullying classification using text mining. In Proceedings of the 2017 1st International Conference on Informatics and Computational Sciences (ICICoS), Semarang, Indonesia, 15–16 November 2017; pp. 241–246. [Google Scholar]
  10. Azeez, N.A.; Idiakose, S.O.; Onyema, C.J.; Van Der Vyver, C. Cyberbullying Detection in Social Networks: Artificial Intelligence Approach. J. Cyber Secur. Mobil. 2021, 10, 745–774. [Google Scholar] [CrossRef]
  11. Jain, V.; Saxena, A.K.; Senthil, A.; Jain, A.; Jain, A. Cyber-Bullying Detection in Social Media Platform using Machine Learning. In Proceedings of the 2021 10th International Conference on System Modeling & Advancement in Research Trends (SMART), Gold Coast, Australia, 16–18 December 2021; pp. 401–405. [Google Scholar]
  12. Bozyiğit, A.; Utku, S.; Nasibov, E. Cyberbullying detection: Utilizing social media features. Expert Syst. Appl. 2021, 179, 115001. [Google Scholar] [CrossRef]
  13. Kumar, A.; Sachdeva, N. Multimodal cyberbullying detection using capsule network with dynamic routing and deep convolutional neural network. Multimed. Syst. 2021, 1–10. [Google Scholar] [CrossRef]
  14. Rezvani, N.; Beheshti, A. Towards Attention-Based Context-Boosted Cyberbullying Detection in social media. J. Data Intell. 2021, 2, 418–433. [Google Scholar] [CrossRef]
  15. Fang, Y.; Yang, S.; Zhao, B.; Huang, C. Cyberbullying Detection in Social Networks Using Bi-GRU with Self-Attention Mechanism. Information 2021, 12, 171. [Google Scholar] [CrossRef]
  16. Rupesh, C.; Gad, R.; Gawali, P.; Gite, M.; Pawa, A.B. Hate Speech Detection on Social Media Using Machine Learning Algorithms. J. Cogn. Hum.-Comput. Interact. 2022, 2, 56–59. [Google Scholar] [CrossRef]
  17. Zhang, C.; Zhang, F.M.; Li, F.; Wu, H.S. Improved artificial fish swarm algorithm. In Proceedings of the 2014 9th IEEE Conference on Industrial Electronics and Applications, Tianjin, China, 14–16 August 2014; pp. 748–753. [Google Scholar]
  18. Dey, A.; Jenamani, M.; Thakkar, J.J. Lexical TF-IDF: An n-gram feature space for cross-domain classification of sentiment reviews. In Pattern Recognition and Machine Intelligence; Springer: Cham, Switzerland, 2017; pp. 380–386. [Google Scholar]
  19. Zhu, H.; Liu, G.; Zhou, M.; Xie, Y.; Abusorrah, A.; Kang, Q. Optimizing Weighted Extreme Learning Machines for imbalanced classification and application to credit card fraud detection. Neurocomputing 2020, 407, 50–62. [Google Scholar] [CrossRef]
  20. Liu, Z.; Tang, D.; Cai, Y.; Wang, R.; Chen, F. A hybrid method based on ensemble WELM for handling multi class imbalance in cancer microarray data. Neurocomputing 2017, 266, 641–650. [Google Scholar] [CrossRef]
  21. Neshat, M.; Sepidnam, G.; Sargolzaei, M.; Toosi, A.N. Artificial fish swarm algorithm: A survey of the state-of-the-art, hybridization, combinatorial and indicative applications. Artif. Intell. Rev. 2014, 42, 965–997. [Google Scholar] [CrossRef]
Figure 1. Overall process of the proposed model for secure social media applications.
Figure 1. Overall process of the proposed model for secure social media applications.
Applsci 12 06375 g001
Figure 2. The structure of ELM.
Figure 2. The structure of ELM.
Applsci 12 06375 g002
Figure 3. Confusion matrix of the proposed model for different training/testing dataset sizes. Where (a) considers 90:10 training/testing data split, (b) considers 80:20 training/testing data split, (c) considers 70:30 training/testing data split, and (d) considers 60:40 training/testing data split.
Figure 3. Confusion matrix of the proposed model for different training/testing dataset sizes. Where (a) considers 90:10 training/testing data split, (b) considers 80:20 training/testing data split, (c) considers 70:30 training/testing data split, and (d) considers 60:40 training/testing data split.
Applsci 12 06375 g003
Figure 4. Precision–recall analysis of the proposed model with different training/testing dataset sizes. Where (a) considers 90:10 training/testing data split, (b) considers 80:20 training/testing data split, (c) considers 70:30 training/testing data split, and (d) considers 60:40 training/testing data split.
Figure 4. Precision–recall analysis of the proposed model with different training/testing dataset sizes. Where (a) considers 90:10 training/testing data split, (b) considers 80:20 training/testing data split, (c) considers 70:30 training/testing data split, and (d) considers 60:40 training/testing data split.
Applsci 12 06375 g004
Figure 5. The ROC analysis of the proposed model with different training/testing dataset sizes. Where (a) considers 90:10 training/testing data split, (b) considers 80:20 training/testing data split, (c) considers 70:30 training/testing data split, and (d) considers 60:40 training/testing data split.
Figure 5. The ROC analysis of the proposed model with different training/testing dataset sizes. Where (a) considers 90:10 training/testing data split, (b) considers 80:20 training/testing data split, (c) considers 70:30 training/testing data split, and (d) considers 60:40 training/testing data split.
Applsci 12 06375 g005
Figure 6. The A c c y analysis of the proposed model and existing algorithms.
Figure 6. The A c c y analysis of the proposed model and existing algorithms.
Applsci 12 06375 g006
Figure 7. The P r e c n analysis of the AFSA-WELM technique and existing algorithms.
Figure 7. The P r e c n analysis of the AFSA-WELM technique and existing algorithms.
Applsci 12 06375 g007
Figure 8. The R e c a l analysis of the proposed model and existing algorithms.
Figure 8. The R e c a l analysis of the proposed model and existing algorithms.
Applsci 12 06375 g008
Table 1. A comparative analysis of the AFSA-WELM technique and existing algorithms.
Table 1. A comparative analysis of the AFSA-WELM technique and existing algorithms.
MethodsAccuracyPrecisionRecall
SVM (linear algorithm)95.1793.7595.07
SVM (poly-algorithm)98.5597.9498.79
SVM (RBF algorithm)87.2888.5789.67
SVM (sigmoid algorithm)98.2998.4398.50
Naïve Bayes algorithm95.5995.2496.70
The proposed model99.1495.2299.53
Table 2. Results of the proposed model with different measures and training/testing datasets sizes.
Table 2. Results of the proposed model with different measures and training/testing datasets sizes.
Class LabelsAccuracyPrecisionRecallF-Score
Training/Testing (90:10)
Non-Cyberbullying99.14100.0099.0699.53
Cyberbullying99.1490.43100.0094.98
Average99.1495.2299.5397.25
Training/Testing (80:20)
Non-Cyberbullying98.5999.0299.4499.23
Cyberbullying98.5993.5089.0591.22
Average98.5996.2694.2595.23
Training/Testing (70:30)
Non-Cyberbullying95.5096.9898.1497.56
Cyberbullying95.5076.7966.7771.43
Average95.5086.8882.4684.49
Training/Testing (60:40)
Non-Cyberbullying97.5398.3298.9998.65
Cyberbullying97.5388.0781.4684.63
Average97.5393.1990.2291.64
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Riad, K.; Elhoseny, M. A New Model for a Secure Social Media Application. Appl. Sci. 2022, 12, 6375. https://doi.org/10.3390/app12136375

AMA Style

Riad K, Elhoseny M. A New Model for a Secure Social Media Application. Applied Sciences. 2022; 12(13):6375. https://doi.org/10.3390/app12136375

Chicago/Turabian Style

Riad, Khaled, and Mohamed Elhoseny. 2022. "A New Model for a Secure Social Media Application" Applied Sciences 12, no. 13: 6375. https://doi.org/10.3390/app12136375

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop