Next Article in Journal
Revolution in Flexible Wearable Electronics for Temperature and Pressure Monitoring—A Review
Previous Article in Journal
Hierarchical Blockchain-Based Multi-Chaincode Access Control for Securing IoT Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Topic Modeling and Sentiment Analysis of Online Education in the COVID-19 Era Using Social Networks Based Datasets

by
Samer Abdulateef Waheeb
,
Naseer Ahmed Khan
and
Xuequn Shang
*
School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(5), 715; https://doi.org/10.3390/electronics11050715
Submission received: 26 December 2021 / Revised: 20 February 2022 / Accepted: 23 February 2022 / Published: 25 February 2022
(This article belongs to the Special Issue Automated Framework for Access Control Systems)

Abstract

:
Sentiment Analysis (SA) is a technique to study people’s attitudes related to textual data generated from sources like Twitter. This study suggested a powerful and effective technique that can tackle the large contents and can specifically examine the attitudes, sentiments, and fake news of “E-learning”, which is considered a big challenge, as online textual data related to the education sector is considered of great importance. On the other hand, fake news and misinformation related to COVID-19 have confused parents, students, and teachers. An efficient detection approach should be used to gather more precise information in order to identify COVID-19 disinformation. Tweet records (people’s opinions) have gained significant attention worldwide for understanding the behaviors of people’s attitudes. SA of the COVID-19 education sector still does not provide a clear picture of the information available in these tweets, especially if this misinformation and fake news affect the field of E-learning. This study has proposed denoising AutoEncoder to eliminate noise in information, the attentional mechanism for a fusion of features as parts where a fusion of multi-level features and ELM-AE with LSTM is applied for the task of SA classification. Experiments show that our suggested approach obtains a higher F1-score value of 0.945, compared with different state-of-the-art approaches, with various sizes of testing and training datasets. Based on our knowledge, the proposed model can learn from unified features set to obtain good performance, better results than one that can be learned from the subset of features.

1. Introduction

Fake news is part of feedback, opinions, and news, and this information is greatly increasing with each minute on social network platforms, such as Facebook, Instagram, and Twitter [1]. In the last decade, social networks worked like a pipeline for many researchers, especially for text mining [2]. One of the challenging topics nowadays is the textual data related to the Coronavirus (COVID-19) based on the education sector. This study investigates the subjective nature of this textual data by employing specialists that can assess E-learning-related textual data, to determine attitudes, sentiment, and fake news about COVID-19 [3]. In today‘s challenging times, there is a dire need to analyze comments and people’s attitudes based on textual data on those platforms. The global education system was impacted by the COVID-19 pandemic, which resulted in the suspension of traditional educational activities. Moreover, conversion of such huge textual unstructured data into structured data and extracting useful knowledge from such data is also a complex task [4]. This study has proposed an approach that overcomes this limitation, leading to promising results that were achieved based on the devised experiments.
Transforming unstructured textual data related to the education sector to a structured one is a way to examine the attitudes and sentiments towards COVID-19, based on people’s opinions and views. Some pre-processing steps to get the structured textual data are text cleaning, text transformation, and dimension reduction for the selection of meaningful and relevant text. The evaluation quality is the main measure for testing the accuracy of this structured data, with the main objective of examining the fake news of COVID-19. Therefore, evaluation of this data has an important role in determining the effect of the domain of education and providing better feedback for educational institutions to enhance the quality of their services [5].
Based on related works, many studies have focused on the COVID-19 topic, [6] testing the Twitter dataset to increase the community awareness of the COVID-19 epidemic. A study conducted by [7] examined the government, organizations, and individuals’ viewpoints on Twitter during the pandemic. The authors in [8] tested Twitter for European countries as a cross-language platform during the lockdown. The study in [9] examined the social network data for guidance by tracking the effect of “life”-related terms on Twitter. In the light of the above, earlier works have numerous limitations; for instance, most of the research focused on testing the community awareness, viewpoints, and lifestyle, but did not lead us to suggest a novel method to extract useful applicable knowledge and salient information from huge social networks’ textual data. Using huge available online data demands new approaches to process vast information to extract the most vital and relevant information, and to obtain fast and efficient information. A year of data collection will help us to get not just a larger picture of the public debate and concerns over COVID-19 for E-learning, but identify significant shifts in main topics and sentiment during the duration of the epidemic. However, several challenges arise when moving from conventional learning methods to online teaching techniques. The E-learning boundary’s high devices and equipment cost and infrastructure for communication technology availability are limitations, as well as teachers’ and students’ technical knowledge being limited, and for profitable and effective online education, a cultural shift is required [10].
In this study, our approach is based on the idea of “deep learning” for the classification of the text into five classes, such as very positive, positive, neutral, negative, and very negative. A fusion of multi-level features helps us evaluate data on the attitudes and sentiment towards COVID-19 that can be tracked almost immediately, by classifying the textual data into those five classes. Our TSA (Text Sentiment Analysis) approach includes multi-level features, such as phrase frequency, sentence score, sentiment shifter rules, sentiment knowledge, medical concept, word and sentence embedding, linguistic, and statistical knowledge. This study has used these nine features as input to our suggested model. It is pertinent to mention here that these features have not been fully examined with SA in the COVID-19 textual data before. The fusion of these sets of features will overcome the following issues: (i) limit the lexicon to a coverage list of words; (ii) adopt a vector space model for converting each sentence or word to a vector and then get vector representation in a low-dimensional space; (iii) word embedding has a problem with word senses, which implies that this model cannot distinguish synonyms for a word and only provides a single word form representation; (iv) the most difficult task is the context polarity within SA, related to SA of the context, where the previous word polarity changes concerning contexts; (v) the lexicon lacks specific fields in COVID-19; (vi) this research will focus on the subjective corpus and SA issues that deal with COVID-19, for instance, data with a large feature set, polarity shift, data sparsity, the lexicon of a specific domain, multi-classification, and accuracy. The word embedding model also faces important issues, such as disregarding the polarity of sentiment words.
The contributions of our suggested approach can be briefed as follows:
  • This study applied a denoising AutoEncoder and used a combination of an AutoEncoder with an attention mechanism for extracting the features of textual data and for extracting more precise features redisplaying the original information.
  • This research proposed a novel cross-feature fuse method that relies on the attention technique, which can efficiently fuse the features of various levels and afford more efficient and precise data for sentiment classifier tasks.
  • This study used multi-strategies, such as Weighted-Support Vector Machine (W-SVM), Chi-Squared Statistics (CSS), Weighted-Principal Component Analysis (W-PCA), and Value Average Weights (VAW), which help us produce the specific lexicon domain.
  • The deep learning-based model with unified features set has been adopted as a multi-classification model by using multi-level features with a COVID-19 dataset that can help us evaluate the attitudes, sentiment, and fake news towards COVID-19, especially when examining parents’, students’, and teachers’ (as of opinion) effect on the education sector.
This research is organized as follows: the next section summarizes the recent related works. Section 3 presents our suggested approach in detail. Section 4 explains our experimental results. Lastly, we give the conclusion and ideas for future works.

2. Related Work

2.1. Text Sentiment Analysis

TSA-related traditional approaches can be divided into three categories, which are hybrid approaches, lexicon, and machine learning. TSA is one of the dynamic research areas related to NLP, used to identify the orientation of a given set of documents, for instance, very positive, positive, neutral, negative, and very negative categorization. The modern TSA-related works are divided into three groups: aspect-based, sentence-based, and document-based.
The study proposed by [11] examines sentiment analysis of tweets to have a clearer understanding of people’s opinions or sentiments about E-learning during the pandemic. The research conducted experiments applying RF (Random Forest), NB (Naïve Bayes), and KNN (K-Nearest Neighbour). Due to its capacity to function effectively with high-dimensional data, like text categorization, the multi-class classification with the RF (Random Forest) approach has the greatest accuracy of the classification task. Shakhovska et al. [12] proposed a hybrid machine learning based-method, which combined rule-based algorithms, Support Vector Machine, XGBoost, and logistic regression. The Ukrainian language was used for their work and they collected 32,007 pieces of feedback from Google Maps and divided the dataset into testing (20%) and training (80%). One of the drawbacks was that they only used the TF-IDF (Term Frequency Inverse Document Frequency) feature, resulting in an accuracy of 0.88. Cagliero et al. [13] developed a method based on the Latent-space model on a sentence-level dataset, related to hotel reviews consisting of 200,000 reviews. They adopted the Google word representation model for transferring the list of words from high to low dimensions, called BERT (Bidirectional Encoder Representations from Transformers) pre-trained language embeddings model. They applied a clustering K-means algorithm based on cosine distance, using K = 8, based on the cities, to classify the reviews into cities. The studies related to TSA are used by many machine learning methods, such as Naïve Bayes (NB) [14], Support Vector Machine (SVM) [15], Logistic Regression Algorithm (LRA) [16], and Maximum Entropy (ME) [17]. However, all these approaches suffer from high, large unstructured textual datasets, with a varying and large number of terms (textual) [18]. There are some approaches applied based on a set of features, such as Part of Speech (POS), unigrams, bigrams, statistical methods, words, and sentence embedding [19]. Recently, researchers have focused more on techniques using word embeddings in Deep Learning models [20]. Doc2Vec and Word2Vec were applied in [21] for the sentiment analysis task related to medical records. The authors of the study also used the Welsh statistic of the WordNet for the evaluation of unsupervised models for the medical domain. Atodiresei et al. [22] proposed a method to classify fake news and users based on Twitter by applying text sentiment analysis, emoji, entity, and hashtag recognition. Rustam et al. [23] investigated COVID-19 tweets based on sentiment analysis, applying lexicon-based techniques and machine learning. TextBlob was used to annotate the data that was collected from Twitter, and for machine learning methods, TF-IDF and BoW (Bag of Word) features were employed. The ETC (Extra Trees Classifier) models perform best with Textblob and BoW features, according to the results.

2.2. Deep Learning Methods

One of the newest areas of machine learning is “deep learning”, which is implanted using layers of neural networks stacked together, which are considered to be working like a human brain processing model. These hidden layers contain several nodes for each layer, starting from the input layer and ending at the output layer, with all layers including the hidden layers joined in a feedforward manner. A back-propagation algorithm is used in training those neural network layers so that parameters of the layers get tuned after successive iterations. These neural network-inspired deep learning models are used in sentiment analysis so that a better representation can be learned on the textual data, using selected features [24]. Deep learning models have gained important attention in different SA-related NLP tasks and several TSA models have been suggested based on deep learning techniques [25,26].
In the study of [27], BERT was proposed to predict the ICD (International Classification of Diseases) codes, resulting in an F1 value of 0.68. Named Entity Recognition (NER) was modeled using the bidirectional LSTM-RNN [28] model and a transfer learning technique was used for the limited availability of labeled data for Chinese medical records.
Two famous techniques in deep learning, called Recurrent Neural Network (RNN) and Convolutional Neural Network (CNN), have gained considerable attention related to NLP tasks [29]. RNN is applied for sequential data models in a wide range of applications. RNN’s approaches are used for sentiment classification by representing a full sentence of words embedding model. A variant of RNN, known as Long Short-Term Memory (LSTM), is more suitable for handling the sequencing problems, and has a considerable improvement in results compared to basic RNN, especially when applied for TSA where long term dependency is more common and problems of vanishing and exploding gradients could hamper the training of RNN models [30].
On the other hand, CNN is used mainly in image processing, computer vision, and text processing [31]. The architecture of CNN consists of a numerical input that converts the textual terms to a vector that is passed from the input layer to a fully connected layer, where various filters to recognize the patterns based on the number of kernel sizes are used. The pooling layer is used later to choose to minimize the number of parameters in the layers [32]. Deepak et al. [33] developed a deep learning based-approach, with LSTM and a Feed-Forward Neural Network (FNN), in combination with various vector representations of the word. The suggested approach mostly concentrates on gathering data from the news article that relies on the title and content.

2.3. Twitter Dataset Evaluation

Twitter is now a very active and challangineg dataset, as the number of tweets that are posted on Twitter is increasing at a staggering rate; for example, based on later statistics, there are around 300 billion tweets. The most complex networks contain nodes that are mutually interconnected to a larger extent than they are connected to the rest of the network. These numerous amounts of textual information are generated by users and so the extraction of useful information from this huge online text data is gaining attention in the NLP community. Recently, studies have evaluated producers’ quality based on people’s opinions on social media [34]. Some studies focused on COVID-19, based textually on SA related to people’s lifestyle [8,35]. Additionally, the researchers in [36] suggested a semi-supervised learning model for capturing fake applications on Twitter, applying confident valuable features.
Sharif et al. [37] applied deep learning and machine techniques, which contained CNN-BiLSTM, BiLSTM, CCN, and SVM methods, with the Word2Vec embedding and TF-IDF approaches. Furthermore, Sharma et al. [38] suggested a sentiment model that categorizes fake news or wrong information on Twitter related to COVID-19. They captured misleading and unreliable content that relied on checking the websites and examined the misinformation in tweets. Furthermore, the work in [39] proposed a text-based method to detect fake news by applying a two-layer classification. The first one was applied to identify the fake topic and the second one was applied to identify the fake event. Similarly, Nyow et al. [40] proposed an approach to determine the most significant features that affected the machine learning approaches’ performance when applied for the fake news classification and used this information to recognize real and fake tweets. However, the research stated above did not use an evolutionary classification approach to illustrate the most important features of the COVID-19 fake news data. Our proposed method goal is to fuse and/or reduce the number of features and attain better accuracy.
Most recently, with the growing computational power and the possibility to connect with huge text corpora, bottom-up methods applying machine learning or text-mining allow for “deeper” and more complex analyses of communication structures on social media, and such text can then be used to assess, for example, a certain text’s narrative coherence, its cognitive complexity, or determine its synchronicity, similarity or “match” with other texts [41].
Deep learning can efficiently achieve non-explicit and nonlinear user–item relations and, furthermore, extracts complex relations between data from available data sources, such as visual, textual, and background information. The challenge with the Twitter dataset is that it is difficult to obtain good estimation. The most complex networks contain nodes that are mutually interconnected to a larger extent than they are connected to the rest of the network; therefore, there is a pressing need for societies to understand data distributed in complex networks to, between other considerations, make predictions about their behavior. While neural networks have the flexibility to handle a larger number of inputs, this research only considered fake news and misinformation related to COVID-19 tweets. Developing more complex networks with more hidden layers, more inputs, and various networks’ parameters, to enhance neural networks’ performance, warrants further investigation, which leads to learning more useful and complex representations in neural networks. The main goal of this work is to predict fake news and misinformation for the education sector during various periods of the year [42,43].

2.4. Topic Modeling

The method that can help to organize a huge amount of collected documents by categorizing these documents into different themes is called Topic Modeling (TM). TM is sometimes referred to as a possibility of clustering, but TM is more robust and usually affords more realistic outcomes than a clustering algorithm like k-means. The clustering algorithm assumes a distance between subjects is measured, and each document is assigned a topic, whereas TM allocates a document to a set of topics, with varying weights or probability, without making any assumptions about the distance between topics. Several TMs are available, of which the most extensively applied is the latent Dirichlet allocation (LDA) model [44]. In order to extract common topics from such a massive number of tweets, we adopted the LDA algorithm for the TM task.
Understanding the nature of humanity is required for analyzing and modeling social networks as neural networks. If we built our understanding of neural networks based on connected relations and nodes only, then we may end up having no network from a human standpoint. Two factors need to be observed very carefully: (1) individuals have natural needs to receive the information or to be its source, (2) individuals are the information sources. the first point leads to the fact that each person in a network has attributes and those attributes can make changes to, or can be changed by, the social environment. The second point confirms that the flow of information is what produces complex networks, or what we call a “living network” [43,45].
The main goal of this research is identifying the themes and sentiments of COVID-19 related to the education sector, which relies on the complexity of social networks’ discussion and discerns the salient changes in themes and sentiments over time, for better understanding the public emotions, concerns, and perceptions, that may affect E-learning quality.

3. Proposed Method

In this study, a novel approach is proposed for SA to evaluate the data based on the attitudes, sentiments, and fake news towards COVID-19, for online learning extracted from opinions and views on Twitter. This study contains five main steps, namely: a collection of data, text pre-processing, lexicon building for a specific domain, model development based on extreme learning machine and denoising AutoEncoder, and a final step is evaluation and assessment. For evaluation and comparison, our proposed method uses metrics such as F1, precision, recall, and accuracy. This study also compared the proposed model with state-of-the-art approaches from the related works. Figure 1 shows the main architecture for the suggested method.
The data collection is the first step and the first dataset is collected using the famous Twitter API (Application Programming Interface). This study used a query based on terms, (COVID-19, education, online learning, E-learning, teacher, and student) and collected 853,000 unique tweets between the dates of 20 December 2020 to 20 October 2021. The second dataset is the Fake News related to COV (https://data.mendeley.com/datasets/zwfdmp5syg/1, accessed on 25 December 2021) ID-191 entered on 1 June 2021, which was published in [46]. The second step is pre-processing and simplification of the text, and this step contains six sub-steps, such as tokenizing, which means dividing the input document into units with various levels for example paragraphs, sentences, and words, then Word2Vec was used for representing each word into a constant length of the numeric vector [47]. Tokenizing processes the text based on punctuation and spaces to separate the text of the tweet into a list of words. The third step is how to solve the polarity issue by assigning the polarity to the list of sentences, building specific lexicon for the medical domain, adopting our lexicon for more information (see [19]), and applying available lexicon ULMS (Unified Medical Language System), Valence Aware Dictionary for Sentiment Reasoning (VADER), TextBlob, and SentiWordNet to investigate the subjectivity and polarity of text. Step four is to apply the nine features (phrase frequency, sentence score, sentiment shifter rules, sentiment knowledge, medical concept, words and sentence embedding, linguistic, and statistical knowledge), extracting these textual features by using a Denoising AutoEncoder, to avoid the over-fitting issues. Synthetic Minority Oversampling Technique (SMOTE) is applied to produce synthetic data for the minority class in order to balance the dataset [48]. The fusion of these features is achieved by applying an attention mechanism. These lists help us train our suggested model ELM-AE which is a Neural Network (NN) of one hidden layer, where the hidden layer weights are set at random and the output layer weights are calculated by training dataset. The fifth step is to evaluate if the proposed method is appropriate for evaluating the trustworthiness and usefulness of our suggested method to examine qualities such as eliminating redundancy, readability, comprehensibility, and noisy information. These approaches applied the following standard metric as follows:
y j = i = 1 φ β i g w i . x i   + b i ,   j = 1 , 2 , , N ,
where, w i = w i , w i , , w i n T relates to the ith input neurons and hidden neuron, bi presents the ith hidden neuron bias, connecting the output and the hidden neuron is presented by βi. Within matrix form:
y = H β ,
where,
y = y 1 , y 2 , y n T ,
H = g w 1 x 1 + b 1 g w φ x 1 + b φ   g w 1 x 1 + b 1 g w φ x N + b φ N × φ ,
β = β 1 , β 2 , β n T ,
Here, H can be a matrix of non-square so there cannot exist βi, bi, wi, where i = 1, 2, …, N such as y = H β. The last square result of this approach is
β = H   y .
Here, H is the Moore–Penrose popularized matrix H opposite [49].

4. Experiments and Results

An important question regards whether our proposed approach-based ELM-AE-LSTM and attention technique help improve the performance of SA to a vaccination textual dataset obtained from Twitter. Therefore, LSTM uses a one sentence-level node. One way of enriching the parameter space is to add more sentence-level nodes, each communicating with word-level nodes in the same way. The dropping probability for the dropout layers is 0.2. The number of nodes in each LSTM layer is always identical to the number of nodes in the input layer. In the following sections, we will discuss in detail the dataset used in this study, our proposed model components, designed experiments, and the results that our study achieved, based on the proposed model in the Dataset to validate the point that our proposed model is better.

4.1. The Dataset Collecting and Setup

This study applied SMMT (Social Media Mining Toolkit) [50] to listen to the Twitter Stream API, and used the Twitter API for tweets with the described keywords, gathering the raw tweets, and it collected 853,000 unique tweets from 565,371 various users, based on six query words (COVID-19, education, online learning, E-learning, teachers, and students). Such a large number of raw tweets poses challenges, such as noise and redundant information, because the people who use social media are from a wide range of social domains and have inherent diversity, as well as local and political dynamics of the area that affect their views and opinions [51].
For this work, there is also a need for a gold standard dataset, and for the creation of such a dataset, this work used the technique proposed in our previous work [19]. This dataset also contains what is called the ground truth or corrected labels and for this task, seven annotators, including three lecturers and teachers, with good skills in understanding and reading the English language, and also an additional four Ph.D. students in English language linguistics, were asked to provide their services. All annotators worked on the dataset at a sentence level and labelled each of those sentences by using terms, “very positive”, “positive”, “neutral”, “negative”, and “very negative”. The list of sentences was labelled with the associated polarity based on the majority votes of the annotators. Finally, we were able to build a gold standard dataset for evaluating the suggested model at the sentence level, with 60,000 tweets or rows of sentences. The age criteria of the people included in the study were that we only used tweets from people aged 18 or above, so that a relatively mature and reliable textual dataset could be obtained. The numeric sentiment tags for each tweet with polarity are as follows: very positive = +1, positive = 0.5, neutral = 1, negative = −0.5, and very negative = −1. Table 1 shows the dataset used for evaluation.
The second dataset, fake news, contains a gathering of fake and true news, correlated to COVID-19. The news articles were collected based on three query words: “COVID-19”, “coronavirus”, and “corona”. Its more than 6000 articles from various website outlets throughout the social network were included in the original dataset and gathered between December 2019 and July 2020. Additionally, the dataset contained three news subcategories: partially false, false, and true news. The data have been given labels of 1 for true news, 0 for partially false, and −1 for false news.

4.2. Text Pre-Processing

Normalization, which is also called “pre-processing” in the domain of NLP, is considered a crucial first step to clean the text and prepare the text in a standardized format that can be used in the training of machine learning or deep learning-based models for classification. Our strategy of developing the automated machine-learning model is divided into various steps. In the first step, this study removed punctuation, filtered stop words stemming, the lemmatizing, fixed-length vocabulary of size 30,000, and length of sentences between the ranges of 3 to 25 words for each of those documents [52]. The part of Speech (POS) tagging method is also applied for classifying the terms of the textual dataset on the main classes of POS they belong to (verb, noun, adjective, and adverb), and the tagging of POS provides suitable information of lexicon [53]. The yield of this stage is plaintext.

4.3. Lexicon Building

In the next step, we fed our processed dataset to the two embedding pre-trained models to get the numerical representation for each sentence. After that, UMLS, VADER, TextBlob, and SentiWordNet, and our statistical technique, were adopted in the method for building a specific lexicon (medical domain), which helps assign the polarity to each of the sentences (the details of this observation are discussed in [19]). This study also used different sizes of the lexicon (number of terms) for examining the reliability of this approach [54]. The capacity of lexicons was tested (10,000, 20,000, 30,000, 40,000, 50,000, 60,000, and 70,000) and this study examined the various size of lexicons to investigate whether the biggest or smallest lexicon of sentiment can lead to promising results [55]. The smallest lexicon resulted poorly and the biggest lexicon resulted in the highest performance [56].
The sentiments ratio, after using the lexicons to assign the polarity, is not equivalent, so the model can be over-fit on the imbalanced dataset. SMOTE is applied to tackle issues with unbalanced datasets by balancing the sample’s number for all of the dataset’s classes. Balancing is performed by creating synthetic samples of the minority class in such a way that the number of samples to the minority class is almost equal to the number of samples to the majority class. As a result, when the SMOTE approach is used to generate synthetic samples of the minority class, models have a high level of accuracy [57]. As a result, when the SMOTE approach is used to generate artificial samples of the minority class, models have a high level of accuracy. The F-measure is based on the number of terms in each lexicon, (lexicon 1 = 10,000, lexicon 2 = 20,000, lexicon 3 = 30,000, lexicon 4 = 40,000, and so on), see Figure 2. Figure 3 shows the accuracy results using our statistical techniques.
This study has also adopted some equations for assigning the polarity for each term and sentence. These equations help solve the coverage of the limited terms. The first equation is used to calculate the polarity of the word, as discussed in [58]:
W P = P s N s ,
where WP = word Polarity, Ps = positive score, and Ns = negative score. Then the second equation is:
S P = P s N s ,
where SP = sentence Polarity, Ps = positive score, and Ns = negative score. Then the third and final equation is:
S s = P s N s P N .  
where Ss = sentiment score, P and N indicate the number of positive and negative terms, respectively. P s = the summation total number of positive words in the sentence, and N s = the summation total number of negative words in the sentence. For the issue of how to calculate this for other classes, we adopted the rounding theory to the nearest number, for example, +0.9 can equal +1, as discussed in [59].
Notes: Weighted by Chi-squared statistic (CSS), Support Vector Machine (SVM), and Principal Component Analysis (PCA).
The result of this phase on the Y-axis is that there is a list of sentences, and on the X-axis, the dimension-1 of the chosen sentences is displayed based on the vector, see Figure 4.
I P F = S I P i I F d      
where I P F refers to important phrases’ frequency, S I P I refers to sentence number that contains important phrases and I F d refers to the total number of the important frequency in documents.
S L S d = m 1 m
where SL refers to sentence location, S d refers to the d t h sentence in the document, and m is the maximum number of a sentence in document d.
S e n t e n c e   S c o r e = w o r d s   i n   s e n t e n s e   w o r d s   i n   t h e   l o n g e s t   s e n t e n c e    

4.4. Vector Space Model

Deep Neural Network (DNN) uses the continuous bag-of-words method with other strategies that will be described in the following discussion to deal with the different problems in the model. We used the Word2Vec model as pre-trained on famous datasets, such as Wikipedia and Google News, with the embedding dimension of size 200. The main parameters of our model are the word minimal vocabulary frequency, the layer size, the text window size, and the negative numbers were set to 30, 200, 5, respectively. The values of 30 and 5 for the minimal vocabulary frequency and the text window size were chosen after doing experiments with a range of values for both parameters. Solving the issue related to shifter words, parameter values of minimum confidence were set to 0.8, gain theta to 2.0, and Laplace k to 1.0 [19,52,56,60].
Finally, one of the primary purposes of word embedding is to extract the meaningful numeric representation of fixed length, corresponding to the words in a language, but still, substantial noise occurs in the context information of social networks, which affects the feature extraction process. For eliminating noise information intrusion and obtain more robust features, this study applied a Denoising AutoEncoder (DAE) for textual feature extraction. The purpose for applying DAE instead of a vanilla AutoEncoder is that DAE adds noise to the input data in the encoder part and in the decoder part, original data are used for reconstruction. After training the DAE, either the embedding layer or the middle layer was extracted, as this layer corresponds to the lower number of neurons and it is believed that a useful representation of the high number of features is learned in a low dimension with a nonlinear representation [61,62].

4.5. Input Features

The list of multi-level features represents the textual characteristic used to capture data specific patterns. Nine key features (phrase frequency, sentence score, sentiment shifter rules, sentiment knowledge, medical concept, word and sentence embedding, linguistic, and statistical knowledge) were fed to the model for the training of the model for the classification task. The set of features with their details of calculation and working are discussed in detail in [63]. Table 2 shows a description of these nine features, presented, for clarity, to the readers. One of the challenging tasks is to use the fused features for SA on the Twitter dataset, and for this issue, an attention mechanism was adopted based on the textual feature extraction model [64]. Figure 5 shows the details of the fusion features method. The results from the extraction method were used as input to the features fusion method.
Our goal in this study was to construct the most prominent or significant list of features that can select the most relevant sentences (tweets) to increase the context coverage and variety between tweets for final classification. It is pertinent to mention here that the outcome of any classifier model is extremely affected by the list of selected features, as these features play an important role in capturing the non-linear relationship between the classes and the features of the dataset. Table 3 shows an improvement in the performance using a fusion of the approaches. This is based on the number of features applied; for example, Experiment 1 means one feature was used, Experiment 2 means two features were used, and Experiment 9 means the full list of features was used.

4.6. Data-Driven Modeling of Neural Networks

The tweets’ Topic modeling, related to the education sector, produced 12 topics, which were grouped into 4 major themes. We divided these topics based on the hottest topic discussed for each week of our selected period. Sentiments about the education sector (254,781/853,000 tweets, 21.73%) were the maximum tweeted theme and remained a hotly debated issue throughout our investigation. The topics are (1) Advantages of E-learning, (2) Effect of economics, (3) Viewpoints on online learning, (4) Collaboration and assistance on a global scale, (5) Instructions for receiving vaccinations, (6) Disadvantages of E-learning, (7) Immunization administration is progressing, (8) Educating communities, (9) Procedures for controlling teachers and students, (10) Vaccines Functioning, (11) Vaccination authorization, and (12) Fake news. In order to see the weekly frequency of the 12 topics from 20 December 2020 to 20 October 2021, see Figure 6. The overlap between each topic is grouped into four major themes, which are: (a) Emotions and opinions around E-learning, which included topics 1, 2, and 12. (b) Information about vaccines and online learning, which included topics 3, 5, and 6. (c) Vaccines as a universal problem, which included topics 4, 8, and 9. (d) Vaccine administration, development, and authorization, which included topics 7, 10, and 11. We applied LDA for selecting these 12 topics, based on the following reasons: the highest coherence score came with 12 topics; the second one is the TM with 12 topics, which achieves a compromise between being too tight and risking eliminating key issues, as well as being too wide and diluting the core focus. In this study, the weekly proportion of emotions is shown in Figure 7. Based on this figure, we observed that individuals change their opinions according to the changes in the general situation around them, and this depends on the 12 topics that were selected, and also it is possible to observe the overlap and change of opinions weekly.

5. Evaluation and Comparison

The ELM-AE was applied with a dataset divided into training, verification, and testing sets (80%), (10%), and (80%), respectively, according to the 8:1:1 ratio, then we used an optimization grid to reduce computation costs, and number of parameters and running time were applied. The main parameters for our model are epoch, learning rate, batch size, and dropout, and the range values are 35–50, 0.001–0.0001, 32–128, and 0.3–0.9, respectively.
The model consists of an input layer, which contains the neurons’ number that represented the list of features number, Nelm parameter, the hidden layer, which contains several nodes, and these parameters need tuning, composing C and λ. Parameters bias and weight for the hidden layer are produced at random, as discussed in [65]. We also suggested an optimized parameters grid search, which has the ability to tune the key parameters for the classifier model in the sub-process for getting the ideal parameters. An equation was used to improve the approach, resulting in the highest F1 score when compared with traditional and state-of-the-art methods [66]. In detail, the results are shown in Figure 8. The running time was also examined for calculating the efficiency for each approach, illustrating the results in Figure 9.
O u r   a p p r o a c h O t h e r   a p p r o a c h   O t h e r   a p p r o a c h . 100
This study is aimed at measuring the attitudes, sentiments, and fake news towards COVID-19, based on the education sector, which can be tracked almost instantaneously, of two different keywords that are used in this epidemic (COVID-19 and E-learning). The patterns of connectivity are required between very positive, positive, neutral, negative, and very negative labels, related to people’s opinions, especially when the people try to evaluate the learning techniques (detail is presented in Figure 10). The second dataset experimented on was based on our proposed model and comparisons with other classification approaches. This experiment will help researchers and other people easily detect between fake and true news, see Table 4.
In the second experiment, we applied precision, recall, F-measure, and accuracy as evaluation measures. A confusion matrix was used to calculate these measures, where:
  • The false negatives (FN) represented the data that were forecasted as not fake but were fake.
  • The false positives (FP) were the data that were forecasted as fake but were non-fake.
  • The true negatives (TN) represented the data that were forecasted as not fake and were non-fake.
  • The true positives (TP) were the data that were forecasted as fake and were fake.
The following metrics were used:
Precision, which is the percentage of news that is successfully detected as fake out of all fake (good) news, is shown as follows, in Equation (14).
Precision = T P F P + T P ,
Recall, which is the model sensitivity, means how the technique can recognize the positive examples (phony) of news, and can be calculated according to Equation (15).
Recall = T P F N + T P ,
To calculate the F-measure by using the weighted average of the precision and recall measures, we provide Equation (16).
F - measure = 2 × Precision × Recall FPrecision + Recall ,
The ratio of properly diagnosed non-fake and fake news incidents overall correctly and incorrectly categorized cases is known as accuracy. Equation (17) is the accuracy equation.
Accuracy   = T P + T N T P + T N + F B + F N .

6. Discussion

This research examined a long period of information based on sentiments and topics, by focusing on the debates in the education sector during the COVID-19 pandemic on Twitter, from 20 December 2020 to 20 October 2021, when all the education institutes transferred from conventional learning to the online teaching system.
Our proposed, semi-supervised technique uses a Word2Vec pre-trained model and LSTM, with attention making them into a hybrid model, ELM-LSTM-EA. We evaluated our model on the Twitter datasets related to vaccines, to measure the data on the attitudes and sentiment towards COVID-19 for E-learning. The dataset created was extracted from the Twitter text corpus, with COVID-19, education, online learning, E-learning, teachers, and students as keywords. A gold standard dataset based on tweets was also constructed, with very positive, positive, neutral, negative, and very negative as labels. Using these labels, we classified each of the documents into one of the categories corresponding to the label classes. The fusion of multi-level features related to the online education text domain, which includes the sentence score, sentiment shifter rules, sentiment knowledge, medical concept, word and sentence embedding, linguistic, and statistical knowledge.
Promising results based on the proposed model, when compared with state-of-the-art methods in the literature, using the metrics of F1, recall, precision, and accuracy were achieved. Equation 10 was used to examine the improvement in percentage when compared (SA-CD19-EL) with other methods; the results are +35.8, +33.6, +32.4, +32.5, +26.5, and +21.4, SVM, Naïve Bayes, Logistic Regression, Bi-LSTM, RNN-LSTM, CNN-LSTM, and LSTM, respectively. The proposed model also outperformed other methods using different testing and training ratios, and with two keywords, COVID-19 and E-learning; the accuracies based on data set evaluation are 0.878 and 0.945, respectively.
ELM was used in this study. Its inherent capability to model complex phenomena and fast training, and the combination of the weighted method and feature selection that was applied in our study, has resulted in better accuracy in the case of the imbalanced dataset because traditional classifiers are biased towards the majority class and are not useful in the case of imbalanced datasets. The chosen features are more compatible and robust with the ELM configuration, as seen in Figure 8.
The computational complexity of generating the threshold for our model is determined based on how the list of sentences was transformed into a list of features, which was then applied to select testing and training segments. If there are 100 sentences in each vaccine category, the computational complexity, a function of the total sentence number in all the vaccines, is a factor of 100, but if we extract the gist of each sentence by using a few sentences based on a list of features, then the model complexity, as well as the density of the resultant classifier’s accuracy, is increased to a greater extent when compared with the gold standard dataset.
Furthermore, this paper applied a pre-trained Word2Vec model, extended the list of features, examined five classes, and tested different sizes of lexicons. The proposed approach is faster than other existing approaches, as explained in the time complexity section. This led us to apply our proposed model for experiments in other available datasets to test the attitudes, sentiment, and fake news towards COVID-19, within an online teaching system based on assigning labels as very positive, positive, neutral, negative, and very negative.
We conclude that applying a vector space model increases the number of robust features, and using the lexicon and applying novel statistical techniques to build lexicon in a specific domain produced better results than existing approaches, even when we applied the same classifier model compared with our previous work.

7. Conclusions and Future Work

In this study, we aimed to identify the sentiments and major topics about E-learning relating to the COVID-19 pandemic that have been raised on social networks, namely Twitter. The E-learning with COVID-19 discussion on Twitter was mainly driven by online teaching and has been the subject of numerous news stories recently and reflected the hot new themes across the majority of social media. Natural Language Processing is a sub-category of machine learning problems, which is considered challenging, with no gold standard algorithm that can be used efficiently for all NLP problems. To evaluate the data on the attitudes, sentiments, and fake news towards the COVID-19 textual dataset, we developed a novel method using the fusion of both unsupervised (association rules, Word2Vec, and SV-BOW), denoising AutoEncoder and semi-supervised (extreme machine learning AutoEncoder), with an attention mechanism. The phrase frequency, sentence score, sentiment shifter rules, sentiment knowledge, medical concept, word and sentence embedding, linguistic, and statistical knowledge, were applied as a set of features, which resulted in improved accuracy. Additionally, from the lexicon, we have identified the most relevant set of features related to our dataset. Our proposed approach obtained good performance, in terms of F1 measure, precision, recall, and accuracy, when compared with traditional state-of-the-art methods. Our approach outperformed other approaches based on the utilization of an SA technique, which was used to examine the attitudes, sentiments, and fake news towards COVID-19 that can be tracked almost instantaneously as a subjective corpus. The results show that combining data balance with SMOTE improves classification accuracy. Moreover, the ELM-AE as a classifier method, with five classes (very positive, positive, neutral, negative, and very negative) has a greater capability for exploring an aspect of feature set learning and has a reduced time complexity, which is encouraging for suggesting this approach as a promising machine learning method for complex data.
For future works, we propose more effective feature extraction approaches and feature fusion, for providing more efficient and relevant information using sentiment analysis. Our future objective will be to enhance the current model and examine further modalities, including images, videos, and audio, as a multimodal model.

Author Contributions

X.S. conceived the idea. S.A.W. proposed the model, architecture, and devised implementing a strategy for the proposed model. N.A.K. helped in editing the manuscript and the implementation of the methodology. All the authors checked the overall progress of the methodology, results, and suggested edits. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. U1811262).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We are very grateful to the Chinese Scholarship Council (CSC) for providing us with financial and moral support.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Spieler, B.; Ballard, D.H.; Mazaheri, P.; Legro, N.; Catanzano, T.; Dey, C.; Prejean, E.; Fontentot, J.; Martin, M.D.; Danrad, R. Social media in radiology: Overview and usefulness of online professional# SoMe profiles. Acad. Radiol. 2021, 28, 526–539. [Google Scholar] [PubMed]
  2. Glowacki, E.M.; Wilcox, G.B.; Glowacki, J.B. Identifying #addiction concerns on twitter during the COVID-19 pandemic: A text mining analysis. Subst. Abus. 2021, 42, 39–46. [Google Scholar] [CrossRef] [PubMed]
  3. Nemes, L.; Kiss, A. Social media sentiment analysis based on COVID-19. J. Inf. Telecommun. 2021, 5, 1–15. [Google Scholar] [CrossRef]
  4. Al-Rawi, A.; Grepin, K.; Li, X.; Morgan, R.; Wenham, C.; Smith, J. Investigating public discourses around gender and COVID-19: A social media analysis of Twitter data. J. Healthc. Inform. Res. 2021, 5, 249–269. [Google Scholar] [CrossRef] [PubMed]
  5. Bhagat, K.K.; Mishra, S.; Dixit, A.; Chang, C.-Y. Public Opinions about Online Learning during COVID-19: A Sentiment Analysis Approach. Sustainability 2021, 13, 3346. [Google Scholar] [CrossRef]
  6. Boon-Itt, S.; Skunkan, Y. Public perception of the COVID-19 pandemic on Twitter: Sentiment analysis and topic modeling study. Health Surveill. 2020, 6, e21978. [Google Scholar] [CrossRef]
  7. Pokharel, B.P. Twitter Sentiment Analysis during COVID-19 Outbreak in Nepal; SSRN: Amsterdam, The Netherlands, 2020. [Google Scholar]
  8. Kruspe, A.; Häberle, M.; Kuhn, I.; Zhu, X.X. Cross-language sentiment analysis of European Twitter messages duringthe COVID-19 pandemic. arXiv 2020, arXiv:2008.12172 v1. [Google Scholar]
  9. Chakraborty, K.; Bhatia, S.; Bhattacharyya, S.; Platos, J.; Bag, R.; Hassanien, A.E. Sentiment Analysis of COVID-19 tweets by Deep Learning Classifiers—A study to show how popularity is affecting accuracy in social media. Appl. Soft Comput. 2020, 97, 106754. [Google Scholar] [CrossRef]
  10. Mpungose, C.B.; Communications, S.S. Emergent transition from face-to-face to online learning in a South African University in the context of the Coronavirus pandemic. Humanit. Soc. Sci. Commun. 2020, 7, 113. [Google Scholar] [CrossRef]
  11. Althagafi, A.; Althobaiti, G.; Alhakami, H.; Alsubait, T. Arabic Tweets Sentiment Analysis about Online Learning during COVID-19 in Saudi Arabia. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 620–625. [Google Scholar] [CrossRef]
  12. Shakhovska, K.; Shakhovska, N.; Veselý, P. The Sentiment Analysis Model of Services Providers’ Feedback. Electronics 2020, 9, 1922. [Google Scholar] [CrossRef]
  13. Cagliero, L.; La Quatra, M.; Apiletti, D. From Hotel Reviews to City Similarities: A Unified Latent-Space Model. Electronics 2020, 9, 197. [Google Scholar] [CrossRef] [Green Version]
  14. Adilah, M.T.; Supendar, H.; Ningsih, R.; Muryani, S.; Solecha, K. Sentiment Analysis of Online Transportation Service using the Naïve Bayes Methods. J. Phys. Conf. Ser. 2020, 1641, 012093. [Google Scholar]
  15. Han, K.-X.; Chien, W.; Chiu, C.-C.; Cheng, Y.-T. Application of Support Vector Machine (SVM) in the Sentiment Analysis of Twitter DataSet. Appl. Sci. 2020, 10, 1125. [Google Scholar] [CrossRef] [Green Version]
  16. Saleena, N. An ensemble classification system for twitter sentiment analysis. Procedia Comput. Sci. 2018, 132, 937–946. [Google Scholar]
  17. Htet, H.; Khaing, S.S.; Myint, Y.Y. Tweets sentiment analysis for healthcare on big data processing and IoT architecture using maximum entropy classifier. In Proceedings of the International Conference on Big Data Analysis and Deep Learning Applications, Miyazaki, Japan, 14–15 May 2018; pp. 28–38. [Google Scholar]
  18. Wang, Y.; Kim, K.; Lee, B.; Youn, H.Y.; Sciences, I. Word clustering based on POS feature for efficient twitter sentiment analysis. Hum.-Cent. Comput. Inf. Sci. 2018, 8, 17. [Google Scholar] [CrossRef]
  19. Waheeb, S.A.; Khan, N.A.; Chen, B.; Shang, X.J. Machine Learning Based Sentiment Text Classification for Evaluating Treatment Quality of Discharge Summary. Information 2020, 11, 281. [Google Scholar] [CrossRef]
  20. Jianqiang, Z.; Xiaolin, G.; Xuejun, Z. Deep convolution neural networks for twitter sentiment analysis. IEEE Access 2018, 6, 23253–23260. [Google Scholar] [CrossRef]
  21. Chen, Q.; Sokolova, M. Specialists, Scientists, and Sentiments: Word2Vec and Doc2Vec in Analysis of Scientific and Medical Texts. SN Comput. Sci. 2021, 2, 414. [Google Scholar] [CrossRef]
  22. Atodiresei, C.-S.; Tănăselea, A.; Iftene, A. Identifying fake news and fake users on Twitter. Procedia Comput. Sci. 2018, 126, 451–461. [Google Scholar] [CrossRef]
  23. Rustam, F.; Khalid, M.; Aslam, W.; Rupapara, V.; Mehmood, A.; Choi, G.S. A performance comparison of supervised machine learning models for Covid-19 tweets sentiment analysis. PLoS ONE 2021, 16, e0245909. [Google Scholar] [CrossRef] [PubMed]
  24. Day, M.-Y.; Lin, Y.-D. Deep learning for sentiment analysis on google play consumer review. In Proceedings of the 2017 IEEE International Conference on Information Reuse and Integration (IRI), San Diego, CA, USA, 4–6 August 2017; pp. 382–388. [Google Scholar]
  25. Alharbi, A.S.M.; de Doncker, E.J. Twitter sentiment analysis with a deep neural network: An enhanced approach using user behavioral information. Cogn. Syst. Res. 2019, 54, 50–61. [Google Scholar] [CrossRef]
  26. Lim, W.L.; Ho, C.C.; Ting, C.-Y. Tweet sentiment analysis using deep learning with nearby locations as features. In Computational Science and Technology; Springer: Berlin/Heidelberg, Germany, 2020; pp. 291–299. [Google Scholar]
  27. Heo, T.-S.; Yoo, Y.; Park, Y.; Jo, B.-C. Medical Code Prediction from Discharge Summary: Document to Sequence BERT using Sequence Attention. In Proceedings of the 20th IEEE International Conference on Machine Learning and Applications (ICMLA), Pasadena, CA, USA, 13–16 December 2021. [Google Scholar]
  28. Dong, X.; Chowdhury, S.; Qian, L.; Li, X.; Guan, Y.; Yang, J.; Yu, Q. Deep learning for named entity recognition on Chinese electronic medical records: Combining deep transfer learning with multitask bi-directional LSTM RNN. PLoS ONE 2019, 14, e0216046. [Google Scholar] [CrossRef] [PubMed]
  29. Habimana, O.; Li, Y.; Li, R.; Gu, X.; Yu, G. Sentiment analysis using deep learning approaches: An overview. Sci. China Inf. Sci. 2020, 63, 111102. [Google Scholar] [CrossRef] [Green Version]
  30. Yadav, A.; Vishwakarma, D.K. Sentiment analysis using deep learning architectures: A review. Artif. Intell. Rev. 2020, 53, 4335–4385. [Google Scholar] [CrossRef]
  31. Basiri, M.E.; Nemati, S.; Abdar, M.; Cambria, E.; Acharya, U.R. ABCDM: An attention-based bidirectional CNN-RNN deep model for sentiment analysis. Future Gener. Comput. Syst. 2020, 115, 279–294. [Google Scholar] [CrossRef]
  32. Ombabi, A.H.; Ouarda, W.; Alimi, A.M. Deep learning CNN–LSTM framework for Arabic sentiment analysis using textual information shared in social networks. Soc. Netw. Anal. Min. 2020, 10, 53. [Google Scholar] [CrossRef]
  33. Deepak, S.; Chitturi, B. Deep neural approach to Fake-News identification. Procedia Comput. Sci. 2020, 167, 2236–2243. [Google Scholar]
  34. Giachanou, A.; Crestani, F. Like it or not: A survey of twitter sentiment analysis methods. ACM Comput. Surv. (CSUR) 2016, 49, 1–41. [Google Scholar] [CrossRef]
  35. Xue, J.; Chen, J.; Chen, C.; Zheng, C.; Li, S.; Zhu, T. Public discourse and sentiment during the COVID 19 pandemic: Using Latent Dirichlet Allocation for topic modeling on Twitter. PLoS ONE 2020, 15, e0239441. [Google Scholar] [CrossRef] [PubMed]
  36. Balaanand, M.; Karthikeyan, N.; Karthik, S.; Varatharajan, R.; Manogaran, G.; Sivaparthipan, C.B. An enhanced graph-based semi-supervised learning algorithm to detect fake users on Twitter. J. Supercomput. 2019, 75, 6085–6105. [Google Scholar] [CrossRef]
  37. Sharif, O.; Hossain, E.; Hoque, M.M. Combating hostility: COVID-19 fake news and hostile post detection in social media. arXiv 2021, arXiv:2101.03291. [Google Scholar]
  38. Sharma, K.; Seo, S.; Meng, C.; Rambhatla, S.; Liu, Y. COVID-19 on social media: Analyzing misinformation in twitter conversations. arXiv 2020, arXiv:2003.12309. [Google Scholar]
  39. Zhang, C.; Gupta, A.; Kauten, C.; Deokar, A.V.; Qin, X. Detecting fake news for reducing misinformation risks using analytics approaches. Eur. J. Oper. Res. 2019, 279, 1036–1052. [Google Scholar] [CrossRef]
  40. Nyow, N.X.; Chua, H.N. Detecting fake news with tweets’ properties. In Proceedings of the 2019 IEEE Conference on Application, Information and Network Security (AINS), Pulau Pinang, Malaysia, 19–21 November 2019; pp. 24–29. [Google Scholar]
  41. Yu, M.; Huang, Q.; Qin, H.; Scheele, C.; Yang, C. Deep learning for real-time social media text classification for situation awareness–using Hurricanes Sandy, Harvey, and Irma as case studies. Int. J. Digit. Earth 2019, 12, 1230–1247. [Google Scholar] [CrossRef]
  42. Al-Tarawneh, A.; Al-Saraireh, J.A.; Systems, F. Efficient detection of hacker community based on twitter data using complex networks and machine learning algorithm. J. Intell. Fuzzy Syst. 2021, 40, 12321–12337. [Google Scholar] [CrossRef]
  43. Villalba-Diez, J.; Molina, M.; Schmidt, D. Geometric Deep Lean Learning: Evaluation Using a Twitter Social Network. Appl. Sci. 2021, 11, 6777. [Google Scholar] [CrossRef]
  44. Blum, A.; Hopcroft, J.; Kannan, R. Foundations of Data Science; Cambridge University Press: Cambridge, UK, 2020. [Google Scholar]
  45. de Oliveira, J.F.; Marques-Neto, H.T.; Karsai, M. Measuring the effects of repeated and diversified influence mechanism for information adoption on Twitter. Soc. Netw. Anal. Min. 2022, 12, 16. [Google Scholar] [CrossRef]
  46. Koirala, A. COVID-19 Fake News Classification with Deep Learning. Master’s Thesis, Asian Institute of Technology, Bangkok, Thailand, 2020. [Google Scholar]
  47. Cai, M.; Luo, H.; Cui, Y. A Study on the Topic-Sentiment Evolution and Diffusion in Time Series of Public Opinion Derived from Emergencies. Complexity 2021, 2021, 23. [Google Scholar] [CrossRef]
  48. Mujahid, M.; Lee, E.; Rustam, F.; Washington, P.B.; Ullah, S.; Reshi, A.A.; Ashraf, I. Sentiment Analysis and Topic Modeling on Tweets about Online Education during COVID-19. Appl. Sci. 2021, 11, 8438. [Google Scholar] [CrossRef]
  49. Huang, G.; Huang, G.-B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef]
  50. Tekumalla, R.; Banda, J.M. Social Media Mining Toolkit (SMMT). Genom. Inform. 2020, 18, e16. [Google Scholar] [CrossRef] [PubMed]
  51. Garcia, K.; Berton, L. Topic detection and sentiment analysis in Twitter content related to COVID-19 from Brazil and the USA. Appl. Soft Comput. 2021, 101, 107057. [Google Scholar] [CrossRef] [PubMed]
  52. Waheeb, S.A.; Husni, H. Multi-document arabic summarization using text clustering to reduce redundancy. Int. J. Adv. Sci. Technol. 2014, 2, 194–199. [Google Scholar]
  53. AlKhwiter, W.; Al-Twairesh, N. Part-of-speech tagging for Arabic tweets using CRF and Bi-LSTM. Comput. Speech Lang. 2021, 65, 101138. [Google Scholar] [CrossRef]
  54. Zaghir, J.; Rodrigues, J.F., Jr.; Goeuriot, L.; Amer-Yahia, S. Real-world Patient Trajectory Prediction from Clinical Notes Using Artificial Neural Networks and UMLS-Based Extraction of Concepts. J. Healthc. Inform. Res. 2021, 5, 474–496. [Google Scholar] [CrossRef]
  55. Bravo-Marquez, F.; Khanchandani, A.; Pfahringer, B. Incremental Word Vectors for Time-Evolving Sentiment Lexicon Induction. Cogn. Comput. 2021, 14, 425–441. [Google Scholar] [CrossRef]
  56. Waheeb, S.A.; Khan, N.A.; Shang, X. An efficient sentiment analysis based deep learning classification model to evaluate treatment quality. Malays. J. Comput. Sci. 2022, 35, 1–20. [Google Scholar]
  57. Rupapara, V.; Rustam, F.; Shahzad, H.F.; Mehmood, A.; Ashraf, I.; Choi, G.S. Impact of SMOTE on Imbalanced Text Features for Toxic Comments Classification using RVVC Model. IEEE Access 2021, 9, 78621–78634. [Google Scholar] [CrossRef]
  58. Gomes, B.G.; Murai, F.; Goussevskaia, O.; da Silva, A.P.C. Sequence-Based Word Embeddings for Effective Text Classification. In Proceedings of the International Conference on Applications of Natural Language to Information Systems, Saarbrücken, Germany, 23–25 June 2021; pp. 135–146. [Google Scholar]
  59. Mukherjee, S. Sentiment analysis. In ML. NET Revealed; Springer: Berlin/Heidelberg, Germany, 2021; pp. 113–127. [Google Scholar]
  60. Waheeb, S.A.; Khan, N.A.; Chen, B.; Shang, X. Multidocument Arabic Text Summarization Based on Clustering and Word2Vec to Reduce Redundancy. Information 2020, 11, 59. [Google Scholar]
  61. Aziguli, W.; Zhang, Y.; Xie, Y.; Zhang, D.; Luo, X.; Li, C.; Zhang, Y.J.S.P. A robust text classifier based on denoising deep neural network in the analysis of big data. Sci. Program. 2017, 2017. [Google Scholar] [CrossRef] [Green Version]
  62. Zhang, K.; Lian, Z.; Li, J.; Li, H.; Hu, X. Short Text Clustering with a Deep Multi-embedded Self-supervised Model. In Proceedings of the International Conference on Artificial Neural Networks, Bratislava, Slovakia, 14–17 September 2021; pp. 150–161. [Google Scholar]
  63. Cayèré, C.; Sallaberry, C.; Faucher, C.; Bessagnet, M.-N.; Roose, P.; Masson, M.; Richard, J. Multi-Level and Multiple Aspect Semantic Trajectory Model: Application to the Tourism Domain. Int. J. Geo-Inf. 2021, 10, 592. [Google Scholar] [CrossRef]
  64. Li, W.; Liu, P.; Zhang, Q.; Liu, W. An improved approach for text sentiment classification based on a deep neural network via a sentiment attention mechanism. Future Internet 2019, 11, 96. [Google Scholar] [CrossRef] [Green Version]
  65. Yao, L.; Ge, Z. Deep learning of semisupervised process data with hierarchical extreme learning machine and soft sensor application. IEEE Trans. Ind. Electron. 2017, 65, 1490–1498. [Google Scholar] [CrossRef]
  66. Sun, C.; Richard, S.; Miyoshi, T. Agent-based model and data assimilation: Analysis of COVID-19 in Tokyo. arXiv 2021, arXiv:2109.00258. [Google Scholar]
Figure 1. Steps of the suggested approach.
Figure 1. Steps of the suggested approach.
Electronics 11 00715 g001
Figure 2. Illustration of the size and performance of each sentiment lexicon.
Figure 2. Illustration of the size and performance of each sentiment lexicon.
Electronics 11 00715 g002
Figure 3. The F1 score results, when applied in three weighing approaches with seven various cases.
Figure 3. The F1 score results, when applied in three weighing approaches with seven various cases.
Electronics 11 00715 g003
Figure 4. The list of sentence mapping by vectors.
Figure 4. The list of sentence mapping by vectors.
Electronics 11 00715 g004
Figure 5. Shows the Attention mechanism method steps.
Figure 5. Shows the Attention mechanism method steps.
Electronics 11 00715 g005
Figure 6. The frequency of each topic on Twitter weekly. Notes: The number from 1 to 12 represent the topics.
Figure 6. The frequency of each topic on Twitter weekly. Notes: The number from 1 to 12 represent the topics.
Electronics 11 00715 g006
Figure 7. The weekly changes of emotional terms are based on 12 topics.
Figure 7. The weekly changes of emotional terms are based on 12 topics.
Electronics 11 00715 g007
Figure 8. Performance statistics based on the comparison between (SA-CD19-EL) and other methods, using Precision (P), Recall (R), and F1 measures (F1).
Figure 8. Performance statistics based on the comparison between (SA-CD19-EL) and other methods, using Precision (P), Recall (R), and F1 measures (F1).
Electronics 11 00715 g008
Figure 9. Illustrated time complexity for each approach.
Figure 9. Illustrated time complexity for each approach.
Electronics 11 00715 g009
Figure 10. The percentages of the polarity (very positive, positive, neutral, negative, and very negative) labels are connectivity based on vaccines, and each color represents a label.
Figure 10. The percentages of the polarity (very positive, positive, neutral, negative, and very negative) labels are connectivity based on vaccines, and each color represents a label.
Electronics 11 00715 g010
Table 1. Gold standard corpus statistics are based on the names of vaccines.
Table 1. Gold standard corpus statistics are based on the names of vaccines.
KeywordsVery PositivePositiveNeutralNegativeVery NegativeTotal
COVID-192500300020001000150010,000
Education1000200030002000200010,000
Online learning5000100020001000100010,000
E-learning50050020003000400010,000
Teachers25025010004000450010,000
Students2502507501250750010,000
Table 2. Presented list of features.
Table 2. Presented list of features.
FeaturesLevelCategoryContributionDescription
X1Sentences embeddingSentencesStatisticalSentence relationThe sentence vectors
X2Words embeddingWordsStatisticalWord relationA list of vectors terms
X3Sentiment shifter Words & sentencesSemanticVariety & CoverageThere is some word that can change the polarity sentence like but
X4medical conceptWords & sentencesSemanticVariety & CoverageProvide the medical terms that can help for building lexicon
X5phrase frequencyWords & sentencesStatisticalVariety & CoverageSelect the most important terms
X6sentiment knowledgeWords & sentencesSemanticVariety & CoverageApply the specific lexicon domain
X7LinguisticWords & sentencesSemanticVariety & CoverageIt is more related to the semantic relation
X8StatisticalWords & sentencesStatisticalVariety & CoverageIt is more related to numerical relation
X9Sentence scoreSentencesStatisticalVariety & CoverageSelect high score sentence for final classifier
Table 3. The outcomes obtained from the experimentations carried out the feature list.
Table 3. The outcomes obtained from the experimentations carried out the feature list.
Our ApproachX1X2X3X4X5X6X7X8X9PrecisionRecallF1-Measure
Experiment 9+++++++++0.98350. 91120.9459
Experiment 8++++++++ 0.85110.93190.889
Experiment 7+++++++ 0.86760.65720.7478
Experiment 6++++++ 0.81770.61940.7048
Experiment 5+++++ 0.84030.63680.7246
Experiment 4++++ 0.78680.59970.6807
Experiment 3+++ 0.78250.60030.6795
Experiment 2++ 0.74680.57320.6487
Experiment 1+ 0.74680.57350.6487
Example: X1 = Feature1, and so on.
Table 4. The final results compared between (SA-CD19-EL) and other methods for the fake news dataset.
Table 4. The final results compared between (SA-CD19-EL) and other methods for the fake news dataset.
ApproachesPrecisionRecallF-MeasureAccuracy
SVM0.55770.60390.57990.7348
Naïve Bayes0.69570.54080.60850.6963
MultiSentiNet0.56330.66220.60880.7543
RNN-LSTM0.59260.61110.60170.7338
CNN-LSTM0.59450.60950.60190.7332
CNN-Multi 0.60300.61940.61110.6639
Our Method
(SA-CD19-EL)
0.86750.65710.74780.7632
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Waheeb, S.A.; Khan, N.A.; Shang, X. Topic Modeling and Sentiment Analysis of Online Education in the COVID-19 Era Using Social Networks Based Datasets. Electronics 2022, 11, 715. https://doi.org/10.3390/electronics11050715

AMA Style

Waheeb SA, Khan NA, Shang X. Topic Modeling and Sentiment Analysis of Online Education in the COVID-19 Era Using Social Networks Based Datasets. Electronics. 2022; 11(5):715. https://doi.org/10.3390/electronics11050715

Chicago/Turabian Style

Waheeb, Samer Abdulateef, Naseer Ahmed Khan, and Xuequn Shang. 2022. "Topic Modeling and Sentiment Analysis of Online Education in the COVID-19 Era Using Social Networks Based Datasets" Electronics 11, no. 5: 715. https://doi.org/10.3390/electronics11050715

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop