You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

21 November 2023

Sentiment Analysis of Arabic Course Reviews of a Saudi University Using Support Vector Machine

,
,
,
and
1
Information Systems, College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
2
SMART Lab, ISG Tunis, University of Tunis, Le Bardo, Tunis 2000, Tunisia
3
College of Information Technology, Kingdom University, Riffa 40434, Bahrain
*
Author to whom correspondence should be addressed.
This article belongs to the Section Computing and Artificial Intelligence

Abstract

This study presents the development of a sentimental analysis system for high education students using Arabic text. There is a gap in the literature concerning understanding the perceptions and opinions of students in Saudi Arabia Universities regarding their education beyond COVID-19. The proposed SVM Sentimental Analysis for Arabic Students’ Course Reviews (SVM-SAA-SCR) algorithm is a general framework that involves collecting student reviews, preprocessing them, and using a machine learning model to classify them as positive, negative, or neutral. The suggested technique for preprocessing and classifying reviews includes steps such as collecting data, removing irrelevant information, tokenizing, removing stop words, stemming or lemmatization, and using pre-trained sentiment analysis models. The classifier is trained using the SVM algorithm and performance is evaluated using metrics such as accuracy, precision, and recall. Fine-tuning is done by adjusting parameters such as kernel type and regularization strength to optimize performance. A real dataset provided by the deanship of quality at Prince Sattam bin Abdulaziz University (PSAU) is used and contains students’ opinions on various aspects of their education. We also compared our algorithm with CAMeLBERT, a state-of-the-art Dialectal Arabic model. Our findings show that while the CAMeLBERT model classified 70.48% of the reviews as positive, our algorithm classified 69.62% as positive which proves the efficiency of the suggested SVM-SAA-SCR. The results of the proposed model provide valuable insights into the challenges and obstacles faced by Arab Universities post-COVID-19 and can help to improve their educational experience.

1. Introduction

The significance of continuous assessment in educational institutions is paramount. It is instrumental for gauging student needs and perspectives, as well as for assessing the efficacy of educational programs. Accreditation agencies, acknowledging this, have increasingly emphasized outcome-based assessments to align educational programs with institutional goals and learning objectives. Additionally, the evolving job market and industry demands necessitate perpetual enhancement of educational programs to ensure students are well-prepared for professional challenges. Historically, one of the most informative methods to assess student learning experiences has been through the collection and analysis of feedback. Feedback serves as a mirror reflecting past behaviors and as a predictor for future improvements [1]. Particularly in education, feedback is critical for comprehending student engagement with curriculum and pedagogical strategies [2]. When students provide formative feedback, they offer essential insights into the course’s strengths, the obstacles they encounter, and the effectiveness of teaching methods. Surveys have traditionally been a primary method for collecting student feedback. They can be deployed anonymously, encouraging candidness, and can range from rating-based questions to open-ended textual responses. While rating-based surveys yield quantitative data, they often lack the depth of understanding provided by textual feedback [3]. Recognizing the limitations of conventional feedback analysis, we propose a novel approach using machine learning (ML) algorithms to analyze student feedback obtained from an online survey platform. Current research indicates several methods for sentiment analysis: rule-based, lexicon-based, and machine learning-based [4]. We adopt a machine learning-based approach, acknowledging its growing importance in extracting sentiment from textual data [5]. The Support Vector Machine (SVM) method is selected for its effectiveness in text classification and its robustness in handling high-dimensional data [6]. However, its application in analyzing educational feedback, particularly in the context of improving program effectiveness, is less explored. This paper aims to bridge this gap by detailing the use of SVM in the sentiment analysis of student feedback. The main contribution of this paper is the development of a methodological framework for extracting and analyzing sentiment from student feedback using SVM. This framework has the potential to enhance educational experiences by providing educators with nuanced insights into student sentiment, thereby improving the design and delivery of educational programs. Moreover, this study pioneers in developing a sentiment analysis system specifically for analyzing high-education students’ feedback in Arabic text, addressing a significant gap in current literature. Our key contributions are the creation of the SVM-SAA-SC algorithm for processing Arabic course reviews; a comprehensive preprocessing and classification technique involving tokenization, stop word removal, and stemming; the employment of SVM for classification fine-tuned for optimal performance; the analysis of a substantive dataset from PSAU reflecting diverse educational aspects; and benchmarking against the advanced CAMeLBERT-DA SA Model to validate our approach. This endeavor is poised to uncover valuable insights into the unique challenges Arab students faced during the transition to online learning post-COVID-19 and to enhance the educational experience within Saudi Arabia’s higher education system.

3. Contributions

To the best of our knowledge, there is currently no attempt to develop a sentimental analysis system for high education students using Arabic text. This is an important gap in the literature as there is a need to understand the perceptions and opinions of students in Saudi Arabia Universities regarding their education beyond COVID-19. Developing such a system would provide valuable insights into the challenges and obstacles faced by Arab students during online learning and help to improve their educational experience. The main contributions of this study include:
  • The proposed SVM Sentimental Analysis for Arabic Students’ Review called SVM-SAA-SC algorithm for students’ Arabic course reviews is a general framework that involves collecting student reviews, preprocessing them, and using a machine learning model to classify them as positive, negative, or neutral.
  • The suggested technique for preprocessing and classifying reviews includes steps such as collecting data, removing irrelevant information, tokenizing, removing stop words, stemming or lemmatization, and using pre-trained sentiment analysis models.
  • To train the classifier, SVM algorithm is used, and performance is evaluated using metrics such as accuracy, precision, and recall.
  • Fine-tuning is done by adjusting parameters such as kernel type and regularization strength to optimize performance.
  • The use of a real dataset provided by the deanship of quality at PSAU, which is ranked 801–900 globally and 6th in Saudi Arabia according to the 2022 Shanghai Rankings.
  • The dataset contains students’ opinions on various aspects of their education, including courses, faculties, assignments, projects, and exams.
  • To validate the suggested SVM-SAA-SC algorithm, a comparison with a sentiment analysis tool called CAMeLBERT Dialectal Arabic model (CAMeLBERT-DA SA Model) [30] is adopted. CAMeLBERT represents an interplay of variant, size, and task type in Arabic pre-trained language models.
  • The use of textual comments instead of grading-based feedback to reveal the exact sentiment of students.
The proposed model provides insights about education at PSAU and contributes to improving the quality of teaching at PSAU.

4. Dataset

The dataset used in this study was collected from PSAU, specifically from the College of Computer Engineering and Sciences (CCES) via student course evaluation surveys post-COVID-19 (refer to Figure 1). The data was gathered by the quality unit of CCES, resulting in a total of 384 records and 30 features from 30 different courses spreading across 66 sections. Figure 2 provides a sample of the dataset used. The features include both categorical and quantitative variables, as well as textual comments. Table 1 provide a translation of the Arabic comments provided in Figure 2. Table 2 provides an overview of the extracted features and their descriptions. We highlight that the proportion of class labels in our dataset refers to the distribution of different categories or classifications within the data. For instance, in a sentiment analysis task, the class labels is maily considered to be “positive”, “negative”, and “neutral”, each corresponding to the sentiment expressed in a piece of text. The proportion of these labels would then indicate how many of the entries in the dataset are positive, negative, or neutral, typically represented as a percentage or a count. The proportions of each class within the dataset are as following:
Figure 1. Sample of students’ course evaluation survey.
Figure 2. Sample of students’ comments.
Table 1. Translation of students comments illustrated in Figure 2.
Table 2. Features extracted from the datase.
  • Positive Class: Significantly more than half of the data, 57.1%.
  • Neutral Class: Roughly around half the number of Positive samples, 28.6%.
  • Negative Class: Roughly around a quarter of the number of Positive samples, 14.3%.

4.1. Dataset Cleaning

Data preprocessing is the method of changing raw data into an understandable format. It is additionally a critical step in any data science project [31]. The quality of the data must be checked before applying machine learning techniques to increase results credibility and reliability. During this phase, we prepare collected data for further processing. This involves four steps, which are described in the following subsections.

4.2. Removing Straightlining

Straightlining is when a respondent chooses the same answer option over and over again along all questions (e.g., the last answer option). Straightliners are often speeders as well, as they race through the survey by answering each question with little to no thought. In our survey we found 60 straightlining responses which represents 15.625% of the dataset. These responses have a high impact and could lead to misleading results. After this step we ended up with 324 observations. We performed several preprocessing steps on the datasets. The objective of these measures was to remove elements of the text that could negatively affect the quality of the results.First, we removed diacritical marks from the text. This is a common text preprocessing step in Arabic NLP. To do this, we simply specified a list of these characters and removed them from both datasets. The second preprocessing step was to normalize several Arabic letters which could be written in different manners but have the same meaning. The reason behind this normalization is because we had observed that the data probably (آ) because consists of data collected from a group of students, which tend to be written informally. For example, all instances of We also noticed that some texts included repeated (ي) were changed to (ى) and all instances of (آإأ) were changed to (ا). In text preprocessing for NLP, normalizing repetitive characters and diacritical marks, as commonly done in Arabic NLP, has influenced our model’s generalization capabilities. This process reduces noise, ensuring consistency across word variants and decreasing the vocabulary size, which can streamline training and potentially improve model performance on new, unseen data by reducing out-of-vocabulary errors. However, such normalization can also remove nuanced information, such as emotional intensity indicated by character elongation, which might impact the model’s sensitivity to the subtleties of language. Therefore, the benefits of normalization, which include helping the model focus on relevant features and achieve better generalization, must be weighed against the potential loss of specific linguistic details, especially when applying the model in varying contexts that may demand recognition of formal versus informal text structures.
  • Tokenization
    To identify the sentiments and extract emotions of sentences, Students’ reviews (e.g., comments) are split into words, or tokens, this task is completed using the tokenize function in the python package natural language toolkit (NLTK).
  • Normalization
    We also detect that some comments included repeated characters. In colloquial written speech, some letters are frequently repeated in order to focus on certain emotions. To normalize how this type of emotive word was written, we removed repeated letters and replaced all instances with one only. For example, the word رااااائع was changed to رائع.
  • Removal of irrelevant content
    Punctuation, stop words, symbols, and prepositions, which are irrelevant for Sentiment analysis, are removed to improve the results, response time and effectiveness.
  • Stemming
    Stemming is the most daunting task due to the need to review lexicons. To further facilitate word matching, words in student comments are converted to their root word. For example, المحاضرات, المحاضر. are all converted to محاضرة.

4.3. EDA

Data exploration, or Exploratory Data Analysis (EDA), is the process of analyzing and understanding the patterns, outliers, and relationships in a dataset. It is a crucial step in any data science project, as it allows us to gain a deeper understanding of the data and make informed hypotheses. One powerful tool for EDA is the pandas profiling library, which is an open-source Python package that allows us to perform EDA quickly and easily with just a few lines of code. Additionally, it generates web reports that can be shared with others, making it an ideal tool for presenting our findings. With pandas profiling, we can save time and effort by easily visualizing and understanding the distribution of each variable in the dataset. It creates a comprehensive report that includes summary statistics (Table 3), tables, and visualizations, making it a valuable tool for data exploration and understanding. Additionally, the report also includes information about missing values, such as the Count, Heatmap, and Dendrogram. The Count chart, shown in Figure 3, illustrates the number of missing values in each feature and the count of it. The Dendrogram, shown in Figure 4, displays the missing values in a hierarchical format. It’s important to note that all missing values are textual in nature, as filling in textual data in the survey is optional. To further understand the relationships between variables (features) in our dataset, we used a correlation matrix. The square matrix is based on the Spearman correlation coefficient, represented by ρ . The correlation value describes the interaction strength between two features. The range of correlation varies from +1 to −1. A positive value indicates a positive association between two features, while a value of 0 means that there is no association between these two features, or the relationship is not linear. Thus, they are independent from other features. The correlation matrix provides a clear picture of the association between features. Figure 5 illustrates the matrix for 24 selected features. Additionally, we created a scatter plot using the Seaborn library to represent the pairwise correlations between the various variables. This is useful in providing a simple summary of how the data is spread and whether it includes outliers. Figure 6 illustrates the scatter plot to display the correlation between the highest positive correlation based on Figure 7 where the red shadow in this figure represents the confidence interval for the regression line, indicating the uncertainty about the line’s true position based on the variability of the data points. This association shows that the teamwork and communication features have a strong correlation coefficient of 0.83.
Table 3. Dataset Statistics.
Figure 3. Missing Values Count.
Figure 4. Missing Values, Dendrogram.
Figure 5. Correlation Matrix.
Figure 6. Correlation between Teamwork and Communication.
Figure 7. Variables Overview.

5. Methods

5.1. SVM-SAA-SCR

To develop our system, we used a C-Support Vector Classification (C-SVC) approach, which is widely used for classification problems and allows for tuning of the kernel type and regularization strength to find the best performing model on the training data [32]. The proposed sentimental analysis algorithm for students’ Arabic course reviews is as follows:
  • Collect a dataset of student reviews for an Arabic course.
  • Preprocess the reviews by removing any irrelevant information such as special characters and numbers.
  • Tokenize the reviews to create a list of words for each review.
  • Create a sentiment dictionary or use pre-trained sentiment lexicon to classify each review as positive, negative, or neutral.
  • Use a machine learning model to train on the preprocessed and classified reviews.
  • Use the trained model to classify new reviews.
  • Analyze the results to identify patterns and areas for improvement in the Arabic course based on the sentiment classification of the reviews.
  • Repeat the process with updated reviews and continue to make improvements to the course based on the analysis of the results.
The suggested algorithm is a general framework and different techniques can be used to preprocess and classify the reviews. Additionally, the algorithm can be improved by using additional techniques such as removing stop words, using stemming or lemmatization, and using advanced machine learning models.

5.1.1. SVM-SAA-SCR Preprocessing and Classification

The proposed technique for preprocessing and classifying students’ Arabic course reviews includes the following steps:
  • Collect a dataset of students’ Arabic course reviews from various sources, such as online forums, social media, and surveys.
  • Preprocess the reviews by removing any irrelevant information, such as URLs, usernames, and special characters.
  • Tokenize the reviews by splitting them into individual words or phrases.
  • Remove stop words, such as “and” and “the”, which do not carry much meaning in the analysis.
  • Perform stemming or lemmatization to reduce words to their base form and improve consistency in the dataset.
  • Use pre-trained sentiment analysis models, such as SentimentIntensityAnalyzer or TextBlob, to classify each review as positive, negative, or neutral.
  • Use the resulting sentiment scores to create visualizations, such as bar charts or pie charts, to show the distribution of sentiments among the reviews.
  • Analyze the results to identify patterns and trends in the students’ sentiments towards the course.
  • Use the insights gained from the analysis to improve the educational process and address areas of improvement.
Please note that this is a general approach and it may require some adjustments to suit the specific requirements of the task. Also, you may need to preprocess the data more deeply to handle the Arabic language specifics such as diacritics, or even use specialized libraries for Arabic text processing.

5.1.2. Model Training Based SVM-SAA-SCR

To train the SVM-SAA-SCR, we used the course reviews of PSAU students provided by the deanship of quality in the College of Computer Engineering and Sciences. Then the following steps are adopted:
  • Collect and preprocess the reviews by removing any irrelevant information such as diacritics, URLs, and names.
  • Split the dataset into training and testing sets.
  • Convert the text data into numerical features using techniques such as bag-of-words or TF-IDF.
  • Train an SVM classifier using the training dataset and the numerical features as input.
  • Test the classifier on the testing dataset and evaluate its performance using metrics such as accuracy, precision, and recall.
  • Fine-tune the model by adjusting parameters such as the kernel type and regularization strength to optimize performance.
  • Use the trained and fine-tuned classifier to classify new reviews and predict their sentiment.

5.1.3. Evaluation of SVM-SAA-SCR Performance

The performance of the suggested SVM-SAA-SCR can be evaluated using several metrics, such as accuracy, precision, and recall. Here is one possible approach to evaluate the performance of the SVM-SAA-SCR:
  • Split the dataset into a training set and a test set.
  • Train the SVM model on the training set.
  • Use the trained model to predict the labels of the samples in the test set.
  • Compare the predicted labels to the true labels of the test set samples.
  • Compute the accuracy, precision, and recall of the model using the following formulas:
    • Accuracy = (True Positives + True Negatives)/Total Samples
    • Precision = True Positives/(True Positives + False Positives)
    • Recall = True Positives/(True Positives + False Negatives)
  • Interpret the results and make any necessary adjustments to the model or the data preprocessing steps.
We note that accuracy is the percentage of correctly classified samples, precision is the percentage of correctly classified positive samples among all the positive samples, recall is the percentage of correctly classified positive samples among all the actual positive samples. The suggested SVM-SAA-SCR has the ability to classify sentiments using a small training set (PSAU dataset). It has also shown strong competition in achieving a high accuracy of 84.7%. The detailed performance results based on the PSAU dataset are detailed in Table 4 and Figure 8
Table 4. SVM-SAA-SCR performance on PSAU Dataset.
Figure 8. SVM-SAA-SCR performance on on PSAU dataset.

5.1.4. Tuning Hyper-Parameters for SVM-SAA-SCR

To tune the hyper-parameters of SVM-SAA-SCR, we adopted the way suggested by Louati et al., (2022) [31], which could be summarized as follows:
  • Begin by initializing the model with a default kernel type (e.g., radial basis function kernel) and regularization strength (e.g., C = 1).
  • Use cross-validation techniques to evaluate the performance of the model on the training data, using metrics such as accuracy, precision, and recall.
  • Experiment with different kernel types (e.g., linear, polynomial, sigmoid) and regularization strengths (e.g., C = 0.1, C = 10) to find the combination that gives the best performance on the training data.
  • Once the optimal kernel type and regularization strength have been identified, use these parameters to train the model on the entire training set.
  • Evaluate the performance of the optimized model on the test set using the same metrics as before.
  • If the performance is not satisfactory, fine-tune the model further by adjusting other parameters such as the degree of the polynomial kernel or the gamma parameter for the radial basis function kernel.
  • Repeat the process of evaluating and fine-tuning the model until the desired level of performance is achieved.
To avoid overfitting L1 or L2 regularizations are adopted, which add a penalty term to the loss function to discourage the model from assigning too much weight to any one feature. Additionally, an early stopping technique is used to prevent the model from memorizing the training data. To prevent underfitting, we adopted data augmentation to expand the size of the training dataset.

5.2. CAMeLBERT-DA SA Model

This CAMeLBERT-DA SA Model is implemented based on the following datasets:
  • The Arabic Speech-Act and Sentiment Corpus of Tweets (ArSAS), where 23,327 tweets have been used [33];
  • The Arabic Sentiment Tweets Dataset (ASTD) [34];
  • SemEval-2017 task 4-A benchmark dataset [35];
  • The Multi-Topic Corpus for Target-based Sentiment Analysis in Arabic Levantine Tweets (ArSenTD-Lev) [36].
These datasets were combined and preprocessed in a similar way to what was done in [37,38]. The preprocessing steps included removing diacritics, URLs, and Twitter usernames from all the tweets.
The experimental setup adopted by [30] for CAMeLBERT-DA SA is summarized as follows:
  • Build three pre-trained language models across three variants of Arabic: Modern Standard Arabic (MSA), dialectal Arabic, and classical Arabic.
  • Build a fourth language model pre-trained on a mix of the three variants.
  • Build additional models pre-trained on a scaled-down set of the MSA variant to examine the importance of pre-training data size.
  • Fine-tune the models on five NLP tasks spanning 12 datasets.
  • Compare the performance of the models to each other and to eight publicly available models.
  • Analyze the results to identify the importance of variant proximity of pre-training data to fine-tuning data and the impact of pre-training data size.
  • Define an optimized system selection model for the studied tasks based on the insights gained from the experiments.
  • Make the created models and fine-tuning code publicly available.

6. Results

Figure 9 and Figure 10 provide an illustration of the sentiment polarity for a total of 1870 reviews. These reviews have been classified into three categories: positive, negative, and neutral. Table 5 represents a comparison between SVM-SAA-SCR and CAMeLBERT-DA SA based on these 1870 students’ reviews. Figure 9 shows that the majority of reviews are positive, with roughly a quarter of the reviews being neutral. This information can be used to identify areas of improvement in the educational process. Similarly, Figure 10 illustrates the sentiment polarity for 1870 reviews, with the majority of these reviews being positive. Again, this information can be used to pinpoint areas of improvement in the educational process. In summary, the CAMeLBERT-DA SA Model shows that 70.48% of the reviews are classified as positive, while the SVM-SAA-SCR shows that 69.62% of the reviews are classified as positive. Additionally, the CAMeLBERT-DA SA Model shows that 22.03% of the reviews are classified as neutral, while the SVM-SAA-SCR shows that 20.74% of the reviews are classified as neutral. Lastly, the CAMeLBERT-DA SA Model shows that 7.48% of the reviews are classified as negative, while the SVM-SAA-SCR shows that 9.62% of the reviews are classified as negative.
Figure 9. CAMeLBERT-DA SA Model result.
Figure 10. SVM-SAA-SCR result.
Table 5. SVM-SAA-SCR vs. CAMeLBERT-DA SA based on PSAU dataset of 1870 students’ reviews.

7. Discussion

The proposed sentimental analysis system for high education students using Arabic text can provide valuable insights into the perceptions and opinions of students at PSAU. The analysis of student reviews, classified as positive, negative, or neutral, can help identify areas of improvement in the educational process. This information can be used to make necessary adjustments to the curriculum and teaching methods to enhance the educational experience for students. Additionally, the system can help identify specific issues that students may be facing and provide targeted solutions to address these issues. By providing a more detailed understanding of student perceptions and opinions, the proposed system can help to improve the overall quality of education at Prince Sattam University. Furthermore, the system can also help in improving the quality of teaching by providing teachers with feedback on their performance, which can help them to make necessary adjustments and improve their teaching methods.
Compared to the literature, our work enhanced sentiment analysis literature by providing improved analytical precision, surpassing prior research outcomes. It captures an extensive array of sentiments, going beyond traditional emotional categories, and broadens the scope of analysis to include a diverse range of text data, with a particular focus on overcoming the intricacies of the Arabic language. The SVM-SAA-SCR technique offers marked advancements, such as enhanced performance indicators, expanded sentiment identification, and increased flexibility for various complex Arabic datasets. Finally, we draw the reader’s attention that this work sought to validate the suggested system with a well-known tool such as CAMeLBERT through a comparison based on accuracy and results similarity. This comparison has shown a very slight difference between both systems and promising results for SVM-SAA-SCR.

8. Conclusions

The study conducted an in-depth development of a sentimental analysis system tailored for higher education students using Arabic text, marking a significant stride in the relevant academic field. Our SVM-SAA-SCR algorithm, applied to the real dataset from Prince Sattam bin Abdulaziz University (PSAU), was set against the backdrop of the sophisticated CAMeLBERT Dialectal Arabic model, allowing us to evaluate student sentiments about their educational experiences in a post-COVID-19 context. Results highlighted a predominantly positive outlook, with both models revealing over 69% positive feedback. Furthermore, when we delve into the performance metrics of the SVM-SAA-SCR model, we find that it not only performs sentiment classification with a limited training set but also exhibits robustness with an impressive accuracy rate of 84.7%. The presence of negative and neutral sentiments, albeit in a smaller proportion, pinpoints areas that are ripe for improvement within the educational framework. The deployment of sentiment analysis thus emerges as an essential tool for educators and administrators, keen on optimizing the educational milieu based on student feedback. In summary, this investigation underscores the potential of advanced sentiment analysis to extract actionable insights from student feedback, thereby enhancing the overall educational quality and experience at PSAU. It serves as a testament to the power of machine learning in transforming qualitative data into a catalyst for educational innovation.

Author Contributions

Methodology, A.L.; Software, A.L., F.A. and A.A.; Validation, A.L.; Data Acquisition: A.L.; Writing—original draft preparation, A.L. and F.A.; Review and editing, A.L. and H.L.; Supervision, A.L.; Project administration, E.K. All authors have read and agreed to the published version of the manuscript.

Funding

Prince Sattam bin Abdulaziz University project number (PSAU/2023/R/1444).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data could be provided on reasonable request from the corresponding author.

Acknowledgments

The authors thank the Deanship of Scientific Research at Prince Sattam bin Abdulaziz University (PSAU) for funding this work via project number (PSAU/2023/R/1444). In addition, the authors thank Luay Assidmi (l.assidmi@psau.edu.sa) the ex-dean of quality in the College of Computer Engineering and Sciences, PSAU for providing the dataset. Finally, The authors would like to acknowledge that this research work was partially financed by Kingdom University, Bahrain from the research grant number 2023-11-015.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Smith, A.; McCarthey, S.; Magnifico, A. Recursive feedback: Evaluative dimensions of e-learning. In E-Learning Ecologies; Routledge: Milton Park, UK, 2017; pp. 118–142. [Google Scholar]
  2. Jones, B.D. Motivating and engaging students using educational technologies. In Handbook of Research in Educational Communications and Technology; Learning Design; Contemporary Educational Technology: London, UK, 2020; pp. 9–35. [Google Scholar]
  3. Galante, J.; Friedrich, C.; Dawson, A.F.; Modrego-Alarcón, M.; Gebbing, P.; Delgado-Suárez, I.; Jones, P.B. Mindfulness-based programmes for mental health promotion in adults in nonclinical settings: A systematic review and meta-analysis of randomised controlled trials. PLoS Med. 2021, 18, e1003481. [Google Scholar] [CrossRef] [PubMed]
  4. Poria, S.; Hazarika, D.; Majumder, N.; Mihalcea, R. Beneath the tip of the iceberg: Current challenges and new directions in sentiment analysis research. IEEE Trans. Affect. Comput. 2020, 14, 108–132. [Google Scholar] [CrossRef]
  5. Kusal, S.; Patil, S.; Choudrie, J.; Kotecha, K.; Mishra, S.; Abraham, A. AI-based conversational agents: A scoping review from technologies to future directions. IEEE Access 2022, 10, 92337–92356. [Google Scholar] [CrossRef]
  6. Mirończuk, M.M.; Protasiewicz, J. A recent overview of the state-of-the-art elements of text classification. Expert Syst. Appl. 2018, 106, 36–54. [Google Scholar] [CrossRef]
  7. Rakshitha, K.H.M.R.; Pavithra, M.H.D.A.; Hegde, M. Sentimental analysis of indian regional languages on social media. Glob. Transit. Proc. 2021, 2, 414–420. [Google Scholar] [CrossRef]
  8. Morshed, S.A.; Khan, S.S.; Tanvir, R.B.; Nur, S. Impact of COVID-19 pandemic on ride-hailing services based on large-scale twitter data analysis. J. Urban Manag. 2021, 10, 155–165. [Google Scholar] [CrossRef]
  9. Shelke, N.; Chaudhury, S.; Chakrabarti, S.; Bangare, S.L.; Yogapriya, G.; Pandey, P. An efficient way of text-based emotion analysis from social media using lra-dnn. Neurosci. Inform. 2022, 2, 100048. [Google Scholar] [CrossRef]
  10. Bonifazi, G.; Cauteruccio, F.; Corradini, E.; Marchetti, M.; Sciarretta, L.; Ursino, D.; Virgili, L. A Space-Time Framework for Sentiment Scope Analysis in Social Media. Big Data Cogn. Comput. 2022, 6, 130. [Google Scholar] [CrossRef]
  11. Bonifazi, G.; Cauteruccio, F.; Corradini, E.; Marchetti, M.; Terracina, G.; Ursino, D.; Virgili, L. A framework for investigating the dynamics of user and community sentiments in a social platform. Data Knowl. Eng. 2023, 146, 102183. [Google Scholar] [CrossRef]
  12. Najar, D.; Mesfar, S. Opinion mining and sentiment analysis for arabic on-line texts: Application on the political domain. Int. J. Speech Technol. 2017, 20, 575–585. [Google Scholar] [CrossRef]
  13. Sghaier, M.A.; Zrigui, M. Sentiment analysis for arabic e-commerce websites. In Proceedings of the 2016 International Conference on Engineering & MIS (ICEMIS), Agadir, Morocco, 22–24 September 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–7. [Google Scholar]
  14. Mourad, A.; Darwish, K. Subjectivity and sentiment analysis of modern standard arabic and arabic microblogs. In Proceedings of the 4th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Atlanta, Georgia, 14 June 2013; pp. 55–64. [Google Scholar]
  15. Alwakid, G.; Osman, T.; Haj, M.E.; Alanazi, S.; Humayun, M.; Sama, N.U. Muldasa: Multifactor lexical sentiment analysis of social-media content in nonstandard arabic social media. Appl. Sci. 2022, 12, 3806. [Google Scholar] [CrossRef]
  16. Tartir, S.; Abdul-Nabi, I. Semantic sentiment analysis in arabic social media. J. King Saud Univ. Comput. Inf. Sci. 2017, 29, 229–233. [Google Scholar] [CrossRef]
  17. Zhou, M.; Mou, H. Tracking public opinion about online education over COVID-19 in China. Educ. Technol. Res. Dev. 2022, 70, 1083–1104. [Google Scholar] [CrossRef] [PubMed]
  18. Toçoğlu, M.A.; Onan, A. Sentiment analysis on students’ evaluation of higher educational institutions. In International Conference on Intelligent and Fuzzy Systems; Springer: Berlin/Heidelberg, Germany, 2020; pp. 1693–1700. [Google Scholar]
  19. Nikolić, N.; Grljević, O.; Kovačević, A. Aspect-based sentiment analysis of reviews in the domain of higher education. Electron. Libr. 2020, 38, 44–64. [Google Scholar] [CrossRef]
  20. Mohiudddin, K.; Rasool, A.M.; Mohd, M.S.; Mohammad, R.H. Skill-centered assessment in an academic course: A formative approach to evaluate student performance and make continuous quality improvements in pedagogy. Int. J. Emerg. Technol. Learn. (Online) 2019, 14, 92. [Google Scholar] [CrossRef]
  21. Dsouza, D.D.; Deepika, D.P.N.; Machado, E.J.; Adesh, N. Sentimental analysis of student feedback using machine learning techniques. Int. J. Recent Technol. Eng. 2019, 8, 986–991. [Google Scholar]
  22. Louati, A. Cloud-assisted collaborative estimation for next-generation automobile sensing. Eng. Appl. Artif. Intell. 2023, 126, 106883. [Google Scholar] [CrossRef]
  23. Louati, A. A hybridization of deep learning techniques to predict and control traffic disturbances. Artif. Intell. Rev. 2020, 53, 5675–5704. [Google Scholar] [CrossRef]
  24. Webb, M.E.; Fluck, A.; Magenheim, J.; Malyn-Smith, J.; Waters, J.; Deschênes, M.; Zagami, J. Machine learning for human learners: Opportunities, issues, tensions and threats. Educ. Technol. Res. Dev. 2021, 69, 2109–2130. [Google Scholar] [CrossRef]
  25. Singh, N.K.; Tomar, D.S.; Sangaiah, A.K. Sentiment analysis: A review and comparative analysis over social media. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 97–117. [Google Scholar] [CrossRef]
  26. Alrehili, A.; Albalawi, K. Sentiment analysis of customer reviews using ensemble method. In Proceedings of the 2019 International Conference on Computer and Information Sciences (ICCIS), Sakaka, Saudi Arabia, 3–4 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  27. Al-Smadi, M.; Qawasmeh, O.; Al-Ayyoub, M.; Jararweh, Y.; Gupta, B. Deep recurrent neural network vs. support vector machine for aspect-based sentiment analysis of arabic hotels’ reviews. J. Comput. Sci. 2018, 27, 386–393. [Google Scholar] [CrossRef]
  28. Al-Horaibi, L.; Khan, M.B. Sentiment analysis of arabic tweets using text mining techniques. In First International Workshop on Pattern Recognition; SPIE: Tokyo, Japan, 2016; Volume 10011, pp. 288–292. [Google Scholar]
  29. Alsayat, A.; Elmitwally, N. A comprehensive study for arabic sentiment analysis (challenges and applications). Egypt. Inform. J. 2020, 21, 7–12. [Google Scholar] [CrossRef]
  30. Inoue, G.; Alhafni, B.; Baimukan, N.; Bouamor, H.; Habash, N. The interplay of variant, size, and task type in arabic pre-trained language models. arXiv 2021, arXiv:210306678. [Google Scholar]
  31. Louati, A.; Lahyani, R.; Aldaej, A.; Aldumaykhi, A.; Otai, S. Price forecasting for real estate using machine learning: A case study on Riyadh city. Concurr. Comput. Pract. Exp. 2022, 34, e6748. [Google Scholar] [CrossRef]
  32. Jiang, L.; Yao, R. Modelling personal thermal sensations using C-Support Vector Classification (C-SVC) algorithm. Build. Environ. 2016, 99, 98–106. [Google Scholar] [CrossRef]
  33. Elmadany, A.; Mubarak, H.; Magdy, W. Arsas: An arabic speech-act and sentiment corpus of tweets. OSACT 2018, 3, 20. [Google Scholar]
  34. Nabil, M.; Aly, M.; Atiya, A. Astd: Arabic sentiment tweets dataset. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 2515–2519. [Google Scholar]
  35. Rosenthal, S.; Farra, N.; Nakov, P. Semeval-2017 task 4: Sentiment analysis in twitter. arXiv 2019, arXiv:191200741. [Google Scholar]
  36. Baly, R.; Khaddaj, A.; Hajj, H.; El-Hajj, W.; Shaban, K.B. Arsentd-lev: A multi-topic corpus for target-based sentiment analysis in arabic levantine tweets. arXiv 2019, arXiv:190601830. [Google Scholar]
  37. Farha, I.A.; Magdy, W. Mazajak: An online arabic sentiment analyser. In Proceedings of the Fourth Arabic Natural Language Processing Workshop, Florence, Italy, 1 August 2019; pp. 192–198. [Google Scholar]
  38. Obeid, O.; Zalmout, N.; Khalifa, S.; Taji, D.; Oudah, M.; Alhafni, B.; Inoue, G.; Eryani, F.; Erdmann, A.; Habash, N. Camel tools: An open source python toolkit for arabic natural language processing. In Proceedings of the 12th Language Resources and Evaluation Conference, Marseille, France, 11–16 May 2020; pp. 7022–7032. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.