Next Article in Journal
Green’s Function for the Cauchy Problem to the Dissipative Linear Evolution Equation of Arbitrary Order
Previous Article in Journal
Enhancing Road Freight Price Forecasting Using Gradient Boosting Ensemble Supervised Machine Learning Algorithm
Previous Article in Special Issue
Enhancing Explainable Artificial Intelligence: Using Adaptive Feature Weight Genetic Explanation (AFWGE) with Pearson Correlation to Identify Crucial Feature Groups
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Stance Detection in Arabic Tweets: A Machine Learning Framework for Identifying Extremist Discourse

Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(18), 2965; https://doi.org/10.3390/math13182965
Submission received: 21 July 2025 / Revised: 1 September 2025 / Accepted: 5 September 2025 / Published: 13 September 2025
(This article belongs to the Special Issue Machine Learning Theory and Applications)

Abstract

Terrorism remains a critical global challenge, and the proliferation of social media has created new avenues for monitoring extremist discourse. This study investigates stance detection as a method to identify Arabic tweets expressing support for or opposition to specific organizations associated with extremist activities, using Hezbollah as a case study. Thousands of relevant Arabic tweets were collected and manually annotated by expert annotators. After extensive preprocessing and feature extraction using term frequency–inverse document frequency (tf-idf), we implemented traditional machine learning (ML) classifiers—Support Vector Machines (SVMs) with multiple kernels, Multinomial Naïve Bayes, and Weighted K-Nearest Neighbors. ML models were selected over deep learning (DL) approaches due to (1) limited annotated Arabic data availability for effective DL training; (2) computational efficiency for resource-constrained environments; and (3) the critical need for interpretability in counterterrorism applications. While interpretability is not a core focus of this work, the use of traditional ML models (rather than DL) makes the system inherently more transparent and readily adaptable for future integration of interpretability techniques. Comparative experiments using FastText word embeddings and tf-idf with supervised classifiers revealed superior performance with the latter approach. Our best result achieved a macro F-score of 78.62% using SVMs with the RBF kernel, demonstrating that interpretable ML frameworks offer a viable and resource-efficient approach for monitoring extremist discourse in Arabic social media. These findings highlight the potential of such frameworks to support scalable and explainable counterterrorism tools in low-resource linguistic settings.

1. Introduction

Social media has become an integral part of modern life, serving not only as a primary platform for content sharing but also as a dominant source of information. As of April 2025, over 5.3 billion people—nearly 65% of the global population—use social media, with the typical user engaging across about 6.9 platforms monthly [1]. Platforms such as Facebook, YouTube, Instagram, TikTok, and Twitter encompass diverse communication forms, including personal updates, news, and educational content. Studies report that users often engage with multiple platforms simultaneously—such as browsing Instagram or YouTube while interacting on Twitter—highlighting the multi-modal and concurrent nature of social media consumption [2,3].
Twitter, one of the most prominent microblogging platforms, allows users to share short posts (tweets) limited to 280 characters. Although the platform has been renamed X, we retain the term “Twitter” throughout this study for clarity and consistency with prior studies. Tweets are published to the user’s “followers”—subscribers to their account—who can interact by replying, retweeting, or marking them as favorites [4].
While sentiment analysis and stance detection are closely related, they address distinct objectives. Sentiment analysis determines the polarity of a text (positive, negative, or neutral) by identifying emotional cues such as “love” or “hate.” However, sentiment alone is often insufficient; understanding a user’s position toward a predefined topic—known as stance detection—is equally crucial [5]. As an emerging research area, which is particularly underexplored for Arabic-language content, stance detection seeks to classify a text as favoring, opposing, or remaining neutral toward a pre-determined target.
A key challenge in stance detection lies in the implicitness of targets. Often, the target is not explicitly mentioned within the text, requiring systems to infer stance from indirect references, rhetorical devices, or context-specific cues that assume shared knowledge with the audience. For instance, consider a synthetic tweet such as “They say progress is being made, but the streets tell a different story” does not explicitly name the government, yet it may reflect opposition to its policies. Conversely, “Finally, someone is taking bold steps to fix years of neglect” might imply support. Addressing such complexities, where stance must be inferred rather than directly observed, is a central objective of stance detection systems [5].
The term terrorism originates from the French word terrorisme, which was derived from the Latin term terror (“great fear” or “dread”), and is related to terrere (“to frighten”). While historically applied to state-imposed terror during revolutionary periods (e.g., the Reign of Terror), contemporary usage more commonly describes subnational violence against states [6], though state-sponsored violence against civilian populations also falls within many scholarly definitions. Despite its widespread use, no universal legal or scientific consensus exists, with over 109 distinct definitions documented [7]. Governments and legal systems employ varying interpretations, and political sensitivities have prevented international agreement. https://en.wikipedia.org/wiki/Definition_of_terrorism (accessed on 9 June 2025). Attempts like Alex Schmid’s 1992 proposal to define terrorism as “peacetime equivalents of war crimes” based on international humanitarian law failed to gain adoption.
Nevertheless, terrorism remains a pressing global issue, with governments dedicating substantial resources to monitoring and countering terrorist activities. Motivated by this, we focus on Hezbollah as the central target of this study due to its classification as a terrorist organization by the Gulf Cooperation Council (GCC) and multiple international bodies. Hezbollah’s prominent role in regional conflicts, particularly in Syria and Lebanon, and its extensive discussion on Arabic social media make it an ideal case for examining stance detection in highly polarized and sensitive discourse.
Critically, we contend that this focused approach is a methodological strength. Unlike prior research, which often examines stance detection in broad or unspecified contexts, our deliberate emphasis on Hezbollah enables a deeper, contextually grounded analysis of the group’s unique discourse patterns, rhetorical strategies, and regional linguistic nuances. To analyze public sentiment, we compiled thousands of Arabic tweets—concise texts by nature—and developed a robust system for detecting stances specifically regarding Hezbollah’s activities in Syria. For brevity, we refer to this subject simply as “Hezbollah” throughout the paper. This specificity not only enhances the validity and reliability of our methodology for Hezbollah-related content but also provides a clear, replicable framework that can be adapted to other organizations in future work. To the best of our knowledge, this is the first stance detection study tailored to a specific organization, setting a precedent for more precise and comparable analyses in this domain.
In this work, we employed traditional machine learning (ML) classifiers—including Support Vector Machines (SVMs), Multinomial Naïve Bayes (MNB), and weighted K-Nearest Neighbors—rather than deep learning (DL) approaches. While state-of-the-art performance in Arabic stance detection is currently achieved using DL models like AraBERT, our decision was driven by three specific objectives critical to counterterrorism contexts: (i) model interpretability, as opaque “black box” predictions are unsuitable for high-stakes scenarios requiring validation and justification, with serious consequences for decision-making; (ii) computational efficiency, given many real-world applications operate in resource-constrained environments; and (iii) effective operation with limited training data, since high-quality annotated Arabic datasets remain scarce and DL models are prone to overfitting.
We recognize this choice may sacrifice some absolute performance compared to modern DL methods. However, it ensures broader applicability in low-resource or explainability-critical settings while establishing a clear, well-documented baseline. This foundation enables future studies to precisely measure improvements when applying DL models while currently providing policymakers with the transparent classification logic essential for counterterrorism analysis.

1.1. Problem Definition

The computational analysis of terrorist organizations on social media involves two complementary tasks: content detection and stance detection. Content detection identifies posts containing extremist material—such as propaganda, recruitment messages, hate speech, or violent imagery—using binary or multi-class classification. This supports content moderation and counterterrorism monitoring by flagging material that explicitly promotes or supports terrorist activities.
Stance detection, in contrast, provides a finer-grained analysis by determining the author’s position (supporting, opposing, or neutral) toward a specific organization like Hezbollah. Unlike content detection, which focuses on the presence of extremist themes, stance detection examines ideological alignment and rhetorical patterns to assess public sentiment. Such analysis is critical for understanding polarization, mapping societal attitudes, and informing policy decisions in conflict-prone regions.
The distinction lies in analytical granularity: content detection broadly scans for terrorism-related material, while stance detection targets individual attitudes toward predefined entities. Together, these tasks offer a comprehensive framework for monitoring terrorist organizations’ digital presence—from operational content identification to public opinion mapping.

1.2. Our Contributions

The main contributions of this work are as follows:
  • Creation of a specialized Arabic dataset: A collection of 7053 tweets focused on Hezbollah’s activities was compiled and annotated, addressing the scarcity of labeled data for Arabic stance detection research. We made the dataset available for non-commercial use.
  • Development of a practical stance detection system: An efficient ML-based framework for classifying stances in Arabic tweets about Hezbollah’s involvement in Syria was developed.
While interpretability is not a core focus of this work, the use of traditional ML models (rather than deep learning) makes the system inherently more transparent and easily expandable to incorporate interpretability techniques for stance analysis in future work.
The remainder of this paper is structured as follows: Section 2 provides the necessary background. Section 3 reviews related work, beginning with research in stance detection and proceeding to studies on terrorism detection. Section 4 details the proposed methodology for stance detection in Arabic tweets. Section 5 presents experimental setups and discusses the results. Finally, Section 6 concludes the study and suggests promising directions for future research.

2. Background

In this section, we provide essential background on the Arabic language, outline its inherent linguistic challenges, and present an overview of stance detection along with the key difficulties it entails.

2.1. Challenges in Handling Arabic Tweets

Twitter has emerged as a vital platform for studying contemporary public discourse, offering researchers unprecedented access to real-time public opinion and diverse societal voices [8]. While the platform’s big data capabilities complement traditional survey methods [9], scholars must navigate significant methodological challenges, including algorithmic biases, echo chambers, and the performative nature of online self-presentation. There is a growing consensus about the limitations of hashtag-centric studies and platform-specific analyses, and the field is witnessing a shift toward more robust theoretical frameworks and interdisciplinary approaches to move beyond descriptive analyses toward a deeper understanding of mediated public discourse [10].
From a computational perspective, this work is hampered by the unique nature of Twitter data itself. The platform is characterized by brevity and inherent noise, and this short length severely restricts the available word co-occurrence statistics and contextual information, which are crucial for deriving reliable semantic similarity measures [11].
These challenges are profoundly exacerbated by the intrinsic complexities of the Arabic language. Modern Standard Arabic (MSA) is typically written without diacritical markings (harakat), leading to pervasive lexical and semantic ambiguity. This issue is compounded by the language’s rich morphology, its extensive use of synonyms, and its highly inflectional and derivational structure.
The problem is further complicated by the diglossic reality of Arabic social media, where tweets are composed in a blend of MSA and various regional colloquial dialects [12]. These dialects exhibit greater variation than those typically found in European languages and lack standardized spelling rules or formal dictionaries [12]. Consequently, a single word can possess contradictory meanings across different dialects; for instance, the word بلش can mean “finished” in one dialect but “started” in another.
The core of the disambiguation problem lies in the absence of short vowel diacritics. These markings are essential for resolving a word’s precise meaning and grammatical function. While a human reader can use context to infer the correct interpretation of an unvowelized word like علم (which can mean “science,” “flag,” “taught,” “knew,” or “knowledge”), this task is exceptionally difficult for automated systems [12]. This ambiguity even extends to vowelized texts, where a word like عَام can mean either “year” or “public,” requiring deep contextual understanding for accurate disambiguation.
The intertwined challenges of ambiguity, dialectal variation, and sparse contexts are not peripheral concerns for social media analysis; they represent a core bottleneck to the progress of Arabic NLP at large. Addressing them demands not incremental refinements but domain-specific innovation and substantial resource investment. For example, the TaSbeeb judicial decision support system [13] illustrates the scale of effort required to adapt NLP to the rigor of legal discourse, while the Arabic essay evaluator in [14,15] underscores the difficulty of simultaneously assessing linguistic proficiency and semantic depth in educational settings. These cases demonstrate that overcoming the inherent linguistic complexity of Arabic is indispensable for building robust and socially impactful AI systems. Nowhere is this more critical than in computational analysis of Arabic Twitter data, where meaningful societal insights can only emerge if models effectively navigate the dense interplay between dialect, implicitness, and limited contexts.

2.2. Stance Detection

Stance detection is the task of automatically determining whether the author of a text is in favor, neutral, or against a given target (e.g., person, organization, movement, or topic). While progress has been made in detecting stance toward explicitly mentioned targets, a persistent challenge arises when the target is implicit—that is, not directly mentioned in the text but only implied through the context, allusion, or shared knowledge. This implicitness, also referred to as target latency, transforms stance detection from a surface-level classification problem into a complex reasoning task that integrates natural language understanding, commonsense inference, and knowledge retrieval.
Several interconnected factors make implicit targets particularly difficult to handle:
  • Disambiguation Problem: A stance-bearing statement may not identify its target explicitly but instead reference related sub-events, policies, or entities. For instance, a tweet stating, “They finally lifted the ban!” conveys a positive stance, but without an explicit target, it could refer to (1) women’s driving rights in Saudi Arabia, (2) COVID-19 travel restrictions, or (3) the reinstatement of a suspended player. Correct interpretation requires deep contextual modeling and extensive world knowledge to resolve that the term “they” refers to a specific authority and “the ban” to a specific policy.
  • Limits of Lexical and Syntactic Patterns: Traditional supervised and feature-based approaches (e.g., relying on tf-idf, n-grams, or syntactic dependencies) struggle when surface lexical cues are absent. If words such as “climate,” “global,” or “warming” never appear, these models lack anchors to tether stance predictions to the correct topic. This illustrates why implicit targets cannot be resolved by shallow statistical associations alone.
  • Knowledge Gap for LLMs: Modern large language models (LLMs) encapsulate vast world knowledge, yet they are not immune to errors with implicit targets. The challenge lies in retrieval and alignment: the model must (a) recall the relevant event (e.g., that lifting the ban on Saudi women driving occurred in 2018), (b) align this knowledge to ambiguous references such as “they” and “ban,” and (c) filter this event from countless similar “ban-lifting” instances. This process is highly sensitive to context length, prompt phrasing, and model biases, which often results in inconsistent stance predictions.
  • Context and Figurative Language: Implicit stance frequently co-occurs with figurative language such as irony, sarcasm, or metaphor, where the literal form diverges from the intended meaning. Statements like “What a brilliant idea!” could signal sarcasm and convey a negative stance, requiring pragmatic inference beyond lexical cues.
  • Data Scarcity and Annotation Difficulty: Constructing datasets for implicit stance detection is particularly challenging. Human annotators can rely on commonsense reasoning to infer implicit targets, but for machine learning, this requires datasets labeled not only with stance but also with supporting evidence and often explicit target resolution. This is expensive, time-intensive, and hard to scale. Consequently, most benchmarks (e.g., SemEval-2016) underrepresent implicit targets, leaving models poorly trained for real-world ambiguity.
  • Evaluation Dilemma: Assessing stance detection performance on implicit targets is also non-trivial. For example, if a model labels a post as “against” a policy, but the gold label associates the stance with the political party responsible for the policy, should this be considered correct? Simple accuracy or F-scores are insufficient, and evaluation frameworks must account for reasoning chains and conceptual proximity to capture model competence fairly.
The implicitness of targets magnifies the difficulty of stance detection because it requires resolving hidden references, integrating discourse-level and contextual information, and applying external world knowledge. While LLMs have advanced the field, reliably bridging the gap between explicit utterances and implicit meaning remains one of the most fundamental challenges.
The challenge of implicit targets compounds the difficulty of stance detection by necessitating the resolution of unstated references, the integration of discourse and contextual cues, and the application of external world knowledge. Although LLMs have propelled the field forward, reliably inferring implicit meaning from explicit utterances remains a fundamental unsolved problem. This challenge is further amplified by dialectal variation when shifting focus across regional groups. For instance, while this work concentrates on Hezbollah—which is associated with a specific dialect—applying these methods to groups using other Arabic dialects (e.g., Maghrebi or Gulf Arabic) [16] introduces additional complexity due to significant linguistic variation. Effective cross-dialect stance detection thus requires more than additional training data; it demands the development of dialect-specific lexical resources to address contradictory word senses, the creation of large-scale annotated corpora capturing unique morphological and syntactic features, and the design of dialect-aware models capable of dynamically adapting to the blended MSA–dialect continuum prevalent in social media.

3. Related Works

Our review of related work follows two key strands of research: (1) studies focusing on stance detection in social media, which provide methodological foundations for analyzing user perspectives and (2) research on terrorism-related content detection, which offers domain-specific insights relevant to our study of Hezbollah-related discourse. This dual perspective allows us to situate our work at the intersection of computational methods and security applications.

3.1. Stance Detection

Stance detection, the computational task of determining a text’s position toward a specific target, has become a cornerstone of computational linguistics. The field gained significant momentum with the establishment of standardized benchmarks through the SemEval-2016 Task 6 competition [17], which continues to influence contemporary methodologies. Prior foundational work by Somasundaran and Wiebe [18] combined sentiment and argument features in supervised classifiers, achieving 63.93% accuracy on contentious topics. Similarly, Addawood et al. [19] leveraged syntactic, lexical, and argumentative features in an SVM model to analyze public reaction to a specific event, attaining a 90.4% F-score.
The SemEval-2016 shared task [17] provided a formal definition and a dataset of 4,163 tweets, catalyzing further research. The best score attained in the shared task was an F-score of 67.8%. This benchmark was later advanced by Al-Ghadir et al. [5], whose pipeline combining tf-idf with sentiment features achieved a state-of-the-art macro F-score of 76.45%. The IberEval-2017 shared task [20] expanded the scope to include multilingual tweets on Catalan’s independence, combining stance detection with author profiling.
The advent of LLMs has transformed the field, as analyzed in the taxonomy introduced by Pangtey et al. [21], which categorizes approaches by learning methods, data modalities, and target relationships. Methodological progress has since advanced along multiple fronts. Architecturally, Garg and Caragea [22] proposed Stanceformer, which integrates a target awareness matrix to amplify attention to target-relevant terms. In the multimodal domain, Barel et al. [23] introduced TASTE, a model that fuses transformer-based textual embeddings with a structural social context using a gated residual network (GRN). A key innovation in prompting is the “Chain of Stance” (CoS) method by Ma et al. [24], which decomposes the detection process into intermediate reasoning steps for superior zero-shot and few-shot performance.
Substantial work has also focused on improving generalization. Zhao et al. [25] designed a framework using multi-expert cooperative learning, where a gating mechanism fuses diverse semantic features to enhance transfer to unseen targets. Similarly, Ma et al. [26] pioneered a zero-shot multi-agent debate (ZSMD) framework, where opposing agents debate a text’s stance with access to external knowledge for more robust predictions.
Research in Arabic stance detection has progressed from initial studies to rich shared tasks. Abdelhade et al. [27] developed a deep neural network to classify Arabic tweets in financial and sports domains, achieving a 90.68% F-score. A significant large-scale analysis was conducted by Azmi and Al-Ghadir [8], who examined Twitter as a real-time barometer of public opinion during Saudi Arabia’s 2016–2017 reform period. Their study collected approximately 200 million Arabic tweets, identifying 50,000 unique hashtags. From this corpus, they performed a two-tiered analysis: first, a behavioral study clustered 4000 prevalent hashtags into 12 primary categories, from which five key topics—four related to gender relations (e.g., women’s driving and male guardianship)—were selected for in-depth stance detection. In the second phase, they employed a dual-labeling process to classify tweets as supportive, neutral, or opposed to these social changes. Among multiple classifiers evaluated, a weighted 9-NN model achieved the highest performance, with an F-score of 72.45%. This work highlighted Twitter’s utility as a complementary tool to traditional surveys for capturing nuanced public sentiment during socio-political transformations.
The creation of new resources like the ArabicStanceX dataset by Alkhathlan et al. [28] and the multi-dialectal corpus by Charfi et al. [29] provided valuable benchmarks for dialect-aware research. This progression culminated in StanceEval 2024 [30], the first Arabic shared task for stance detection, where fine-tuned AraBERT by Badran et al. [31] achieved the highest F-score (84.38%), underscoring the transformative impact of LLMs in Arabic NLP.
The foundational landscape of the field is thoroughly mapped in the comprehensive survey by Zhang et al. [32], which details core definitions, datasets, and the methodological progression from traditional models to LLM-based approaches, while outlining emergent challenges. For a more focused examination of recent algorithmic advances, the survey by Gera and Neal [33] provides a specialized overview of deep learning techniques within stance detection.

3.2. Terrorism Detection

Recent studies have developed various computational approaches to identify terrorism-related content on social media. Omer [34] proposed a machine learning framework to classify radical vs. non-radical tweets collected using certain ideological hashtags, accepting both Arabic and non-Arabic content. After preprocessing (removing retweet tags, URLs, and annotations), the text was lemmatized and tokenized. Experiments with stylometric, sentiment-based, and time-based features showed that AdaBoost outperformed SVM and NB classifiers when features were combined.
Alsaedi and Burnap [35] developed a robust framework for detecting disruptive events through real-time Twitter analysis. Their methodology leveraged Twitter’s streaming API to collect geolocated data using region-specific keywords and hashtags, with MongoDB employed for efficient data storage. The preprocessing pipeline incorporated stop word removal and content filtering, followed by feature extraction using tf-idf and dual-language stemming (Khoja for Arabic and Porter for English). The system achieved notable performance (85.43% F-score) through a novel combination of NB classification in R and online clustering for topic identification, culminating in a voting-based summarization mechanism. This approach demonstrated particular effectiveness in handling multilingual social media data during crisis events.
For terrorist affiliation detection, Ali [36] collected tweets via Twitter API and preprocessed them by eliminating numbers, stop words, and punctuation. An unsupervised clustering algorithm grouped the data, with cluster numbers determined by frequency histograms. Visualization tools (R, NodeXL, and Gephi) revealed network patterns, demonstrating that combining basic data mining tools with linguistic analysis effectively identifies terrorist affiliations.
The ALT-TERROS system [37] is an AI-enabled framework designed to detect terrorist behavior in Arabic social media. It analyzed 10,000 tweets, assigning suspicion probabilities based on hostile emotions; links to terrorist organizations; and violent imagery. Microsoft Cognitive Services supported semantic tagging, while a risk-scoring mechanism generated alerts. By integrating heuristic rules with advanced analytics and visualizations, ALT-TERROS offers an end-to-end workflow for extracting and evaluating textual and visual data. The system also addresses challenges such as limited Arabic API support and the need for robust language resources to enhance sentiment analysis.
Al-Shawakfa et al. [38] proposed RADAR#, a system combining two neural network approaches—a CNN-BiLSTM model and the AraBERT transformer—using an ensemble technique to identify radicalization indicators in tweets. When tested on nearly 90,000 Arabic tweets, the system achieved high detection accuracy (98%) while revealing important insights through component analysis. Key findings show that the model struggles most with sarcastic posts (a third of the errors) and regional dialect variations (one-fifth of errors), highlighting areas needing improvement for real-world counterterrorism applications. The technical approach demonstrates how combining different AI methods can improve detection of harmful online content in Arabic language contexts.
Another model to detect extremist content in Arabic social media using AraBERT was proposed by [39]. The model combined AraBERT with sentiment features (SFs) and multilayer perceptrons (MLPs). The study evaluated the model on three datasets: (1) 89,816 Arabic tweets with 56% extremist content collected via X’s APIs; (2) 24,078 ISIS-related Arabic tweets having 10,000 extremist contents, analyzed via a 3400-tweet subset; and (3) 100 professionally translated Kazakh texts, half of which is extremist content for cross-lingual validation. These were, respectively, sourced from [40,41,42]. The highest performance achieved was an F-score of 96% on the first dataset using AraBERT with SF and MLP.

3.3. Concluding Remarks

This review examined two interrelated research areas: stance detection and terrorism-related content identification in social media. Current work in both domains demonstrates a clear progression toward deep learning methods, particularly LLMs, as evidenced by recent surveys of advances in stance detection [32,33] and prominent NLP tasks such as abstractive summarization [43] and question answering [44]. However, our analysis reveals a significant research gap in stance detection specifically applied to terrorism-related Arabic tweets. Sabri and Abdullah [45]’s comprehensive survey on Arabic extremism content detection highlights two key findings: (1) most studies focused on MSA despite widespread use of regional dialects in social media, and (2) among traditional machine learning approaches (excluding deep learning), SVM classifiers consistently outperform other algorithms (e.g., Naïve Bayes and Random Forests) in detection accuracy.

4. Our Proposed System

In this section, we present the key aspects of our proposed stance detection system for terrorism-related Arabic tweets. The discussion begins with the processes of data collection and annotation and concludes with the classification phase. This study focuses on terrorism as the target of interest. Recognizing that the definition of terrorism varies across individuals and governments, we adopt the organizations designated as terrorist by the Gulf Cooperation Council (GCC) as our targets of interest. Specifically, this research centers on Hezbollah-related terrorism in Syria. As noted earlier, stance detection is an emerging area of research, and this work represents one of the few studies addressing stance detection in the Arabic language. Our proposed system comprises four main components: data collection and annotation, preprocessing, feature extraction and selection, and stance classification. These components are illustrated in Figure 1.

4.1. Dataset and Annotation

We collected over 8000 tweets using Twitter’s API, which after filtering redundant and irrelevant content resulted in about 7000 usable tweets. The final dataset comprises 7053 tweets discussing Hezbollah, categorized as 3201 supporting, 2352 opposing, and 1500 neutral.
For supervised classification, we employed manual annotation by two native Arabic speakers who labeled tweets as “favor,” “against,” or “neither” (following [46]’s combined neutral/no-stance category). Although labor-intensive, this approach ensured reliable and trustworthy annotations given the task’s sensitivity. Crowdsourcing alternatives like ClickWorker or Amazon Mechanical Turk proved problematic for Arabic texts due to quality control issues [47], making professional annotation essential for this high-stakes domain. Table 1 shows some samples of tweets and their stance.
The dataset was partitioned into the training (70%) and testing (30%) subsets, preserving the original distribution of all stance categories (see Table 2). This standard split ratio, consistent with SemEval-2016’s methodology [17], ensures balanced model development and evaluation.

4.2. Tweet Preprocessing

The preprocessing phase cleans the dataset by removing noise, reducing features, and improving classification performance. We first eliminate non-Arabic content, including special symbols (“@” and “#”), URLs, retweet markers (RTs), and end-of-line characters. Next, we normalize Arabic words to their standard forms (e.g., مرررحبا to مرحبا) and remove Arabic stop words using Python’s NLTK 3.6.5 library. Finally, we tokenize the text into individual word units using standard tokenization methods [47,48].

4.3. Stemming

Stemming reduces words to their base or root form [48]. For instance, various Arabic verb conjugations like قرأ (“he read”), قرأت (“she read”), and سيقرأ (“he will read”) all share the same root meaning of reading. A stemming system consolidates these variants into the single-root form قرأ, significantly simplifying text processing.
This study employs the ISRI stemmer from Python’s NLTK library, an Arabic root-based stemmer developed by the Information Science Research Institute [49]. The stemmer first attempts to find the word’s root in its dictionary, returning the normalized form if the root is not found, rather than the original word.

5. Experiments and Discussion

In this section, we present the experimental setup and discuss the results obtained. Two distinct experimental configurations were designed for the dataset: (i) classification without the application of SMOTE and (ii) classification with SMOTE applied to address class imbalance. Across all experiments, we evaluated the performance of four SVM kernels (linear, RBF, polynomial, and sigmoid), alongside Multinomial Naive Bayes (MNB) and weighted K-Nearest Neighbors (K-NN) classifiers.
The synthetic minority over-sampling technique (SMOTE) is a widely used method for addressing class imbalance in datasets [50]. It generates synthetic samples for the minority class by interpolating between existing minority instances and their nearest neighbors. For example, consider two minority class instances x 1 = ( 2 ,   3 ) and x 2 = ( 4 ,   5 ) in a two-dimensional feature space. SMOTE can create a new synthetic instance x by interpolating between them as follows: x = x 1 + δ · ( x 2 x 1 ) , where δ [ 0 ,   1 ] is a random number. If, for example, δ = 0.5 , then x = ( 3 ,   4 ) . This process reduces the bias towards the majority class during training and enhances classifier performance, particularly in imbalanced classification tasks. First, we cover the evaluation metrics.

5.1. Evaluation Metrics

We evaluated classification performance using standard metrics, namely accuracy, precision (P), recall (R), and the F-score, which are derived from fundamental prediction outcomes. True positives (TPs) and true negatives (TNs) represent correct classifications of positive and negative samples, respectively, while false positives (FPs) and false negatives (FNs) indicate misclassified negative and positive samples. These measures form the basis for all subsequent performance calculations, with FPs and FNs being particularly critical for assessing model reliability in imbalanced datasets.
Precision represents the fraction of correctly predicted positives among all predicted positives: P = TP / ( TP + FP ) , while recall indicates the fraction of correctly identified positives among all actual positives: R = TP / ( TP + FN ) . The F-score provides their harmonic mean:
F 1 = 2 P R P + R .
Following the evaluation protocol established in the SemEval-2016 stance detection shared tasks [17], we employed the macro-averaged F-score evaluation metric, focusing exclusively on the “favor” and “against” classes while treating “neutral” as negative. The F-average ( F avg ) metric is computed as
F avg = F favor + F against 2 ,
where F favor and F against are calculated as
F favor = 2 P favor R favor P favor + R favor ,
F against = 2 P against R against P against + R against ,
where P favor and P against denote the precision values for the classes “favor” and “against”, respectively, while R favor and R against represent the corresponding recall values for these two classes.
This evaluation framework aligns with established stance detection competitions and enables direct performance comparisons across systems [17].

5.2. Word Embedding Using FastText

The word embedding approach represents words as real-valued vectors in a predefined space, capturing semantic relationships by assigning similar vectors to words with related meanings [51].
We employed Facebook’s FastText library, which offers pretrained models for over 290 languages and supports both supervised and unsupervised word vector generation [51]. For our experiment, we trained a supervised FastText model on the datasets, using the default vector size of 300.
Model tuning focused on two key parameters:
  • Epochs: The term refers to the number of times the instance is seen by FastText. The standard range for the number of epochs is 5–50.
  • Learning rate (LR): It refers to the amount of change performed by the model after processing each instance. The standard range for LR is 0.1–1.0.
Our experimental results demonstrate that the optimal configuration of an LR = 0.1, and 25 epochs yielded peak performance with an F-score of 70.86%, as detailed in Table 3.

5.3. Supervised Classifiers

Our feature selection process began with term frequency–inverse document frequency (tf-idf) weighting, which evaluates word importance through a balance of two factors: term frequency tf (how often a word appears in a document) and inverse document frequency df (how rare it is across all documents) [52,53]. It is calculated by
w i j = t f i j · log ( D / df i ) ,
where tf i j denotes occurrences of word i in document j, df i tracks how many documents contain word i, and D is our total document count. This formula automatically boosts distinctive terms while suppressing overly common ones.
When we cross-validated these tf-idf features using Chi-square selection, we found complete agreement—both methods identified the same set of meaningful features. This gave us confidence to proceed with these optimized word representations.
To determine the optimal number of features, we evaluated our system using 100, 500, 1000, 1500, and 2000 features, measuring the impact on F avg across three classifiers: Multinomial Naïve Bayes (MNB), an SVM with the RBF kernel, and weighted K-NN (with K = 1 ). The dataset was partitioned into 70% for training and 30% for testing, adhering to the established evaluation framework used in the SemEval-2016 stance detection shared tasks [17]. We assessed all combinations of unigrams, bigrams, and trigrams across these feature set sizes, leveraging their established effectiveness for Arabic text processing.
The results, as detailed in Table 4, reveal a clear performance trend. The SVM-RBF classifier achieved its highest F avg of 0.7243 using a combination of 1000 unigram and trigram features. A close second result of F avg = 0.7238 was attained using 1000 unigram and bigram features for the same classifier. Both values appear as 0.724 in the table due to rounding. MNB performed fairly well, but weighted 1NN yielded the least favorable performance.
Figure 2 presents the confusion matrix heatmaps for the top-performing feature configuration (1000 unigram and trigram features, N = 1 + 3 from Table 4) under imbalanced data conditions. Several observations can be made. SVM-RBF achieves the most favorable trade-off, yielding both higher TP for the favor class (454) and higher TN for the against class (317). MNB performs reasonably well but suffers from a larger number of false positives (157 cases). In contrast, W-1NN exhibits the weakest performance, misclassifying a substantial portion of both favor and against examples.
Interestingly, the number of TPs for the favor class is almost identical for MNB and SVM-RBF, despite the latter being the stronger classifier overall. The key advantage of SVM-RBF lies not in identifying favor instances but in avoiding the misclassification of against examples as favor. Specifically, MNB produces 157 FPs (against predicted as favor), whereas SVM-RBF reduces this to only 107 FPs. Consequently, although both models capture approximately 79% of the favor class (similar recall), SVM-RBF achieves superior precision by minimizing false positives.
In summary, the strength of SVM-RBF is evident in its effective management of true negatives and false positives rather than true positives alone. This capability leads to consistently higher precision and F avg scores, underscoring its advantages over MNB in handling imbalanced stance detection tasks.

5.4. Hyper-Parameter Optimization

The goal of a learning algorithm A is to find a function f that minimizes the expected loss L ( x ; f ) over samples x drawn from the true data distribution G x . The algorithm A maps the training set X ( train ) to f and depends on hyper-parameters λ , which must be tuned to produce the best-performing model. Thus, the key challenge is selecting the optimal λ that minimizes the generalization error:
λ = argmin λ Λ E x G x L x ; A λ ( X ( train ) ) .
This selection process is known as hyper-parameter optimization [54].
Among the available optimization methods, grid search is particularly effective in low-dimensional spaces due to its simplicity [54]. In this study, we employed grid search to tune the hyper-parameters of the SVM classifier.
The SVM’s performance is highly sensitive to its hyper-parameters, particularly C (regularization parameter) and γ (kernel coefficient). The C parameter controls the trade-off between maximizing training accuracy and ensuring a smooth decision boundary—higher values enforce strict classification of training samples, while lower values prioritize smoother decision surfaces [55]. The γ parameter determines the influence range of a single training instance; higher values mean only nearby points significantly affect the decision boundary [56].
To optimize C and γ , we employed a grid search strategy combined with 10-fold cross-validation on the training dataset. The search space for both parameters was defined as [ 0.01 ,   0.1 ,   1 ,   10 ,   100 ] . The results, presented in Table 5, was evaluated in terms of classification accuracy [54].
Table 6 presents a comparative analysis of stance detection performance on the Hezbollah dataset, evaluating results both without and with the application of SMOTE. The table details the precision, recall, and F-score metrics for the “favor” and “against” classes, in addition to the overall macro-average F-score ( F avg ). Complementary to these results, Figure 3 and Figure 4 illustrate the comparative F avg performance across different classifiers and n-gram feature combinations under both experimental conditions.
Employing a combination of unigram and trigram features, the SVM-RBF classifier yielded the strongest performance in our experiments, achieving an F avg of 72.43% without SMOTE (Figure 3e; see also Table 4) and improving further to 78.62% with SMOTE (Figure 4e).
The analysis of feature combinations revealed that integrating multiple n-grams consistently equaled or exceeded the performance of any single feature type. For instance, without SMOTE, combining unigram and bigram features increased the MNB classifier’s F avg from 68.85% (unigrams alone) and 63.92% (bigrams alone) to 69.41%, demonstrating a clear synergistic effect. Among the classifiers we evaluated—MNB, the SVM (with multiple kernels), and weighted K-NN—MNB delivered competitive results approaching the SVM’s performance while requiring significantly less computation time. The weighted K-NN, while inefficient for our dataset, demonstrated good performance on the SemEval-2016 English dataset, as reported in [5]. In kernel comparisons, RBF consistently outperformed polynomial and sigmoid kernels for our Hezbollah dataset.

5.5. Discussion

In the context of stance detection on extremist groups, recall is a more critical metric than precision. The primary objective is to ensure that supportive or sympathetic stances are comprehensively detected, as overlooking such content (FNs) may result in missed signals of radicalization, recruitment, or propaganda dissemination, with direct implications for national security. While precision remains important for reducing false alarms and avoiding misallocation of resources, its trade-off is less consequential than recall in high-stakes domains. From an operational standpoint, it is preferable to flag more potential threats—even at the cost of some FPs—than to miss genuine supportive expressions. Thus, models for this task must be optimized for high recall while maintaining an acceptable level of precision to ensure feasibility in deployment.
The application of SMOTE underscores a critical divergence in how classifiers balance the precision–recall trade-off in stance detection, as evidenced by the confusion matrices in Figure 5. When compared to the non-SMOTE performance (Figure 2), both MNB and SVM-RBF exhibit improved specificity against the “against” class, with FPs decreasing from 157 to 103 and from 107 to 77, respectively. However, their behavior diverges sharply in terms of recall. MNB undergoes a conservative shift that severely compromises its sensitivity: TPs for the “favor” class drop from 453 to 405, while FNs increase from 123 to 171. This significant loss of recall—wherein numerous genuine supportive stances are overlooked—renders MNB unsuitable for applications such as counter-extremism, where detecting subtle or early signals of radicalization depends on high recall. In such contexts, failing to capture authentic “favor” instances poses a far greater risk than misclassifying a limited number of negative examples.
At the same time, the ethical implications of FPs cannot be ignored. Each misclassification of an “against” stance as “favor” constitutes a false accusation, potentially stigmatizing individuals as sympathizers of extremist groups when in fact they are not. This risk of creating “false victims” underscores the necessity of maintaining strong precision to protect individuals from wrongful suspicion and to preserve trust in automated systems. In this respect, SVM-RBF demonstrates a more desirable profile: it maintains high recall with only a marginal decrease in TPs (454 to 441) while simultaneously achieving greater precision by substantially reducing FPs and increasing TNs (317 to 347). By balancing the need to capture supportive stances (high recall) with the responsibility to avoid wrongful accusations (high precision), SVM-RBF emerges as the more robust and socially responsible classifier. This dual consideration highlights that effective stance detection in sensitive domains must optimize not only for security imperatives but also for ethical safeguards.
We observed an interesting performance gap between FastText ( F avg = 70.86 % , Table 3) and SVM-RBF ( F avg = 78.62 % ). While we included FastText for its distinct neural approach and proven effectiveness with morphologically rich languages like Arabic, two main factors contributed to this difference: (1) Our dataset size was relatively limited for optimal FastText performance, which typically benefits from larger corpora or pretrained embeddings, while SVM-RBF excels with smaller datasets when using well-engineered features like tf-idf and n-grams. (2) FastText’s shallow architecture is less capable of modeling complex decision boundaries compared to SVM-RBF’s kernel-based approach.
We maintained FastText in our comparisons to provide a comprehensive methodological perspective and demonstrate how different approaches perform under identical conditions. Although it underperformed relative to SVM-RBF in this specific context, FastText remains a valuable baseline that could potentially achieve better results with larger-scale pretraining or enhanced preprocessing techniques.

6. Conclusions

This study aims to develop an automated system for stance detection in Arabic tweets, focusing on Hezbollah’s activities in Syria as a case study. The proposed algorithm addresses the unique challenges of Arabic tweets, including dialectal variations in stance identification, as noted by the annotators. Despite these complexities, the system demonstrates strong performance by first preprocessing tweets using the ISRI stemmer, then converting them into feature vectors via tf-idf vectorization. These vectors serve as input for multiple classifiers, including MNB, SVMs, and weighted K-NN.
The experiments were conducted on a dataset of approximately 7000 tweets from Hezbollah-related discussions. The results indicate that the system performs effectively, particularly when the dataset is balanced. We achieved an F avg score of 78.63%—a competitive result compared to state-of-the-art performance in English tweet stance detection, where the highest reported F-average score for the SemEval-2016 dataset was 76%. This underscores the system’s capability to handle the linguistic nuances of Arabic social media content.
For future work we plan to incorporate a profiler, similar to the one in [5]. This profiler would enable the system to infer demographic attributes of users, such as gender and age group, thereby providing richer contexts for stance detection and content analysis.

Author Contributions

Conceptualization, A.K.A. and A.M.A.; methodology, A.K.A.; software, A.K.A.; validation, A.K.A.; formal analysis, A.M.A. and A.K.A.; investigation, A.K.A.; resources, A.M.A.; data curation, A.K.A.; writing—original draft preparation, A.K.A.; writing—review and editing, A.M.A.; supervision, A.M.A.; funding acquisition, A.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors would like to thank the Ongoing Research Funding Program (ORFFT-2025-006-4), King Saud University, Riyadh, Saudi Arabia for financial support.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author. They are available for non-commercial research purposes.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Backlinko Team. Social Media Users & Growth Statistics. 2024. Available online: https://backlinko.com/social-media-users (accessed on 9 June 2025).
  2. DataReportal/Kepios. Global Social Media Statistics. 2025. Available online: https://datareportal.com/social-media-users (accessed on 12 June 2025).
  3. Boczkowski, P.J.; Matassi, M.; Mitchelstein, E. How Young Users Deal With Multiple Platforms: The Role of Meaning-Making in Social Media Repertoires. J. Comput.-Mediat. Commun. 2018, 23, 245–259. [Google Scholar] [CrossRef]
  4. Zubiaga, A.; Aker, A.; Bontcheva, K.; Liakata, M.; Procter, R. Detection and resolution of rumours in social media: A survey. Acm Comput. Surv. (Csur) 2018, 51, 1–36. [Google Scholar] [CrossRef]
  5. Al-Ghadir, A.I.; Azmi, A.M.; Hussain, A. A novel approach to stance detection in social media tweets by fusing ranked lists and sentiments. Inf. Fusion 2020, 67, 29–40. [Google Scholar] [CrossRef]
  6. Williamson, M. Terrorism, War and International Law: The Legality of the Use of Force Against Afghanistan in 2001; Routledge: London, UK, 2016. [Google Scholar]
  7. Kruglanski, A.W.; Fishman, S. Terrorism between “syndrome” and “tool”. Curr. Dir. Psychol. Sci. 2006, 15, 45–48. [Google Scholar] [CrossRef]
  8. Azmi, A.M.; Al-Ghadir, A.I. Using Twitter as a digital insight into public stance on societal behavioral dynamics. J. King Saud-Univ.-Comput. Inf. Sci. 2024, 36, 102078. [Google Scholar] [CrossRef]
  9. Callegaro, M.; Yang, Y. The role of surveys in the era of “big data”. In The Palgrave Handbook of Survey Research; Palgrave Macmillan: Cham, Switzerland, 2017; pp. 175–192. [Google Scholar]
  10. Bruns, A.; Stieglitz, S. Twitter Data: What do they represent? IT—Inf. Technol. 2014, 56, 240–245. [Google Scholar] [CrossRef]
  11. Alruily, M.; Manaf Fazal, A.; Mostafa, A.M.; Ezz, M. Automated Arabic long-tweet classification using transfer learning with BERT. Appl. Sci. 2023, 13, 3482. [Google Scholar] [CrossRef]
  12. Azmi, A.M.; Aljafari, E.A. Universal web accessibility and the challenge to integrate informal Arabic users: A case study. Univers. Access Inf. Soc. 2018, 17, 131–145. [Google Scholar] [CrossRef]
  13. Almuzaini, H.A.; Azmi, A.M. TaSbeeb: A judicial decision support system based on deep learning framework. J. King Saud-Univ.-Comput. Inf. Sci. 2023, 35, 101695. [Google Scholar] [CrossRef]
  14. Al-Jouie, M.F.; Azmi, A.M. Automated Evaluation of School Children Essays in Arabic. In Proceedings of the 3rd International Conference on Arabic Computational Linguistics (ACLing 2017), Dubai, United Arab Emirates, 5–6 November 2017; Volume 117, pp. 19–22. [Google Scholar]
  15. Alqahtani, A.; Al-Saif, A. Automated Arabic essay evaluation. In Proceedings of the 17th International Conference on Natural Language Processing (ICON), Patna, India, 18–21 December 2020; pp. 181–190. [Google Scholar]
  16. AlShenaifi, N.; Azmi, A. Arabic dialect identification using machine learning and transformer-based models. In Proceedings of the Seventh Arabic Natural Language Processing Workshop (WANLP 2022), Abu Dhabi, United Arab Emirates, 8 December 2022; pp. 464–467. [Google Scholar]
  17. Mohammad, S.; Kiritchenko, S.; Sobhani, P.; Zhu, X.; Cherry, C. Semeval-2016 Task 6: Detecting stance in tweets. In Proceedings of the SemEval, San Diego, CA, USA, 16–17 June 2016; pp. 31–41. [Google Scholar]
  18. Somasundaran, S.; Wiebe, J. Recognizing stances in ideological on-line debates. In Proceedings of the NAACL HLT 2010 Workshop on Computational Approaches to Analysis and Generation of Emotion in Text, Los Angeles, CA, USA, 1–6 June 2010; pp. 116–124. [Google Scholar]
  19. Addawood, A.; Schneider, J.; Bashir, M. Stance Classification of Twitter Debates: The Encryption Debate as A Use Case. In Proceedings of the 8th International Conference on Social Media & Society, Toronto, ON, Canada, 28–30 July 2017. [Google Scholar]
  20. Taulé, M.; Martí, A.; Rangel, F.; Rosso, P.; Bosco, C.; Patti, V. Overview of the Task on Stance and Gender Detection in Tweets on Catalan Independence at IberEval 2017. In Proceedings of the 2nd Workshop on Evaluation of Human Language Technologies for Iberian Languages, Murcia, Spain, 19 September 2017. [Google Scholar]
  21. Pangtey, L.; Choudhary, R.; Mehta, P. Large Language Models Meet Stance Detection: A Survey of Tasks, Methods, Applications, Challenges and Future Directions. arXiv 2025, arXiv:2505.08464. [Google Scholar] [CrossRef]
  22. Garg, K.; Caragea, C. Stanceformer: Target-Aware Transformer for Stance Detection. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, FL, USA, 12–16 November 2024; pp. 4969–4984. [Google Scholar]
  23. Barel, G.; Tsur, O.; Vilenchik, D. Acquired TASTE: Multimodal Stance Detection with Textual and Structural Embeddings. In Proceedings of the 31st International Conference on Computational Linguistics, Abu Dhabi, United Arab Emirates, 19–24 January 2025; pp. 6492–6504. [Google Scholar]
  24. Ma, J.; Wang, C.; Xing, H.; Zhao, D.; Zhang, Y. Chain of stance: Stance detection with large language models. In Natural Language Processing and Chinese Computing (NLPCC 2024); Lecture Notes in Computer Science; Wong, D.F., Wei, Z., Yang, M., Eds.; Springer: Singapore, 2024; Volume 15363, pp. 82–94. [Google Scholar]
  25. Zhao, X.; Li, F.; Wang, H.; Zhang, M.; Liu, Y. Zero-Shot Stance Detection Based on Multi-Expert Collaboration. Sci. Rep. 2024, 14, 15923. [Google Scholar] [CrossRef]
  26. Ma, J.; Qian, Y.; Yang, J. Exploring Multi-Agent Debate for Zero-Shot Stance Detection: A Novel Approach. Appl. Sci. 2025, 15, 4612. [Google Scholar] [CrossRef]
  27. Abdelhade, N.; Soliman, T.; Ibrahim, H. Detecting Twitter users’ opinions of Arabic comments during various time episodes via deep neural network. In Proceedings of the International Conference on Advanced Intelligent Systems and Informatics, Cairo, Egypt, 1–3 September 2018. [Google Scholar]
  28. Alkhathlan, A.; Alahmadi, F.; Kateb, F.; Al-Khalifa, H. Constructing and evaluating ArabicStanceX: A social media dataset for Arabic stance detection. Front. Artif. Intell. 2025, 8, 1615800. [Google Scholar] [CrossRef]
  29. Charfi, A.; Khene, M.; Belguith, L.H. Stance Detection in Arabic with a Multi-Dialectal Cross-Domain Stance Corpus. Soc. Netw. Anal. Min. 2024, 14, 69. [Google Scholar] [CrossRef]
  30. Alturayeif, N.; Luqman, H.; Alyafeai, Z.; Yamani, A. StanceEval 2024: The first Arabic stance detection shared task. In Proceedings of the Second Arabic Natural Language Processing Conference, Bangkok, Thailand, 16 August 2024; pp. 774–782. [Google Scholar]
  31. Badran, M.; Hamdy, M.; Torki, M.; El-Makky, N.M. AlexUNLP-BH at StanceEval2024: Multiple contrastive losses ensemble strategy with multi-task learning for stance detection in Arabic. In Proceedings of the Second Arabic Natural Language Processing Conference, Bangkok, Thailand, 16 August 2024; pp. 823–827. [Google Scholar]
  32. Zhang, B.; Jiang, Y.; Wang, L.; Wang, Z.; Li, X. A Survey of Stance Detection on Social Media: New Directions and Perspectives. arXiv 2024, arXiv:2409.15690. [Google Scholar] [CrossRef]
  33. Gera, P.; Neal, T. Deep Learning in Stance Detection: A Survey. ACM Comput. Surv. 2025, 58, 1–37. [Google Scholar] [CrossRef]
  34. Omer, E. Using Machine Learning to Identify Jihadist Messages on Twitter. In Proceedings of the 2015 European Intelligence and Security Informatics Conference, Manchester, UK, 7–9 September 2015. [Google Scholar]
  35. Alsaedi, N.; Burnap, P. Arabic event detection in social media. In Computational Linguistics and Intelligent Text Processing (CICLing 2015); Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2015; Volume 9041, pp. 384–401. [Google Scholar]
  36. Ali, G. Identifying Terrorist Affiliations through Social Network Analysis Using Data Mining Techniques. Master’s Thesis, Valparaiso University, Valparaiso, IN, USA, 2016. [Google Scholar]
  37. Alhalabi, W.; Jussila, J.; Jambi, K.; Visvizi, A.; Qureshi, H.; Lytras, M.; Malibari, A.; Adham, R. Social mining for terroristic behavior detection through Arabic tweets characterization. Future Gener. Comput. Syst. 2021, 116, 132–144. [Google Scholar] [CrossRef]
  38. Al-Shawakfa, E.M.; Alsobeh, A.M.; Omari, S.; Shatnawi, A. RADAR#: An Ensemble Approach for Radicalization Detection in Arabic Social Media Using Hybrid Deep Learning and Transformer Models. Information 2025, 16, 522. [Google Scholar] [CrossRef]
  39. Himdi, H.; Alhayan, F.; Shaalan, K. Neural Networks and Sentiment Features for Extremist Content Detection in Arabic Social Media. Int. Arab J. Inf. Technol. (IAJIT) 2025, 22, 522–534. [Google Scholar] [CrossRef]
  40. Aldera, S.; Emam, A.; Al-Qurishi, M.; Alrubaian, M.; Alothaim, A. Annotated Arabic Extremism Tweets. IEEE Dataport, 8 August 2021. [Google Scholar] [CrossRef]
  41. Fraiwan, M. Identification of Markers and Artificial Intelligence-Based Classification of Radical Twitter Data. Appl. Comput. Inform 2022. [Google Scholar] [CrossRef]
  42. Mussiraliyeva, S.; Bolatbek, M.; Omarov, B.; Bagitova, K. Detection of extremist ideation on social media using machine learning techniques. In Proceedings of the International Conference on Computational Collective Intelligence (ICCCI 2020); Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2020; Volume 12496, pp. 743–752. [Google Scholar]
  43. Almohaimeed, N.; Azmi, A.M. Abstractive text summarization: A comprehensive survey of techniques, systems, and challenges. Comput. Sci. Rev. 2025, 57, 100762. [Google Scholar] [CrossRef]
  44. Abdel-Nabi, H.; Awajan, A.; Ali, M.Z. Deep learning-based question answering: A survey. Knowl. Inf. Syst. 2023, 65, 1399–1485. [Google Scholar] [CrossRef]
  45. Sabri, R.F.; Abdullah, N.A. A Review for Arabic Extremism Detection Using Machine Learning. Iraqi J. Sci. 2024, 65, 6617–6630. [Google Scholar] [CrossRef]
  46. Mohammad, S.; Kiritchenko, S.; Sobhani, P.; Zhu, X.; Cherry, C. A dataset for detecting stance in tweets. In Proceedings of the 10th Edition of the the Language Resources and Evaluation Conference (LREC), Portorož, Slovenia, 23–28 May 2016. [Google Scholar]
  47. Almuzaini, H.A.; Azmi, A.M. An unsupervised annotation of Arabic texts using multi-label topic modeling and genetic algorithm. Expert Syst. Appl. 2022, 203, 117384. [Google Scholar] [CrossRef]
  48. Al-Zyoud, A.; Al-Rabayah, W. Arabic Stemming Techniques: Comparisons and New Vision. In Proceedings of the 8th IEEE GCC Conference and Exhibition, Muscat, Oman, 1–4 February 2015. [Google Scholar]
  49. Taghva, K.; Elkhoury, R.J.; Coombs, J. Arabic stemming without a root dictionary. In Proceedings of the International Conference on Information Technology: Coding and Computing (ITCC’05)-Volume II, Las Vegas, NV, USA, 4–6 April 2005; Volume 1, pp. 152–157. [Google Scholar]
  50. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  51. Bojanowski, P.; Grave, E.; Joulin, A.; Mikolov, T. Enriching Word Vectors with Subword Information. Trans. Assoc. Comput. Linguist. 2017, 5, 135–146. [Google Scholar] [CrossRef]
  52. Sparck Jones, K. A Statistical Interpretation of Term Specificity and Its Application in Retrieval. J. Doc. 1972, 28, 11–21. [Google Scholar] [CrossRef]
  53. Salton, G.; Buckley, C. Term-weighting approaches in automatic text retrieval. Inf. Process. Manag. 1988, 24, 513–523. [Google Scholar] [CrossRef]
  54. Bergstra, J.; Bengio, Y. Random Search for Hyper-Parameter Optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  55. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  56. Schölkopf, B.; Burges, C.J.; Vapnik, V.N. Comparing Support Vector Machines with Gaussian Kernels to Radial Basis Function Classifiers. In Proceedings of the 9th International Conference on Artificial Neural Networks (ICANN), Edinburgh, UK, 7–10 September 1997; pp. 63–72. [Google Scholar]
Figure 1. Our proposed terrorism stance detection system.
Figure 1. Our proposed terrorism stance detection system.
Mathematics 13 02965 g001
Figure 2. The confusion matrix heapmap for 1000 features (unigram and trigram) for three different classifiers.
Figure 2. The confusion matrix heapmap for 1000 features (unigram and trigram) for three different classifiers.
Mathematics 13 02965 g002
Figure 3. The behavior of F avg without SMOTE across different n-gram combinations and classifiers: (a) unigram, (b) bigram, (c) trigram, (d) unigram + bigram, (e) unigram + trigram, and (f) bigram + trigram.
Figure 3. The behavior of F avg without SMOTE across different n-gram combinations and classifiers: (a) unigram, (b) bigram, (c) trigram, (d) unigram + bigram, (e) unigram + trigram, and (f) bigram + trigram.
Mathematics 13 02965 g003
Figure 4. The behavior of F avg with SMOTE (Synthetic Minority Over-sampling Technique) for identical n-gram combinations and classifiers as in the previous figure. To enable direct comparison, the y-axis limits are kept identical.
Figure 4. The behavior of F avg with SMOTE (Synthetic Minority Over-sampling Technique) for identical n-gram combinations and classifiers as in the previous figure. To enable direct comparison, the y-axis limits are kept identical.
Mathematics 13 02965 g004
Figure 5. The confusion matrix heatmaps for unigram and trigram features using two classifiers trained with SMOTE. For comparability with the earlier results, the dataset size is normalized to 1000 instances.
Figure 5. The confusion matrix heatmaps for unigram and trigram features using two classifiers trained with SMOTE. For comparability with the earlier results, the dataset size is normalized to 1000 instances.
Mathematics 13 02965 g005
Table 1. Sample tweets from the dataset, accompanied by their translations and associated stance labels.
Table 1. Sample tweets from the dataset, accompanied by their translations and associated stance labels.
TweetStance
أصلا حزب الله وداعش حلفاء ولم يتعرض حزب الله لداعش في سوريا أبدا وكلاهما كان يقتل في الشعب السوري وكلاهما صناعة إسرائيلية وما وراء ذلك كله تمثيلAgainst
In fact, Hezbollah and ISIS are allies, and Hezbollah never confronted ISIS in Syria; both were killing the Syrian people, both are an Israeli creation, and everything beyond that is mere theatrics.
قَسَم لو ما قَسَم مايفرق.. بعد لو يحط ايد عالقرآن وإيد على عيونه عساها العمى ويحلف حلف مايفيده،حزب الله بيرد بيرد والسيد ما يثنّي كلمته، بس اعطوا هالنتن كلينكس لايصيحFavor
Whether he swears an oath or not, it makes no difference. Even if he puts one hand on the Qur’an and the other over his eyes—may they go blind—and swears the strongest oath, it won’t help him. Hezbollah will definitely respond, and “the Sayyid” doesn’t go back on his word. Just give that stinker a Kleenex so he doesn’t start crying.
والله مو كاسر خاطري الا الايتام الي بالدول العربيه الي فقدوا اهلهم بالربيع العربي تخيل فجاه تصير بدون عائله وبدون وطن وعمرك صغير وعندك اخوات كنت اتمنى من سفارات دول الخليج انها تتبنى الايتام وتوفر لهم سكن وملبس وتعليم خصوصا دول الخليج مقتدره ماديا #سورياNeutral
By God, what breaks my heart most are the orphans in Arab countries who lost their families during the Arab Spring. Imagine suddenly being without a family or a homeland, being so young and having siblings. I wish the embassies of the Gulf states would take in these orphans and provide housing, clothing, and education especially the Gulf countries which are financially capable. #Syria
هههههه تمثيلياتكم وأفلامكم أنتهت والجيش السوري بالتعاون مع روسيا وإيران يحررون سوريا من ما تبقي من جرذان الغرب و المدن التي يحررها دائما ما يمهل فترة للمدنيين ومن يبقون هم الخونة والعملاء فقط وعلي أساس الخونة مسالمون وليسوا بلطجية وإرهابيين ومسلحينFavor
Hahaha—your theatrics and films are over. The Syrian army, in cooperation with Russia and Iran, is liberating Syria from what remains of the West’s “rats.” In the cities it liberates, it always gives civilians a grace period, and those who remain are only the traitors and collaborators. As if those traitors were “peaceful” and not thugs, terrorists, and armed men.
الاف القتلى المسلمين السنة في سوريا و تهجير الملايين المسلمين السنة السوريين بمساعدة الحرس الثوري الايراني و الميليشيات الشيعية العنصرية الارهابية من حشد شعبي و ما يسمى حزب الله و حتى الجيش الروسي و انت تستغرب !!‏Against
Thousands of Sunni Muslims have been killed in Syria, and millions of Syrian Sunni Muslims have been displaced with the help of the Iranian Revolutionary Guard, sectarian terrorist Shiite militias from the Popular Mobilization Forces, the so-called Hezbollah, and even the Russian army, and you’re surprised?!
Table 2. The distribution of tweets in our Hezbollah dataset across the three stance categories: in favor, against, and neutral.
Table 2. The distribution of tweets in our Hezbollah dataset across the three stance categories: in favor, against, and neutral.
Training (70%)Testing (30%)Total
Favor22419603201
Against16467062352
Neither10504501500
Total493721167053
Table 3. Results of word embedding using FastText. We boldface the best entry.
Table 3. Results of word embedding using FastText. We boldface the best entry.
Learning RateEpochsF-Score
0 (default)5 (default)0.1361
0.1250.7086
0.2250.7006
0.3250.6957
1.0250.6851
1.0500.6847
Table 4. Performance tuples ( P favor , R favor , F favor , P against , R against , F against , and F avg ) for MNB, the SVM with the RBF kernel, and W-1NN across different feature counts and n-gram configurations. The first six tuples are, respectively, designated by P + , R + , F + , P , R , and F . The N refers to the size of the N-gram, with 1 refering to unigram, 1 + 3 refering to unigram and trigram (combined), and so on. For each classifier we boldface the best F avg .
Table 4. Performance tuples ( P favor , R favor , F favor , P against , R against , F against , and F avg ) for MNB, the SVM with the RBF kernel, and W-1NN across different feature counts and n-gram configurations. The first six tuples are, respectively, designated by P + , R + , F + , P , R , and F . The N refers to the size of the N-gram, with 1 refering to unigram, 1 + 3 refering to unigram and trigram (combined), and so on. For each classifier we boldface the best F avg .
MNBSVM-RBFW-1NN
N # Feature P + R + F + P R F F avg P + R + F + P R F F avg P + R + F + P R F F avg
11000.6750.8060.7350.5460.5590.5520.6440.8020.7710.7860.6430.6060.6240.7050.6640.6830.6730.5020.4510.475 0.574
5000.7140.7960.7530.5730.6630.6150.6840.8140.7790.7960.6390.6400.6390.7180.5270.6380.5770.3880.4470.4150.496
10000.7350.7880.7610.5620.6820.6160.6890.8070.7890.7980.6500.6410.6460.7220.5180.5570.5370.3540.5020.4150.476
15000.7450.7810.7630.5620.7130.6290.6960.7990.7820.7900.6440.6560.6500.7200.9280.1570.2690.3560.9770.5220.395
20000.7410.7790.7590.5490.7120.6200.6900.8040.7810.7920.6420.6560.6490.7200.9490.1520.2620.3560.9840.5230.393
21000.5780.8750.6960.5880.2880.3870.5420.5950.8450.6980.5430.4160.4710.5850.6720.3070.4220.3420.7170.4630.442
5000.6590.8030.7240.5510.5010.5250.6240.6550.7900.7160.5590.5370.5480.6320.6790.4690.5550.4100.6160.4920.523
10000.6800.7850.7290.5530.5460.5500.6390.6720.7800.7220.5530.5680.5610.6410.6520.5880.6190.4190.5040.4570.538
15000.6950.7760.7330.5460.5750.5600.6470.6770.7770.7230.5560.5990.5770.6500.6490.6200.6340.4380.4900.4630.548
20000.7140.7720.7420.5440.6110.5760.6590.6870.7780.7290.5520.6030.5770.6530.6250.6400.6330.4350.4350.4350.534
31000.4810.9620.6410.5000.0820.1410.3910.4820.9610.6420.4960.0950.1600.4010.7690.1350.2300.3320.9220.4880.359
5000.5060.9490.6600.5830.1860.2820.4710.5060.9510.6610.6010.1960.2960.4780.8010.2000.3200.3440.9000.4970.409
10000.5130.9320.6620.5710.2280.3260.4940.5130.9350.6630.5860.2240.3240.4930.7650.2410.3670.3530.8810.5040.435
15000.5270.9210.6700.5820.2690.3680.5190.5240.9250.6690.5990.2530.3560.5130.7680.2560.3840.3530.8550.4990.441
20000.5340.9130.6740.5910.2900.3890.5310.5310.9160.6720.6030.2800.3820.5270.7700.2790.4100.3600.8490.5060.458
1 + 21000.6780.8040.7360.5530.5550.5540.6450.7960.7720.7840.6360.5930.6140.6990.6670.6880.6770.4970.4420.4680.572
5000.7160.7970.7550.5850.6570.6190.6870.8190.7790.7980.6410.6460.6430.7210.5400.6590.5940.4100.4700.4380.516
10000.7420.7900.7650.5780.6760.6230.6940.8070.7870.7970.6530.6490.6510.7240.5330.5890.5590.3670.5040.4240.492
15000.7590.7870.7730.5780.7040.6350.7040.8090.7820.7950.6450.6560.6500.7230.9150.1640.2790.3560.9690.5200.400
20000.7600.7810.7700.5790.7280.6450.7080.8030.7870.7950.6440.6500.6470.7210.9350.1610.2750.3560.9770.5220.399
1 + 31000.6780.8040.7360.5530.5550.5540.6450.7960.7720.7840.6360.5930.6140.6990.6670.6880.6770.4970.4420.4680.572
5000.7170.7930.7530.5820.6590.6180.6860.8210.7800.8000.6430.6490.6460.7230.5410.6560.5930.4110.4760.4410.517
10000.7430.7870.7640.5770.6780.6240.6940.8090.7890.7990.6530.6470.6500.7240.5260.5900.5560.3590.4850.4120.484
15000.7580.7860.7720.5770.7030.6340.7030.8080.7830.7950.6450.6540.6500.7220.9170.1670.2830.3560.9690.5210.402
20000.7660.7800.7730.5820.7350.6490.7110.8050.7880.7970.6420.6520.6470.7220.9360.1630.2780.3560.9770.5220.400
2 + 31000.5780.8670.6930.5690.3020.3940.5440.5930.8430.6960.5350.4200.4710.5840.6560.2980.4100.3390.7230.4610.436
5000.6590.8000.7230.5500.4930.5200.6210.6510.7920.7150.5570.5140.5350.6250.6930.4590.5520.4110.6250.4960.524
10000.6790.7730.7230.5400.5450.5420.6320.6720.7710.7180.5460.5680.5570.6370.6600.5600.6060.4130.5270.4630.534
15000.6900.7660.7260.5420.5710.5560.6410.6810.7720.7240.5550.5990.5760.6500.6640.5870.6230.5100.4170.4590.541
20000.7040.7700.7350.5340.5920.5620.6480.6900.7790.7320.5490.6000.5730.6530.6580.6170.6370.4530.5110.4800.558
Table 5. Hyper-parameter tuning results for our dataset for the SVM classifier.
Table 5. Hyper-parameter tuning results for our dataset for the SVM classifier.
Kernel #Best CBest γ
Linear1.00.01
RBF1.01.0
Polynomial0.0110
Sigmoid100.01
Table 6. Results for our dataset without SMOTE and after applying it. N refers to the size of the N-gram combinations (same as in Table 4). For each N, we boldface the best F avg .
Table 6. Results for our dataset without SMOTE and after applying it. N refers to the size of the N-gram combinations (same as in Table 4). For each N, we boldface the best F avg .
Without SMOTEWith SMOTE
N Model P favor R favor F favor P against P favor F against F avg P favor R favor F favor P against P favor F against F avg
1MNB0.7350.7880.76060.5620.6820.61640.68850.7890.6880.73540.6490.7160.68090.7081
SVM linear0.7940.7770.78520.6380.6220.63000.70760.8470.7180.77720.7200.6980.70890.7430
SVM RBF0.8070.7890.79790.6500.6410.64550.72170.8450.7690.80530.7740.7490.7613 0.7833
SVM poly0.6250.8900.73470.6530.5650.60590.67030.9730.3290.49160.5630.9320.70170.5966
SVM sigmoid0.7280.8200.77120.6000.6570.62750.69930.8600.7030.77340.7230.6610.69060.7320
W-1NN0.5180.5570.53660.3540.5020.41500.47580.9180.1720.28910.5740.7620.65490.4720
W-5NN0.6410.7790.70320.5090.5900.54680.62500.9110.1040.18690.5170.7330.60670.3968
W-10NN0.6640.7800.71730.5510.6030.57580.64650.9590.0730.13480.5190.7670.61860.3767
2MNB0.6800.7850.72880.5530.5460.54970.63920.7870.5400.64080.5050.6910.58370.6122
SVM linear0.6420.8170.71930.6040.4880.53970.62950.7680.5650.65090.6830.5380.60160.6263
SVM RBF0.6720.7800.72200.5530.5680.56060.64130.7940.5640.65940.6780.5880.6298 0.6446
SVM poly0.5390.8770.66790.5260.2500.33920.50360.8210.4300.56430.7450.4270.54270.5535
SVM sigmoid0.5030.9610.66080.5970.2040.30340.48210.8120.5380.64740.7140.4180.52740.5874
W-1NN0.6520.5880.61860.4190.5040.45740.53800.7530.4940.59670.6720.5460.60260.5997
W-5NN0.6130.6360.62410.4140.4600.43580.52990.6090.6410.62480.6630.5080.57490.5998
W-10NN0.5840.6760.62650.4060.4060.40580.51620.6020.6210.61130.6750.4570.54480.5781
3MNB0.5130.9320.66210.5710.2280.32630.49420.8390.2340.36580.3600.9200.51720.4415
SVM linear0.5070.9430.65930.5910.2010.29940.47930.8500.2310.36300.6060.2240.32670.3449
SVM RBF0.5130.9350.66250.5860.2240.32410.49330.8090.2510.38340.6150.2270.33150.3574
SVM poly0.5050.9430.65740.5740.1760.26900.46320.8420.2070.33270.6470.1990.30420.3185
SVM sigmoid0.4670.9940.63530.4120.0100.02000.32760.9790.0970.17650.3410.9850.50690.3417
W-1NN0.7650.2410.36700.3530.8810.50390.43540.3790.9020.53360.5870.1890.28610.4099
W-5NN0.7570.2430.36830.3520.8890.50410.43620.7470.2530.37830.3580.8960.5119 0.4451
W-10NN0.5100.9210.65670.5630.2210.31750.48710.7630.2490.37560.3570.8860.50870.4422
1 + 2MNB0.7420.7900.76520.5780.6760.62300.69410.7980.7050.74830.6700.7140.69130.7198
SVM linear0.7950.7760.78540.6380.6310.63430.70990.8540.7300.78740.7340.7170.72530.7564
SVM RBF0.8070.7870.79670.6530.6490.65090.72380.8500.7660.80600.7750.7550.7647 0.7853
SVM poly0.6340.8940.74190.6480.5680.60530.67360.9760.3390.50340.5650.9240.70120.6023
SVM sigmoid0.7310.8200.77340.6010.6430.62090.69720.8640.6990.77240.7130.6590.68490.7286
W-1NN0.5330.5890.55940.3670.5040.42440.49190.9320.1810.30280.5660.7520.64600.4744
W-5NN0.6420.7780.70370.5180.5960.55410.62890.9130.1180.20970.5290.7530.62110.4154
W-10NN0.6740.7890.72680.5470.5970.57100.64890.9620.0780.14360.5130.7660.61450.3791
1 + 3MNB0.7430.7870.76410.5770.6780.62350.69380.7970.7040.74720.6720.7110.69070.7189
SVM linear0.7950.7730.78390.6380.6300.63370.70880.8480.7300.78480.7330.7180.72560.7552
SVM RBF0.8090.7890.79870.6530.6470.65000.72430.8520.7650.80620.7750.7580.7663 0.7862
SVM poly0.6350.8940.74220.6450.5650.60260.67240.9740.3380.50180.5680.9260.70420.6030
SVM sigmoid0.7330.8220.77470.5990.6380.61800.69630.8590.6960.76860.7120.6530.68080.7247
W-1NN0.5260.5900.55610.3590.4850.41240.48430.9330.1840.30710.5670.7490.64530.4762
W-5NN0.6440.7780.70460.5180.5970.55470.62960.9230.1230.21640.5280.7490.61920.4178
W-10NN0.6770.7870.72790.5530.6060.57860.65320.9650.0840.15410.5130.7620.61330.3837
2 + 3MNB0.6790.7730.72260.5400.5450.54220.63240.7800.5280.62970.4960.6840.57530.6025
SVM linear0.6450.8090.71760.5890.4800.52900.62330.7510.5570.63920.6740.5190.58650.6129
SVM RBF0.6720.7710.71790.5460.5680.55660.63730.7860.5560.65110.6820.5970.6364 0.6437
SVM poly0.5480.8630.67000.5390.2740.36310.51650.7900.4460.57040.7560.4390.55510.5627
SVM sigmoid0.5060.9580.66240.5940.2170.31750.49000.7950.5300.63600.7120.4350.54030.5881
W-1NN0.6600.5600.60590.4130.5270.46300.53440.5900.6730.62880.6880.5470.60950.6191
W-5NN0.6360.6030.61910.4140.4880.44780.53350.5900.6520.61940.6700.5030.57490.5971
W-10NN0.5970.6410.61830.3930.4190.40560.51200.5930.6500.62010.6750.4670.55180.5859
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alkhraiji, A.K.; Azmi, A.M. Stance Detection in Arabic Tweets: A Machine Learning Framework for Identifying Extremist Discourse. Mathematics 2025, 13, 2965. https://doi.org/10.3390/math13182965

AMA Style

Alkhraiji AK, Azmi AM. Stance Detection in Arabic Tweets: A Machine Learning Framework for Identifying Extremist Discourse. Mathematics. 2025; 13(18):2965. https://doi.org/10.3390/math13182965

Chicago/Turabian Style

Alkhraiji, Arwa K., and Aqil M. Azmi. 2025. "Stance Detection in Arabic Tweets: A Machine Learning Framework for Identifying Extremist Discourse" Mathematics 13, no. 18: 2965. https://doi.org/10.3390/math13182965

APA Style

Alkhraiji, A. K., & Azmi, A. M. (2025). Stance Detection in Arabic Tweets: A Machine Learning Framework for Identifying Extremist Discourse. Mathematics, 13(18), 2965. https://doi.org/10.3390/math13182965

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop