You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

4 July 2025

Emoji-Driven Sentiment Analysis for Social Bot Detection with Relational Graph Convolutional Networks

,
and
College of Computer Science, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Sensor Networks

Abstract

The proliferation of malicious social bots poses severe threats to cybersecurity and social media information ecosystems. Existing detection methods often overlook the semantic value and emotional cues conveyed by emojis in user-generated tweets. To address this gap, we propose ESA-BotRGCN, an emoji-driven multi-modal detection framework that integrates semantic enhancement, sentiment analysis, and multi-dimensional feature modeling. Specifically, we first establish emoji–text mapping relationships using the Emoji Library, leverage GPT-4 to improve textual coherence, and generate tweet embeddings via RoBERTa. Subsequently, seven sentiment-based features are extracted to quantify statistical disparities in emotional expression patterns between bot and human accounts. An attention gating mechanism is further designed to dynamically fuse these sentiment features with user description, tweet content, numerical attributes, and categorical features. Finally, a Relational Graph Convolutional Network (RGCN) is employed to model heterogeneous social topology for robust bot detection. Experimental results on the TwiBot-20 benchmark dataset demonstrate that our method achieves a superior accuracy of 87.46%, significantly outperforming baseline models and validating the effectiveness of emoji-driven semantic and sentiment enhancement strategies.

1. Introduction

Social media platforms such as Twitter (now renamed X) and Facebook have emerged as pivotal channels for global users to access news, engage in public discourse, and foster emotional interactions, owing to their real-time dissemination capabilities and low-barrier communication mechanisms [1]. According to Meta Group’s financial reports [2], its monthly active users surged exponentially from 1.5 billion in 2015 to 3 billion in 2023. However, this thriving social ecosystem coexists with the rampant proliferation of malicious social bot accounts. A study by the University of Southern California revealed that 15% of Twitter accounts are bots [3], which are automated programs designed to perform social media activities [4], with 78.6% implicated in disseminating disinformation [5]. These bots mimic genuine user behaviors to establish a systematic attack chain: “fabricating pseudo-consensus → inducing group polarization → manipulating decision-making processes” [6]. A notable example occurred during the 2016 U.S. presidential election, where the Botometer system [7] identified over 280,000 bot accounts persistently spreading political misinformation, accounting for one-third of candidate-endorsing tweets [8]. Such activities distorted voter perceptions and severely undermined electoral integrity. Therefore, the detection of social bots has attracted the attention of relevant researchers.
Current mainstream detection methodologies face multifaceted technical challenges. First, their capability to identify highly anthropomorphic behaviors remains limited; social bots now emulate human-like information propagation patterns, significantly degrading the performance of detection models reliant on static shallow statistical features [9]. Second, bots increasingly exploit social network analysis techniques to construct and optimize their relational networks, rendering their interactions indistinguishable from authentic users [10]. These technological advancements have made existing detection methods ineffective in identifying malicious bot accounts, thus creating an urgent demand for a new detection framework to tackle this challenge.
In social bot detection, tweets represent one of the most critical feature sources. However, most existing detection methods dismiss emojis as noise and routinely eliminate them during preprocessing [11]. However, in reality, emojis not only significantly enhance the semantic interpretability of tweets but also provide emotional cues absent in plain text [12,13,14,15,16]. According to a report from Emojipedia, the proportion of tweets containing emojis on Twitter has surged from 3.24% in 2013 to 26.7% in 2023 [17]. As illustrated in Figure 1, emoji usage frequency increased by 724% over this decade, underscoring their growing role as core carriers of nuanced semantics and affective signals. Taking the tweet “Let’s enjoy some Sensors 25 04179 i001 and watch Sensors 25 04179 i002 at the Sensors 25 04179 i003 later!” as an example, when analyzed solely through its textual content, the meaning becomes ambiguous. However, integrating emojis with the text resolves semantic uncertainties, revealing a clear contextual narrative. This contrast demonstrates that emojis encode rich semantic information, thereby proving their irreplaceable value in tweet interpretation.
Figure 1. Emoji use in tweets (2013 vs. 2023).
Beyond semantic enrichment, emojis excel at conveying emotional information that text alone often fails to capture. Particularly in informal short texts such as comments, they serve as critical indicators of user sentiment. Taking the tweet “It’s raining on my picnic day. Sensors 25 04179 i004” as an example, if the emojis are ignored, it expresses the negative emotions of disappointment and chagrin after the plan is ruined. However, after adding the “Sensors 25 04179 i004” emojis, the sentiment of the sentence undergoes a reversal, conveying an easygoing and adaptable attitude of “Since it’s raining and the picnic can’t happen, might as well have a drink and then go to sleep”, thus transforming into a positive sentiment. To meet escalating demands for emotional articulation, the Unicode Consortium has expanded emoji categories through iterative updates. As illustrated in Figure 2, official data reveal a 35-fold increase in emoji diversity from 2005 to 2025 [18], with an annual growth rate of 42.7%. This expansion substantially elevates emojis’ research value in sentiment analysis.
Figure 2. Growth of emoji types (2005–2025).
To address the aforementioned challenges, we propose ESA-BotRGCN, an emoji-driven multi-modal detection framework, with its core innovations spanning five key dimensions:
  • We utilize the Emoji Library to convert emojis into textual descriptions, refine the coherence of transformed tweets via GPT-4, and generate semantically and sentiment-enriched tweet embeddings using RoBERTa.
  • We propose a seven-dimensional sentiment feature quantification framework, leveraging foundational sentiment metrics, including positive, negative, and neutral sentiment values and complexity measures, to further capture statistical discrepancies between bot accounts and genuine users in terms of sentiment polarity span, dynamic volatility, and expression consistency.
  • An attention gating mechanism is developed to adaptively integrate sentiment features, user descriptions, tweet content, numerical attributes, and categorical features into a globally context-aware unified feature representation.
  • To capture bot-specific behavioral patterns, a Relational Graph Convolutional Network (RGCN) is constructed to model user-following relationships based on graph topological structures.
  • We conduct extensive experiments on the TwiBot-20 and Cresci-15 datasets, demonstrating our model’s superiority through significant improvements in accuracy, F1-score, and MCC over mainstream baselines.

3. ESA-BotRGCN Model Architecture

According to the aforementioned discussion, existing social bot detection methods predominantly suffer from two major limitations. First, the rich semantic and emotional expressive value of emojis is often overlooked, as most studies treat them as noise and directly remove them during preprocessing, leading to the loss of critical features [11]. Second, there exist significant statistical deviations in emotional expression patterns between bot accounts and genuine users, including differences in emotional complexity, polarity span, dynamic fluctuations, and expression consistency. However, these discriminative features have yet to be fully explored [42]. To address these issues, this study innovatively proposes ESA-BotRGCN, a Graph Convolutional Network model enhanced by emoji-driven sentiment and semantics, which dynamically integrates multi-dimensional features to improve the accuracy and robustness of social bot detection. Figure 3 illustrates the architecture of our proposed model.
Figure 3. ESA-BotRGCN model diagram.
The detection method proposed in this study achieves end-to-end modeling through multi-stage collaborative processing. First, tweets undergo preprocessing; emojis are converted into corresponding textual descriptions using the Emoji Library. The GPT-4 model is then employed to refine the textual coherence of the tweets. Next, the RoBERTa model is utilized to generate tweet embeddings that integrate emoji semantics, thereby addressing the limitations of traditional word vectors in representing the semantic information conveyed by emojis and enhancing the expressiveness of sentiment features. Subsequently, seven sentiment-based features are extracted from the preprocessed tweets to capture statistical differences in emotional expression patterns between bot and human accounts. ESA-BotRGCN then jointly encodes tweet features, sentiment-based features, user description features, numerical attributes, and categorical features. Through an attention gating mechanism, the model dynamically adjusts the weights of these heterogeneous features to generate a globally context-aware unified feature vector. This comprehensive feature vector is fed into an RGCN, where two layers of graph convolution operations integrate the topological structure of user social relationships, and dropout is applied to suppress overfitting. Finally, the output features are mapped to a high-dimensional space via a linear layer, and the detection probability of bot accounts is computed using the Softmax function. The Adam optimizer is employed to perform gradient backpropagation based on the cross-entropy loss function, optimizing model parameters to obtain the final classification results.

3.1. Emoji Preprocessing

According to the aforementioned discussion, existing social bot detection studies often treat emojis as noise and directly remove them during preprocessing [11], resulting in severe loss of semantic and emotional cues in tweets. Particularly in short tweets, a single emoji can significantly alter the meaning of the entire tweet. To address this issue, this study employs emoji-to-text mapping and language model optimization to fully preserve the semantic and emotional value of emojis. First, each tweet t is defined as an ordered joint structure of textual tokens and emojis:
t = { w , e }
In this formula, w = { w 1 , w 2 , , w M } denotes the ordered sequence of textual tokens in the tweet, with M being the total number of words and e = { e 1 , e 2 , , e K } denoting the ordered sequence of emojis in the tweet, where K is the total number of emojis. The preprocessing workflow includes the following steps:

3.1.1. Emoji Mapping

First, emoji detection is performed on the tweet. If no emojis are detected, no further processing is applied. If emojis are present, each emoji e k is converted into its corresponding textual description text k using a predefined mapping table M :
f ( e k ) = text k ( k = 1 , 2 , , K )
This study compared the Unicode coverage, multilingual description support, semantic mapping accuracy, and update frequency and found that Emoji Library outperformed both Emojiswitch Library and EmojiBase Library: it fully covers the entire Unicode 15.0 version, including the emojis added in 2025, whereas Emojiswitch only covers up to Unicode 13.0 and EmojiBase covers up to 14.0 with some missing variants. Additionally, Emoji Library provides long-text descriptions in multiple languages, such as English and Chinese, while the other two only support English descriptions. Finally, Emoji Library maintains quarterly updates to keep pace with official Unicode releases, which is more adaptable to the semantic evolution of emojis compared to Emojiswitch’s annual updates and EmojiBase’s irregular updates. Therefore, this study selected the Emoji Library to construct the emoji–text mapping table M . As shown in Table 1, which lists 10 typical mapping examples.
Table 1. Examples of popular emoji and their text mapping.
The mapped tweet t is expressed as follows:
t = w { f ( e 1 ) , f ( e 2 ) , , f ( e K ) }

3.1.2. Text Coherence Optimization

Tweets converted from emojis often suffer from poor textual coherence and semantic ambiguity. For example, the tweet “Let’s go to the beach! Sensors 25 04179 i003” becomes “Let’s go to the beach! :beach_with_umbrella:” after conversion, resulting in semantic discontinuity. To address this, this study introduces the GPT-4 model for semantic reconstruction; the prompt is: “You are a language analysis expert. Analyze the tweet provided to you. The emojis in the tweet have been converted into corresponding text label embeddings. Please make the tweet coherent without changing its original meaning, and ensure the length of the generated tweet does not exceed 200% of the original tweet.”. As illustrated in Figure 4, this process ensures that tweets fully express their semantic and emotional information. For instance, the aforementioned tweet is reconstructed as “Let’s go to the beach and enjoy the sun under an umbrella!”.
Figure 4. Tweet processing flow: from emojis to Unicode and descriptions using GPT-4.

3.1.3. Handling of Abnormal Emojis

For rare or custom emojis not included in the Emoji Library, this study retains their original Unicode encoding and appends the special token “[UNK]” as a placeholder to avoid information loss. Subsequently, all modified tweets are fed into the model for training instead of the original tweets.

3.2. Sentiment-Based Features Processing

Sentiment-based features r p senti capture statistical differences in emotional expression patterns between bot and human accounts by quantifying the sentiment dynamics in user tweets. The innovation of this study lies in proposing seven sentiment features, including four foundational ones—positive, negative, and neutral sentiment values, as well as sentiment complexity—and three advanced features derived from these, including sentiment polarity span, volatility, and consistency. These features significantly enhance discriminative power. The specific definitions are as follows:

3.2.1. Basic Sentiments and Complexity

This study employs VADER [43] for analyzing basic sentiment values. The selection of VADER is attributed to three advantages it possesses. First, VADER is adaptable to the characteristics of social media texts. It can effectively handle fragmented content such as short texts and informal language (e.g., slang) in tweets. Its built-in sentiment dictionary contains a large number of commonly used social media words (e.g., “OMG”). Second, VADER does not require a large amount of labeled data. It is lightweight and efficient, enabling sentiment analysis solely through predefined sentiment dictionaries and grammatical rules. It has low computational costs and is easy to deploy, making it suitable for processing large-scale tweet data. Third, VADER can calculate sentiment scores by decomposing elements such as sentiment words, negation words, and intensifier words in the text. The results have clear interpretability, facilitating the analysis of differences in sentiment expression between bots and real users. In recent years, VADER has seen many typical and effective application cases. For example, Francesco et al. [44] applied VADER to investigate the sentiment experiences among e-sports spectators, and Anny et al. [45] used VADER to classify sentiments and discovered the correlation between Twitter sentiments and the price movements of Bitcoin. VADER is highly compatible with the multimodal sentiment feature extraction requirements of this study. Therefore, VADER is adopted to analyze basic sentiment values, and the specific process is as follows:
Conduct sentiment analysis on each tweet of the user. Based on a predefined sentiment dictionary, which contains the positive, negative, and neutral sentiment intensity values for each word, combined with context-adjustment rules (such as negative words, emphasis words, punctuation marks, etc.), extract the positive sentiment value ( S Pos ( i ) ), negative sentiment value ( S Neg ( i ) ), neutral sentiment value ( S Neu ( i ) ), and sentiment complexity ( S Com ( i ) ):
S P o s ( i ) = j = 1 M I P o s ( w j ) C ( w j )
S N e g ( i ) = j = 1 M I N e g ( w j ) C ( w j )
In this formula, i denotes the i-th tweet, j represents the j-th word in the i-th tweet, and M is the total number of words in the i-th tweet. I P o s ( w j ) and I N e g ( w j ) , respectively, denote the positive and negative sentiment intensities of each word w j , and C ( w j ) represents the contextual adjustment factor.
S N e u ( i ) = 1 ( S P o s ( i ) + S N e g ( i ) )
Given that both S Pos ( i ) + S Neg ( i ) + S Neu ( i ) = 1 and S Pos ( i ) + S Neg ( i ) 1 always hold, the neutral sentiment value corresponds to the residual portion not covered by positive or negative sentiments.
Sentiment complexity S Com ( i ) , which quantifies the diversity and volatility of emotional expression, is measured via entropy or standard deviation as follows:
S C o m ( i ) = c { P o s , N e g , N e u } P c log ( P c )
In this formula, P c = S c ( i ) S P o s ( i ) + S N e g ( i ) + S N e u ( i ) is the normalized proportion of each sentiment category. Finally, the mean of sentiment features across all tweets from users is computed to derive the user-level sentiment feature vector:
S ¯ Pos = 1 N i = 1 N S Pos ( i )
S ¯ Neg = 1 N i = 1 N S Neg ( i )
S ¯ Neu = 1 N i = 1 N S Neu ( i )
S ¯ Com = 1 N i = 1 N S Com ( i )
In this formula, i denotes the i-th tweet; N is the total number of tweets by the user; S ¯ Pos and S ¯ Neg reflect the overall positive and negative emotional tendencies of the user, respectively; S ¯ Neu characterizes the degree of emotional neutrality; and S ¯ Com measures the complexity of emotional expression.

3.2.2. Sentiment Polarity Span

To capture the range of sentiment fluctuations in sentiment expressions, the sentiment polarity span S Span is defined as the absolute difference between the maximum positive sentiment value and the maximum negative sentiment value across all tweets from the same user:
S Span = max 1 i N S Pos ( i ) + max 1 i N S Neg ( i )
Here, S Pos ( i ) and S Neg ( i ) are both positive values. This metric quantifies the extremity of a user’s emotional expression. Real users, due to the diversity of life events (such as celebrating victories and encountering setbacks), will have significant positive and negative emotional fluctuations, and S Span is usually relatively high; however, robots, restricted by task objectives (such as conducting concentrated propaganda or attacks), have their emotional expressions constrained within a specific polarity range, and S Span is significantly low. Even if attackers attempt to simulate the human emotional range, it will reduce the attack efficiency due to disrupting task consistency. Therefore, this feature can significantly improve the performance of robot detection.

3.2.3. Sentiment Volatility

To further quantify the volatility of emotional expression, sentiment volatility S Var is introduced, measured as the standard deviation of sentiment complexity across a user’s tweets:
S Var = 1 N 1 i = 1 N ( S Com ( i ) S ¯ Com ) 2
Sentiment volatility reveals the differences in randomness during the evolution of emotions. Human emotions, triggered by immediate events, exhibit irregular fluctuations (such as a shift from anger to surprise caused by breaking news). Thus, the sentiment complexity of real users shows relatively large fluctuations ( S Var is high). In contrast, robots generate content in a programmatic manner, so the distribution of their sentiment complexity is more concentrated ( S Var is low). If robots forcibly inject random fluctuations to evade detection, it will lead to a decline in semantic coherence and expose traces of generation.

3.2.4. Sentiment Consistency

Sentiment consistency S Consist is defined as the mean cosine similarity of sentiment complexity between adjacent tweets, capturing the continuity of a user’s emotional expression:
S Consist = 1 N 1 i = 1 N 1 v i v i + 1 v i v i + 1
In this formula, v i = [ S Pos ( i ) , S Neg ( i ) , S Com ( i ) ] represents the sentiment vector of the tweet. Robot accounts generate purposeful content in batches. For example, sending 50 consecutive advertising tweets will lead to a high degree of similarity in sentiment between adjacent tweets ( S Consist approaches 1). In contrast, human beings switch topics more frequently than robots (such as shifting from sports discussions to food sharing), and their sentiment expressions are more random, resulting in low sentiment continuity ( S Consist is relatively low). If a robot deliberately reduces consistency, it needs to sacrifice task focus, significantly weakening its propagation effect.
The sentiment polarity span, sentiment volatility, and sentiment consistency form a triangular constraint on sentiment expression, bringing about a joint defense advantage. If a robot attempts to adjust one feature (such as increasing the polarity span to simulate human beings), it will simultaneously disrupt other features (such as causing an abnormal increase in volatility or a break in consistency), forming an “impossible triangle” for detection evasion.
Finally, the sentiment feature vector is constructed by concatenating the above metrics:
r p senti = [ S ¯ Pos , S ¯ Neg , S ¯ Neu , S ¯ Com , S Span , S Var , S Consist ] 7

3.3. User Node Feature Processing

ESA-BotRGCN aims to leverage multi-modal user information to address the challenge of bot account camouflage, enabling accurate identification of malicious bot accounts and preventing the spread of misinformation. Specifically, ESA-BotRGCN jointly encodes five categories of features: sentiment-based features, user description features, tweet content features, numerical attributes, and categorical features. In this section, the encoding process of the latter four types of features will be elaborated in detail.

3.3.1. User Description Feature r b

User description features are extracted from the personal bio on the account homepage. To capture their semantic information, a pre-trained RoBERTa model is employed for encoding. First, the user description text is input into RoBERTa to generate word-level embedding vectors:
b ¯ = R o B E R T a ( { b i } i = 1 L ) , b ¯ D s × 1
In this formula, b ¯ represents the user description embedding, i denotes the i-th word in the description, L is the total number of words, and D s is the embedding dimension of RoBERTa. The user description representation vector is then derived as follows:
r b = ϕ ( W B b ¯ + b B ) , r b D / 4 × 1
In this formula, W B and b B are learnable parameters, ϕ is the activation function, and D is the embedding dimension for Twitter users. In subsequent sections of this paper, Leaky-ReLU is adopted as ϕ .

3.3.2. Tweet Content Feature r t

After emoji preprocessing (Section 3.1), the semantic and emotional information of tweets is enhanced. The processed tweets are encoded using RoBERTa in a manner similar to the above. First, RoBERTa generates an embedding vector t i D for each tweet. The mean of all tweet embeddings for a user is then computed to obtain the user-level tweet representation:
r t = 1 N i = 1 N t i
In this formula, N is the total number of tweets by the user.

3.3.3. Numerical Features r p n u m

Numerical attributes refer to measurable metrics that quantify user behaviors and attributes, as listed in Table 2. These features provide numerical descriptions of user behavioral patterns and preferences. The numerical attributes are processed using Multi-Layer Perceptrons (MLPs) and graph neural networks (GNNs). Specifically, 10 numerical features are first directly retrieved from the Twitter API, with the first three being emoji-related. Z-score normalization is then applied to eliminate scale differences, followed by a fully connected layer to derive the user’s numerical feature representation r p n u m .
Table 2. User’s numerical features.

3.3.4. Categorical Features r p c a t

Categorical features are qualitative indicators describing user attributes, as shown in Table 3. These features are typically represented in binary form (yes/no). Similar to numerical attributes, we avoid manual feature engineering and instead encode them using MLPs and GNNs. Specifically, 11 categorical features are directly retrieved from the Twitter API. One-hot encoding is applied to convert these features into sparse vectors, which are then concatenated and transformed via a fully connected layer with Leaky-ReLU activation, yielding the user categorical feature representation r p c a t .
Table 3. User’s categorical features.

3.4. Features Fusion

The proposed model achieves multi-modal feature fusion through feature encoding and an attention mechanism. Given a set of user node features { r p senti , r b , r t , r p n u m , r p c a t } , representing sentiment-based features, user descriptions, tweet content, numerical attributes, and categorical features, respectively; the fusion process is defined as follows:

3.4.1. Feature Alignment and Nonlinear Transformation

Heterogeneous features are projected into a unified dimensional space via fully connected layers to eliminate dimensional discrepancies and introduce nonlinear expressive capabilities:
r ˜ k = LeakyReLU W k r k + b k , k { senti , b , i , n u m , c a t }
In this formula, W k D × d k is the learnable weight matrix, b k D is the bias term, D = 256 is the unified embedding dimension, and Leaky-ReLU mitigates gradient vanishing.

3.4.2. Single-Head Attention Weight Allocation

A single-head attention mechanism dynamically evaluates feature importance through query–key interactions:
α k = exp q T ( W a r ˜ k ) j { s e n t i , b , i , n u m , c a t } exp q T ( W a r ˜ j )
In this formula, W a D × D is a parameter matrix and q D is a learnable global query vector that captures cross-feature global dependencies.

3.4.3. Context-Aware Feature Fusion

A globally integrated feature vector r i D is generated via weighted summation and fed into the subsequent RGCN module for social graph modeling:
r i = k { s e n t i , b , i , n u m , c a t } α k r ˜ k
Through this approach, the model dynamically adjusts the influence of each feature during training, thereby enhancing the discernment of feature importance. This fusion method enables the model to effectively integrate information from diverse sources and deliver superior performance in complex tasks.

3.5. Heterogeneous Social Graph Modeling

To address the limitations of traditional graph neural networks in modeling multi-relational data, this study employs RGCN [46] for deep feature extraction from heterogeneous social graphs. We construct a heterogeneous information network G = ( V , E , R ) , where the node set V represents user accounts, the edge set E includes two relationship types R = { r 1 , r 2 } = { f o l l o w i n g , f o l l o w e r } , characterizing active and passive follow behaviors between users. Each user node v i V is associated with a feature vector r i , derived from the attention-based fusion in Section 3.3. RGCN integrates rich contextual information from nodes through efficient message passing and aggregation mechanisms. During this process, RGCN not only considers the intrinsic features r i of node v i but also incorporates neighborhood information to generate context-aware node representations. This multi-layer stacking progressively expands the receptive field, capturing global topological patterns. Furthermore, social networks are typically large-scale and sparse, with limited node connections, posing challenges for information propagation. RGCN overcomes these limitations via a dynamic relation-aware edge weighting mechanism, which updates relationship weights and relation-specific embeddings to address potential information gaps in the graph. The graph convolution operations of RGCN are implemented as follows.

3.5.1. Node Initialization

The user feature vector r i undergoes linear transformation and activation function to generate the initial hidden representation:
x i ( 0 ) = ϕ ( W 1 r i + b 1 ) , x i ( 0 ) D × 1
In this formula, W 1 is a learnable weight matrix, b 1 is a bias term, and ϕ is the Leaky-ReLU activation function.

3.5.2. Relational Graph Convolution

At the l-th layer, the representation of node v i is updated as follows:
x i ( l + 1 ) = Θ self x i ( l ) + r R j N r ( i ) 1 | N r ( i ) | Θ r x j ( l ) , x i ( l + 1 ) D × 1
Here, N r ( i ) denotes the set of neighbors of v i under relation r, and Θ self and Θ r are the weight matrices for self-connections and relationship r, respectively.

3.5.3. Multi-Layer Perceptron (MLP) Enhancement

Stacked RGCN layers are interleaved with MLPs to further refine node representations and expand the receptive field:
h i = MLP ( l ) x i ( l ) = ϕ ( W MLP ( l ) x i ( l ) + b MLP ( l ) )
In this formula, W MLP ( l ) and b MLP ( l ) are learnable parameters. The final node representation h i is mapped to a high-dimensional discriminative space via the MLP.

3.6. Learning and Optimization

ESA-BotRGCN formulates social bot detection as a binary classification task, where the label y i { 0 , 1 } of a user account denotes a genuine user (0) or a social bot (1). This section details the classification mechanism and loss function design.

3.6.1. Classifier Design

The model leverages context-aware node representations h i 128 , extracted by the RGCN, to construct the classifier. First, the node representation h i is mapped to a 2D space via a fully connected layer:
z i = W O h i + b O
In this formula, W O 128 × 2 is the weight matrix and b O 2 is the bias term. The class probability distribution is then computed using the Softmax function:
y ^ i = Softmax ( z i ) = e z i 0 e z i 0 + e z i 1 , e z i 1 e z i 0 + e z i 1

3.6.2. Loss Function

The total loss combines cross-entropy loss and a weight decay term to balance classification accuracy and model complexity:
L = i Y y i log ( y ^ i ) + ( 1 y i ) log ( 1 y ^ i ) + λ w θ w 2
The first part is the cross-entropy loss, which measures the deviation between the predicted probability and the true label to evaluate the accuracy of the prediction. The second part is the weight decay term. i y represents the summation over all labeled users, w θ represents the summation over all learnable parameters in the ESA-BotRGCN framework, and λ is the decay coefficient used to control the model parameters and suppress overfitting.

4. Experiments

4.1. Experimental Setup

4.1.1. Dataset

This study uses the publicly available Twitter bot detection benchmark datasets TwiBot-20 [47] and Cresci-15 [48] to conduct experimental verification.
The TwiBot-20 dataset contains 229,580 user nodes, 33,488,192 tweet texts, 8,723,736 user attribute data entries, and 455,958 bidirectional follow relationships. Following the original data partitioning strategy, we employed a stratified sampling method to divide the dataset into training (80%), validation (10%), and test (10%) sets, ensuring balanced class distribution. Furthermore, based on user nodes and follow relationships, we constructed a heterogeneous graph structure where nodes represent user entities and edges represent unidirectional follow behaviors, ultimately generating a social network topology comprising 229,580 nodes and 227,979 directed edges.
The Cresci-15 dataset consists of five sub-datasets, with a total of 5301 users and 7,086,134 edges. All accounts are labeled, among which 3351 accounts are marked as bot accounts. Similar to the TwiBot-20 dataset, we divide it into a training set (80%), a validation set (10%), and a test set (10%).

4.1.2. Evaluation Metrics

To comprehensively evaluate model performance, this study adopts the following three metrics.
  • Accuracy: Measures the overall proportion of correct classifications.
Accuracy = TP + TN TP + TN + FP + FN
In this formula, TP (true positive) denotes the number of correctly identified bot accounts, TN (true negative) represents the number of correctly identified genuine users, and FP (false positive) and FN (false negative) indicate the numbers of misclassified positive and negative samples, respectively.
2.
F1-score [49]: The harmonic mean of precision and recall, suitable for scenarios requiring balanced class distributions.
F 1 - score = 2 × Precision × Recall Precision + Recall
In this formula, the meanings of precision and recall are as follows:
Precision = TP TP + FP
Recall = TP TP + FN
3.
MCC (Matthews Correlation Coefficient) [50]: A statistical measure that synthesizes all four elements of the confusion matrix, robust to class imbalance.
MCC = TP × TN FP × FN ( TP + FP ) ( TP + FN ) ( TN + FP ) ( TN + FN )
MCC’s range is [−1, 1], where a value closer to 1 indicates superior model performance.
The aforementioned metrics evaluate the model’s classification capabilities from different dimensions. Accuracy reflects overall correctness, the F1-score balances class sensitivity, and MCC mitigates the impact of class distribution skew. Their combination can comprehensively validate the robustness of the detection framework.

4.1.3. Baseline Methods

We selected 10 mainstream social bot detection methods as baseline models for performance comparison with ESA-BotRGCN, covering diverse technical approaches including traditional machine learning, deep learning, and graph neural networks. The specific methods are as follows.
  • Lee et al. [51]: A Random Forest-based model that integrates multi-dimensional features such as user-following networks and tweet content for detection.
  • Yang et al. [52]: Employs lightweight metadata combined with a Random Forest algorithm to achieve efficient tweet stream analysis.
  • Kudugunta et al. [53]: Leverages a contextual Long Short-Term Memory (LSTM) network to jointly model tweet content and metadata for tweet-level bot detection.
  • Wei et al. [54]: Utilizes a three-layer Bidirectional Long Short-Term Memory (BiLSTM) with word embeddings to identify social bots on Twitter.
  • Miller et al. [55]: Treats bot detection as an anomaly detection problem rather than classification, introducing 95 lexical features extracted from tweet text.
  • Cresci et al. [26]: Innovatively encodes user behaviors into DNA-like sequences and distinguishes bots from humans via Longest Common Subsequence (LCS) similarity.
  • Botometer [7]: A widely used public Twitter detection tool that employs over 1000 features.
  • Alhosseini et al. [56]: Detects social bots using a Graph Convolutional Neural Network (GCNN) by leveraging node features and aggregated neighborhood node features.
  • SATAR [57]: A self-supervised representation learning framework capable of generalizing by jointly utilizing user semantics, attributes, and neighborhood information.
  • BotRGCN [36]: Addresses bot detection by constructing a heterogeneous graph from Twitter user-follow relationships and applying an RGCN.

4.2. Comparative Experiments

This section validates the effectiveness of each module in the ESA-BotRGCN model through four sets of ablation experiments, including the impact of emoji preprocessing, sentiment feature integration, the attention mechanism, and the selection of graph neural network architecture and its layer count on detection performance. Additionally, we have conducted statistical significance analysis under different feature settings.

4.2.1. Emoji Preprocessing Ablation Study

To validate the importance of emojis in tweet semantics and sentiment analysis, this section compares the impact of four tweet input formats on detection performance under the framework of using seven-dimensional advanced sentiment features and a two-layer RGCN-based graph neural network architecture.
  • Original tweets: Raw tweets without any emoji processing.
  • Emoji-mapped tweets: Tweets where emojis are replaced with corresponding textual descriptions from the Emoji Library.
  • GPT-4 optimized tweets: Emoji-mapped tweets further refined by GPT-4 for textual coherence.
  • Qwen-2.5 Optimizes Tweets: Use Qwen-2.5, owned by Alibaba, to optimize the text coherence of tweets after mapping with the Emoji library, and keep the prompt consistent with that of GPT-4.
As shown in Table 4, the original tweets yielded the lowest detection performance due to the neglect of emoji semantics and emotional cues. Mapping emojis to textual descriptions via the Emoji Library enabled the model to capture semantic and emotional associations, significantly improving performance. Tweets optimized by large language models (LLMs) further enhance text coherence, and GPT-4, in particular, achieves even better results. This enables the model to more accurately extract semantic and sentiment features, thereby attaining optimal performance. Experimental results demonstrate that semantic reconstruction and coherence optimization of emojis play a critical role in improving detection accuracy.
Table 4. Emoji preprocessing ablation study results.
Notably, while LLMs improve text coherence and enhance detection accuracy, their computational cost should be taken into account: the inference time for 10,000 tweets on an A100 GPU is 2.1 h (Qwen-2.5) and 2.6 h (GPT-4). This trade-off between performance and cost makes LLMs suitable for scenarios where detection accuracy is prioritized over real-time deployment.

4.2.2. Sentiment Feature Ablation Study

To validate the contribution of sentiment features, this section conducts ablation experiments under three configurations, each employing preprocessed tweets and a two-layer RGCN-based graph neural network architecture.
  • No sentiment features: Only user descriptions, tweet content, numerical attributes, and categorical features are used.
  • Basic sentiment features: Incorporates positive, negative, and neutral sentiment polarities, as well as sentiment complexity.
  • Advanced sentiment features: On the basis of basic sentiment features, three additional features are added—polarity span, volatility, and consistency.
As shown in Table 5, the model performance significantly declined when no sentiment features were used, indicating that sentiment features are critical clues for distinguishing bots from genuine users. After introducing four basic sentiment features, accuracy was partially improved, and further incorporating polarity span, volatility, and consistency into the seven-dimensional sentiment features significantly enhanced the model’s ability to capture emotional expression patterns, ultimately achieving an accuracy of 87.46%. The experiments demonstrate that these seven sentiment features can effectively reflect the differences in emotional expression between bot accounts and genuine users, thereby significantly enhancing the model’s discriminative capability.
Table 5. Sentiment feature ablation study results.

4.2.3. Selection of Graph Neural Network Architecture and Its Layers

To validate the advantages of RGCN in modeling heterogeneous social graphs, this study compares the detection performance of different graph neural network architectures, each employing preprocessed tweets and seven-dimensional advanced sentiment features.
  • Graph Attention Network (GAT): Dynamically assigns edge weights through an attention mechanism.
  • Graph Convolutional Network (GCN): Models homogeneous social relationships without distinguishing node or edge types.
  • Fully Connected Neural Network: Applies nonlinear transformations via Multi-Layer Perceptrons.
As shown in Table 6, while GAT introduces an attention mechanism, its stability in modeling complex relationships remains insufficient. The GCN model ignores relational heterogeneity, failing to differentiate between active following and passive following behaviors, resulting in suboptimal performance. The fully connected neural network achieves better classification performance through effective feature learning and processing. Although the neural network achieves high detection accuracy through feature fusion, it essentially belongs to a node-level feature processing model and cannot capture the implicit topological structure information in social networks. With the increase in social interaction density, the connections between users are becoming closer and closer, and the dynamic changes in network topology put forward higher requirements for the model. The multi-layer message-passing mechanism of RGCN can gradually aggregate multi-order neighbor information and maintain the stable extraction of topological features in dense networks, which elevates the social bot detection from “individual behavior recognition” to the level of “group structure understanding”. In the experiment, RGCN achieved optimal performance, thus demonstrating its indispensability.
Table 6. GNN architecture selection ablation study results.
Furthermore, to further investigate the impact of RGCN layer depth on model performance, this study conducted comparative experiments on RGCN architectures with varying numbers of layers. As shown in Figure 5, the 1-layer RGCN failed to adequately aggregate global topological information due to its limited receptive field, while models with three or more layers suffered from performance degradation caused by over-smoothing. Ultimately, the 2-layer RGCN achieved an optimal balance between parameter efficiency and feature representation capability, validating the rationality of the current architectural design.
Figure 5. Performance comparison of RGCN with varying layer numbers.

4.2.4. Ablation Experiment on Attention Mechanism

To verify the effectiveness of the attention gating mechanism in multi-modal feature fusion, this section designs the following two sets of comparative experiments while keeping other variables unchanged (that is, all use preprocessed tweets, seven-dimensional advanced emotional features, and a two-layer RGCN architecture).
  • Without-attention mechanism: Directly concatenate five types of feature vectors (sentiment features, user descriptions, tweet content, numerical features, and categorical features) to replace attention-weighted fusion.
  • Complete-attention mechanism: Use the attention gating network proposed in Section 3.4 of this paper for dynamic feature fusion.
The experimental results are shown in Table 7. After removing the attention mechanism, the performance of ESA-BotRGCN declines comprehensively, indicating that simple feature concatenation is difficult to effectively distinguish feature importance. The attention mechanism dynamically allocates weights, enabling ESA-BotRGCN to dynamically adjust the influence of various features during the training process, thereby improving the ability to recognize feature importance. This fusion method allows the model to more effectively integrate information from different sources and provide better performance in complex tasks.
Table 7. Attention mechanism ablation experiment results.

4.2.5. Statistical Significance Analysis of Different Feature Settings

To further validate the independent and joint contributions of emotional features and emoji preprocessing to performance enhancement, this section compares the performance of three model variants against the complete ESA-BotRGCN model and applies the Welch’s t-test to assess whether the observed differences are statistically significant. Specifically, the following three simplified model settings were used for comparison:
  • Without emotion and emojis: sentiment features and emoji preprocessing were removed, retaining only the other feature modalities.
  • Use only sentiment features: Retain the seven-dimensional sentiment features and remove the text enhancement processing for emojis.
  • Use only emojis: Perform emoji-to-text mapping and GPT-4 semantic optimization, but do not introduce sentiment features.
As shown in Table 8, the average performance of all three simplified variants differs significantly from that of the complete model. The “without emotion and emojis” version exhibited the largest performance gap (t = −16.82, p < 0.00001), indicating an extremely significant difference and highlighting the importance of integrating both sentiment and emoji-related features. In addition, introducing either sentiment features alone (t = −9.25, p = 0.000099) or emoji preprocessing alone (t = −8.09, p = 0.000045) also resulted in statistically significant improvements, thereby confirming that both information sources contribute independently to model performance.
Table 8. Significance analysis of different feature settings.
It is worth noting that the t-value in a Welch’s t-test measures the standardized difference between the means of two groups relative to their variance. A negative t-value indicates that the complete model outperforms the comparison version. The corresponding p-value represents the probability of observing such a result under the null hypothesis (i.e., no significant difference). When the p-value is substantially lower than the commonly accepted threshold (e.g., 0.05), the likelihood that the observed effect is due to random variation is minimal, thus providing strong statistical evidence to reject the null hypothesis and support the alternative—that the proposed feature processing yields a positive and significant impact on model performance.

4.3. Performance Comparison

Table 9 shows the social bot detection performance of ESA-BotRGCN compared with the baseline models from Section 4.1.3 on the TwiBot-20 dataset. Table 10 shows the social bot detection performance of ESA-BotRGCN and seven mainstream baseline models on the Cresci-15 dataset. The results in Figure 6 demonstrate that ESA-BotRGCN significantly outperforms all other methods on TwiBot-20, confirming its general effectiveness in Twitter social bot detection tasks. Additionally, ESA-BotRGCN achieves superior performance compared to baseline methods that also leverage user-follow relationships, such as Alhosseini et al. and SATAR, with MCC improvements of 39.25% and 6.05%, respectively. This indicates that ESA-BotRGCN better exploits follow relationships by situating users within their social contexts.
Table 9. Bot detection performance on the TwiBot-20 benchmark.
Table 10. Bot detection performance on the Cresci-15 benchmark.
Figure 6. Comparison of classification performance metrics for various models on TwiBot-20 [27,51,52,53,54,55,56].

5. Conclusions

To address the issue that existing social bot detection methods commonly overlook the semantic and emotional value of emojis, we innovatively propose ESA-BotRGCN, an emoji-driven end-to-end detection framework. This framework introduces an emoji–text enhancement strategy, combining GPT-4-based text coherence optimization and RoBERTa embeddings, effectively resolving the loss of emoji semantics and emotional value in traditional methods. Second, we design seven sentiment quantification metrics to systematically capture statistical differences in emotional expression patterns between bot and human accounts. Furthermore, an attention gating fusion mechanism is constructed to dynamically integrate diverse user features. Finally, by modeling heterogeneous social topologies with an RGCN, the detection robustness is substantially improved through the four aforementioned processes. Experiments on the TwiBot-20 benchmark dataset demonstrate that ESA-BotRGCN outperforms ten mainstream baseline models, highlighting the effectiveness of our emoji-driven multi-modal feature fusion and heterogeneous relationship modeling. Future work will focus on expanding multi-modal features by incorporating user-generated images, audio, and video content to enhance the model’s ability to identify camouflaged behaviors.

Author Contributions

Conceptualization, K.Z. and X.W.; methodology, K.Z. and X.W.; software, Z.L. and K.Z.; validation, Z.L.; formal analysis, K.Z. and Z.L.; investigation, Z.L.; resources, X.W.; data curation, Z.L.; writing—original draft preparation, K.Z.; writing—review and editing, K.Z., X.W., and Z.L.; visualization, K.Z.; supervision, X.W.; project administration, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ESA-BotRGCNEmoji-Driven Sentiment Analysis for Social Bot Detection with Relational Graph Convolutional Network
RGCNRelational Graph Convolutional Network
MCCMatthews Correlation Coefficient
CNNsConvolutional Neural Networks
RNNsRecurrent Neural Networks
GNNgraph neural network
URLUniform Resource Locator
LRlogistic regression
SVMsupport vector machine
GBgradient boosting
LCSLongest Common Subsequence
LSTMLong Short-Term Memory
BiGRUbidirectional gated recurrent unit
UNKunknown (placeholder for rare emojis)
MLPsMulti-Layer Perceptrons
APIApplication Programming Interface
TPtrue positive
TNtrue negative
FPfalse positive
FNfalse negative
BiLSTMBidirectional Long Short-Term Memory
GCNNGraph Convolutional Neural Network
GATGraph Attention Network
GCNGraph Convolutional Network
AdamAdaptive Moment Estimation
Leaky-ReLULeaky Rectified Linear Unit

References

  1. Krithiga, R.; Ilavarasan, E. A Comprehensive Survey of Spam Profile Detection Methods in Online Social Networks. J. Phys. Conf. Ser. 2019, 1362, 12111. [Google Scholar] [CrossRef]
  2. Meta Investor Platforms. Meta Reports Fourth Quarter and Full Year 2023 Results; Initiates Quarterly Dividend: New York, NY, USA, 2024. [Google Scholar]
  3. Varol, O.; Ferrara, E.; Davis, C.A.; Menczer, F.; Flammini, A. Online human-bot interactions: Detection, estimation, and characterization. Proc. Int. AAAI Conf. Web Soc. Media 2017, 11, 280–289. [Google Scholar] [CrossRef]
  4. Shafahi, M.; Kempers, L.; Afsarmanesh, H. Phishing through social bots on Twitter. In Proceedings of the 4th IEEE International Conference on Big Data, Big Data 2016, Washington, DC, USA, 5–8 December 2016; Institute of Electrical and Electronics Engineers Inc.: Washington, DC, USA, 2016; pp. 3703–3712. [Google Scholar]
  5. Hajli, N.; Saeed, U.; Tajvidi, M.; Shirazi, F. Social Bots and the Spread of Disinformation in Social Media: The Challenges of Artificial Intelligence. Br. J. Manag. 2022, 33, 1238–1253. [Google Scholar] [CrossRef]
  6. Chinnaiah, V.; Kiliroor, C.C. Heterogeneous feature analysis on twitter data set for identification of spam messages. Int. Arab. J. Inf. Technol. 2022, 19, 38–44. [Google Scholar] [CrossRef]
  7. Yang, K.-C.; Ferrara, E.; Menczer, F. Botometer 101: Social bot practicum for computational social scientists. J. Comput. Soc. Sci. 2022, 5, 1511–1528. [Google Scholar] [CrossRef]
  8. Weedon, J.; Nuland, W.; Stamos, A. Information Operations and Facebook; Facebook, Inc.: Menlo Park, CA, USA, 2017. [Google Scholar]
  9. Ferrara, E. Social bot detection in the age of ChatGPT: Challenges and opportunities. First Monday 2023, 28, 13185. [Google Scholar] [CrossRef]
  10. Yang, Y.; Yang, R.; Li, Y.; Cui, K.; Yang, Z.; Wang, Y.; Xu, J.; Xie, H. RoSGAS: Adaptive Social Bot Detection with Reinforced Self-supervised GNN Architecture Search. ACM Trans. Web 2023, 17, 1–31. [Google Scholar] [CrossRef]
  11. Cappallo, S.; Svetlichnaya, S.; Garrigues, P.; Mensink, T.; Snoek, C.G.M. New Modality: Emoji Challenges in Prediction, Anticipation, and Retrieval. IEEE Trans. Multimed. 2019, 21, 402–415. [Google Scholar] [CrossRef]
  12. Zhao, J.; Dong, L.; Wu, J.; Xu, K. MoodLens: An Emoticon-Based Sentiment Analysis System for Chinese Tweets; ACM: New York, NY, USA, 2012; pp. 1528–1531. [Google Scholar]
  13. Liu, K.-L.; Li, W.-J.; Guo, M. Emoticon Smoothed Language Models for Twitter Sentiment Analysis. Proc. AAAI Conf. Artif. Intell. 2021, 26, 1678–1684. [Google Scholar] [CrossRef]
  14. Alexander, H.; Bal, D.; Flavius, F.; Bal, M.; Franciska De, J.; Kaymak, U. Exploiting emoticons in polarity classification of text. J. Web Eng. 2015, 14, 22. [Google Scholar]
  15. Boia, M.; Faltings, B.; Musat, C.-C.; Pu, P. A :) Is Worth a Thousand Words: How People Attach Sentiment to Emoticons and Words in Tweets. In Proceedings of the 2013 International Conference on Social Computing, Alexandria, VA, USA, 8–14 September 2013; pp. 345–350. [Google Scholar]
  16. Wang, Q.; Wen, Z.; Ding, K.; Liang, B.; Xu, R. Cross-Domain Sentiment Analysis via Disentangled Representation and Prototypical Learning. IEEE Trans. Affect. Comput. 2025, 16, 264–276. [Google Scholar] [CrossRef]
  17. Share of Posts on Twitter Containing Emojis Worldwide in July 2013 and March 2023. Available online: https://www.statista.com/statistics/1399380/tweets-containing-emojis/ (accessed on 2 July 2023).
  18. In 2025, Global Emoji Count Could Grow to 3790. Available online: https://www.statista.com/chart/17275/number-of-emojis-from-1995-bis-2019/ (accessed on 17 July 2024).
  19. Sivanesh, S.; Kavin, K.; Hassan, A.A. Frustrate Twitter from automation: How far a user can be trusted? In Proceedings of the 2013 International Conference on Human Computer Interactions (ICHCI), Chennai, India, 23–24 August 2013; pp. 1–5. [Google Scholar]
  20. Kantepe, M.; Ganiz, M.C. Preprocessing framework for Twitter bot detection. In Proceedings of the 2017 International Conference on Computer Science and Engineering (UBMK), Antalya, Turkey, 5–8 October 2017; pp. 630–634. [Google Scholar]
  21. Sayyadiharikandeh, M.; Varol, O.; Kai-Cheng, Y.; Flammini, A.; Menczer, F. Detection of Novel Social Bots by Ensembles of Specialized Classifiers. arXiv 2020. [Google Scholar] [CrossRef]
  22. Grimme, C.; Preuss, M.; Adam, L.; Trautmann, H. Social Bots: Human-Like by Means of Human Control? Big Data 2017, 5, 279–293. [Google Scholar] [CrossRef]
  23. Grimme, C.; Assenmacher, D.; Adam, L. Changing Perspectives: Is It Sufficient to Detect Social Bots? In Proceedings of the Social Computing and Social Media, Las Vegas, NV, USA, 15–20 July 2018; Meiselwitz, G., Ed.; Springer International Publishing AG: Cham, Switzerland, 2018; Volume 10913, pp. 445–461. [Google Scholar]
  24. Xin, R.; Zhenyu, W.; Haining, W.; Jajodia, S. Profiling Online Social Behaviors for Compromised Account Detection. IEEE Trans. Inf. Forensics Secur. 2016, 11, 176–187. [Google Scholar] [CrossRef]
  25. Amato, F.; Castiglione, A.; De Santo, A.; Moscato, V.; Picariello, A.; Persia, F.; Sperlí, G. Recognizing human behaviours in online social networks. Comput. Secur. 2018, 74, 355–370. [Google Scholar] [CrossRef]
  26. Cresci, S.; Di Pietro, R.; Petrocchi, M.; Spognardi, A.; Tesconi, M. DNA-Inspired Online Behavioral Modeling and Its Application to Spambot Detection. IEEE Intell. Syst. 2016, 31, 58–64. [Google Scholar] [CrossRef]
  27. Cresci, S.; Pietro, R.D.; Petrocchi, M.; Spognardi, A.; Tesconi, M. Social Fingerprinting: Detection of Spambot Groups Through DNA-Inspired Behavioral Modeling. IEEE Trans. Dependable Secur. Comput. 2018, 15, 561–576. [Google Scholar] [CrossRef]
  28. Peng, H.; Zhang, J.; Huang, X.; Hao, Z.; Li, A.; Yu, Z.; Yu, P.S. Unsupervised Social Bot Detection via Structural Information Theory. ACM Trans. Inf. Syst. 2024, 42, 1–42. [Google Scholar] [CrossRef]
  29. Ilias, L.; Roussaki, I. Detecting malicious activity in Twitter using deep learning techniques. Appl. Soft Comput. 2021, 107, 107360. [Google Scholar] [CrossRef]
  30. Wu, Y.; Fang, Y.; Shang, S.; Wei, L.; Jin, J.; Wang, H. Detecting Social Spammers in Sina Weibo Using Extreme Deep Factorization Machine; Zhang, Y., Wang, H., Zhou, R., Beek, W., Huang, Z., Eds.; Springer International Publishing AG: Cham, Switzerland, 2020; Volume 12342, pp. 170–182. [Google Scholar]
  31. Wu, Y.; Fang, Y.; Shang, S.; Jin, J.; Wei, L.; Wang, H. A novel framework for detecting social bots with deep neural networks and active learning. Knowl.-Based Syst. 2021, 211, 106525. [Google Scholar] [CrossRef]
  32. Arin, E.; Kutlu, M. Deep Learning Based Social Bot Detection on Twitter. IEEE Trans. Inf. Forensics Secur. 2023, 18, 1763–1772. [Google Scholar] [CrossRef]
  33. Fazil, M.; Sah, A.K.; Abulaish, M. DeepSBD: A Deep Neural Network Model With Attention Mechanism for SocialBot Detection. IEEE Trans. Inf. Forensics Secur. 2021, 16, 4211–4223. [Google Scholar] [CrossRef]
  34. Najari, S.; Salehi, M.; Farahbakhsh, R. GANBOT: A GAN-based framework for social bot detection. Soc. Netw. Anal. Min 2022, 12, 1–11. [Google Scholar] [CrossRef]
  35. Feng, S.; Wan, H.; Wang, N.; Tan, Z.; Luo, M.; Tsvetkov, Y. What Does the Bot Say? Opportunities and Risks of Large Language Models in Social Media Bot Detection. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024. [Google Scholar]
  36. Feng, S.; Wan, H.; Wang, N.; Luo, M. BotRGCN: Twitter bot detection with relational graph convolutional networks. In Proceedings of the 2021 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Eindhoven, The Netherlands, 8–11 November 2021; Cuzzocrea, A., Coscia, M., Shu, K., Eds.; ACM: New York, NY, USA, 2021; pp. 236–239. [Google Scholar]
  37. Feng, S.; Tan, Z.; Li, R.; Luo, M. Heterogeneity-Aware Twitter Bot Detection with Relational Graph Transformers. Proc. AAAI Conf. Artif. Intell. 2022, 36, 3977–3985. [Google Scholar] [CrossRef]
  38. Li, S.; Yang, J.; Liang, G.; Li, T.; Zhao, K. SybilFlyover: Heterogeneous graph-based fake account detection model on social networks. Knowl.-Based Syst. 2022, 258, 110038. [Google Scholar] [CrossRef]
  39. Peng, H.; Zhang, Y.; Sun, H.; Bai, X.; Li, Y.; Wang, S. Domain-Aware Federated Social Bot Detection with Multi-Relational Graph Neural Networks. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
  40. Wang, X.; Wang, K.; Chen, K.; Wang, Z.; Zheng, K. Unsupervised twitter social bot detection using deep contrastive graph clustering. Knowl.-Based Syst. 2024, 293, 111690. [Google Scholar] [CrossRef]
  41. Wang, B.; Huang, S.; Liang, B.; Tu, G.; Yang, M.; Xu, R. What do they “meme”? A metaphor-aware multi-modal multi-task framework for fine-grained meme understanding. Knowl.-Based Syst. 2024, 294, 111778. [Google Scholar] [CrossRef]
  42. Ma, X.; Yan, Z.; Li, Y. Social robot detection based on user behavioral representation. Inf. Sci. 2025, 690, 121537. [Google Scholar] [CrossRef]
  43. Borg, A.; Boldt, M. Using VADER sentiment and SVM for predicting customer response sentiment. Expert Syst. Appl. 2020, 162, 113746. [Google Scholar] [CrossRef]
  44. Cauteruccio, F.; Kou, Y. Investigating the emotional experiences in eSports spectatorship: The case of League of Legends. Inf. Process. Manag. 2023, 60, 103516. [Google Scholar] [CrossRef]
  45. Mardjo, A.; Choksuchat, C. HyVADRF: Hybrid VADER-Random Forest and GWO for Bitcoin Tweet Sentiment Analysis. IEEE Access 2022, 10, 101889–101897. [Google Scholar] [CrossRef]
  46. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks; Navigli, R., Gangemi, A., Hollink, L., Vidal, M.-E., Alam, M., Hitzler, P., Tordai, A., Troncy, R., Eds.; Springer International Publishing AG: Cham, Switzerland, 2018; Volume 10843, pp. 593–607. [Google Scholar]
  47. Feng, S.; Wan, H.; Wang, N.; Li, J.; Luo, M. TwiBot-20: A Comprehensive Twitter Bot Detection Benchmark; ACM: New York, NY, USA, 2021; pp. 4485–4494. [Google Scholar]
  48. Cresci, S.; Di Pietro, R.; Petrocchi, M.; Spognardi, A.; Tesconi, M. Fame for sale: Efficient detection of fake Twitter followers. Decis. Support Syst. 2015, 80, 56–71. [Google Scholar] [CrossRef]
  49. Lipton, Z.C.; Elkan, C.; Narayanaswamy, B. Thresholding Classifiers to Maximize F1 Score. arXiv 2014. [Google Scholar] [CrossRef]
  50. Chicco, D.; Jurman, G. The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genom. 2020, 21, 6. [Google Scholar] [CrossRef]
  51. Lee, K.; Eoff, B.D.; Caverlee, J. Seven Months with the Devils: A Long-Term Study of Content Polluters on Twitter. In Proceedings of the 5th International AAAI Conference on Weblogs and Social Media, ICWSM 2011, Barcelona, Spain, 17–21 July 2011; AAAI Press: Barcelona, Spain, 2011; pp. 185–192. [Google Scholar]
  52. Yang, K.-C.; Varol, O.; Hui, P.-M.; Menczer, F. Scalable and Generalizable Social Bot Detection through Data Selection. Proc. AAAI Conf. Artif. Intell. 2020, 34, 1096–1103. [Google Scholar] [CrossRef]
  53. Kudugunta, S.; Ferrara, E. Deep neural networks for bot detection. Inf. Sci. 2018, 467, 312–322. [Google Scholar] [CrossRef]
  54. Wei, F.; Nguyen, U.T. Twitter Bot Detection Using Bidirectional Long Short-Term Memory Neural Networks and Word Embeddings. In Proceedings of the 2019 First IEEE International Conference on Trust, Privacy and Security in Intelligent Systems and Applications (TPS-ISA), Los Angeles, CA, USA, 12–14 December 2019; pp. 101–109. [Google Scholar]
  55. Miller, Z.; Dickinson, B.; Deitrick, W.; Hu, W.; Wang, A.H. Twitter spammer detection using data stream clustering. Inf. Sci. 2014, 260, 64–73. [Google Scholar] [CrossRef]
  56. Alhosseini, S.A.; Najafi, P.; Tareaf, R.B.; Meinel, C. Detect me if you can: Spam bot detection using inductive representation learning. In Proceedings of the 2019 World Wide Web Conference, WWW 2019, San Francisco, CA, USA, 13–17 May 2019; Association for Computing Machinery: San Francisco, CA, USA, 2019; pp. 148–153. [Google Scholar]
  57. Feng, S.; Wan, H.; Wang, N.; Li, J.; Luo, M. SATAR: A Self-Supervised Approach to Twitter Account Representation Learning and Its Application in Bot Detection. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, New York, NY, USA, 1–5 November 2021; ACM: New York, NY, USA, 2021; pp. 3808–3817. [Google Scholar]
  58. Mazza, M.; Cresci, S.; Avvenuti, M.; Quattrociocchi, W.; Tesconi, M. RTbust: Exploiting Temporal Patterns for Botnet Detection on Twitter. In Proceedings of the 10th ACM Conference on Web Science, New York, NY, USA, 27–30 May 2018; ACM: New York, NY, USA, 2019; pp. 183–192. [Google Scholar]
  59. Shi, S.; Qiao, K.; Yang, J.; Song, B.; Chen, J.; Yan, B. RF-GNN: Random Forest Boosted Graph Neural Network for Social Bot Detection. arXiv 2023. [Google Scholar] [CrossRef]
  60. Hayawi, K.; Mathew, S.; Venugopal, N.; Masud, M.M.; Ho, P.-H. DeeProBot: A hybrid deep neural network model for social bot detection based on user profile data. Soc. Netw. Anal. Min. 2022, 12, 43. [Google Scholar] [CrossRef]
  61. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised Contrastive Learning. arXiv 2020. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.