Next Article in Journal
Bifurcation and Optimal Control Analysis of an HIV/AIDS Model with Saturated Incidence Rate
Previous Article in Journal
Enhancing Stock Price Forecasting with CNN-BiGRU-Attention: A Case Study on INDY
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Backfire Effect Reveals Early Controversy in Online Media

1
Institute of Cyberspace Security, Zhejiang University of Technology, Hangzhou 310023, China
2
Binjiang Institute of Artificial Intelligence, Zhejiang University of Technology, Hangzhou 310056, China
3
Center for Computational Communication Research, Beijing Normal University, Zhuhai 519087, China
4
School of Journalism and Communication, Beijing Normal University, Beijing 100875, China
*
Authors to whom correspondence should be addressed.
Mathematics 2025, 13(13), 2147; https://doi.org/10.3390/math13132147
Submission received: 3 June 2025 / Revised: 24 June 2025 / Accepted: 26 June 2025 / Published: 30 June 2025

Abstract

The rapid development of online media has significantly facilitated the public’s information consumption, knowledge acquisition, and opinion exchange. However, it has also led to more violent conflicts in online discussions. Therefore, controversy detection becomes important for computational and social sciences. Previous research on detection methods has primarily focused on larger datasets and more complex computational models but has rarely examined the underlying mechanisms of conflict, particularly the psychological motivations behind them. In this paper, we propose a lightweight and language-independent method for controversy detection by introducing two novel psychological features: ascending gradient (AG) and tier ascending gradient (TAG). These features capture psychological signals in user interactions—specifically, the patterns where controversial comments generate disproportionate replies or replies outperform parent comments in likes. We develop these features based on the theory of the backfire effect in ideological conflict and demonstrate their consistent effectiveness across models and platforms. Compared with structural, interaction, and text-based features, AG and TAG show higher importance scores and better generalizability. Extensive experiments on Chinese and English platforms (Reddit, Toutiao, and Sina) confirm the robustness of our features across languages and algorithms. Moreover, the features exhibit strong performance even when applied to early-stage data or limited “one-page” scenarios, supporting their utility for early controversy detection. Our work highlights a new psychological perspective on conflict behavior in online discussions and bridges behavioral patterns and computational modeling.

1. Introduction

The definition of controversy or conflict in communication is the interaction of interdependent people who perceive the opposition of goals, aims, and/or values and who see the other party as potentially interfering with the realization of these goals [1]. Debates about religion, politics, intellectual ideology, and even personal preferences occur online and offline daily, and they can disrupt effective communication, leading to cognitive polarization, extremist behavior, and even war [2]. Therefore, ideological conflict has been one of the stresses in modern society that permeates all human activities.
The proliferation of online media has created the potential for more conflicts, and this can be attributed to the nature of the Internet. First, the Internet breaks through traditional geographical restrictions, allowing users with different cultural traditions, religious beliefs, and living habits to contact each other directly, resulting in unprecedented conflicts [3]. Second, the Internet makes users feel anonymous, resulting in the lack of necessary personal restraint in the online discussion process. Third, online media, especially social media based on user-generated content, are leaner and lack the necessary regulation, which has led to the proliferation of disinformation and further intensified conflicts [4]. Finally, artificial intelligence technology in online media could also spark more conflict [5]. These characteristics may result in volatile conversations that are more dynamic in the number of participants and opinions. While this can allow users to engage with each other and discuss differing viewpoints respectfully, it can also lead to escalated disagreements among users.
Given the universality and importance of online conflict, its research has attracted attention from both the social and computational sciences. Research from the social sciences has usually aimed at special controversial issues around critical politics and ideological topics, such as liberalism, equity, and environmental protection [6]. Social science mainly pays attention to the social, psychological, and cultural extension and connotation of these controversial issues and how users participate in discussing these issues [7]. In contrast, computational science has primarily focused on creating detection models in online media [8]. Facing the vast content of online media, efficient and accurate controversy detection methods are the technical basis for carrying out related research and the premise for effective conflict management.
Current controversy detection models are trained using features such as natural language, social networks, and user interactions to identify conflicting content at different levels, including topic-level, post level, and interaction-level [9]. However, the existing research still has certain limitations. With the development of artificial intelligence technology, natural language processing (NLP) methods have become the most natural choice for controversy detection. Due to natural languages’ inherent complexity and uncertainty, these models depend highly on language types and the pertinence of corpus [10]. As a result, these models cannot be used directly in other languages or topics. Unlike language-based models, detection models based on social networks and user interaction have better generalizability [11,12,13]. However, they usually exhibit severe data dependence, such as large-scale friend relationships or complete discussion processes. In multilingual and cross-platform environments, the challenge becomes more complex as models must cope with both behavioral variance and linguistic diversity. Prior studies, such as the use of artificial intelligence techniques for multilingual web traffic forecasting [14], have shown the potential of AI to generalize across languages. Inspired by this, our work also aims to design features that are robust across linguistic contexts, avoiding dependency on language-specific processing while preserving detection effectiveness. Meanwhile, the dilemma between universality, usability, and effectiveness requires further development of existing detection models, especially to discover more effective representations of online conflicts. The backfire effect, a cognitive phenomenon in which exposure to opposing views strengthens rather than weakens one’s original beliefs, has been extensively studied in recent years. It has been shown to amplify polarization in social media through mechanisms such as biased assimilation and homophily, as formalized in recent opinion dynamics models [15,16]. We hypothesize that the backfire effect may also manifest in measurable patterns of user behavior in real-world media platforms.
We aim to resolve this dilemma at the post level (whether a piece of news or a post will create conflict) by finding new conflicting features (Figure 1). Instead of analyzing and detecting a post using all the information of the post, we use the data that can be obtained within the “one page” containing the content and related comments of the post to avoid relying on massive data. At the same time, we use the post and the comment data with the earlier posting time to achieve early controversy detection. More thoughtfully, we consider the diversity of online media (including forums and news websites) and only select commonly available data for the various platforms. By considering the psychological mechanisms behind features, we have introduced two novel features for controversy detection across media platforms from different languages and types. We quantify the value of the new features by comparing the discernibility with other frequently used features. The new features can facilitate the development of better detection models, and this feature-mining method based on psychology and behavior also provides new pathways for future social media research. The main contributions of this paper include the following:
(1)
Propose two novel psychological features—ascending gradient (AG) and tier ascending gradient (TAG)—based on the backfire effect to capture ideological resistance in user interactions for effective controversy detection.
(2)
Demonstrate the robustness and generalizability of AG and TAG across multiple platforms (Reddit, Toutiao, Sina), languages (English and Chinese), and classification algorithms, outperforming traditional features.
(3)
Validate the practical utility of the proposed features in early-stage and limited-data scenarios (e.g., “one-page” content), enabling efficient and interpretable controversy detection for real-world applications.
The remainder is organized as follows. Section 2 provides an overview of the related work on controversy detection, focusing on different categories of features. Section 3 outlines multilingual and multi-platform datasets, candidate features, and the detection method. Section 4 presents the analysis of the experiments. Finally, Section 5 summarizes the key findings and highlights future research directions.

2. Related Work

Many features have been used for controversy detection, which can be divided into four categories: user characteristics, text-based features, social networks, and discussion interactions. In this section, we briefly summarize the research on these features.

2.1. User Characteristics

Reviewing previous research, distinct population characteristics can lead to different types of participation in conflict. For example, Emily and Davinia showed that there was a big difference between males and females when it came to aggressive behavior [17], and Holt and DeVore found that men were more likely to report using force than women in individualistic cultures and with regard to organizational role, men were more likely than women to choose a forcing style with their superiors [18]. Triandis demonstrated that the differences between culture and race had profound implications for the probability of conflict and the type of conflict [19]. Cheng et al. [20] showed that antisocial behavior in online groups is a relatively stable individual difference. Levy et al. [21] also proved that user activity is also closely related to conflict behaviors. In addition, the previous behavior of the users participating in the discussion can also be used to detect a future conflict. Saveski et al. [12] found that the structural characteristics of the conversation were also predictive of whether the next reply posted by a specific user would be toxic or not. These works emphasize the predictive capacity of user characteristics. However, they suffer from practical limitations: such characteristics often require access to external data beyond the discussion page, may violate user privacy, and are not always available in anonymous environments.

2.2. Text-Based Features

In controversy detection, the text data of posts or comments are usually easy to obtain, which is conducive to fully mining sentiments, emotions, and opinions [9,22,23,24]. These papers have analyzed the intensity of emotions and sentiments, showing that these text-based features are strong indicators of controversy. Weingart et al. [25] analyzed conflict spirals in the workplace that were influenced by conflict expressions varying in directness and oppositional intensity. Zhang et al. [26] predicted from the very start of a conversation whether it would get out of hand within Wikipedia talk page discussions through various text-based features such as politeness strategies and prompt types. Text-based methods rely fundamentally on NLP techniques, which are inherently limited by language specificity and corpus dependency. Moreover, compounded by social media’s informal expressions and short texts, it reduce sentiment analysis accuracy across languages and platforms—motivating language-independent alternatives.

2.3. Social Networks

In addition to user characteristics, researchers study the conflict commenter’s social network structure. Conover et al. [27] demonstrate the existence of controversial structures in Twitter through network modularity and graph partitioning based on user relationships and behavioral relationships in social networks. Morales et al. [28] quantified the controversial nature of topics through opinion leaders in the network structure. Akoglu et al. [29] constructed a bipartite graph based on user relationships to quantify the contentiousness of information. Coletto et al. [11] studied users’ connections in the context of controversial threads. To do so, they analyzed local network patterns of user-follower and user-reply graphs. Their findings showed that controversial interactions are less likely between users who follow each other on social media. de Dreu argued that conflicts occurred more often in intergroups than in interpersonal [30]. Many of these conflicts are identity-based, in which people believe that the group or subgroup they identify with (e.g., ethnicity, race, religion, or political party) is superior to an outgroup [31]. While many conflicts originate from differences between intergroups, these intergroup conflicts can evolve into interpersonal ones [32], and frequent pleasant interactions can reduce intergroup conflicts [33].
Furthermore, some scholars combine network structure with text semantics. Zhong et al. [34] proposed a controversial microblog detection algorithm based on a Graph Convolutional Network. This algorithm builds a topic-microblog-comment network based on behavioral relationships and combines the textual information of topics and comments to detect controversial details. Benslimane et al. [35] proposed a method for contentious information detection based on graph structure and textual features. The technique embeds graph representations (including text features) into feature vectors [36], considers the attention mechanism when calculating user nodes, fuses the importance of neighbor nodes, and finally performs the classification task through the Graph Neural Network. These methods uncover underlying group dynamics but often rely on large-scale and structured social network data, which is unavailable on many platforms, such as news websites. Our study seeks to avoid such dependency by proposing lightweight features extracted from visible interactions.

2.4. Discussion Interactions

In addition to social relations, users will also discuss and interact through comments below the post. The structure of this interaction has also been shown to be related to the conflict. Coletto et al. [11] not only analyzed the relationship between user-follower and user-reply networks but also investigated intensity through a different perspective by assuming controversial topics may generate “dense” discussions in time so that the inter-reply rate for these conversations is lower (i.e., more rapid replies) than those of non-controversial ones. Such interaction structures capture valuable behavioral signals that reflect emotional involvement, disagreement, and user polarization. Discussion interaction contains the psychological and behavioral characteristics of users expressing opinions on conflicting issues, and its potential needs further exploration. However, existing methods based on interaction features often rely on manually defined structural heuristics, which may not fully capture the latent psychological patterns associated with ideological conflict. Additionally, few works explicitly connect these behaviors to cognitive theories, such as the backfire effect, which explains how users react to opposing views. Motivated by this gap, our study aims to uncover and formalize behavioral patterns within discussion interactions that are theoretically grounded and computationally lightweight.

3. Methodology

Motivated by the limitations of existing approaches and the need to better capture psychologically behavioral signals in online interactions, we construct a lightweight, effective controversy detection model, as illustrated in Figure 2. The proposed model is designed to uncover latent indicators of ideological disagreement embedded in discussion interactions. The framework consists of three major components: Data Processing, Feature Extraction, and Model Building. This section will elaborate on each component of the proposed model in detail. The effectiveness of the model will be further validated through specific tasks.

3.1. Data Processing

To evaluate the discriminative power of features in different languages and types of online media, we collected three datasets from various sources: Reddit (English), Toutiao (Chinese), and Sina (Chinese). Each dataset contains posts from multiple sets of conflicting and non-conflicting topics from their respective platforms. Our methodology is not language-dependent, as it focuses on user behaviors such as likes and replies. As long as a platform supports these interactions, the approach is applicable. In this study, we use Chinese and English media as representative examples.
The English-based media platform is Reddit. For the Reddit platform, we use the May 2015 public dataset from the kaggle.com platform. We filtered controversial topics from two subreddits using keywords according to the list from the library of Shippensburg University. The two topics mentioned above, Gun and War, and the content under the “DebateReligion” subreddit make up the three controversial topics. Meanwhile, we get three mild topics of Shopping, Scenery, and Music around the three subreddits “Random_Acts_Of_Amazon”, “Watches”, and “Guitar”. The dataset contains the complete information of posts, comment information, and the relationship between each comment. To ensure that the posts under each topic have a complete relationship structure and diverse information attributes, we mainly focus on posts with more than 50 comments. Finally, the dataset contains 542 news items and 110,176 comments (The main data in Table 1).
The Chinese-based media platforms are Toutiao and Sina. Toutiao, or Jinri Toutiao, is a news and information platform based on user-generated content and recommendation algorithms. The dataset includes news information, comment information, and the relationship of comments. The period of the data is from January 2019 to December 2019. In Toutiao, we choose the topics of Huawei, NBA, and Football that have apparent controversy and the topics of Life, Food, and Travel with minor controversy. The raw data of this dataset is collected by the web crawler, and then we clean the raw data to remove the empty and duplicate values. In the end, there were 647 news articles and 119,725 comments in the dataset. Sina is one of the most influential news platforms on the Chinese Internet. The period of crawled data is from January 2019 to December 2019. We select the same controversial topics as the Toutiao, as well as mildly controversial topics such as Science, Music, and Movies, for a total of 728 news items and 104,462 comments (the test data in Table 1).
Our dataset has a clear hierarchical structure. Topics (subreddits) allow users to come together to discuss topics of interest, share relevant content and resources, interact with others, and build a community focused on a specific topic. Posts are the body for our subsequent analysis, containing information such as the post topic, the post time, and the user ID. Comments are the key unit of our analysis and contain information such as comment text, comment time, and number of likes.
This paper uses P to denote a post, that is
P = Topic , P i d , P t , T
where P i d is the individual ID of a post, P t is the time when a post was published, and T is the post content.
We define a comment as C, that is:
C = P i d , C i d , C t , C l , T , C p i d
where P i d is the ID of a post to which the comment belongs, C i d is the comment ID, C t is the time that the comment was posted, C l is the number of likes for a comment, T is the textual content, and  C p i d is the ID of the parent text of the comment (i.e., the previous level of text to which the current comment belongs).
The form of a graph can be good for representing and understanding posts. We define a node set V = v 0 , v 1 , v 2 , , v n , where v 0 represents the post P, v i ( i 0 ) denotes the comments, and n is the number of comments on the post. Then, we define the edge set E = e 1 , e 2 , , e n . Where, if the parent text ID C p i d of node v j points to node v i , then the connected edge e j = < v i , v j > is formed. From this, we can construct a comment graph G = ( V , E ) for different data in terms of posts, and Figure 2a is a simple schematic diagram.

3.2. Feature Extraction

Based on multilingual and multi-platform datasets, we choose the common features among them as comparison objects. This web page usually contains the post title, content, and comments on the post and their replies. Therefore, we consider structural, interactive, and textual features. Structural features describe the tree-like organization of comments and replies, quantifying the discussion’s shape and connections. Interactive features capture user engagement and temporal dynamics within the discussion, such as reply times and comment frequency. Textual features analyze the emotions in both the main post and its comments. Note that we did not consider user characteristics, as this was far beyond the data presented by the information presentation page. We aim to identify conflicting posts with limited information quickly.
First, we consider features related to comment structure. A post and its comments can form a tree structure, where the root is the post and other nodes are comments, and these nodes are connected by comments or reply relationships. Based on the tree structure, we can extract the following features [37,38]:
  • Size (n). The size of a tree corresponds to the number of nodes (without root) in that tree. For example, the size of the demo post in Figure 2 is 5.
  • Depth (d). The depth of a node is the number of links from the node to the root. The depth of a tree is the maximum depth of the node in all nodes. In other words, the depth of a tree, d, with n nodes is defined as
    d = m a x ( d i ) , 0 i n
    where d i denotes the depth of node i. For example, in Figure 2, d = 3 .
  • Breadth (b). The breadth of a cascade is a function of its depth. At each depth, the breadth is the number of nodes at that depth. The breadth of a tree is its maximum breadth at all depths. For a tree with depth d, the breadth, b, is defined as
    b = m a x ( b i ) , 0 i d
    where b i denotes the breadth at depth i. For example, in Figure 2, b = 2 .
  • Average degree (k). In a tree, the degree of a node is only the number of direct children of a node. Therefore, the average degree of a tree is defined as
    k = i = 1 n k i n , 0 i n
    where k i denotes the degree of node i. For example, in the Figure 2, k = 1.2 .
  • Structural Virality (v). The structural virality of a cascade, is the average distance between all pairs of nodes in a cascade. For a cascade with n > 1 nodes, the virality v is defined as
    v = 1 n ( n 1 ) i = 1 n j = 1 n d i j
    where d i j denotes the length of the shortest path between nodes i and j. For example, in Figure 2, v = 1.1 .
Second, we considered text-based features, including the following:
  • Intensity of comment emotion ( s c ). We combine two sentiment methods, Vader and SnowNLP, to achieve efficient semantic analysis of Chinese and English [39,40]. Based on the sentiment dictionary, a sentiment score c i can be calculated for each comment (node i), and then the comment emotional intensity of a post can be obtained by the following equation:
    s c = i = 1 n c i n , 0 i n
  • Intensity of post emotion ( s p ). The emotional intensity of a post is also based on the above sentiment dictionary, and the sentiment score of an entire post’s content is calculated as s p .
Third, the features of discussion interaction are also considered [11]:
  • Reply time ( t ). For each reply link e x = i , j in a post, the time elapsed between the parent comment (node i) and its child (node j) is reply time ( t i j ). We consider average ( t a v g ) and minimum ( t m i n ) values of all e x .
  • Comment density ( c ). For a post, the time interval from its publication to the latest comment is δ ; then, the density of comments can be defined as
    c = n δ
    where n denotes the number of comments.
Unlike the previous features, we consider the number of likes and comments [21]. We first define L i to denote the number of likes received by comment i and consider the average number of likes received by all comments:
q = i = 1 n L i n
Second, we hypothesize that when users view a comment with an opposing opinion, they are more willing to press “like” to reply to criticize the comment. So, we define each reply edge as e x = i , j , where i is the parent comment and j is the reply. The ascending gradient function is then defined based on whether the reply comment receives more likes than its parent:
g a ( e x ) = 1 , if L i L j < 0 ; 0 , otherwise .
For all edges, we can obtain the proportion of ascending gradient as
p a = x = 1 m g a ( e x ) m
where m is the number of reply edges in a post.
Furthermore, we expand the concept of ascending gradient to encompass the difference in the number of replies across comments, which we refer to as a “tier ascending gradient”. First define R i for each comment, i.e., the total number of comments that directly reply to comment i. Then, similar to Equations (10) and (11), we can obtain the functional representation of the tier-ascending gradient as
g t ( e x ) = 1 , if R i R j < 0 ; 0 , otherwise .
p t = x = 1 m g t ( e x ) m
All features are also summarized and described in Figure 2b. For convenience, we subsequently abbreviate the ascending gradient as AG and the tier ascending gradient as TAG and refer to them uniformly as psychological features (this will be analyzed in the experimental section to explain the psychological property of our proposed features). Finally, all the features and their descriptions that we have compiled are shown in Table A1.

3.3. Model Construction

LightGBM is a gradient enhancement framework based on the decision tree algorithm proposed by Microsoft Research in 2017 [41], which is widely used in classification or regression tasks. LightGBM optimizes the defects of Gradient Boosting Decision Tree (GBDT) that exist in high time and space consumption, supports efficient parallel training, and has the advantages of faster training speed, lower memory consumption, and a wider amount of processed data. It can also be called the GBDT algorithm using Gradient-based One-Side Sampling (GOSS) [41] and Exclusive Feature Bundling (EFB) [41], so first is a brief introduction to the core strategies of GOSS and EFB.
GOSS: The GOSS algorithm starts from the point of view of reducing samples, excludes most of the samples with small gradients, and uses only the remaining samples to calculate the information gain, realizing the reduction in the data volume while guaranteeing the balance of accuracy. Therefore, GOSS only retains the data with larger gradients when sampling data, but if all the data with smaller gradients are directly discarded, it will definitely affect the overall distribution of the data. Therefore, GOSS firstly arranges all the values of the features to be split in descending order of absolute value size and selects a 100 % of the data with the largest absolute value. Then b 100 % of the remaining smaller gradient data are randomly selected. This b 100 % of the data is then multiplied by a constant ( 1 a ) / b so that the algorithm focuses more on the under-trained samples and does not change the distribution of the original dataset too much. Finally, this ( a + b ) 100 % of the data is used to calculate the information gain.
EFB: In order to solve the sparsity of high-dimensional data, mutually exclusive features are bundled to achieve feature dimensionality reduction without losing information. The process is mainly divided into four steps:
(1)
Take features as the node of a graph, for the features that are not mutually exclusive are connected (i.e., there exist samples that are not 0 at the same time), and the number of samples whose features are not 0 at the same time are used as the weights of the edges.
(2)
Sort the features in descending order based on the degree of the nodes, with a larger degree indicating that the features are in greater conflict with other features (the less it will be bundled with other features).
(3)
Set the maximum conflict threshold K, traverse the existing feature clusters, and if it is found that the number of conflicts for the feature to be added to the feature cluster will not exceed the maximum threshold K, then the feature is added to the cluster. Otherwise, create a new feature cluster and add the feature to the newly created cluster.
(4)
Separate the original feature from the merged feature by adding an offset constant.
Among them, the first three steps solve the problem of determining which features should be bundled together, and the last step solves the problem of how to bundle (merge) multiple features into one.
In addition to this, the histogram-based decision tree strategy allows the algorithm to run with a smaller memory footprint and less computational cost. The leaf-wise strategy with depth constraints ensures the high efficiency of the algorithm while preventing overfitting. So, LightGBM is our main algorithmic model. To further clarify the implementation, we provide the pseudo-code of the entire method, as shown in Algorithm 1.
Algorithm 1 Controversy Detection Model.
Require:  dataset = { post 1 , post 2 , , post m } ▹ Collection of posts with comment trees and labels
Ensure: trained_model▹ LightGBM classifier for controversy detection
1: // Initialize LightGBM with optimization techniques
2: model ← LightGBM
3: SetParam(model, “objective”, “binary”)▹ Binary classification
4: SetParam(model, “boosting”, “gbdt”)▹ Gradient Boosting Decision Trees
5: SetParam(model, “sampling”, “GOSS”)▹ Gradient-based One-Side Sampling
6: SetParam(model, “bundling”, “EFB”)▹ Exclusive Feature Bundling
7: // Process each post in the dataset
8: for each post ∈ dataset do
9:      features ← []▹ Initialize feature vector
10:     // Extract structural features
11:     features[0] ← CommentCount(post)▹ Number of comments (n)
12:     features[1] ← MaxDepth(post.tree)▹ Maximum depth (d)
13:     features[2] ← MaxBreadth(post.tree)▹ Maximum breadth (b)
14:     features[3] ← AverageDegree(post.tree)▹ Average node degree (k)
15:     features[4] ← StructuralVirality(post.tree)▹ Avg. path length (v)
16:     // Extract textual features
17:     features[5] ← AvgCommentSentiment(post)▹ Avg. comment sentiment ( s c )
18:     features[6] ← PostSentiment(post)▹ Post sentiment intensity ( s p )
19:     // Extract interaction features
20:     features[7] ← AverageReplyTime(post)▹ Avg. reply time ( t a v g )
21:     features[8] ← MinimumReplyTime(post)▹ Min. reply time ( t m i n )
22:     features[9] ← CommentDensity(post)▹ Comments per time unit (c)
23:     // Extract psychological features (novel contribution)
24:     features[10] ← LikeAscendingGradient(post)▹ AG proportion ( p a )
25:     features[11] ← ReplyAscendingGradient(post)▹ TAG proportion ( p t )
26:     AddTrainingSample(model, features, post.label)
27end for
28// Configure optimization parameters
29SetGOSSParam(model, top_rate=0.2, other_rate=0.1)
30SetEFBParam(model, max_conflict=5)
31SetParam(model, “num_leaves”, 31)
32SetParam(model, “learning_rate”, 0.05)
33// Train model with GOSS and EFB optimizations
34Train(model)
35// Predict controversy probability
36: prob ← Predict(model, features)
37: return model, prob

4. Results and Discussion

In this section, detailed experiments are conducted around the task realization designed in Section 3, and the results are presented in three parts as follows: algorithm evaluation results, importance evaluation results, and application evaluation results.

4.1. Controversy Detection Results

If suitable features are chosen, even simple classification models can achieve excellent detection results. Based on this viewpoint, we first validate the enhancement effect of psychological features on different controversial topic detection algorithms, which include interval-based SVM, distance-based KNN, probability-based LR, tree-structure-based DT, integrated learning-based GBDT, and LightGBM. The detailed results, as shown in Table 2, reveal the detection performance across four types of features—structure, interaction, text, and psychology—under each algorithm. The values in the table represent the classification accuracy (%) of each feature group under each algorithm. Across all classical classifiers, the best performance is achieved when using psychological features, with LightGBM and SVM both reaching 93.25% accuracy on the Reddit dataset. Compared with structure, interaction, and text features, psychological features consistently contribute the highest performance across almost all algorithms. When psychological features are removed, all algorithms show a clear performance drop, with a maximum reduction of over 10%, indicating their irreplaceable contribution. We further validate this trend on Chinese platforms using the current best-performing LightGBM algorithm. As shown in Table 2, psychological features again lead to the highest detection accuracy: 90.21% on Toutiao and 83.66% on Sina, outperforming all other feature groups. In the English platform (Reddit), text-based features also show strong support due to the maturity of NLP techniques in English, while in the Chinese datasets, text features perform less effectively, partly due to the limitations in sentiment analysis for short Chinese texts. This highlights the language sensitivity of text features, whereas our psychological features show cross-lingual robustness and generalizability. Unlike the volatility seen with structural, interactive, or text-based features—whose effectiveness can fluctuate with data scale or language—the psychological features maintain a stable and strong enhancement across algorithms and platforms. This stability underlines their potential as reliable indicators and suggests value for further exploration in future research.

4.2. Importance Evaluation Results

4.2.1. Quantitative Analysis

At the post level, we mixed all posts regardless of topic and assessed the importance of features using three classification algorithms. The results show that in all algorithms and platforms, the two features ( p a , p t ) based on the ascending gradient are far more critical than the others in distinguishing controversies and are generally in the top three positions of importance (Figure 3a–c). In the RF algorithm, the two features are in the first and second positions, respectively, with values up to 0.27 and 0.14. In the XGBoost algorithm, the gap between the first feature and the second feature can even reach nearly three times. The SHAP [42] values of each feature also show the same effect (Figure 3d), i.e., p a is generally in the first position, and psychology-based features are commonly in the top three positions. It is worth noting that in Figure 3a, some features (such as post-emotion, comment emotion, and structural breadth) show negative values. This is because the linear regression-based importance measure allows negative coefficients, indicating that those features are negatively correlated with the likelihood of a post being controversial. This also reflects the inherent volatility in the emotional and structural aspects of online content. Overall, for the discriminative ability of conflicting posts, the value of psychological features is significantly better than other features, while p a is generally better than p t .
Second, we used the Kolmogorov–Smirnov test to compare the differences in the distribution of each feature across the three conflicting and non-conflicting topics of the Reddit dataset (Appendix A Figure A1). The results showed that among all five comment structure features, only breadth and structural virality differed significantly between conflicting and non-conflicting topics ( p < 0.05 ). Also, there were significant differences between specific topics in terms of size and depth. For text-based features, the sentiment intensity of comments mostly showed certain differences when comparing conflicting and non-conflicting topics, which is consistent with previous findings [23]. We found a significant difference in the distribution of average replay time and average ups for discussions between conflicting and non-conflicting topics, with slightly shorter phrases for controversial topics and non-controversial topics.
Among the different types of topics in the Reddit platform, for the gradient-related features (AG features in blue and TAG features in orange in Figure 4), there is a clear distributional difference, i.e., the feature values of non-controversial topics are generally lower than those of controversial features. Controversial topics such as “Gun”, “War”, and “Religion” have significantly higher AG and TAG values than non-controversial ones like “Shopping” or “Music”, indicating that user engagement—measured by likes and replies—increases more rapidly and persistently as discussions deepen. In particular, the average value of the AG feature in controversial topics is about 20% higher than that of non-controversial topics, which shows that the feature is a common situation among different types of topics. To verify that the difference in the average number of likes and replies to comments is also a common phenomenon across different platforms, we conducted a validation on two news platforms, Toutiao and Sina, whose results show a very obvious difference as well. Although these platforms mainly support flat comment structures, controversial topics such as “Huawei” and “NBA” still show noticeably higher AG and TAG values compared with neutral topics like “Food” or “Life”, while the latter often approach zero. Meanwhile, because the news platforms are mostly dominated by first-level comments, the feature magnitude of the controversial topics is a few times different from that of Reddit platforms, and the gradient feature values in terms of non-controversial topics are dominated by 0. In summary, the controversial topic gradient feature values are generally higher than the non-controversial topics, with significant fluctuations. These results demonstrate that gradient-based features effectively distinguish controversial content, even across platforms with differing comment structures.
To summarize, psychological features are significant and valuable in detecting controversies, and they can be applied to various media platforms in both Chinese and English.

4.2.2. Psychological Interpretation

Why does the ascending gradient perform well in discriminating between conflicting topics and posts? From a psychological point of view, the backfire effect could explain the emergence of the ascending gradient of likes and replies. A recent wave of studies suggested exposure to those with opposing political views may create backfire effects that stimulate the user’s cognitive resistance and exacerbate political polarization [15,43]. In theory, the deeper the level of comments, the more difficult it is to be exposed and the lower the likelihood of getting likes or replies. But in controversial posts, when some users (who may be silent by default) first see a comment that challenges their opinion at a lower level, and then, due to the backfire effect, they will be more insistent on their opinions. They will be motivated to seek psychological compensation. Therefore, when these users find that other users are refuting or fighting against this comment, they are more likely to be encouraged to express their desire to express it through likes or replies. In the upper level, comments that are consistent with the user’s point of view are not easy to get likes or replies from the type of users because they are not stimulated by the backfire effect. Since this excitation is consistent with user groups with opposing opinions, it leads to the emergence of ascending gradient situations in controversial posts. Due to this, we refer to the proposed gradient features as psychological features.
The backfire effect reminds us that in the discussion of controversial topics, users will be repeatedly exposed to opposite opinions, which will inevitably change their psychological state. It can be considered that the increase in emotional intensity and negativity in the discussion is related to this, but emotion recognition relies on natural language processing technology, which has certain limitations. In a word, the user’s mental state and behavior leading to conflict is a causal relationship, while identifying conflict based on techniques such as natural language processing is a process of data analysis. Next, we apply psychological features to several real-life application scenarios.

4.3. Application Evaluation Results

In view of the above sufficient validation, we focus on more efficient application scenarios applied to topic controversy detection for a small amount of hot information (one page) or partial information (early detection). We demonstrate that our approach achieves efficient controversy detection with minimal information while providing an interpretable analysis of its performance.

4.3.1. One Page

For each post page under the social platform, the first thing that is displayed is the highly liked comments to get an idea of the hot topics being discussed at the moment. As a result, most of the debates are also centered around the hot comments, while the hidden comments, due to low attention, have less influence on the judgment of topic conflicts. Therefore, we propose “one-page”, that is, using a small amount of intuitively visible hot information to achieve highly accurate controversy detection. For the experimental validation of hot comments, we focus only on comments with a high number of likes (top 20%) and their reply pairs in the current post. Meanwhile, the four structural features are not used because the global tree structure is lost. With only 20% of the hot comment information used, the LightGBM algorithm still has 87.12% detection accuracy with nine features. We again used the above three methods to assess the importance of features, and the results are shown in Figure 5a. Psychological features are generally in the top position, and their values are significantly better than other features.
As shown in Figure 5b, it is worth noting that the results suggest 0.2 may serve as a critical threshold where psychological features begin to exhibit strong differences. Although this effect is most prominent in the religion topic, it reveals a general pattern in feature dynamics. Specifically, TAG features dominate in the early stages, when only hot comments are available, due to their ability to capture surface-level engagement patterns such as reply density. In contrast, AG features become increasingly effective as more data accumulates, capturing deeper endorsement patterns such as the evolution of likes across reply layers. As the comment ratio increases, the inter-class variance of AG features expands while that of TAG features contracts. This demonstrates the complementary roles of AG and TAG: TAG is more informative with limited early data, while AG excels in full-data conditions, ensuring robust performance across different time stages. This dynamic also explains the difference in feature rankings between Figure 3 (full information) and Figure 5a (partial information).
We also explain the analysis in terms of the distribution of feature values (Figure 5c shows the case of posts with complete information on the left and the case of posts with hot information on the right), where the horizontal and vertical coordinates of the figure represent the two psychological features, and each point represents a post. It can be clearly seen that when the information is complete, the boundary between the two types of posts is clear, and the AG feature brings more distinctive differentiation. When the information is incomplete, a large number of non-controversial posts (in orange), are distributed all over the bottom, which makes the vertical TAG feature work more clearly than the horizontal AG feature. Overall, both AG features and TAG features can play an important role in controversy detection while also complementing each other in different situations.

4.3.2. Early Detection

The important findings described above also provide additional theoretical support for our realization of earlier conflict discovery. Because each comment of the data carries a posting time, we analyze the posts for different time periods. To do so, we sort all comments of each post by time, use the first 3 h of comments for the computation of features, and evaluate their controversy detection performance. In the time interval, each round linearly increases the number of comments by 3 h for multiple rounds of experiments, and in the features, each round adds different types of features on top of the existing ones, and the results are shown in Table 3. It can be seen that around 3 h after the post is sent out, the accuracy of detecting whether it is a controversial article or not can reach 85.28% with the possession of psychological features, which is close to reaching the late state of other feature modes. Using only 6 h of data approximated the performance of using 24 h of data to achieve 93.25% controversy detection performance. In addition to this, the detection performance of the psychological mode has a clear advantage in almost all stages, and the improvement becomes more obvious the further back we go (after “+” in the table are the values of the current column’s improvement relative to the best accuracy). Our feature model generally achieves its best detection performance in the first and middle phases and remains stable until the event is unattended, reflecting the model’s ability to recognize conflicts earlier as well as the stability of its features.

5. Conclusions

In this study, we collected data from multiple media platforms across English (Reddit) and Chinese (Toutiao, Sina). Our primary objective was to achieve high-performance controversy detection with minimal information, validated through “one-page” (20% hot comments) and early detection (3–6 h) experiments. By quantitatively evaluating feature discriminability in LR, RF, and XGBoost algorithms, we identified generalized features for controversy detection: the ascending gradient in likes/replies between superior and inferior comments proves to be an efficient and cross-lingual indicator. Crucially, conflicting posts exhibit significantly higher AG ratios, with >93.25% detection accuracy achieved using psychological features on Reddit (LightGBM/SVM). These novel AG/TAG psychological features demonstrated consistent algorithmic enhancement, boosting performance by >10% when added to baseline models and maintaining cross-platform robustness (90.21% on Toutiao, 83.66% on Sina). We interpret this through the backfire effect mechanism in ideological conflicts. For applications, even 20% hot comments (TAG-dominated) achieve 87.12% accuracy, with AG features gaining dominance beyond the critical 0.2 comment ratio threshold. Early detection using just 3–6 h of data reached 85.28–93.25% accuracy, proving our approach enables efficient controversy identification without complex semantic analysis or global network structures.

6. Future Work

This study has some limitations. Firstly, the algorithms used in this work are relatively basic, and we have not explored more advanced or diverse modeling approaches (e.g., deep learning architectures or ensemble strategies) that might further enhance performance or reveal deeper insights. Then, our current methodology primarily identifies controversial content based on the manifestation of conflict in user interactions. We have not, however, provided a direct causal correlation between the observed ascending gradient patterns and the emergence of conflict.
Despite these limitations, they point to several promising directions for future research. One key direction is to explore hybrid models that combine the advantages of our lightweight behavioral signals with the richness of content and structural information for more comprehensive controversy detection. Furthermore, future work could expand to more diverse platforms beyond Reddit and Toutiao to validate and refine our models, ensuring broader applicability. Investigating the underlying causal mechanisms that lead to the observed patterns, perhaps through controlled experimental designs, would also be a valuable direction to establish a direct causal link. Finally, while this study focuses on early controversy detection, expanding the temporal scope of analysis to understand the long-term evolution of controversies would offer deeper insights into online dynamics. These unresolved questions collectively challenge future computational sociology research, prompting us to explore how to solve problems in online media more quickly and effectively with limited data and computing power and whether psychological mechanisms can indeed yield local features more powerful than global ones.

Author Contributions

Conceptualization, S.P.; methodology, S.P. and T.J.; validation, S.P., T.J. and K.Z.; formal analysis, S.P., T.J. and K.Z.; investigation, S.P., T.J. and K.Z.; writing—original draft preparation, S.P.; writing—review and editing, S.P., T.J., Q.X. and Y.M.; visualization, S.P., T.J. and K.Z.; supervision, Q.X. and Y.M.; project administration, Q.X. and Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by the General Project of the Ministry of Education Foundation on Humanities and Social Sciences 23YJA860011; the Fundamental Research Funds for the Central Universities 1243200012; the Guangdong Philosophy and Social Science Foundation Regular Project GD24XXW02; the China Postdoctoral Science Foundation under Grant Number 2024M762912; the Postdoctoral Science Preferential Funding of Zhejiang Province of China under Grant ZJ2024060, the Key R&D Program of Zhejiang under Grant 2022C01018, and the National Natural Science Foundation of China under Grants U21B2001.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Features Comparison

We provide distributions of all features in topic levels at Figure A1 as an appendix.
Figure A1. The box plot of each topic in 11 features (except ascending gradient and tier ascending gradient). Compare the differences in the distribution of controversial and non-controversial topics under the same features.
Figure A1. The box plot of each topic in 11 features (except ascending gradient and tier ascending gradient). Compare the differences in the distribution of controversial and non-controversial topics under the same features.
Mathematics 13 02147 g0a1

Appendix B. Features Description

Table A1. Feature description.
Table A1. Feature description.
No.NameCategorySymbolDetailed Description
1SizeStructurenNumber of nodes in the comment tree
2DepthStructuredMaximum depth of the comment tree
3BreadthStructurebMaximum breadth of comment tree
4Average degreeStructurekThe average degree of the comment tree
5Structural viralityStructurevThe average shortest path of the comment tree
6Minimum reply timeInteraction t m i n The shortest time of all reply pairs
7Average reply timeInteraction t a v g Average time for all reply pairs
8Comment densityInteractioncNumber of comments per second on post
9Average upsInteractionqAverage of the number of likes of all comments
10Comment emotional intensityText s c Average of the emotional intensity of all comments
11Post emotional intensityText s p The emotional intensity of post content
12Ascending gradientPsychology p a Percentage of difference in likes for all reply pairs
13Tier ascending gradientPsychology p t The difference in the number of reply to all comments

References

  1. Folger, J.P.; Poole, M.S.; Stutman, R.K. Working Through Conflict: Strategies for Relationships, Groups, and Organizations; Routledge: London, UK, 2024. [Google Scholar]
  2. Jost, J.T.; Baldassarri, D.S.; Druckman, J.N. Cognitive–motivational mechanisms of political polarization in social-communicative contexts. Nat. Rev. Psychol. 2022, 1, 560–576. [Google Scholar] [CrossRef]
  3. Zhang, L. The Digital Age of Religious Communication: The Shaping and Challenges of Religious Beliefs through Social Media. Stud. Relig. Philos. 2025, 1, 25–41. [Google Scholar] [CrossRef]
  4. Vicario, M.D.; Quattrociocchi, W.; Scala, A.; Zollo, F. Polarization and fake news: Early warning of potential misinformation targets. ACM Trans. Web (TWEB) 2019, 13, 1–22. [Google Scholar] [CrossRef]
  5. Stella, M.; Ferrara, E.; De Domenico, M. Bots increase exposure to negative and inflammatory content in online social systems. Proc. Natl. Acad. Sci. USA 2018, 115, 12435–12440. [Google Scholar] [CrossRef] [PubMed]
  6. Littlejohn, S.W.; Foss, K.A. Encyclopedia of Communication Theory; Sage: New York, NY, USA, 2009; Volume 1. [Google Scholar]
  7. De Wit, F.R.; Greer, L.L.; Jehn, K.A. The paradox of intragroup conflict: A meta-analysis. J. Appl. Psychol. 2012, 97, 360. [Google Scholar] [CrossRef]
  8. Dori-Hacohen, S.; Allan, J. Automated controversy detection on the web. In Proceedings of the European Conference on Information Retrieval, Vienna, Austria, 29 March–2 April 2015; Springer: Cham, Switzerland, 2015; pp. 423–434. [Google Scholar]
  9. Al-Ayyoub, M.; Rabab’ ah, A.; Jararweh, Y.; Al-Kabi, M.N.; Gupta, B.B. Studying the controversy in online crowds’ interactions. Appl. Soft Comput. 2018, 66, 557–563. [Google Scholar] [CrossRef]
  10. Bide, P.; Dhage, S. Similar event detection and event topic mining in social network platform. In Proceedings of the 2021 6th International Conference for Convergence in Technology (I2CT), Maharashtra, India, 2–4 April 2021; pp. 1–11. [Google Scholar]
  11. Coletto, M.; Garimella, K.; Gionis, A.; Lucchese, C. Automatic controversy detection in social media: A content-independent motif-based approach. Online Soc. Netw. Media 2017, 3, 22–31. [Google Scholar] [CrossRef]
  12. Saveski, M.; Roy, B.; Roy, D. The structure of toxic conversations on Twitter. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 1086–1097. [Google Scholar]
  13. Hu, L.; Han, G.; Liu, S.; Ren, Y.; Wang, X.; Yang, Z.; Jiang, F. Dual-Aspect Active Learning with Domain-Adversarial Training for Low-Resource Misinformation Detection. Mathematics 2025, 13, 1752. [Google Scholar] [CrossRef]
  14. Samaan, S.S.; Korial, A.E.; Sarra, R.R.; Humaidi, A.J. Multilingual Web Traffic Forecasting for Network Management Using Artificial Intelligence Techniques. Results Eng. 2025, 26, 105262. [Google Scholar] [CrossRef]
  15. Nyhan, B.; Reifler, J. When corrections fail: The persistence of political misperceptions. Political Behav. 2010, 32, 303–330. [Google Scholar] [CrossRef]
  16. Dandekar, P.; Goel, A.; Lee, D.T. Biased assimilation, homophily, and the dynamics of polarization. Proc. Natl. Acad. Sci. USA 2013, 110, 5791–5796. [Google Scholar] [CrossRef] [PubMed]
  17. Theophilou, E.; Hernández-Leo, D.; Gómez, V. Gender-based learning and behavioural differences in an educational social media platform. J. Comput. Assist. Learn. 2024, 40, 2544–2557. [Google Scholar] [CrossRef]
  18. Holt, J.L.; DeVore, C.J. Culture, gender, organizational role, and styles of conflict resolution: A meta-analysis. Int. J. Intercult. Relat. 2005, 29, 165–196. [Google Scholar] [CrossRef]
  19. Triandis, H.C. Culture and conflict. Int. J. Psychol. 2000, 35, 145–152. [Google Scholar] [CrossRef]
  20. Cheng, J.; Danescu-Niculescu-Mizil, C.; Leskovec, J. Antisocial behavior in online discussion communities. In Proceedings of the International AAAI Conference on Web and Social Media, Oxford, UK, 26–29 May 2015; Volume 9, pp. 61–70. [Google Scholar]
  21. Levy, S.; Kraut, R.E.; Yu, J.A.; Altenburger, K.M.; Wang, Y.C. Understanding Conflicts in Online Conversations. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; pp. 2592–2602. [Google Scholar]
  22. Kumar, S.; Hamilton, W.L.; Leskovec, J.; Jurafsky, D. Community interaction and conflict on the web. In Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018; pp. 933–943. [Google Scholar]
  23. Choi, M.; Aiello, L.M.; Varga, K.Z.; Quercia, D. Ten social dimensions of conversations and relationships. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 1514–1525. [Google Scholar]
  24. Li, J.; Zhang, C.; Jiang, L. Innovative Telecom Fraud Detection: A New Dataset and an Advanced Model with RoBERTa and Dual Loss Functions. Appl. Sci. 2024, 14, 11628. [Google Scholar] [CrossRef]
  25. Weingart, L.R.; Behfar, K.J.; Bendersky, C.; Todorova, G.; Jehn, K.A. The directness and oppositional intensity of conflict expression. Acad. Manag. Rev. 2015, 40, 235–262. [Google Scholar] [CrossRef]
  26. Zhang, J.; Chang, J.P.; Danescu-Niculescu-Mizil, C.; Dixon, L.; Hua, Y.; Thain, N.; Taraborelli, D. Conversations gone awry: Detecting early signs of conversational failure. arXiv 2018, arXiv:1805.05345. [Google Scholar]
  27. Conover, M.; Ratkiewicz, J.; Francisco, M.; Gonçalves, B.; Menczer, F.; Flammini, A. Political polarization on twitter. In Proceedings of the International AAAI Conference on Web and Social Media, Barcelona, Spain, 17–21 July 2011; Volume 5, pp. 89–96. [Google Scholar]
  28. Garimella, K.; De Francisci Morales, G.; Gionis, A.; Mathioudakis, M. Reducing controversy by connecting opposing views. In Proceedings of the Tenth ACM International Conference on Web Search and Data Mining, Cambridge, UK, 6–10 February 2017; pp. 81–90. [Google Scholar]
  29. Akoglu, L. Quantifying political polarity based on bipartite opinion networks. In Proceedings of the International AAAI Conference on Web and Social Media, Ann Arbor, MI, USA, 1–4 July 2014; Volume 8, pp. 2–11. [Google Scholar]
  30. De Dreu, C.K. Social Conflict: The Emergence and Consequences of Struggle and Negotiation. In Handbook of Social Psychology; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2010. [Google Scholar]
  31. McKeown, S.; Haji, R.; Ferguson, N. Understanding Peace and Conflict Through Social Identity Theory: Contemporary Global Perspectives; Springer International Publishing: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  32. Labianca, G.; Brass, D.J.; Gray, B. Social networks and perceptions of intergroup conflict: The role of negative relationships and third parties. Acad. Manag. J. 1998, 41, 55–67. [Google Scholar] [CrossRef]
  33. Pettigrew, T.F.; Tropp, L.R.; Wagner, U.; Christ, O. Recent advances in intergroup contact theory. Int. J. Intercult. Relations 2011, 35, 271–280. [Google Scholar] [CrossRef]
  34. Zhong, L.; Cao, J.; Sheng, Q.; Guo, J.; Wang, Z. Integrating semantic and structural information with graph convolutional network for controversy detection. arXiv 2020, arXiv:2005.07886. [Google Scholar]
  35. Benslimane, S.; Azé, J.; Bringay, S.; Servajean, M.; Mollevi, C. Controversy Detection: A Text and Graph Neural Network Based Approach. In Proceedings of the International Conference on Web Information Systems Engineering, Melbourne, VIC, Australia, 26 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 339–354. [Google Scholar]
  36. Xuan, Q.; Wang, J.; Zhao, M.; Yuan, J.; Fu, C.; Ruan, Z.; Chen, G. Subgraph networks with application to structural feature space expansion. IEEE Trans. Knowl. Data Eng. 2019, 33, 2776–2789. [Google Scholar] [CrossRef]
  37. Goel, S.; Anderson, A.; Hofman, J.; Watts, D.J. The structural virality of online diffusion. Manag. Sci. 2016, 62, 180–196. [Google Scholar] [CrossRef]
  38. Vosoughi, S.; Roy, D.; Aral, S. The spread of true and false news online. Science 2018, 359, 1146–1151. [Google Scholar] [CrossRef] [PubMed]
  39. Min, K.; Ma, C.; Zhao, T.; Li, H. BosonNLP: An ensemble approach for word segmentation and POS tagging. In Natural Language Processing and Chinese Computing; Springer: Berlin/Heidelberg, Germany, 2015; pp. 520–526. [Google Scholar]
  40. Chou, E.; Tramer, F.; Pellegrino, G. Sentinet: Detecting localized universal attacks against deep learning systems. In Proceedings of the 2020 IEEE Security and Privacy Workshops (SPW), San Francisco, CA, USA, 21 May 2020; pp. 48–54. [Google Scholar]
  41. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  42. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  43. Bail, C.A.; Argyle, L.P.; Brown, T.W.; Bumpus, J.P.; Chen, H.; Hunzaker, M.F.; Lee, J.; Mann, M.; Merhout, F.; Volfovsky, A. Exposure to opposing views on social media can increase political polarization. Proc. Natl. Acad. Sci. USA 2018, 115, 9216–9221. [Google Scholar] [CrossRef]
Figure 1. A general illustration of a news or information presentation page in online media. Diverse online media, such as news websites, social media, and forums, can be abstracted into such a page. This page generally includes two parts: the post and related comments. We color-coded features related to controversy detection on this page and showed the tree structure of comments (the orange elements represent the phenomenon of ascending gradient).
Figure 1. A general illustration of a news or information presentation page in online media. Diverse online media, such as news websites, social media, and forums, can be abstracted into such a page. This page generally includes two parts: the post and related comments. We color-coded features related to controversy detection on this page and showed the tree structure of comments (the orange elements represent the phenomenon of ascending gradient).
Mathematics 13 02147 g001
Figure 2. The four key parts of the methodology.
Figure 2. The four key parts of the methodology.
Mathematics 13 02147 g002
Figure 3. The comparison of importance of size ( n ), depth ( d ), breadth ( b ), average degree ( k ), structural virality ( v ), comment emotion ( s c ), post emotion ( s p ), minimum reply time ( t min ), average reply time ( t avg ), comment density ( c ), average ups ( q ), ascending gradients ( p a ), and tier ascending gradients ( p t ) in Reddit dataset.
Figure 3. The comparison of importance of size ( n ), depth ( d ), breadth ( b ), average degree ( k ), structural virality ( v ), comment emotion ( s c ), post emotion ( s p ), minimum reply time ( t min ), average reply time ( t avg ), comment density ( c ), average ups ( q ), ascending gradients ( p a ), and tier ascending gradients ( p t ) in Reddit dataset.
Mathematics 13 02147 g003
Figure 4. Comparing the distribution of ascending gradients ( p a ) and tier ascending gradients ( p t ) of different topics in three media platforms.
Figure 4. Comparing the distribution of ascending gradients ( p a ) and tier ascending gradients ( p t ) of different topics in three media platforms.
Mathematics 13 02147 g004
Figure 5. Analysis of feature importance and interpretation based on hot comments. (a) Three algorithms are used to evaluate the importance of 9 features under hot comments. (b) The trend of the ascending gradient and tier ascending gradient feature values at different scale reviews is assessed using the mean and median, respectively. (c) Evaluate the differences in the distribution of ascending gradient and tier ascending gradient feature values under hot comments (ratio = 0.2) and all comments (ratio = 1.0), respectively.
Figure 5. Analysis of feature importance and interpretation based on hot comments. (a) Three algorithms are used to evaluate the importance of 9 features under hot comments. (b) The trend of the ascending gradient and tier ascending gradient feature values at different scale reviews is assessed using the mean and median, respectively. (c) Evaluate the differences in the distribution of ascending gradient and tier ascending gradient feature values under hot comments (ratio = 0.2) and all comments (ratio = 1.0), respectively.
Mathematics 13 02147 g005
Table 1. All datasets. Three Chinese and English media platforms (Reddit, Toutiao, Sina) are included, and six topics are extracted from each platform. Each topic contains the attributes of controversy, the total number of posts, and comments. Reddit is the main experimental data. Toutiao and Sina are used for experimental testing.
Table 1. All datasets. Three Chinese and English media platforms (Reddit, Toutiao, Sina) are included, and six topics are extracted from each platform. Each topic contains the attributes of controversy, the total number of posts, and comments. Reddit is the main experimental data. Toutiao and Sina are used for experimental testing.
TypeMediaTopicControversyPostComment/Like
Main
Data
Reddit
(English)
Gun7919,371/150,772
War6617,556/139,894
Religion12225,663/57,355
Shopping11126,503/40,831
Scenery717,805/38,476
Music9313,278/83,951
TypeControversyToutiao (Chinese)Sina (Chinese)
TopicPost/Comment/LikeTopicPost/Comment/Like
Test
Data
Huawei111/25,317/63,838Huawei118/23,578/635,701
NBA104/22,695/84,768NBA145/20,608/378,094
Football96/21,747/44,494Football171/23,617/542,666
Life104/13,434/96,734Science84/9,561/206,664
Food115/26,518/21,068Music78/10,943/94,213
Travel117/10,014/33,769Movie132/16,155/197,196
Table 2. Algorithm evaluation results. The detection performance of different features in different classification models is evaluated on the Reddit dataset. The detection performance of different features in the LightGBM model is verified on the Toutiao and Sina datasets (The underlined numbers represent the optimal results with the different datasets).
Table 2. Algorithm evaluation results. The detection performance of different features in different classification models is evaluated on the Reddit dataset. The detection performance of different features in the LightGBM model is verified on the Toutiao and Sina datasets (The underlined numbers represent the optimal results with the different datasets).
PlatformMethodFeature
StructureInteractionTextPsychology
RedditSVM80.9880.3780.9893.25
KNN82.8282.8285.2889.57
LR80.3779.7580.3791.41
DT74.2382.8276.6885.27
GBDT82.8284.6686.5090.80
LightGBM84.0584.0585.2893.25
ToutiaoLightGBM77.8489.1888.6690.21
SinaLightGBM78.2282.1882.6783.66
Table 3. Evaluate the controversy detection performance in different feature modes using comment information over different time periods (The bolded numbers indicate the size of the detection performance improvement, representing the best performance enhancement compared to previous results).
Table 3. Evaluate the controversy detection performance in different feature modes using comment information over different time periods (The bolded numbers indicate the size of the detection performance improvement, representing the best performance enhancement compared to previous results).
Mode
(Feature Num)
Time
3 h6 h9 h24 h
Structure (5)68.7173.6282.2184.05
Interaction (9)73.6283.4484.6684.05
Text (11)73.6287.7384.0585.28
Phychology (13)73.62 + 11.66
(85.28)
87.73 + 5.52
(93.25)
84.66 + 9.21
(93.87)
85.28 + 8.59
(93.87)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, S.; Jin, T.; Zhu, K.; Xuan, Q.; Min, Y. Backfire Effect Reveals Early Controversy in Online Media. Mathematics 2025, 13, 2147. https://doi.org/10.3390/math13132147

AMA Style

Peng S, Jin T, Zhu K, Xuan Q, Min Y. Backfire Effect Reveals Early Controversy in Online Media. Mathematics. 2025; 13(13):2147. https://doi.org/10.3390/math13132147

Chicago/Turabian Style

Peng, Songtao, Tao Jin, Kailun Zhu, Qi Xuan, and Yong Min. 2025. "Backfire Effect Reveals Early Controversy in Online Media" Mathematics 13, no. 13: 2147. https://doi.org/10.3390/math13132147

APA Style

Peng, S., Jin, T., Zhu, K., Xuan, Q., & Min, Y. (2025). Backfire Effect Reveals Early Controversy in Online Media. Mathematics, 13(13), 2147. https://doi.org/10.3390/math13132147

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop