Next Article in Journal
Preparation and Performance of Core–Shell Structured B@NiF2/AP Composite Micro-Units
Previous Article in Journal
Optimal Design of Anti-Collision Spacer Ring for Power Transmission Lines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Irony and Sarcasm Detection in Turkish Texts: A Comparative Study of Transformer-Based Models and Ensemble Learning

by
Murat Eser
and
Metin Bilgin
*
Department of Computer Engineering, Bursa Uludag University, Bursa 16059, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(23), 12498; https://doi.org/10.3390/app152312498
Submission received: 27 October 2025 / Revised: 10 November 2025 / Accepted: 21 November 2025 / Published: 25 November 2025

Abstract

Irony and sarcasm are forms of expression that emphasize the inconsistency between what is said and what is meant. Correctly classifying such expressions is an important text mining problem, especially on user-centered platforms such as social media. Due to the increasing prevalence of implicit expressions, this topic has become a significant area of research in Natural Language Processing (NLP). However, the simultaneous detection of ironic and sarcastic expressions is highly challenging, as both types of implicit sentiments often convey closely related meanings. To address the detection of irony and sarcasm, this study compares the performance of transformer-based models and an ensemble learning method on Turkish texts, using five textual datasets—monogram, bigram, trigram, quadrigram, and omnigram—that share the same textual content but differ in context length. To improve classification performance, an ensemble learning approach based on the Artificial Rabbit Optimization (ARO) algorithm was implemented, combining the outputs of the models to produce final predictions. The experimental results indicate that as the context width of the datasets increases, the models achieve better predictions, leading to improvements across all performance metrics. The ensemble learning method outperformed individual models in all metrics, with performance increasing as the context expanded, achieving the highest success in the omnigram dataset with 76.71% accuracy, 74.64% precision, 73.29% sensitivity, and 73.96% F-Score. This study demonstrates that both model architecture and data structure are decisive factors in text classification performance, showing that community methods can make significant contributions to the effectiveness of deep learning solutions in low-resource languages.

1. Introduction

Humanity lives in an age where data flows every second around the world and only a small fraction of this data can be understood using manual methods [1]. With the advent of the internet, large amounts of textual data have become available [2,3]. While numerical data within this information can be visualized and analyzed using specific statistical tools, the same is not possible when it comes to texts [1].
As technology advances, people have increasingly more opportunities for interaction, ranging from simple text messages and message boards to more engaging and immersive channels such as images and videos [2]. It is becoming increasingly difficult to absorb the mass of information that exists in sources such as websites, news, blogs, books, scientific articles and social media [4]. Due to the inadequacy of human capacity to process large volumes of text data, Natural Language Processing (NLP) is used to read, interpret and understand such data in a much larger volume than human intelligence can perceive at a time in order to make such data more insightful and quickly processable [1]. Therefore, over the past few decades, text summarization has emerged as a popular field of research, aiming to simplify and make the process of obtaining more efficient [4]. NLP is used to overcome many NLP problems, especially sentence classification, sentence-pair classification, asking and answering questions, and named entity-relationship classification [1].
Oral [5,6] or written [7,8] communication is not a simple process. On the contrary, it requires sharing a common experience and history, as well as being able to extract information beyond contextual meaning. Many communicative acts imply information that is not grammatically expressed in order to decode the full meaning: if the listener is unable to extract this information, it means the communicative process is incomplete [6]. These texts contain content with intense figures of speech [9,10,11]. Figurative language includes concepts such as irony, sarcasm, satire, parody, etc. [12]. However, human understanding of figurative language is a difficult matter even using automated text and data mining systems [13] because the stunning feature of sarcastic expressions such as irony and sarcasm is that they make it difficult to bridge the gap between the literal meaning and the intended meaning. For the detection of irony and sarcasm, the fundamental problem in the context of such expressions is to identify the contradiction between what is said and what is meant [14]. As sarcasm is prevalent on social media, online forums and review websites, it is critical that models attempting to analyze online texts are able to detect and understand it. Without this ability, the nature of sarcasm can interfere with the effectiveness of certain types of tasks [15].
Sarcastic expressions are defined as the use of words that carry a meaning different than the speaker’s true intention, expressing the explicit opposite of what they say in order to hurt a person’s feelings or criticize them humorously. The main purpose of sarcasm is to implicitly say derogatory and upsetting words through the use of comments that are made to hurt the feelings of others or to attack something in a humorous way and mean the opposite of what is said; with this definition, sarcasm is very important in our daily lives [16]. Sarcasm is an offensive type of irony used with a cutting tone [17]. Across the world, there is a noticeable increasing trend in the incorporation of sarcasm into everyday life. This trend can be attributed to the frequent use of sarcasm in everyday life, but more specifically to social media and the internet [18].
Ironic expressions are the different expression of a situation than the actual feeling. It can also be expressed as the opposite situation when there is an expectation. In conceptual terms, irony is a sarcastic and indirect expression technique in which sarcasm is evident and contains different meanings. The most common use of irony is to express something that is not serious as if it were serious [19]. Irony, in a simple definition, is the creative use of language [20] and is a form of communication often used to express contempt or ridicule, where the speaker conveys a message that is the opposite of its true meaning, usually with the intention of ridiculing or belittling a specific target [21].
Irony and sarcasm are prominent features of human language, deeply rooted in nuanced expression and social interaction. Both phenomena involve saying something that contradicts the intended meaning, but they are used differently and carry different implications. Context plays an indispensable role in the perception and understanding of both irony and sarcasm. Rohanian et al. point to the complex nature of situational irony, which involves not conforming to expectations and then eliciting emotional responses [22]. Sarcasm and irony are sometimes thought to be interchangeable. However, Sulis et al. distinguish between them, stating that sarcasm is usually more overtly aggressive and sharp in tone, while irony can have a more subtle, more sarcastic quality [10]. Sarcasm is a form of irony that occurs when there is a discrepancy between the actual meaning of a text or expression and its intended meaning. This inconsistency typically manifests as dislike, disdain, or contempt [15].
The use of sarcasm and sarcastic expression, such as irony and sarcasm, is present everywhere, in facial expressions, gestures and even in texts. It is subjective, based on a situation or circumstances, people, language and even one’s culture; it is often a combination of positive and negative expressions [23]. For this, the need to understand the context and to identify emotions in this context has emerged [14]. In this direction, such analyses have recently attracted the attention of computer scientists [20].
The concepts of irony and sarcasm are used as an umbrella term that includes sarcasm and sarcastic communication. In the literature, since there is scientific evidence that sarcasm tends to be more aggressive than irony [24,25,26], expressions that tend to be harsher, hurtful, aggressive are considered as sarcasm, while expressions in which what is said and what is intended to be said are opposite to each other [27] are considered as irony.
As machine learning models become increasingly popular for natural language processing applications, there is a growing need for these models to be able to recognize sarcasm in online texts [15]. At this point, although it has limitations in its current form [13], NLP emerges as a field with serious solutions to this problem [1]. Most of these applications involve the use of natural language and research to date has shown successful results in tackling such tasks [28]. However, there has not been exactly the same success for the figurative arts because understanding figurative language is something in which even human intelligence has difficulty [29,30,31,32,33]. One of the communicative phenomena that best represents this problem is the identification of irony [21,34,35,36,37,38,39,40,41,42,43] and sarcasm [18,21,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50].
A significant part of the data of the texts containing such mocking, satirical, sarcastic expressions that people produce or face are encountered in social media environments [51,52,53]. Various studies have used Twitter, Reddit and Facebook for sarcasm detection [27,54,55,56]. Identifying mocking behaviors such as irony and sarcasm in the field of online social networks such as Facebook, Twitter, Instagram, surveys, etc. has become an important responsibility as they affect social and personal relationships [14]. To this end, significant efforts have been made to develop effective computational methods that can be used to detect expressions such as irony and sarcasm in various contexts, including social media [28,57,58], news [59,60,61] and public speeches [62]. However, it has been observed that there is a significant lack of research focusing on the analysis of figurative expressions such as irony and sarcasm in sources that contain content on specific topics, such as column texts. In this regard, there is a noticeable gap in new approaches and research studies aimed at analyzing satirical content such as irony and sarcasm in column texts.
Sentiment analysis is of vital importance in areas such as decision-making, data mining, opinion mining, and information management, and it is widely used across various domains—including telecommunications, medical, environmental, and political fields—as well as in multiple languages [63,64,65,66]. Since social media allows individuals to share their personal opinions on various topics in real time to express themselves to others, there exists a vast amount of raw data that needs to be analyzed. Consequently, manually classifying this data becomes infeasible [20].
Due to the prevalence of sarcasm on social media and the difficulty machines face in detecting it, the topic has attracted significant attention in the field of NLP [21]. Irony and sarcasm detection is a significant computational challenge in NLP and is essential for better understanding, as it serves as an interface for effective communication between machines and humans [14]. Sarcasm detection is an interesting yet challenging task when it comes to text, as the absence of tone, expression, or gestures can result in a discrepancy between what is said and what is actually meant. A sarcastic text may appear humorous or harsh, critical or praising, but all such instances potentially convey sarcasm in some form [67,68]. Sarcasm is an implicit emotional state and requires additional information such as context and multimodality for more accurate detection at the text level [16]. Furthermore, studies in the literature are generally conducted in English, meaning that the results obtained fall into a language classification that has a different linguistic structure to Turkish [20,69], which, compared to other languages on the internet, has limited resources [4].
Studies in the literature have focused on concepts that can be categorized as (i) sarcasm, (ii) irony, (iii) simile, (iv) metaphor, (v) satire, (vi) exaggeration and (vii) humour, depending on the datasets and analysis methods of the studies [28]. Due to the ambiguity of ironic and sarcastic expressions in textual data, they are difficult to detect. Research into the detection of sarcasm and irony has spanned more than a decade, but recent progress has been made. The existing studies have typically focused on languages such as English [40,70,71,72,73,74,75,76], Arabic [76,77], Chinese [78,79] and other languages [71,80,81,82,83,84,85,86,87]. However, research focusing simultaneously on irony and sarcasm expressions in Turkish column texts offering political, economic, cultural, and social analyses has not been encountered. In studies on irony and sarcasm in the Turkish language [88,89,90,91] a binary classification process has been carried out, as in other languages. The similarities between ironic and sarcastic expressions make it difficult to identify them simultaneously. This study aims to identify ironic and sarcastic expressions simultaneously.
When the studies are analyzed in terms of dataset, it is seen that datasets containing short sentences obtained from web platforms such as Twitter, Reddit and Amazon are used [2,9,92,93,94,95,96,97,98]. There are also datasets consisting of a few sentences obtained through web platforms [11,32,99,100,101]. Datasets obtained through web platforms are usually labelled with hashtag field control. There are also manual labelling methods. The existing studies are presented in Table 1 in terms of methodology as rule-based approaches, machine learning approaches, deep learning-based approaches, ensemble learning approaches, and transformer-based approaches. A review of the existing studies reveals that the data is predominantly collected from social media platforms, typically consisting of short sentences and retrieved using hashtag-based search methods. It is considered that the nature of these datasets may contribute to the seemingly high performance of the proposed methodologies. Upon examining the employed approaches, transformer-based methods have been observed to yield more successful outcomes. The literature largely focuses on single-label classification tasks, particularly targeting phenomena such as irony or sarcasm.

2. Materials and Methods

In this section, information about the methods used in this study and the dataset used will be presented. In this study, the classification processes were performed with a computing cluster with Nvidia V100 GPU and Xeon 6148 2.40 GHz processor in TÜBİTAK ULAKBİM, High Performance and Grid Computing Centre, Ankara, Turkey. The performance of the models was measured by 10-fold cross-validation and train-split strategy of the dataset used in this study. After the model results were obtained using both 10-fold cross-validation and the train-test split technique, comparisons were conducted based on standard machine learning evaluation metrics. Figure 1 presents a step-by-step representation of the proposed approach. In this study, a high-complexity three-class dataset was utilized, and transformer-based Turkish fine-tuned models were employed in conjunction with an optimization algorithm designed to determine the optimal weighting of these models. The resulting outcomes were analyzed according to standard machine learning evaluation metrics.

2.1. Dataset

The dataset used in this study was derived from column texts, manually annotated, and categorized into three classes: irony, sarcasm, and normal. The dataset consists of paragraphs containing multiple sentences, with 429, 354, and 526 instances belonging to irony, sarcasm, and normal classes, respectively. The column texts were obtained from publicly accessible online newspaper websites such as www.milliyet.com.tr (access date: 7 November 2025) [119] and www.cumhuriyet.com.tr (access date: 7 November 2025) [120]. During this study, a total of 3000 column texts written by different authors were examined.
In this study, to investigate the influence of contextual information on the target class, several datasets were constructed based on the number of sentences included. The labelling of the dataset occurs in four stages: the artificial intelligence assistant performs pre-labelling, the human user reviews the labels and filters them, the linguist processes the labelled examples, and any incorrect labels are corrected. In the first stage, due to the time and resource costs of traditional data labelling, AI-supported labelling comes to the fore. Examples of labelling types include zero-shot labelling, few-shot labelling, and instruction-based labelling. Instruction-based labelling is also referred to in the literature as instruction-based fine-tuning or supervised fine-tuning. This fine-tuning is used in example labelling by large language models trained to predict the next word or generate information, following specific instructions set by users. The method consists of instruction and output sections and relies on supervised fine-tuning over specific examples. This ensures that the fine-tuned language model produces the desired output within certain constraints, in line with the instructions [121] Due to the size of the texts to be examined and analysed, the instruction-based labelling method was preferred. An assistant named SarcaTurkGPT, created using a large language model (OpenAI GPT-4o) by a linguist, was used to detect expressions containing irony and sarcasm. During the development of the assistant, the definitions of irony and sarcasm were provided, their differences were highlighted, and the assistant was designed to analyze user-submitted texts and return sentences that potentially contain irony or sarcasm. The temperature parameter was set to 0.5 to ensure more balanced and consistent responses, while the other parameters were kept at their default values (top_p: 1, frequency_penalty: 0, and presence_penalty: 0). The assistant analyses the column texts provided to it and ensures pre-labelling by flagging texts that may contain irony and sarcasm. In the second stage, the examples pre-labelled by SarcaTurkGPT are checked by a human user and marked as belonging to the appropriate class. In the third stage, after the labelling process by SarcaTurkGPT and human users, the examples are checked by the linguist. In the fourth stage, if there are errors in the labelling of the examples checked by the linguist, they are re-examined and labelled by the human user according to the linguist’s recommendations.
The monogram dataset contains only the ironic, sarcastic, or neutral sentence itself. The bigram dataset includes the target sentence along with the immediately preceding sentence. The trigram, quadrigram, and omnigram datasets incorporate two, three, and all preceding sentences, respectively, prior to the target sentence. This configuration enables the examination of how the sentences preceding the ironic or sarcastic sentence affect classification performance. The dataset architecture is presented in Figure 2.
Examples of the bigram and trigram datasets prepared for use in this study are presented in Table 2.
Table 2. Example of the bigram and trigram datasets used in this study.
Table 2. Example of the bigram and trigram datasets used in this study.
ClassMonogram SampleBigram SampleTrigram SampleParagraph
IronyBöylece siz de bir Danimarkalı kadar mutlu olabilirsiniz. (So you too can be as happy as a Dane.)Yumuşak bir aydınlatma, hoş kokulu mumlar, lezzetli yemekler, samimi sohbetler, sıcak tutan yumuşak giysiler… Böylece siz de bir Danimarkalı kadar mutlu olabilirsiniz. (Soft lighting, fragrant candles, delicious food, heartfelt conversations, warm and cozy clothes… So you too can be as happy as a Dane.)Üstelik yapılması gerekenler hiç de zor sayılmazdı. Yumuşak bir aydınlatma, hoş kokulu mumlar, lezzetli yemekler, samimi sohbetler, sıcak tutan yumuşak giysiler… Böylece siz de bir Danimarkalı kadar mutlu olabilirsiniz. (Moreover, what needed to be done was not difficult at all. Soft lighting, fragrant candles, delicious food, heartfelt conversations, warm and cozy clothes… So you too can be as happy as a Dane.)Kitabı çok satanlar listesine taşıyan ve Türkçe dahil pek çok dilde yayınlanıp popülerleşmesini sağlayan da kuşkusuz insanların bu mutluluk formüllerini öğrenme ihtiyacıydı. Üstelik yapılması gerekenler hiç de zor sayılmazdı. Yumuşak bir aydınlatma, hoş kokulu mumlar, lezzetli yemekler, samimi sohbetler, sıcak tutan yumuşak giysiler… Böylece siz de bir Danimarkalı kadar mutlu olabilirsiniz. (What propelled the book onto bestseller lists and led to its publication and popularity in many languages, including Turkish, was undoubtedly people’s need to learn these formulas for happiness. Moreover, what needed to be done was not difficult at all. Soft lighting, fragrant candles, delicious food, heartfelt conversations, warm and cozy clothes… So you too can be as happy as a Dane.)
SarcasmNe de olsa “Alamancı” onun bakışı bir başka… (After all, as an “Alamancı”, her perspective is different…)Bir de dizileri uzun tutuyorlar, kısa yapsınlar! diyordu, o sevimli Türkçesiyle. Ne de olsa “Alamancı” onun bakışı bir başka… (Also, they make the TV series too long—they should make them shorter!” After all, as an “Alamancı”, her perspective is different…)Üç yıldır Türkiye’deyim, ilk defa bir ödül alıyorum. Bir de dizileri uzun tutuyorlar, kısa yapsınlar! diyordu, o sevimli Türkçesiyle. Ne de olsa “Alamancı” onun bakışı bir başka… (“I’ve been in Turkey for three years, and this is the first time I’m receiving an award. Also, they make the TV series too long—they should make them shorter!”
After all, as an “Alamancı”, her perspective is different…)
Oysa “Hürrem Sultan” Meryem Uzerli’nin dediği başka. Üç yıldır Türkiye’deyim, ilk defa bir ödül alıyorum. Bir de dizileri uzun tutuyorlar, kısa yapsınlar! diyordu, o sevimli Türkçesiyle. Ne de olsa “Alamancı” onun bakışı bir başka… (Yet what “Hürrem Sultan” Meryem Uzerli says is different.
She was saying, in her cute Turkish: “I’ve been in Turkey for three years, and this is the first time I’m receiving an award. Also, they make the TV series too long—they should make them shorter!”
After all, as an “Alamancı”, her perspective is different…)
NormalKanalların ve yönetimlerinin geleceğini belirliyor, yapım şirketlerinin milyonlarca dolar kâr veya zarar etmesine neden oluyor. (It determines the future of TV channels and their management, and causes production companies to make millions of dollars in profit or loss.)Her bir rating altın değil pırlanta değerinde. Kanalların ve yönetimlerinin geleceğini belirliyor, yapım şirketlerinin milyonlarca dolar kâr veya zarar etmesine neden oluyor. (Every single rating point is not just worth gold, but diamonds. It determines the future of TV channels and their management, and causes production companies to make millions of dollars in profit or loss.)Her gün rating karnesi alan onlarca yapım şirketi kıyasıya mücadele ediyorlar. Her bir rating altın değil pırlanta değerinde. Kanalların ve yönetimlerinin geleceğini belirliyor, yapım şirketlerinin milyonlarca dolar kâr veya zarar etmesine neden oluyor. (Dozens of production companies, each receiving daily ratings reports, compete fiercely. Every single rating point is not just worth gold, but diamonds. It determines the future of TV channels and their management, and causes production companies to make millions of dollars in profit or loss.)Şaşkınlık verici çünkü Türkiye’de rekabetin en sert ve en görünür olduğu yerlerden biri dizi sektörü. Her gün rating karnesi alan onlarca yapım şirketi kıyasıya mücadele ediyorlar. Her bir rating altın değil pırlanta değerinde. Kanalların ve yönetimlerinin geleceğini belirliyor, yapım şirketlerinin milyonlarca dolar kâr veya zarar etmesine neden oluyor. (It is astonishing, because one of the places where competition is the toughest and most visible in Turkey is the TV series sector. Dozens of production companies, each receiving daily ratings reports, compete fiercely. Every single rating point is not just worth gold, but diamonds. It determines the future of TV channels and their management, and causes production companies to make millions of dollars in profit or loss.)
The basic statistics of the dataset used this study according to classes are presented in Figure 3.
Figure 3. Average and Median Distributions of Words by Classes.
Figure 3. Average and Median Distributions of Words by Classes.
Applsci 15 12498 g003
The thematic distribution statistics of the dataset used in this study are presented in Figure 4. The class distributions are provided under the headings of politics, economy, health, cultural, and others, with the examples predominantly derived from politically themed column texts.

2.2. Modeling

This section will provide information regarding the methods used in this study.

2.2.1. Model Selection

BERTurk, ConvBERTurk, DistilBERTurk, RoBERTaTurk and ELECTRATurk, which are transformer-based models fine-tuned with Turkish language texts, were used in this study. In addition, the Artificial Rabbit Optimization Algorithm (ARO) was used to optimise the weights of these models for ensemble prediction. During the training of the transformer-based models, the learning rate was set to 2 × 10−5, the batch size to 8, and the maximum sequence length to 512. The number of training epochs was dynamically adjusted according to the learning process, typically ranging between 15 and 20.
BERTurk
Bidirectional Encoder Representations from Transformers (BERT) is a model introduced by Google in 2018 that revolutionized the field of NLP [122] One of its main features is its two-way understanding of words. In this way, it produces effective results in meaning transfer by better understanding the context. The model shows high success rates in various tasks such as text classification, sentiment analysis, and question-answer systems. The pre-training and fine-tuning processes of the BERT model are presented in Figure 5. BERTurk is an NLP model with Turkish language understanding and processing capabilities [123]. This model, based on the BERT architecture, has been specially developed to effectively analyze the language structures and features of Turkish. The agglutinative structure of Turkish requires a special approach due to the formation of words with roots and affixes. In this context, BERTurk has been trained in accordance with the morphological and syntactic features of the language. Trained with a large corpus of 35 GB of Turkish text containing about 4 billion tokens, the model successfully understands different language usages and contexts thanks to the diversity of the datasets.
ConvBERTurk
ConvBERT is a language model developed in the field of NLP by combining the BERT architecture with conventional layers [124]. This model aims to provide more meaningful and effective language processing capabilities by combining contextual and local features. The structure of the ConvBERT model is presented in Figure 6. ConvBERTurk is a model developed for Turkish NLP with Turkish comprehension and processing capabilities [123].
DistilBERTurk
DistilBERT is a lightened version of the BERT model [125]. Trained with ‘knowledge distillation’ techniques, this model runs faster and consumes fewer resources, providing efficiency in NLP applications. DistilBERT offers significant advantages in terms of performance and speed by reducing the number of parameters while maintaining the basic building blocks of BERT. In the training process, it benefits from BERT’s large datasets and develops a lighter structure without losing its basic capabilities by learning from the BERT model in the ‘teacher’ position [126]. The architectural structure of the model is presented in Figure 7. In this study, the DistilBERTurk model trained with the Turkish dataset was used.
ELECTRATurk
Efficiently Learning an Encoder that Classifies Token Replacements Accurately (ELECTRA) is a pre-trained NLP model developed in 2020 [128]. ELECTRA aims to combine the speed and efficiency of models such as BERT with high performance and has excelled in numerous NLP tasks [129]. The model is based on two main components: Generator and Discriminator. The Generator is used to replace certain words, while the Discriminator analyses these replaced words and tries to make correct predictions. Thanks to this mechanism, ELECTRA works on a task called “Replaced Token Detection” (RTD) instead of “Masked Language Modelling” (MLM). This process improves the training efficiency of the language model and is particularly advantageous for learning syntactic contexts [130]. In the training process of ELECTRA, the model not only understands the data, but also the context of the data and makes accurate word predictions. Research shows that ELECTRA has a distinct advantage over traditional BERT training methods and learns more information with less computational power [131]. The structure used by the model is presented in Figure 8. The ELECTRATurk model, which was trained with the Turkish dataset and prepared for Turkish NLP, was used in this study.
RoBERTaTurk
RoBERTa (A Robustly Optimized BERT Pretraining Approach) is an advanced transformer-based bidirectional training mechanism that has shown significant success in the field of NLP by improving the pretraining procedure of BERT [132]. This model offers improvements such as training with longer and larger datasets, removing the next sentence prediction (NSP) target, working with longer sequences and using dynamic masking. RoBERTa aims to achieve higher performance on various NLP tasks by making systematic changes to the BERT architecture and training method. It has also achieved the best results on standard NLP benchmarks such as GLUE, SQuAD and RACE. In this study, the RoBERTaTurk model trained with Turkish dataset and adapted for Turkish NLP is used.
Artificial Rabbits Optimization Algorithm
In the work presented by [133] a new bio-inspired meta-heuristic algorithm called ‘Artificial Rabbits Optimization’ (ARO) is proposed to solve engineering optimization problems. The algorithm is inspired by the survival strategies of rabbits in nature: foraging away from their burrows to avoid predators (‘detour foraging’) and choosing a random burrow to hide from enemies (‘random hiding’). By modelling these mechanisms, the algorithm significantly improves the global and local search capabilities and performs well in solving engineering problems. The flowchart of the algorithm adapted for ensemble learning is presented in Figure 9.

2.2.2. Ensemble Learning Model

BERTurk, ConvBERTurk, DistilBERTurk, RoBERTaTurk and ELECTRATurk, which are transformer-based models fine-tuned with Turkish language texts, were used in this study. In addition, the Artificial Rabbit Optimization Algorithm (ARO) was used to optimize the weights of these models for ensemble prediction.

2.3. Evaluation Metrics

An example with a positive label in the dataset is predicted as positive, which is referred to as a True Positive (TP). When considering the irony class, it means that an expression that is ironic is also predicted as ironic by the model. An example that is labelled as negative but is predicted as positive is referred to as a False Positive (FP). This is the situation where an expression that is not actually ironic is incorrectly labelled as ironic by the model. When an example that is positive is predicted to be negative, this is referred to as a False Negative (FN). This is when an expression that is ironic is evaluated by the model as not being ironic. When an example that is negative is predicted to be negative, this is referred to as a True Negative (TN). This is when an expression that is not actually ironic is evaluated by the model as not being ironic. The same evaluation method is applied to the sarcasm and normal classes as well; for each class, TP, FP, FN, and TN are defined similarly based on whether the model correctly or incorrectly predicts the corresponding label. To evaluate the results of this study, accuracy, precision, recall and F-score metrics used in similar tasks in the literature were used. A three-class classification was performed, and all evaluation metrics were calculated over these three classes. Related metrics can be seen in Equations (1)–(4). The Cross-Entropy Loss function, as presented in Equation (5), was utilized during model training. In this formulation, C represents the total number of classes, yi denotes the ground-truth label for the i-th class, and pi corresponds to the predicted probability of the i-th class. This loss function guides the model toward correct classification by minimizing the divergence between the predicted probability distribution and the actual class labels.
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
L = i = 1 C y i log ( p i )

3. Results and Discussion

In this study, five different pre-trained models—BERTurk, ConvBERTurk, DistilBERTurk, ELECTRATurk, and RoBERTaTurk—were fine-tuned using a custom-designed dataset and evaluated for a text classification task. Additionally, an Artificial Rabbit Optimization (ARO) method was employed to combine the outputs of these models, resulting in a final prediction that leverages the strengths of the individual models. To achieve the optimal performance of the ensemble learning approach, the number of rabbits and iterations in the ARO method were tuned using the grid search optimization technique.
This study conducted various analyses on monogram, bigram, trigram, quadrigram, and omnigram data set architectures. Accordingly, experimental studies were carried out exclusively using these two pieces of architecture. The classification performance of the fine-tuned models was comprehensively evaluated using Accuracy, Precision, Recall, and F-Score metrics, providing a detailed assessment of their effectiveness in the given task. The fine-tuned models were first evaluated using the 10-fold cross-validation method, and the corresponding metric values are presented in Table 3.
The results in Table 3 show the comparative classification performances of Transformer-based models in the Turkish language when used with context width dataset architectures. When the results of the dataset were analyzed, BERTurk was the most successful model according to performance metrics, achieving an accuracy value of 65.63% on the monogram dataset, 66.72% on the bigram dataset, and 67.42% on the trigram dataset. ConvBERTurk was the most successful model, achieving an accuracy value of 68.76% on the quadrigram dataset and an accuracy value of 70.74% on the omnigram dataset. The accuracy, precision, sensitivity, and F-Score metric values achieved by the BERTurk model on the monogram dataset and the metric values achieved by the ConvBERTurk model on the omnigram dataset increased by 7.79% for the accuracy metric, 7.60% for the precision metric, 9.45% for the sensitivity metric, and 8.41% for the F-Score metric, respectively. and 8.41% for the F-score metric. This performance increase demonstrates the positive effect of contextual history on classification in detecting irony and sarcasm. A consistent performance increase is observed across the most successful models, from the most basic monogram dataset architecture to the omnigram dataset.
Based on the results obtained from the monogram dataset, the most successful model was BERTurk, achieving 65.63% accuracy, 63.19% precision, 61.89% recall, and an F-score of 62.58%. In contrast, RoBERTaTurk ranked lowest in terms of performance, with 54.32% accuracy, 50.93% precision, 50.72% recall, and an F-score of 50.80%. Based on the results obtained from the bigram dataset, the most successful model was BERTurk, achieving 66.72% accuracy, 64.72% precision, 63.50% recall, and an F-score of 64.10%. In contrast, RoBERTaTurk ranked lowest in terms of performance, with 61.24% accuracy, 60.99% precision, 56,83% recall, and an F-score of 58.84%. Based on the results obtained from the trigram dataset, the most successful model was BERTurk, achieving 67.42% accuracy, 64.73% precision, 64.32% recall, and an F-score of 64.52%. In contrast, RoBERTaTurk ranked lowest in terms of performance, with 61.16% accuracy, 56.84% precision, 56.61% recall, and an F-score of 56.72%. Based on the results obtained from the quadrigram dataset, the most successful model was ConvBERTurk, achieving 68.76% accuracy, 66.29% precision, 65.19% recall, and an F-score of 65.75%. In contrast, RoBERTaTurk ranked lowest in terms of performance, with 61.11% accuracy, 57.62% precision, 56.51% recall, and an F-score of 57.05%. Based on the results obtained from the omnigram dataset, the most successful model was ConvBERTurk, achieving 70.74% accuracy, 67.99% precision, 67.74% recall, and an F-score of 67.84%. In contrast, RoBERTaTurk ranked lowest in terms of performance, with 61.88% accuracy, 52.00% precision, 57.03% recall, and an F-score of 54.39%.
The metric values achieved by the two most successful models, BERTurk and ConvBERTurk, on the datasets are presented in Table 4. Although the BERTurk model did not achieve a performance increase when transitioning from the trigram dataset to the quadrigram dataset, an increase in metrics was observed with the expansion of context. The metric values obtained in the omnigram dataset show an increase of 7.09% for the accuracy metric, 8.16% for the precision metric, 8.90% for the sensitivity metric, and 8.50% for the F-Score metric compared to the monogram dataset. The ConvBERTurk model achieved performance improvements at every step with the expansion of context between the monogram and omnigram datasets. It was observed that the metrics achieved in the omnigram dataset increased by 16.19% for the accuracy metric, 14.16% for the precision metric, 16.41% for the sensitivity metric, and 15.22% for the F-Score metric compared to the monogram dataset.
When examining the impact of the dataset architecture, it is observed that the BERTurk and ConvBERTurk models, when evaluated on the monogram and omnigram datasets, yield improvements in performance metrics. Although the RoBERTaTurk model showed the lowest performance across all datasets, its performance increased as the context of the dataset increased. The findings indicate that both the model architecture and the preferred dataset type play a decisive role in classification performance. Performance comparisons of monogram and omnigram dataset architectures are presented in Figure 10 and Figure 11.
Secondly, after splitting the dataset into training and testing sets, model-specific weights were determined via ARO based on their predictions, and a final prediction was generated by capitalizing on the strengths of the individual models. The metric values computed for the predictions made through ensemble learning are presented in Table 5.
When Table 5 is examined, it is seen that the Ensemble approach reached the highest values in all metrics in all datasets. In the monogram dataset, the results obtained from the models were presented. According to these results, the ensemble learning model outperformed the others, achieving 65.25% accuracy, 63.75% precision, 60.55% recall, and an F-Score of 62.12%. Compared to the second-best performing model, DistilBERTurk demonstrated improvements of 6.19% in accuracy, 18.97% in precision, 16.87% in recall, and 18.38% in F-Score, respectively. When examining the average performance metrics of the transformer-based models, the ensemble learning approach yielded superior results, with increases of 11.86% in accuracy, 20.87% in precision, 12.27% in recall, and 17.54% in F-Score compared to the average values. In the ensemble learning method, the ARO algorithm was employed, where the number of rabbits was searched within the range of 50 to 1000, and the model was trained using 50 rabbits and 50 iterations.
In the bigram dataset, the results obtained from the models were presented. According to these results, the ensemble learning model outperformed the others, achieving 69.84% accuracy, 66.42% precision, 65.18% recall, and an F-Score of 65.80%. Compared to the second-best performing model, ELECTRATurk, it demonstrated improvements of 7.64% in accuracy, 12.25% in precision, 9.18% in recall, and 10.72% in F-Score, respectively. When examining the average performance metrics of the transformer-based models, the ensemble learning approach yielded superior results, with increases of 15.55% in accuracy, 16.03% in precision, 15.38% in recall, and 15.74% in F-Score compared to the average values. In the ensemble learning method, the ARO algorithm was employed using 130 rabbits and 50 iterations.
In the trigram dataset, the results obtained from the models were presented. According to these results, the ensemble learning model outperformed the others, achieving 71.37% accuracy, 69.78% precision, 67.70% recall, and an F-Score of 68.73%. Compared to the second-best performing model, ConvBERTurk demonstrated improvements of 9.34% in accuracy, 13.45% in precision, 10.77% in recall, and 12.10% in F-Score, respectively. When examining the average performance metrics of the transformer-based models, the ensemble learning approach yielded superior results, with increases of 12.25% in accuracy, 12.84% in precision, 11.79% in recall, and 12.32% in F-Score compared to the average values. In the ensemble learning method, the ARO algorithm was employed using 190 rabbits and 50 iterations.
In the quadigram dataset, the results obtained from the models were presented. According to these results, the ensemble learning model outperformed the others, achieving 72.13% accuracy, 70.78% precision, 67.81% recall, and an F-Score of 69.27%. Compared to the second-best performing model, ELECTRATurk, it demonstrated improvements of 8.00% in accuracy, 18.41% in precision, 12.45% in recall, and 15.39% in F-Score, respectively. When examining the average performance metrics of the transformer-based models, the ensemble learning approach yielded superior results, with increases of 11.31% in accuracy, 17.15% in precision, 12.49% in recall, and 14.78% in F-Score compared to the average values. In the ensemble learning method, the ARO algorithm was employed using 610 rabbits and 50 iterations.
In the omnigram dataset, the results obtained from the models were presented. According to these results, the ensemble learning model achieved the best performance, with 76.71% accuracy, 74.64% precision, 73.29% recall, and an F-Score of 73.96%. Compared to the second-best performing model, ConvBERTurk, it showed improvements of 6.90% in accuracy, 9.25% in precision, 7.29% in recall, and 8.25% in F-Score, respectively. Based on the average performance metric values of the transformer-based models, the ensemble learning approach produced superior results, outperforming the averages by 13.93% in accuracy, 18.14% in precision, 16.96% in recall, and 17.56% in F-Score. In the ensemble learning method, the ARO algorithm was employed using 390 rabbits and 50 iterations.
As shown in Figure 11, the ensemble learning method achieved an accuracy of 66.41% on the monogram dataset, and it is observed that as the contextual width increased in the subsequent datasets, the obtained accuracy performance also improved. The increasing trend in performance reached its highest level in the omnigram dataset, which represents the dataset with the widest contextual range, achieving 76.71% accuracy. Similarly, when the results of the monogram and omnigram datasets are compared, a noticeable improvement in performance is observed across the other evaluation metrics as well.

4. Conclusions

In this study, we evaluate the performance of five different pre-trained Transformer-based models—BERTurk, ConvBERTurk, DistilBERTurk, ELECTRATurk, and RoBERTaTurk—on a Turkish text classification task and test these models on specially prepared bigram and trigram datasets with fine-tuning. In the experimental process, an ARO-based ensemble learning approach that combines the outputs of the models to improve classification performance was also applied.
When the results obtained by the models on the datasets using the cross-validation technique are examined, it is observed that as the contextual width increased—from the monogram dataset with the narrowest context to the omnigram dataset with the widest context—the accuracy, precision, recall, and F-score metrics respectively improved from 65.63%, 63.19%, 61.89%, and 62.58% to 70.74%, 67.99%, 67.74%, and 67.84%. Among the models, BERTurk achieved the best performance on the monogram, bigram, and trigram datasets, while ConvBERTurk performed best on the quadrigram and omnigram datasets. Across all datasets, the RoBERTaTurk model ranked last in terms of performance. The BERTurk model achieved 65.63% accuracy on the monogram dataset and 70.28% on the omnigram dataset. Similarly, in terms of accuracy, ConvBERTurk improved from 60.89% to 70.74%, DistilBERTurk from 57.45% to 66.47%, ELECTRATurk from 59.59% to 67.76%, and RoBERTaTurk from 54.32% to 61.66%, all demonstrating an upward trend in performance. In general, as the context expanded from bigram to trigram and quadrigram datasets, the models exhibited either a limited performance improvement or results similar to those of the previous dataset. However, in the omnigram dataset, which possesses the widest contextual range, all models showed a distinct increase in performance.
When the results obtained by the transformer-based models on the datasets were examined according to the 80:20 training–test split, it was observed that as the contextual width of the datasets increased, the performance metrics of the individual models also improved. This performance improvement of the individual models was reflected in the ensemble learning method, which combined their predictions to produce the final outcomes. As the performance of the individual models increased, the ensemble learning method achieved 65.25% accuracy on the monogram dataset with the narrowest context, and 76.71% accuracy on the omnigram dataset with the widest context. The obtained results indicate that the ensemble learning approach successfully and effectively integrated the predictions of individual models, achieving higher overall performance across all datasets compared to any single model.
The ensemble learning method, which enhances the performance of individual models, may pose challenges for real-time applications due to its requirements for extensive computational resources, processing time, and hardware capacity, as it involves the training and performance evaluation of all models.
Overall, this study demonstrates that both the type of model employed, and the architecture of the dataset play a decisive role in text classification performance, and that when these two factors are supported by ensemble methods such as ARO, performance can be significantly enhanced. The findings indicate that, particularly in low-resource languages such as Turkish, the joint consideration of model selection, data structure, and ensemble strategies is essential when developing deep learning-based solutions.

Author Contributions

Conceptualization, M.E. and M.B.; methodology M.E. and M.B.; software, M.E. and M.B.; formal analysis, M.E. and M.B.; investigation, M.E. and M.B.; resources, M.E. and M.B.; writing—original draft preparation, M.E. and M.B.; writing—review and editing, M.E. and M.B.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Bursa Uludağ University Science and Technology Centre (BAP) under Grant FGA-2025-2505.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Junaid, T.; Sumathi, D.; Sasikumar, A.; Suthir, S.; Manikandan, J.; Khilar, R.; Kuppusamy, P.; Raju, M.J. A comparative analysis of transformer based models for figurative language classification. Comput. Electr. Eng. 2022, 101, 108051. [Google Scholar] [CrossRef]
  2. Bharti, S.K.; Vachha, B.; Pradhan, R.K.; Babu, K.S.; Jena, S.K. Sarcastic sentiment detection in tweets streamed in real time: A big data approach. Digit. Commun. Netw. 2016, 2, 108–121. [Google Scholar] [CrossRef]
  3. Reyes, A.; Rosso, P.; Buscaldi, D. From humor recognition to irony detection: The figurative language of social media. Data Knowl. Eng. 2012, 74, 1–12. [Google Scholar] [CrossRef]
  4. Baykara, B.; Güngör, T. Turkish abstractive text summarization using pretrained sequence-to-sequence models. Nat. Lang. Eng. 2023, 29, 1275–1304. [Google Scholar] [CrossRef]
  5. Kumar, H.M.K.; Harish, B.S. Sarcasm classification: A novel approach by using Content Based Feature Selection Method. in Procedia Comput. Sci. 2018, 143, 378–386. [Google Scholar] [CrossRef]
  6. Reyes, A.; Rosso, P. Making objective decisions from subjective data: Detecting irony in customer reviews. Decis. Support Syst. 2012, 53, 754–760. [Google Scholar] [CrossRef]
  7. Fersini, E.; Messina, E.; Pozzi, F.A. Sentiment analysis: Bayesian Ensemble Learning. Decis. Support. Syst. 2014, 68, 26–38. [Google Scholar] [CrossRef]
  8. Wang, G.; Sun, J.; Ma, J.; Xu, K.; Gu, J. Sentiment classification: The contribution of ensemble learning. Decis. Support Syst. 2014, 57, 77–93. [Google Scholar] [CrossRef]
  9. Poria, S.; Cambria, E.; Hazarika, D.; Vij, P. A deeper look into sarcastic tweets using deep convolutional neural networks. In COLING 2016—26th International Conference on Computational Linguistics, Proceedings of COLING 2016: Technical Papers, Osaka, Japan, 11–16 December 2016; The COLING 2016 Organizing Committee: Osaka, Japan, 2016. [Google Scholar]
  10. Sulis, E.; Farías, D.I.H.; Rosso, P.; Patti, V.; Ruffo, G. Figurative messages and affect in Twitter: Differences between #irony, #sarcasm and #not. Knowl. Based Syst. 2016, 108, 132–143. [Google Scholar] [CrossRef]
  11. Filatova, E. Irony and sarcasm: Corpus generation and analysis using crowdsourcing. In Proceedings of the 8th International Conference on Language Resources and Evaluation, LREC 2012, Istanbul, Turkey, 23–25 May 2012; European Language Resources Association: Istanbul, Turkey, 2012. [Google Scholar]
  12. Ravi, K.; Ravi, V. Irony Detection Using Neural Network Language Model, Psycholinguistic Features and Text Mining. In Proceedings of the 2018 IEEE 17th International Conference on Cognitive Informatics and Cognitive Computing, ICCI*CC 2018, Berkeley, CA, USA, 16–18 July 2018. [Google Scholar] [CrossRef]
  13. Ravi, K.; Ravi, V. A novel automatic satire and irony detection using ensembled feature selection and data mining. Knowl. Based Syst. 2017, 120, 15–33. [Google Scholar] [CrossRef]
  14. Ashwitha, A.; Shruthi, G.; Shruthi, H.R.; Upadhyaya, M.; Ray, A.P.; Manjunath, T.C. Sarcasm detection in natural language processing. Mater. Today Proc. 2020, 37, 3324–3331. [Google Scholar] [CrossRef]
  15. Lad, R. Sarcasm Detection in English and Arabic Tweets Using Transformer Models. 2023. Available online: https://digitalcommons.dartmouth.edu/cs_senior_theses/20/ (accessed on 12 November 2023).
  16. Chauhan, D.S.; Singh, G.V.; Arora, A.; Ekbal, A.; Bhattacharyya, P. An emoji-aware multitask framework for multimodal sarcasm detection. Knowl. Based Syst. 2022, 257, 109924. [Google Scholar] [CrossRef]
  17. Attardo, S. Irony as relevant inappropriateness. J. Pragmat. 2000, 32, 793–826. [Google Scholar] [CrossRef]
  18. Goel, P.; Jain, R.; Nayyar, A.; Singhal, S.; Srivastava, M. Sarcasm detection using deep learning and ensemble learning. Multimed. Tools Appl. 2022, 81, 43229–43252. [Google Scholar] [CrossRef]
  19. Kader, F.B.; Nujat, N.H.; Sogir, T.B.; Kabir, M.; Mahmud, H.; Hasan, K. Computational Sarcasm Analysis on Social Media: A Systematic Review. arXiv 2022. Available online: https://arxiv.org/abs/2209.06170 (accessed on 13 November 2023). [CrossRef]
  20. Taşlıoğlu, H. Ironi Detection on Turkish Microblog Texts; Middle East Technical University: Ankara, Turkey, 2014. [Google Scholar]
  21. Yue, T.; Mao, R.; Wang, H.; Hu, Z.; Cambria, E. KnowleNet: Knowledge fusion network for multimodal sarcasm detection. Information Fusion. 2023, 100, 101921. [Google Scholar] [CrossRef]
  22. Rohanian, O.; Taslimipoor, S.; Evans, R.; Mitkov, R. WLV at SemEval-2018 Task 3: Dissecting Tweets in Search of Irony. In NAACL HLT 2018—International Workshop on Semantic Evaluation, SemEval 2018—Proceedings of the 12th Workshop, New Orleans, LA, USA, 5–6 June 2018; Association for Computational Linguistics: New Orleans, LA, USA, 2018; pp. 553–559. [Google Scholar] [CrossRef]
  23. Ren, L.; Xu, B.; Lin, H.; Liu, X.; Yang, L. Sarcasm Detection with Sentiment Semantics Enhanced Multi-level Memory Network. Neurocomputing 2020, 401, 320–326. [Google Scholar] [CrossRef]
  24. Chaudhari, P.; Chandankhede, C. Literature survey of sarcasm detection. In Proceedings of the 2017 International Conference on Wireless Communications, Signal Processing and Networking, WiSPNET 2017, Chennai, India, 22–24 March 2017. [Google Scholar] [CrossRef]
  25. Lee, C.J.; Katz, A.N. The Differential Role of Ridicule in Sarcasm and Irony. Metaphor. Symb. 1998, 13, 1–15. [Google Scholar] [CrossRef]
  26. Clift, R. Irony in conversation. Lang. Soc. 1999, 28, 523–553. [Google Scholar] [CrossRef]
  27. De Freitas, L.A.; Vanin, A.A.; Hogetop, D.N.; Bochernitsan, M.N.; Vieira, R. Pathways for irony detection in tweets. In Proceedings of the ACM Symposium on Applied Computing, Gyeongju, Republic of Korea, 24–28 March 2014. [Google Scholar] [CrossRef]
  28. Abulaish, M.; Kamal, A.; Zaki, M.J. A survey of figurative language and its computational detection in online social networks. ACM Trans. Web 2020, 14, 1–52. [Google Scholar] [CrossRef]
  29. Harared, N.; Nurani, S. SARCASM: MOCK POLITENESS in THE BIG BANG THEORY. Elite Engl. Lit. J. 2020, 7, 186–199. [Google Scholar] [CrossRef]
  30. Kaya, S.; Alatas, B. Sarcasm Detection with A New CNN + BiLSTM Hybrid Neural Network and BERT Classification Model. Int. J. Adv. Netw. Appl. 2022, 14, 5436–5443. [Google Scholar] [CrossRef]
  31. Nagwanshi, P.; Madhavan, C.E.V. Sarcasm detection using sentiment and semantic features. In Proceedings of the International Conference on Knowledge Discovery and Information Retrieval—KDIR 2014, Rome, Italy, 21–24 October 2014. [Google Scholar] [CrossRef]
  32. Oraby, S.; Harrison, V.; Misra, A.; Riloff, E.; Walker, M. Are you serious? Rhetorical questions and sarcasm in social media dialog. In SIGDIAL 2017—18th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Proceedings of the Conference, Saarbrücken, Germany, 15–17 August 2017; Association for Computational Linguistics: Saarbrücken, Germany, 2017. [Google Scholar] [CrossRef]
  33. Savini, E.; Caragea, C. Intermediate-Task Transfer Learning with BERT for Sarcasm Detection. Mathematics 2022, 10, 844. [Google Scholar] [CrossRef]
  34. Frenda, S.; Patti, V.; Rosso, P. When Sarcasm Hurts: Irony-Aware Models for Abusive Language Detection; Springer: Cham, Switzerland, 2023. [Google Scholar] [CrossRef]
  35. Hamza, M.A.; Alshahrani, H.J.; Hassan, A.Q.A.; Gaddah, A.; Allheeib, N.; Alsaif, S.A.; Al-Onazi, B.B.; Mohsen, H. Computational Linguistics with Optimal Deep Belief Network Based Irony Detection in Social Media. Comput. Mater. Contin. 2023, 75, 4137–4154. [Google Scholar] [CrossRef]
  36. Maladry, A.; Lefever, E.; Van Hee, C.; Hoste, V. A Fine Line Between Irony and Sincerity: Identifying Bias in Transformer Models for Irony Detection. In Proceedings of the 13th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis, Toronto, ON, Canada, 14 July 2023. [Google Scholar] [CrossRef]
  37. Malik, M.; Tomás, D.; Rosso, P. How Challenging is Multimodal Irony Detection? In Language Processing and Information Systems; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinfor-matics); Springer: Cham, Switzerland, 2023. [Google Scholar] [CrossRef]
  38. Ortega-Bueno, R.; Rosso, P.; Fersini, E. Cross-Domain and Cross-Language Irony Detection: The Impact of Bias on Models’ Generalization. In Natural Language Processing and Information Systems; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinfor-matics); Springer: Cham, Switzerland, 2023. [Google Scholar] [CrossRef]
  39. Saroj, A.; Pal, S. Ensemble-based domain adaptation on social media posts for irony detection. Multimed. Tools Appl. 2023, 83, 23249–23268. [Google Scholar] [CrossRef]
  40. Tasnia, R.; Ayman, N.; Sultana, A.; Chy, A.N.; Aono, M. Exploiting stacked embeddings with LSTM for multilingual humor and irony detection. Soc. Netw. Anal. Min. 2023, 13, 43. [Google Scholar] [CrossRef]
  41. Tomás, D.; Ortega-Bueno, R.; Zhang, G.; Rosso, P.; Schifanella, R. Transformer-based models for multimodal irony detection. J. Ambient. Intell. Humaniz. Comput. 2023, 14, 7399–7410. [Google Scholar] [CrossRef]
  42. Bettelli, G.; Panzeri, F. ‘Irony is easy to understand’: The role of emoji in irony detection. Intercult. Pragmat. 2023, 20, 467–493. [Google Scholar] [CrossRef]
  43. da Rocha Junqueira, J.; Junior, C.L.; Silva, F.L.V.; Côrrea, U.B.; de Freitas, L.A. Albertina in action: An investigation of its abilities in aspect extraction, hate speech detection, irony detection, and question-answering. In Proceedings of the Simpósio Brasileiro de Tecnologia da Informação e da Linguagem Humana (STIL), SBC, Belo Horizonte, Brazil, 25–29 September 2023; pp. 146–155. [Google Scholar]
  44. Băroiu, A.C.; Trăușan-Matu, Ș. Automatic Sarcasm Detection: Systematic Literature Review. Information 2022, 13, 399. [Google Scholar] [CrossRef]
  45. Govindan, V.; Balakrishnan, V. A machine learning approach in analysing the effect of hyperboles using negative sentiment tweets for sarcasm detection. J. King Saud Univ. —Comput. Inf. Sci. 2022, 34, 5110–5120. [Google Scholar] [CrossRef]
  46. Misra, R. News Headlines Dataset for Sarcasm Detection. arXiv 2022. Available online: http://arxiv.org/abs/2212.06035 (accessed on 14 November 2023). [CrossRef]
  47. Pandey, R.; Singh, J.P. BERT-LSTM model for sarcasm detection in code-mixed social media post. J. Intell. Inf. Syst. 2023, 60, 235–254. [Google Scholar] [CrossRef]
  48. Qiao, Y.; Jing, L.; Song, X.; Chen, X.; Zhu, L.; Nie, L. Mutual-Enhanced Incongruity Learning Network for Multi-Modal Sarcasm Detection. In Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023, Washington, DC, USA, 7–14 February 2023. [Google Scholar] [CrossRef]
  49. Sharma, D.K.; Singh, B.; Agarwal, S.; Kim, H.; Sharma, R. Sarcasm Detection over Social Media Platforms Using Hybrid Auto-Encoder-Based Model. Electronics 2022, 11, 2844. [Google Scholar] [CrossRef]
  50. Vinoth, D.; Prabhavathy, P. An intelligent machine learning-based sarcasm detection and classification model on social networks. J. Supercomput. 2022, 78, 10575–10594. [Google Scholar] [CrossRef]
  51. González-Ibáñez, R.; Muresan, S.; Wacholder, N. Identifying sarcasm in Twitter: A closer look. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies—ACL-HLT 2011, Portland, OR, USA, 19–24 June 2011. [Google Scholar]
  52. Justo, R.; Corcoran, T.; Lukin, S.M.; Walker, M.; Torres, M.I. Extracting relevant knowledge for the detection of sarcasm and nastiness in the social web. Knowl. Based Syst. 2014, 69, 124–133. [Google Scholar] [CrossRef]
  53. del Pilar Salas-Zárate, M.; Paredes-Valverde, M.A.; Rodriguez-García, M.Á.; Valencia-García, R.; Alor-Hernández, G. Automatic detection of satire in Twitter: A psycholinguistic-based approach. Knowl. Based Syst. 2017, 128, 20–33. [Google Scholar] [CrossRef]
  54. Al-Ghadhban, D.; Alnkhilan, E.; Tatwany, L.; Alrazgan, M. Arabic sarcasm detection in Twitter. In Proceedings of the 2017 International Conference on Engineering and MIS, ICEMIS 2017, Monastir, Tunisia, 8–10 May 2017. [Google Scholar] [CrossRef]
  55. da Rocha Junqueira, J.; Da Silva, F.; Costa, W.; Carvalho, R.; Bender, A.; Correa, U.; Freitas, L. BERTimbau in Action: An Investigation of its Abilities in Sentiment Analysis, Aspect Extraction, Hate Speech Detection, and Irony Detection. In Proceedings of the International Florida Artificial Intelligence Research Society Conference, FLAIRS, Clearwater Beach, FL, USA, 14–17 May 2023. [Google Scholar] [CrossRef]
  56. Ramteke, J.; Shah, S.; Godhia, D.; Shaikh, A. Election result prediction using Twitter sentiment analysis. In Proceedings of the International Conference on Inventive Computation Technologies, ICICT 2016, Coimbatore, India, 26–27 August 2016. [Google Scholar] [CrossRef]
  57. Bouazizi, M.; Otsuki, T. A Pattern-Based Approach for Sarcasm Detection on Twitter. IEEE Access 2016, 4, 5477–5488. [Google Scholar] [CrossRef]
  58. Forslid, E.; Wikén, N. Automatic Irony- and Sarcasm Detection in Social Media; Uppsala Universitet: Uppsala, Sweden, 2015. [Google Scholar]
  59. Nayak, D.K.; Bolla, B.K. Efficient Deep Learning Methods for Sarcasm Detection of News Headlines. In Machine Learning and Autonomous Systems; Smart Innovation, Systems and Technologies; Springer: Singapore, 2022. [Google Scholar] [CrossRef]
  60. Shrikhande, P.; Setty, V.; Sahani, A. Sarcasm Detection in Newspaper Headlines. In Proceedings of the 2020 IEEE 15th International Conference on Industrial and Information Systems, ICIIS 2020—Proceedings, Rupnagar, India, 26–28 November 2020. [Google Scholar] [CrossRef]
  61. Zanchak, M.; Vysotska, V.; Albota, S. The sarcasm detection in news headlines based on machine learning technology. In Proceedings of the International Scientific and Technical Conference on Computer Sciences and Information Technologies, Lviv, Ukraine, 22–25 September 2021. [Google Scholar] [CrossRef]
  62. Zuhri, A.T.; Sagala, R.W. Irony and Sarcasm Detection on Public Figure Speech. J. Elem. Sch. Educ. 2022, 1, 41–45. Available online: https://journal.berpusi.co.id/index.php/joese/article/view/13 (accessed on 12 November 2023).
  63. Kandasamy, I.; Vasantha, W.B.; Obbineni, J.M.; Smarandache, F. Sentiment analysis of tweets using refined neutrosophic sets. Comput. Ind. 2020, 115, 103180. [Google Scholar] [CrossRef]
  64. Basiri, M.E.; Nemati, S.; Abdar, M.; Cambria, E.; Acharya, U.R. ABCDM: An Attention-based Bidirectional CNN-RNN Deep Model for sentiment analysis. Future Gener. Comput. Syst. 2021, 115, 279–294. [Google Scholar] [CrossRef]
  65. Abdar, M.; Basiri, M.E.; Yin, J.; Habibnezhad, M.; Chi, G.; Nemati, S.; Asadi, S. Energy choices in Alaska: Mining people’s perception and attitudes from geotagged tweets. Renew. Sustain. Energy Rev. 2020, 124, 109781. [Google Scholar] [CrossRef]
  66. Almuqren, L.; Cristea, A. AraCust: A Saudi Telecom Tweets corpus for sentiment analysis. PeerJ Comput. Sci. 2021, 7, e510. [Google Scholar] [CrossRef]
  67. Zhu, N.; Wang, Z. The paradox of sarcasm: Theory of mind and sarcasm use in adults. Pers. Individ. Dif. 2020, 163, 110035. [Google Scholar] [CrossRef]
  68. Sekharan, S.C.; Vadivu, G.; Rao, M.V. A Comprehensive Study on Sarcasm Detection Techniques in Sentiment Analysis. Int. J. Pure Appl. Math. 2018, 118, 433–442. [Google Scholar]
  69. Özkan, M.; Kar, G. TÜRKÇE DİLİNDE YAZILAN BİLİMSEL METİNLERİN DERİN ÖĞRENME TEKNİĞİ UYGULANARAK ÇOKLU SINIFLANDIRILMASI. Mühendislik Bilim. Tasarım Derg. 2022, 10, 504–519. [Google Scholar] [CrossRef]
  70. Dimovska, J.; Angelovska, M.; Gjorgjevikj, D.; Madjarov, G. Sarcasm and Irony Detection in English Tweets. In ICT Innovations 2018. Engineering and Life Science; Communications in Computer and Information Science; Springer: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  71. Samonte, M.J.C.; Dollete, C.J.T.; Capanas, P.M.M.; Flores, M.L.C.; Soriano, C.B. Sentence-level sarcasm detection in English and Filipino tweets. In Proceedings of the 4th International Conference on Industrial and Business Engineering, Macao, 24–26 October 2018. [Google Scholar] [CrossRef]
  72. Yan, Y.; Zhang, B.-W.; Ding, G.; Li, W.; Zhang, J.; Li, J.-J.; Gao, W. O2-Bert: Two-Stage Target-Based Sentiment Analysis. Cognit. Comput. 2023, 16, 158–176. [Google Scholar] [CrossRef]
  73. Gedela, R.T.; Meesala, P.; Baruah, U.; Soni, B. Identifying sarcasm using heterogeneous word embeddings: A hybrid and ensemble perspective. Soft Comput. 2023, 28, 13941–13954. [Google Scholar] [CrossRef]
  74. Quintero, Y.C.; García, L.A. Irony detection based on character language model classifiers. In Progress in Artificial Intelligence and Pattern Recognition; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  75. Bölücü, N.; Can, B. Sarcasm target identification with lstm networks. In Proceedings of the 2020 28th Signal Processing and Communications Applications Conference (SIU), Türkiye, Gaziantep, 5–7 October 2020; IEEE: Piscataway, NJ, USA; pp. 1–4. [Google Scholar]
  76. Mohammed, P.; Eid, Y.; Badawy, M.; Hassan, A. Evaluation of Different Sarcasm Detection Models for Arabic News Headlines. In Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2019, Cairo, Egypt, 26–28 October 2019; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2020. [Google Scholar] [CrossRef]
  77. Talafha, B.; Za’ter, M.E.; Suleiman, S.; Al-Ayyoub, M.; Al-Kabi, M.N. Sarcasm Detection and Quantification in Arabic Tweets. In Proceedings of the International Conference on Tools with Artificial Intelligence, ICTAI, Washington, DC, USA, 1–3 November 2021. [Google Scholar] [CrossRef]
  78. Zhang, L.; Zhao, X.; Song, X.; Fang, Y.; Li, D.; Wang, H. A Novel Chinese Sarcasm Detection Model Based on Retrospective Reader. In MultiMedia Modeling; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2022. [Google Scholar] [CrossRef]
  79. Liu, L.; Chen, X.; He, B. End-to-End Multi-task Learning for Allusion Detection in Ancient Chinese Poems. In Knowledge Science, Engineering and Management; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2020. [Google Scholar] [CrossRef]
  80. Okimoto, Y.; Suwa, K.; Zhang, J.; Li, L. Sarcasm Detection for Japanese Text Using BERT and Emoji. In Database and Expert Systems Applications; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2021. [Google Scholar] [CrossRef]
  81. Sharma, Y.; Mandalam, A.V. Irony Detection in Non-English Tweets. In Proceedings of the 2021 6th International Conference for Convergence in Technology, I2CT 2021, Pune, India, 2–4 April 2021. [Google Scholar] [CrossRef]
  82. Bharti, S.K.; Naidu, R.; Babu, K.S. Hyperbolic Feature-based Sarcasm Detection in Telugu Conversation Sentences. J. Intell. Syst. 2020, 30, 73–89. [Google Scholar] [CrossRef]
  83. Bharti, S.K.; Babu, K.S.; Jena, S.K. Harnessing Online News for Sarcasm Detection in Hindi Tweets. In Pattern Recognition and Machine Intelligence; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2017. [Google Scholar] [CrossRef]
  84. Yunitasari, Y.; Musdholifah, A.; Sari, A.K. Sarcasm Detection for Sentiment Analysis in Indonesian Tweets. IJCCS (Indones. J. Comput. Cybern. Syst.) 2019, 13, 53–62. [Google Scholar] [CrossRef]
  85. Lunando, E.; Purwarianti, A. Indonesian social media sentiment analysis with sarcasm detection. In Proceedings of the 2013 International Conference on Advanced Computer Science and Information Systems, ICACSIS 2013, Sanur Bali, Indonesia, 28–29 September 2013. [Google Scholar] [CrossRef]
  86. Anan, R.; Apon, T.S.; Hossain, Z.T.; Modhu, E.A.; Mondal, S.; Alam, M.D.G.R. Interpretable Bangla Sarcasm Detection using BERT and Explainable AI. In Proceedings of the 2023 IEEE 13th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 8–11 March 2023; pp. 1272–1278. [Google Scholar] [CrossRef]
  87. Jain, D.; Kumar, A.; Garg, G. Sarcasm detection in mash-up language using soft-attention based bi-directional LSTM and feature-rich CNN. Appl. Soft Comput. 2020, 91, 106198. [Google Scholar] [CrossRef]
  88. Karabaş, A.; Dırı, B. Irony Detection with Deep Learning in Turkish Microblogs. In Proceedings of the 2020 28th Signal Processing and Communications Applications Conference (SIU), IEEE, Gaziantep, Turkey, 5–7 October 2020; pp. 1–4. [Google Scholar]
  89. Onan, A.; Toçoğlu, M.A. Satire identification in Turkish news articles based on ensemble of classifiers. Turk. J. Electr. Eng. Comput. Sci. 2020, 28, 1086–1106. [Google Scholar] [CrossRef]
  90. Bölücü, N.; Can, B. Semantically-Informed Graph Neural Networks for Irony Detection in Turkish. ACM Trans. Asian Low-Resour. Lang. Inf. Process. 2025, 24, 1–20. [Google Scholar] [CrossRef]
  91. Öztürk, A.U. Designing and Debiasing Binary Classifiers for Irony and Satire Detection; Middle East Technical University: Ankara, Turkey, 2024. [Google Scholar]
  92. Riloff, E.; Qadir, A.; Surve, P.; De Silva, L.; Gilbert, N.; Huang, R. Sarcasm as contrast between a positive sentiment and negative situation. In Proceedings of the EMNLP 2013—2013 Conference on Empirical Methods in Natural Language Processing, Seattle, WA, USA, 18–21 October 2013. [Google Scholar]
  93. Amir, S.; Wallace, B.C.; Lyu, H.; Carvalho, P.; Silva, M.J. Modelling context with user embeddings for sarcasm detection in social media. In Proceedings of the CoNLL 2016—20th SIGNLL Conference on Computational Natural Language Learning, Berlin, Germany, 7–12 August 2016. [Google Scholar] [CrossRef]
  94. Bamman, D.; Smith, N.A. Contextualized sarcasm detection on twitter. In Proceedings of the 9th International Conference on Web and Social Media, ICWSM 2015, Oxford, UK, 26–29 May 2015. [Google Scholar] [CrossRef]
  95. Barbieri, F.; Saggion, H.; Ronzano, F. Modelling Sarcasm in Twitter, a Novel Approach. In Proceedings of the 5th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Baltimore, MD, USA, 27 June 2014; Association for Computational Linguistics: Baltimore, MD, USA, 2014. [Google Scholar] [CrossRef]
  96. Ptáček, T.; Habernal, I.; Hong, J. Sarcasm detection on Czech and English twitter. In COLING 2014—25th International Conference on Computational Linguistics, Proceedings of COLING 2014: Technical Papers, Dublin, Ireland, 23–29 August 2014; Dublin City University and Association for Computational Linguistics: Dublin, Ireland, 2014. [Google Scholar]
  97. Schifanella, R.; De Juan, P.; Tetreault, J.; Cao, L. Detecting sarcasm in multimodal social platforms. In Proceedings of the 2016 ACM Multimedia Conference—MM 2016, Amsterdam, The Netherlands, 15–19 October 2016. [Google Scholar] [CrossRef]
  98. Joshi, A.; Tripathi, V.; Patel, K.; Bhattacharyya, P.; Carman, M. Are word embedding-based features useful for sarcasm detection? In EMNLP 2016—Conference on Empirical Methods in Natural Language Processing, Proceedings; Association for Computational Linguistics: Austin, TX, USA, 2016. [Google Scholar] [CrossRef]
  99. Son, L.H.; Kumar, A.; Sangwan, S.R.; Arora, A.; Nayyar, A.; Abdel-Basset, M. Sarcasm detection using soft attention-based bidirectional long short-term memory model with convolution network. IEEE Access 2019, 7, 23319–23328. [Google Scholar] [CrossRef]
  100. Subramanian, J.; Sridharan, V.; Shu, K.; Liu, H. Exploiting emojis for sarcasm detection. In Social, Cultural, and Behavioral Modeling; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics); Springer: Cham, Switzerland, 2019. [Google Scholar] [CrossRef]
  101. Tay, Y.; Tuan, L.A.; Hui, S.C.; Su, J. Reasoning with sarcasm by reading in-between. In ACL 2018—56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers), Melbourne, Australia, 15–20 July 2018; Association for Computational Linguistics: Melbourne, Australia, 2018. [Google Scholar] [CrossRef]
  102. Davidov, D.; Tsur, O.; Rappoport, A. Semi-supervised recognition of sarcastic sentences in twitter and Amazon. In CoNLL 2010—Fourteenth Conference on Computational Natural Language Learning, Proceedings of the Conference, Uppsala, Sweden, 15–16 July 2010; Association for Computational Linguistics: Uppsala, Sweden, 2010. [Google Scholar]
  103. Sundararajan, K.; Palanisamy, A. Multi-rule based ensemble feature selection model for sarcasm type detection in Twitter. Comput. Intell. Neurosci. 2020, 2020, 2860479. [Google Scholar] [CrossRef]
  104. Thakur, S.; Singh, S.; Singh, M. Detecting Sarcasm in Text. In Intelligent Systems Design and Applications; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2020. [Google Scholar] [CrossRef]
  105. Mukherjee, S.; Bala, P.K. Detecting sarcasm in customer tweets: An NLP based approach. Ind. Manag. Data Syst. 2017, 117, 1109–1126. [Google Scholar] [CrossRef]
  106. Abercrombie, G.; Hovy, D. Putting sarcasm detection into context: The effects of class imbalance and manual labelling on supervised machine classification of twitter conversations. In 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016—Student Research Workshop; Association for Computational Linguistics: Berlin, Germany, 2016. [Google Scholar] [CrossRef]
  107. Parde, N.; Nielsen, R. Detecting Sarcasm is Extremely Easy. In Proceedings of the Workshop on Computational Semantics Beyond Events and Roles, Valencia, Spain, 4 April 2017; Association for Computational Linguistics: New Orleans, LA, USA, 2018. [Google Scholar] [CrossRef]
  108. Liu, Y.; Wang, Y.; Sun, A.; Meng, X.; Li, J.; Guo, J. A Dual-Channel Framework for Sarcasm Recognition by Detecting Sentiment Conflict. In Findings of the Association for Computational Linguistics: NAACL 2022—Findings; Association for Computational Linguistics: Seattle, WA, USA, 2022. [Google Scholar] [CrossRef]
  109. Kumar, A.; Narapareddy, V.T.; Srikanth, V.A.; Malapati, A.; Neti, L.B.M. Sarcasm Detection Using Multi-Head Attention Based Bidirectional LSTM. IEEE Access 2020, 8, 6388–6397. [Google Scholar] [CrossRef]
  110. Salim, S.S.; Ghanshyam, A.N.; Ashok, D.M.; Mazahir, D.B.; Thakare, B.S. Deep LSTM-RNN with word embedding for sarcasm detection on twitter. In Proceedings of the 2020 International Conference for Emerging Technology, INCET 2020, Belgaum, India, 5–7 June 2020. [Google Scholar] [CrossRef]
  111. Mehndiratta, P.; Soni, D. Identification of sarcasm using word embeddings and hyperparameters tuning. J. Discret. Math. Sci. Cryptogr. 2019, 22, 465–489. [Google Scholar] [CrossRef]
  112. Potamias, R.A.; Siolas, G.; Stafylopatis, A. A robust deep ensemble classifier for figurative language detection. In Engineering Applications of Neural Networks; Communications in Computer and Information Science; Springer: Cham, Switzerland, 2019. [Google Scholar] [CrossRef]
  113. Jain, T.; Agrawal, N.; Goyal, G.; Aggrawal, N. Sarcasm detection of tweets: A comparative study. In Proceedings of the 2017 10th International Conference on Contemporary Computing, IC3 2017, Noida, India, 10–12 August 2017. [Google Scholar] [CrossRef]
  114. Kumar, A.; Narapareddy, V.T.; Gupta, P.; Srikanth, V.A.; Neti, L.B.M.; Malapati, A. Adversarial and Auxiliary Features-Aware BERT for Sarcasm Detection. In Proceedings of the 3rd ACM India Joint International Conference on Data Science & Management of Data (8th ACM IKDD CODS & 26th COMAD), Bangalore India, 2–4 January 2021. [Google Scholar] [CrossRef]
  115. Kalaivani, A.; Thenmozhi, D. Sarcasm identification and detection in conversion context using BERT. In Second Workshop on Figurative Language Processing; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020. [Google Scholar] [CrossRef]
  116. Srivastava, H.; Varshney, V.; Kumari, S.; Srivastava, S. A Novel Hierarchical BERT Architecture for Sarcasm Detection. In Proceedings of the Second Workshop on Figurative Language Processing, Online, 9 July 2020. [Google Scholar] [CrossRef]
  117. Gregory, H.; Li, S.; Mohammadi, P.; Tarn, N.; Draelos, R.; Rudin, C. A Transformer Approach to Contextual Sarcasm Detection in Twitter. In Proceedings of the Second Workshop on Figurative Language Processing, Online, 9 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020. [Google Scholar] [CrossRef]
  118. Javdan, S.; Minaei-Bidgoli, B. Applying Transformers and Aspect-based Sentiment Analysis approaches on Sarcasm Detection. In Proceedings of the Second Workshop on Figurative Language Processing, Online, 9 July 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 67–71. [Google Scholar] [CrossRef]
  119. Milliyet. “Milliyet Gazetesi,” Milliyet Gazetecilik ve Yayıncılık A.Ş. Available online: https://www.milliyet.com.tr (accessed on 7 November 2025).
  120. Cumhuriyet. “Cumhuriyet Gazetesi,” Yeni Gün Haber Ajansı Basın ve Yayıncılık A.Ş. Available online: https://www.cumhuriyet.com.tr (accessed on 7 November 2025).
  121. Zhang, S.; Dong, L.; Li, X.; Zhang, S.; Sun, X.; Wang, S.; Li, J.; Hu, R.; Zhang, T.; Wu, F.; et al. Instruction Tuning for Large Language Models: A Survey. arXiv 2023. Available online: https://arxiv.org/pdf/2308.10792 (accessed on 21 August 2025). [CrossRef]
  122. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the NAACL HLT 2019—2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; Volume 1, pp. 4171–4186. Available online: https://arxiv.org/abs/1810.04805v2 (accessed on 10 April 2025).
  123. Schweter, S. BERTurk-BERT models for Turkish. Zenodo 2020. [Google Scholar] [CrossRef]
  124. Jiang, Z.; Yu, W.; Zhou, D.; Chen, Y.; Feng, J.; Yan, S. ConvBERT: Improving BERT with Span-based Dynamic Convolution. arXiv 2020. Available online: https://arxiv.org/abs/2008.02496v3 (accessed on 10 April 2025).
  125. Sanh, V.; Debut, L.; Chaumond, J.; Wolf, T. DistilBERT, a distilled version of BERT: Smaller, faster, cheaper and lighter. arXiv 2019. Available online: https://arxiv.org/abs/1910.01108v4 (accessed on 10 April 2025).
  126. Abadeer, M. Assessment of DistilBERT performance on Named Entity Recognition task for the detection of Protected Health Information and medical concepts. In Proceedings of the 3rd Clinical Natural Language Processing Workshop, Online, 19 November 2020; Association for Computational Linguistics: Stroudsburg, PA, USA, 2020; pp. 158–167. [Google Scholar] [CrossRef]
  127. Li, J.; Zhu, Y.; Sun, K. A novel iteration scheme with conjugate gradient for faster pruning on transformer models. Complex Intell. Syst. 2024, 10, 7863–7875. [Google Scholar] [CrossRef]
  128. Clark, K.; Luong, M.T.; Le, Q.V.; Manning, C.D. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020; Available online: https://arxiv.org/abs/2003.10555v1 (accessed on 11 April 2025).
  129. Husein, A.M.; Livando, N.; Andika, A.; Chandra, W.; Phan, G. Sentiment Analysis Of Hotel Reviews on Tripadvisor with LSTM and ELECTRA. Sink. J. Dan. Penelit. Tek. Inform. 2023, 7, 733–740. [Google Scholar] [CrossRef]
  130. Xu, Z.; Gong, L.; Ke, G.; He, D.; Zheng, S.; Wang, L.; Bian, J.; Liu, T.-Y. MC-BERT: Efficient Language Pre-Training via a Meta Controller. arXiv 2020. Available online: https://arxiv.org/abs/2006.05744v2 (accessed on 10 April 2025).
  131. Gargiulo, F.; Minutolo, A.; Guarasci, R.; Damiano, E.; De Pietro, G.; Fujita, H.; Esposito, M. An ELECTRA-Based Model for Neural Coreference Resolution. IEEE Access 2022, 10, 75144–75157. [Google Scholar] [CrossRef]
  132. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019. Available online: https://arxiv.org/abs/1907.11692v1 (accessed on 9 November 2023).
  133. Wang, L.; Cao, Q.; Zhang, Z.; Mirjalili, S.; Zhao, W. Artificial rabbits optimization: A new bio-inspired meta-heuristic algorithm for solving engineering optimization problems. Eng. Appl. Artif. Intell. 2022, 114, 105082. [Google Scholar] [CrossRef]
Figure 1. Artificial Rabbit Optimization-Based Ensemble of Fine-Tuned Transformers for Turkish NLP.
Figure 1. Artificial Rabbit Optimization-Based Ensemble of Fine-Tuned Transformers for Turkish NLP.
Applsci 15 12498 g001
Figure 2. Datasets architecture.
Figure 2. Datasets architecture.
Applsci 15 12498 g002
Figure 4. Categories of texts used in the dataset.
Figure 4. Categories of texts used in the dataset.
Applsci 15 12498 g004
Figure 5. Pre-training and fine-tuning of the BERT model [123].
Figure 5. Pre-training and fine-tuning of the BERT model [123].
Applsci 15 12498 g005
Figure 6. ConvBERT Architecture Diagram [125].
Figure 6. ConvBERT Architecture Diagram [125].
Applsci 15 12498 g006
Figure 7. Schematic diagram of BERTBASE and DistilBERT model architecture [127].
Figure 7. Schematic diagram of BERTBASE and DistilBERT model architecture [127].
Applsci 15 12498 g007
Figure 8. An overview of replaced token detection [128].
Figure 8. An overview of replaced token detection [128].
Applsci 15 12498 g008
Figure 9. Flowchart of the ensemble model created with the ARO algorithm [133].
Figure 9. Flowchart of the ensemble model created with the ARO algorithm [133].
Applsci 15 12498 g009
Figure 10. Performance comparison according to (a) monogram and (b) omnigram dataset architectures.
Figure 10. Performance comparison according to (a) monogram and (b) omnigram dataset architectures.
Applsci 15 12498 g010
Figure 11. Performance comparison with ensemble model according to (a) monogram and (b) omnigram dataset architectures.
Figure 11. Performance comparison with ensemble model according to (a) monogram and (b) omnigram dataset architectures.
Applsci 15 12498 g011
Table 1. Existing sarcasm and irony identification studies.
Table 1. Existing sarcasm and irony identification studies.
ReferenceMethodologyHighlights
[102]Rule BasedThe algorithm employs two modules: semi supervised pattern acquisition for identifying sarcastic patterns that serve as features for a classifier, and a classification stage that classifies each sentence to a sarcastic class. To assign a score to new examples in the test set, we use a k-nearest neighbors strategy.
[103]Rule BasedAfter detecting sarcastic tweets, a multi-rule-based approach was used to classify them into four sarcasm types: polite, rude, raging, and deadpan. This approach relies on predefined rules and classifies the tweets based on lexical features, syntactic features, semantic features, and neutral tone with implicit contradiction.
[104]Machine LearningA supervised machine learning approach was used for sarcasm detection; the Naive Bayes classification algorithm was applied with different feature categories such as content words, function words, POS-based sarcastic patterns, and their combinations.
[105]Machine LearningTo detect sarcasm in customer tweets, supervised learning algorithms using feature sets combining function words and content words were applied with Naive Bayes and maximum entropy classifiers.
[106]Machine LearningHuman vs. machine learning classification performance has been compared under varying amounts of contextual information, and machine performance has been evaluated on balanced and unbalanced, and manually labelled and automatically retrieved datasets.
[107]Machine LearningThe method uses Naive Bayes classification on a feature space augmented using domain adaptation technique. Features are extracted uniformly from both domains and mapped into source, target, and general versions before training.
[108]Deep LearningThe study employs a Dual-Channel Network (DC-Net) that separates the input text into emotional words and the remaining context, encodes these parts separately through the literal and implied channels, and identifies sarcasm in the analyzer module by detecting emotional contradictions.
[101]Deep LearningIn the study, a hybrid deep learning approach was employed that combines features extracted from a soft attention–based bidirectional Long Short-Term Memory (BiLSTM) layer utilizing GloVe word embeddings with a convolutional neural network.
[109]Deep LearningA deep learning model has been developed that performs sarcasm classification by utilizing a multi-head self-attention–based BiLSTM network, which takes both automatically learned and handcrafted features as input.
[110]Deep LearningThe proposed approach is a deep learning method that performs sarcasm classification using a Recurrent neural network (RNN)- Long Short-Term Memory (LSTM) model after applying word embedding to the preprocessed dataset. The model cleans and encodes the training dataset with label and integer encoding, constructs a word vector matrix, feeds this matrix into the LSTM to record the weights, and calculates the accuracy using the test dataset.
[111]Deep LearningThe study identifies sarcasm in textual data by training and evaluating Convolutional Neural Networks (CNN), RNN, and a hybrid combination of these models, while systematically analyzing the effects of training data size, number of epochs, dropout rates, and different word embeddings on classification performance using the large-scale Reddit corpus.
[112]Ensemble LearningThe study detects sarcasm in tweets through a two-phase approach: first extracting sentiment and punctuation features followed by chi-square feature selection, and second combining the top 200 TF-IDF features with the selected sentiment and punctuation features, then applying Support Vector Machine in the first phase and a voting classifier in the second phase for classification.
[113]Ensemble LearningThe study detects sarcasm by leveraging positive sentiment attached to negative situations through two ensemble-based approaches—a voted ensemble classifier and a random forest classifier—trained on a corpus generated via a seeding algorithm, while also incorporating a pragmatic classifier to detect emoticon-based sarcasm.
[114]Transformer-BasedThe study detects sarcasm using the Adversarial and Auxiliary Features-Aware BERT (AAFAB) model, which encodes sentences with Bidirectional Encoder Representations from Transformers (BERT) contextual embeddings, combines them with manually extracted auxiliary features, and applies adversarial training by adding perturbations to input word embeddings for improved generalization.
[115]Transformer-BasedThe study identifies sarcasm in social media conversation texts using the BERT model and compares its performance with alternative approaches on Twitter and Reddit datasets with combined context-response and isolated response texts.
[116]Transformer-BasedIn the study, sarcasm detection was performed by encoding the context and the response through separate BERT layers, processing the response with a BiLSTM, summarizing the context using convolution and BiLSTM, and finally classifying the output through a multi-channel CNN and a fully connected layer.
[117]Transformer-BasedIn the study, LSTM, Gated Recurrent Unit (GRU), and Transformer models were applied to detect sarcasm in Twitter posts, and the best performance was achieved through an ensemble combination of BERT, RoBERTa, XLNet, RoBERTa-large, and ALBERT models.
[118]Transformer-BasedIn the study, BERT and aspect-based sentiment analysis approaches were employed to extract the relationship between the contextual dialogue sequence and the response, and to determine whether the response is sarcastic or not.
Table 3. Comparative Evaluation of Transformer-Based Models for Turkish Language Tasks.
Table 3. Comparative Evaluation of Transformer-Based Models for Turkish Language Tasks.
ModelDatasetAccuracy (%)Precision (%)Recall (%)F-Score (%)
BERTurkmonogram65.63 63.19 61.89 62.58
ConvBERTurk60.89 59.56 58.19 58.88
DistilBERTurk57.45 55.61 54.89 55.22
ELECTRATurk59.59 55.30 56.56 55.87
RoBERTaTurk54.32 50.93 50.72 50.80
BERTurkbigram66.7264.7263.564.12
ConvBERTurk65.364.3962.0063.18
DistilBERTurk60.9157.8157.5257.66
ELECTRATurk65.5364.0861.9362.99
RoBERTaTurk61.2460.9956.8358.84
BERTurktrigram67.4264.7364.3264.52
ConvBERTurk65.6261.5861.9761.77
DistilBERTurk62.7160.2359.6359.93
ELECTRATurk66.1763.7462.9663.35
RoBERTaTurk61.1656.8456.6156.72
BERTurkquadrigram67.0765.08 63.99 64.50
ConvBERTurk68.76 66.29 65.19 65.75
DistilBERTurk66.39 62.70 62.40 62.57
ELECTRATurk67.00 62.22 63.72 62.93
RoBERTaTurk61.11 57.62 56.51 57.05
BERTurkomnigram70.2868.3567.4067.90
ConvBERTurk70.7467.9967.7467.84
DistilBERTurk66.47 63.2263.2063.22
ELECTRATurk67.76 65.05 64.87 64.96
RoBERTaTurk61.8852.0057.0354.39
Table 4. A Comparative Evaluation of the Two Most Successful Models, BERTurk and ConvBERTurk, Across the Datasets.
Table 4. A Comparative Evaluation of the Two Most Successful Models, BERTurk and ConvBERTurk, Across the Datasets.
ModelDatasetAccuracy (%)Precision (%)Recall (%)F-Score (%)
BERTurkmonogram 65.63 63.19 61.89 62.58
BERTurkbigram 66.72 64.72 63.50 64.10
BERTurktrigram 67.4264.7364.3264.52
BERTurkquadrigram 67.0765.08 63.99 64.50
BERTurkomnigram 702868.3567.4067.90
ConvBERTurkmonogram 60.89 59.56 58.19 58.88
ConvBERTurkbigram 65.30 64.39 62.00 63.17
ConvBERTurktrigram 65.6261.5861.9761.77
ConvBERTurkquadrigram 68.76 66.29 65.19 65.75
ConvBERTurkomnigram 70.7467.9967.7467.84
Table 5. Comparative Evaluation of Transformer-Based Models and the Proposed Ensemble Learning Method for Turkish Language Tasks.
Table 5. Comparative Evaluation of Transformer-Based Models and the Proposed Ensemble Learning Method for Turkish Language Tasks.
ModelDatasetAccuracy (%)Precision (%)Recall (%)F-Score (%)Weights
BERTurkmonogram58.02 54.5554.6254.580.2916
ConvBERTurk54.20 60.3157.9759.120.0671
DistilBERTurk61.45 53.5851.8052.480.0086
ELECTRATurk56.49 40.8355.6247.020.6122
RoBERTaTurk61.45 52.4849.7151.060.0205
Average58.3252.7553.9452.85-
Ensemble Model65.25 63.75 60.55 62.12-
BERTurkbigram61.0756.2656.5056.380.0525
ConvBERTurk62.2159.2658.4558.850.0369
DistilBERTurk57.6352.6853.2252.950.0601
ELECTRATurk64.8959.1759.7059.430.4594
RoBERTaTurk56.4958.9154.6056.680.3912
Average60.4557.2556.4956.85-
Ensemble Model69.8466.4265.1865.80-
BERTurktrigram62.5061.6860.4461.060.0434
ConvBERTurk65.2761.5161.1261.310.0761
DistilBERTurk64.5064.3261.5862.930.4505
ELECTRATurk66.4163.0462.6662.850.4267
RoBERTaTurk59.2458.6557.0357.830.0032
Average63.5861.8460.5661.19-
Ensemble Model71.37 69.78 67.70 68.73-
BERTurkquadrigram64.1259.9059.6059.750.1992
ConvBERTurk61.8362.2061.7061.950.2988
DistilBERTurk66.0362.9563.0062.980.3102
ELECTRATurk66.7959.7760.3060.030.0991
RoBERTaTurk65.2757.2956.8057.040.0927
Average64.8060.4260.2860.35
Ensemble Model72.1370.7867.8169.27-
BERTurkomnigram69.8566.9866.3166.640.0639
ConvBERTurk71.7668.3268.3168.320.4383
DistilBERTurk66.4161.1060.6460.870.0346
ELECTRATurk70.2369.2465.7667.460.4601
RoBERTaTurk58.4050.2952.2951.270.0031
Average67.3363.1862.6662.91-
Ensemble Model76.7174.6473.2973.96-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Eser, M.; Bilgin, M. Irony and Sarcasm Detection in Turkish Texts: A Comparative Study of Transformer-Based Models and Ensemble Learning. Appl. Sci. 2025, 15, 12498. https://doi.org/10.3390/app152312498

AMA Style

Eser M, Bilgin M. Irony and Sarcasm Detection in Turkish Texts: A Comparative Study of Transformer-Based Models and Ensemble Learning. Applied Sciences. 2025; 15(23):12498. https://doi.org/10.3390/app152312498

Chicago/Turabian Style

Eser, Murat, and Metin Bilgin. 2025. "Irony and Sarcasm Detection in Turkish Texts: A Comparative Study of Transformer-Based Models and Ensemble Learning" Applied Sciences 15, no. 23: 12498. https://doi.org/10.3390/app152312498

APA Style

Eser, M., & Bilgin, M. (2025). Irony and Sarcasm Detection in Turkish Texts: A Comparative Study of Transformer-Based Models and Ensemble Learning. Applied Sciences, 15(23), 12498. https://doi.org/10.3390/app152312498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop