1. Introduction
In recent years, the spread of disinformation in the digital media space has reached unprecedented levels, significantly influencing public opinion, political processes, and information security. This issue is particularly pressing in multilingual societies such as Kazakhstan and the broader post-Soviet region, where lexical, syntactic, and stylistic variability complicates the identification of fake content. Existing fake news detection models were primarily developed for English-language content and exhibit limited applicability in Kazakh–Russian contexts, underscoring the urgent need to adapt annotation approaches to local linguistic and cultural realities, unlike the English-language media landscape, which benefits from large-scale corpora and mature fake news detection models (e.g., FEVER, LIAR, FakeNewsNet), the Kazakh–Russian information space is characterized by a set of specific features not reflected in current annotation schemes. First, there are linguistic differences related to the agglutinative nature of the Kazakh language, free word order, and unique morphosyntactic structures, which hinder the direct transfer of English-based models without adaptation. Second, the Kazakhstani media environment is dominated by localized forms of disinformation, including ethnopolitical narratives and manipulations on interethnic and socioeconomic grounds, phenomena that differ substantially from Western patterns of fake news. Third, the mode of information delivery also diverges: local texts frequently employ emotional and culturally specific arguments, distort context, or target narrow social groups (e.g., rural populations or ethnic minorities), aspects that are not captured by existing English-language annotation schemes, which are predominantly tailored to the political discourse of the US or Europe. Due to these factors, the unadapted application of existing models and annotation schemes results in reduced accuracy and interpretability of outcomes. The development of a new multi-level annotation framework that incorporates the cultural, pragmatic, and discursive characteristics of Kazakh–Russian fake news is therefore essential for constructing relevant and robust models of automatic disinformation detection in this context.
As Shu et al. [
1] noted, disinformation refers to false or misleading information disseminated with the intent to deceive. During critical events such as elections, pandemics, and economic crises, the volume of fake news increases, fueled by fear, uncertainty, and ideological polarization. The internet and social media serve as catalysts, accelerating the spread of unverified content and replacing professional journalism with user-generated content on social media platforms (e.g., Facebook, Instagram, VK, and Twitter) and online forums (e.g., Reddit and regional discussion boards). Given the scale of the issue, researchers in Natural Language Processing (NLP) have developed various approaches, ranging from manual fact-checking to rhetorical and source credibility analysis [
2].
Singh et al. [
3] emphasized that political agendas, social polarization, and economic incentives are key drivers behind the spread of fake news. Moreover, the low barrier to publication and the viral nature of digital content facilitate the rapid diffusion of disinformation in communities with limited media literacy and critical thinking.
One of the most persistent challenges remains the lack of high-quality annotated corpora, especially for low-resource languages such as Kazakh. Existing datasets typically rely on binary classification of content as fake or real, which fails to account for manipulation techniques, authorial intent, or target audience [
4].
This study aims to develop and implement a multi-level annotation scheme for detecting fake news in Kazakh and Russian. The research objectives include analyzing existing annotation approaches for disinformation; identifying methodological limitations in multilingual contexts; designing a custom annotation framework implemented in Label Studio [
5]; building a pilot corpus; and evaluating annotation consistency and model performance.
The object of the study is fake news discourse in Kazakh and Russian, while the subject is the annotation features of disinformation, its structure and mechanisms of influence.
The discourse of fake news represents a distinct type of media discourse characterized by the creation, dissemination, and consumption of messages containing deliberately false, distorted, or manipulative information aimed at shaping or altering public opinion, emotional states, or audience behavior. This discourse is marked not only by factual inaccuracy (such as false claims or fabricated events) but also by the use of specific linguistic, stylistic, and pragmatic strategies, such as emotional manipulation, clickbait tactics, and contextual distortion, designed to enhance persuasive impact and reduce the recipient’s critical perception. In multilingual environments, such as the Kazakh–Russian context, the discourse of fake news is further complicated by sociocultural factors and differences in linguistic registers. These complexities necessitate a comprehensive linguistic and semantic analysis to effectively capture the nuances of such content.
The study is based on the hypothesis that an adapted, multi-level annotation model provides higher interpretability and detection accuracy than monolingual, binary schemes developed for English. The hypothesis that an adapted multi-level annotation scheme enhances the interpretability and accuracy of fake news detection in the Kazakh–Russian media space is grounded in several empirical and theoretical premises. First, findings from existing studies indicate that binary annotation schemes (i.e., fake/real) fail to capture the complex pragmatic and manipulative dimensions of fake messages, particularly in multilingual and culturally diverse contexts. Second, a preliminary content analysis of Kazakh and Russian-language news texts revealed the presence of stable local disinformation patterns that are not accounted for in existing English-language models. Third, drawing on the principles of cognitive pragmatics, discourse theory, and media manipulation studies, it is posited that a multi-level annotation framework, one that incorporates authorial intent, target audiences, and manipulation techniques, can provide a more structured representation of fake news content and improve the performance of downstream automated detection systems.
The relevance of this research lies in the urgent need for linguistically and culturally sensitive annotation resources that can support effective disinformation detection in multilingual digital environments. The proposed scheme contributes to both applied and theoretical dimensions of media annotation and fake news mitigation.
Literature Review and Conceptual Background
In recent years, a growing body of research has emerged focusing on the detection of fake information in multilingual settings, including approaches based on transformer models. For instance, Ragab et al. [
6] demonstrated the effectiveness of multilingual models such as XLM-RoBERTa and mT5 in detecting propaganda and disinformation across Arabic, English, and Hebrew. Their findings show that multilingual models outperform monolingual counterparts due to their ability to capture cross-lingual semantic patterns and facilitate knowledge transfer between languages. Pretrained multilingual transformer models, including bert-base-multilingual-cased and xlm-roberta-base, have also shown high performance across a range of natural language processing tasks and are particularly promising for the automatic detection of disinformation in low-resource language settings. These models are trained on large-scale multilingual corpora and are capable of identifying universal markers of disinformation, such as manipulative constructions, emotional tone, and rhetorical strategies that persist across languages. In the context of low-resource languages like Kazakh, multilingual models can effectively leverage knowledge acquired from high-resource languages (e.g., English) and apply it to recognize similar phenomena in different linguistic environments. This approach offers flexibility, resilience to data scarcity, and scalability, allowing a single model to be used across multiple languages without the need to develop separate solutions for each language group. Wang et al. [
7] emphasized the challenges of annotation and automated disinformation detection in low-resource language contexts, highlighting the importance of manual annotation for such tasks. Similarly, the EXClaim model proposed by Dutta et al. [
8] integrates entity knowledge and cross-lingual transformers for claim detection—a methodology closely aligned with our annotation labels
CLAIM and
SOURCE.
Multilingual corpora such as MuMiN, Multiverse, and NewsPolygraph highlight the importance of cross-cultural analysis and multi-layered annotation. However, these datasets lack a manual typological structure like the one proposed in this study, which includes not only factuality but also disinformation techniques, authorial intent, and target audience. This is primarily due to the fact that creating multilingual datasets requires substantial resources to recruit annotators proficient in multiple languages and equipped with the linguistic and cultural competence necessary to identify typological features of disinformation specific to each language. Moreover, most existing corpora have prioritized the rapid generation of large-scale training data for transformer models over detailed manual typologization, largely due to the time and financial constraints inherent in many research projects. This makes our schema particularly relevant for the multilingual media landscape of Kazakhstan.
Manual annotation is essential because automated methods for dataset construction often depend on heuristics, distant supervision, or weak indicators such as user reactions or clickbait features. These approaches tend to produce superficial labels and fail to capture the more intricate dimensions of disinformation, including implicit authorial intent or pragmatic manipulation strategies. In multilingual settings, where linguistic and cultural nuances vary widely, automated systems frequently misinterpret local idioms, pragmatic cues, and contextual meanings, leading to reduced annotation accuracy. In contrast, expert manual annotation enables the identification of complex categories and subtypes, taking into account semantic, stylistic, and cultural factors. This is crucial for developing interpretable and reliable corpora that serve as a solid foundation for training effective disinformation detection models.
The most cited resources belong to LIAR by Wang [
9] and FakeNewsNet by Shu et al. [
10], which enriched their annotations with social metadata. In this context, social metadata refers to elements such as the number of shares and likes on a post, user comments, information about the accounts that disseminated the news, as well as temporal and geographic data related to its spread. These parameters enable the analysis of how a piece of news circulates on social media and how audiences engage with the content, providing additional context for assessing the credibility of the information. The LIAR dataset includes the text of each claim, a six-point truthfulness scale, source, date, topic, party affiliation, speaker’s position, and explanatory notes. This structure allows for nuanced analysis of the influence of source credibility, temporal factors, and political context on the spread of disinformation, and supports the development of multi-level annotation frameworks that go beyond binary to multiclass classification. FakeNewsNet complements news content with social media metadata, including the number of reposts, likes, user comments, account profiles of those sharing the news, as well as temporal and geographic information. These features enable the study of audience interaction patterns and the real-time dissemination mechanisms of fake content. Together, LIAR and FakeNewsNet serve as important reference points for the design and validation of multi-level annotation schemes and automatic fake news detection models, providing robust frameworks for both content analysis and user behavior modeling in the digital information landscape.
However, even these corpora offer annotations only at the global claim level, without addressing the semantic and structural components of the text.
Faustini et al. [
11] advocate for deeper linguistic analysis and show the effectiveness of fake news detection systems when built upon well-annotated corpora. Nonetheless, these resources are limited to English and do not account for regional or cultural differences in how disinformation is perceived.
More recently, corpora in Spanish by Silva et al. [
12], Portuguese by Khalil et al. [
13], Arabic by Assaf & Saheb [
14], and other languages have been introduced. Gruppi et al. [
15] emphasize the importance of interpretability in fake news classification, especially when considering stylistic and cultural dimensions. However, none of these resources offer a multi-level annotation structure like the one proposed here, which captures disinformation type, manipulation technique, authorial motivation, and audience targeting. A particularly important contribution of our work is the inclusion of regional features: annotation results show that a substantial portion of disinformation in Kazakh and Russian targets vulnerable groups—particularly elderly individuals and rural populations.
Since our annotation scheme draws on journalistic models such as the Inverted Pyramid, we also consider studies that apply similar methods. Norambuena et al. [
16] developed an evaluation method based on event descriptors (5W1H) for analyzing journalistic structure, while Chakma & Das [
17] annotated a corpus using semantic roles and structural cues for event extraction. These works support the relevance of journalistic frameworks for annotation design.
Another important dimension involves the linguistic characteristics of text as indicators of credibility. Studies by Zhang et al. [
18], Santia et al. [
19], Alam et al. [
20] and Mottola [
21] demonstrated that stylistic, syntactic, and psychological features can aid fake news classification. These findings inform the inclusion of linguistic cues in our annotation framework.
Our multi-level annotation scheme uniquely integrates detailed semantic relations, disinformation techniques, author intent, and target audience, specifically adapted for Kazakh and Russian. In contrast, existing schemes are mostly monolingual, limited in granularity, and lack structured typologies and semantic interconnections. This enables more nuanced analysis of disinformation mechanisms in multilingual and culturally specific contexts. This is further illustrated in
Table 1.
The literature review highlights a clear research gap by identifying the following key limitations in current fake news detection approaches:
- 1.
A predominant focus on English-language content and political topics;
- 2.
Lack of annotation for manipulative techniques, authorial intent, and target audience;
- 3.
Insufficient adaptation to multilingual information environments, particularly for the Kazakh language;
- 4.
Simplistic annotation schemes that do not allow for structural or impact-based analysis of fake news.
The present study’s research novelty lies in the development of a multi-level annotation framework that integrates concepts of factual accuracy, information structure, manipulation strategies, and linguistic features. The framework has been developed for texts in Kazakh and Russian, and addresses the multilingual and sociocultural characteristics of the Central Asian media space.
Our system is distinguished, first and foremost, by its multi-level structure and deep semantic design. Unlike traditional annotation schemes limited to binary labels (fake/real), we propose an expanded typology of disinformation patterns that includes manipulative techniques, authorial intent, and target audiences. The second key advantage lies in the introduction of formalized semantic relationships between entities, enabling the modeling of the internal structure of fake news messages and their pragmatic orientation, i.e., an aspect largely absent in most existing models. The third significant distinction is local adaptation: our scheme accounts for the linguistic characteristics of both Kazakh and Russian, as well as the sociocultural contexts of Kazakhstan. This makes it highly relevant and applicable for detecting disinformation in regional media environments. Taken together, these features not only enhance interpretability but also provide a robust foundation for developing more accurate and resilient models of automatic fake news detection in multilingual settings.
The diagram shows how the annotator iterates over the input text, assigns labels and subtypes to tokens or text spans, establishes semantic relations, and interacts with both the annotation tool and moderator to ensure quality and consistency. This structured, stepwise workflow ensures reproducibility and high-quality data for downstream NLP tasks (see
Figure 1).
The proposed scheme addresses significant gaps in existing datasets and provides a robust foundation for developing more accurate and interpretable manual detection systems for fake news in multilingual environments. It is important to note that this approach encompasses not only the content itself but also the sociocultural context, thereby ensuring a greater degree of responsiveness to the unique dynamics of disinformation in under-resourced language communities.
For the present study, Label Studio was selected as the primary annotation tool, i.e., a modern platform for manual text data labeling, known for its flexible interface and support for multi-level annotation schemes. The choice of this platform was motivated by its widespread use in academic research, its customizable annotation logic, and its ease of integration into machine learning pipelines.
2. Materials and Methods
The use of Label Studio enabled the implementation of the proposed annotation framework with nested label levels and subtypes, capturing structural, semantic, and pragmatic features of disinformation. Over 5000 texts were manually annotated during the study, providing a practical validation of the framework’s applicability to real-world data.
To enrich the corpus, major social media platforms such as Instagram, Facebook, and Twitter were used as primary data sources, given their central role in the dissemination of news content and disinformation within the Kazakhstani and Russian-speaking media landscapes. These platforms provide access to a wide range of genres, including news articles, user-generated posts, and comments, enabling the identification of how fake news circulates among different demographic groups. The inclusion of social media data allowed the corpus to capture stylistic features characteristic of these platforms (e.g., emotionally charged headlines, hashtags, and visually oriented content) as well as patterns of audience interaction with disinformation. This, in turn, enhanced the practical value of the corpus for training models aimed at analyzing content in social media environments.
The annotation pipeline is visualized in
Figure 2, illustrating the core stages of processing news data—from initial text collection to integration with machine learning models. This modular and reproducible approach ensures transparency at each stage, including manual quality control and export of annotations for further model fine-tuning. The diagram demonstrates how manual annotation can be seamlessly embedded into a full NLP workflow, ranging from data collection to optional ML training.
As previously discussed, the process of manually identifying fake news in a multilingual media environment is more complex than a simple binary classification. In the context of linguistic and cultural diversity, it is imperative to annotate texts based on the structure of disinformation, its techniques, authorial intent, and target audience. This transformation of the classical detection task into a problem of structured extraction of semantic and pragmatic features associated with disinformation impact is a significant development in the field.
In this study, a multi-level annotation scheme is introduced, comprising nine primary categories and twenty-three subtypes. Each label corresponds to a semantic or pragmatic component of a disinformation text, ranging from simple elements (e.g., the presence of a source or a claim) to more complex attributes (e.g., disinformation technique or author’s intent).
The UML sequence diagram illustrates the step-by-step token annotation process, where each token is selected, processed, and assigned a corresponding label within the multi-level annotation scheme (see
Figure 3).
The process begins with the annotator (either a human expert or an automated model) iterating through the input text sequence in Equation (1):
where each token xᵢ represents a word, punctuation mark, or other textual element.
The practical significance of introducing Equation (1) lies in the representation of text as a sequence of tokens, which enables annotation at the word and phrase level, i.e., an essential requirement for implementing a multi-level scheme with precise identification of entities and subtypes in Label Studio. This approach ensures compatibility with data formats used in modern NLP frameworks (e.g., SpaCy v3.7.2 and HuggingFace v4.39.3) and facilitates downstream model training for automatic classification, as each annotation is linked to a specific token or token sequence within the text.
The set of labels Y is defined as the union of all categories and subtypes in the annotation scheme, as illustrated in Equation (2):
where each category may be further specified by a subtype (e.g., DISINFORMATION_TECHNIQUE: clickbait or AUTHOR_INTENT: spread_fear).
Formally, the task is to construct a function:
as demonstrated in Equation (3), the process of mapping involves the allocation of each token (or text fragment) to its corresponding label from the annotation scheme.
In practical terms, the task reduces to constructing a classifier that, based on a training corpus of annotated texts, predicts both categorical labels and subtypes for previously unseen texts. Additionally, annotation consistency is evaluated typically using Cohen’s Kappa coefficient to ensure the reproducibility of the annotation scheme and the validity of the assigned labels.
To address the task of manual fake news identification using the proposed multi-level annotation framework, we propose a method grounded in modern approaches to Natural Language Processing (NLP) and Machine Learning (ML). Let the training dataset be defined as follows, in Equation (4):
where X is a sequence of tokens and Y is the corresponding set of labels from the annotation scheme. The objective is to construct a model f: X → Y that approximates the function φ and is capable of accurately predicting labels for new, unseen texts.
The news corpus was intentionally compiled to create a representative sample of Kazakh- and Russian-language publications, covering a wide range of sources, genres, and styles. It includes materials from official online media outlets, regional publications, Telegram channels, social media platforms, and personal blogs. Both reliable and fake news items were selected, as well as hybrid texts containing a mixture of truthful and misleading content.
The balanced distribution of the corpus across languages, comprising 50% Russian-language and 50% Kazakh-language texts, was a deliberate choice aimed at ensuring linguistic balance within the model and preventing potential bias toward one language. This consideration is particularly important in the context of Kazakhstan, a constitutionally bilingual country. Such an approach enabled comparable processing quality across both languages, ensured the validity of cross-linguistic analysis, and provided equal opportunities for subsequent model training and evaluation. Moreover, during the corpus construction phase, we encountered a limited amount of annotated data for each language group, which necessitated a symmetrical distribution to achieve stable and reproducible experimental results (see
https://github.com/baiangali/fake_news (accessed on 19 August 2025)).
Annotation was carried out using the Label Studio platform. Annotators were provided with detailed annotation guidelines that outlined clear criteria for identifying entities, claim types, sources, disinformation techniques, and other components. In cases of disagreement, final decisions were made by a moderator to ensure consistency.
The annotation quality was further evaluated using Cohen’s Kappa, a widely accepted metric for measuring inter-annotator agreement in sequential and categorical labeling tasks.
The proposed annotation scheme is grounded in the principles of discourse analysis and content analysis, incorporating both core categories and a hierarchical subtype structure, enabling a deeper semantic interpretation of text. The primary annotation labels and their subtypes are presented below (see
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6).
To assess the reliability of the annotation scheme, inter-annotator agreement was calculated. This step confirmed the reproducibility of results and the validity of the label structure. Despite the successful implementation of the scheme, several limitations were identified, including the complexity of interpreting subjective labels (e.g., AUTHOR_INTENT) when context is limited; the relatively small volume of Kazakh-language content, due to the limited accessibility of sources; and the high cognitive load on annotators when working with a multi-level structure without automation support.
The absence of model testing on the annotated corpus at this stage. The limited volume of Kazakh-language content affects annotation outcomes by reducing the representation of certain thematic and genre-specific patterns unique to the Kazakh-language segment of the media landscape. This limitation may lead to imbalances in the distribution of categories and subtypes, as well as to lower recall in detecting specific disinformation techniques characteristic of local sources. From the perspective of generalizability, this constraint implies that the results obtained from the current dataset should be interpreted with caution when extrapolating to the broader spectrum of Kazakh-language content. Consequently, there is a clear need for further corpus expansion, including the integration of sources from regional and local media outlets.
Nevertheless, the results demonstrate the high applicability of the scheme and its potential for scaling within broader NLP tasks. The annotated data can serve as a foundation for training automated disinformation detection models, including in multilingual environments.
In developing the custom annotation scheme for fake news in Kazakh and Russian, we analyzed a range of authoritative academic sources and existing datasets containing established categories and labels. The resulting multi-level structure is grounded in both theoretical and empirical foundations, as proposed in prior research.
The binary labels FAKE_NEWS and REAL_NEWS were adapted from the LIAR dataset by Wang [
9], in which political statements are annotated on a credibility scale (true, half-true, pants-on-fire, etc.).
The CLAIM label (main assertion) is widely used in corpora, such as NELA-GT, and is suggested by Gravanis et al. [
26] and is formally included in the ClaimReview specification. It identifies the central statement subject to verification.
The SOURCE label was adopted from FakeNewsNet by Li et al. [
27] and NELA-GT, where sources are annotated with attributes like credibility, reputation, and political orientation [
28].
The EVIDENCE label, indicating support for or refutation of the claim, is used in the MuMiN corpus by Nielsen et al. [
22] and in COVID-related datasets such as CoAID and ReCOVery [
23].
DISINFORMATION_TECHNIQUE captures methods of dissemination (e.g., clickbait, context distortion, emotional pressure) and was developed based on contemporary research on disinformation tactics, particularly in the study by Broniatowski et al. [
29] on information disorder.
AUTHOR_INTENT, which reflects the author’s presumed motivation (e.g., destabilization, satire, and propaganda), draws inspiration from recent studies on strategic information operations and coordinated disinformation campaigns, such as the work by Starbird et al. [
30], who analyzed intent through linguistic pattern analysis.
TARGET_AUDIENCE was derived from Bovet & Makse [
31], whose work examined disinformation’s impact on different social and political groups. This label was included to capture audience targeting, which is especially relevant in the Kazakhstani media landscape.
The TIMESTAMP label was included based on research in temporal annotation for event corpora and fact-checking. The idea of linking claims to specific timeframes comes from ClaimReview and Jiang et al. [
32], which focuses on methods for claim detection and verification based on supporting evidence. Additionally, temporal normalization practices were adopted from the TimeML annotation framework by Jha et al. [
33], widely used in corpora like TimeBank, EventTimeBank, and MEANTIME [
24] for temporal event extraction in NLP.
All labels used in our annotation scheme are based on validated methodologies and have been adapted into a multi-level framework for analyzing fake news, taking into account linguistic and cultural specificities.
Figure 4 illustrates a manual annotation of a Russian-language text in Label Studio (see
Figure 4). The text highlights key components such as TIME (temporal context), CLAIM (main assertion), SOURCE (information source), and EVIDENCE (supporting or refuting evidence). The annotation is further enriched with the type of disinformation (fabricated_news), the dissemination technique (clickbait), the author’s intent (influence_opinion), and the target audience (general_public).
Figure 5 presents a similar annotation of a text in Kazakh. Key components such as TIME, CLAIM, and SOURCE are also annotated, and the message is classified as misleading_content with a political motivation (political_gain) targeting youth as the primary audience (see
Figure 5).
This study employs a comprehensive methodological framework that combines both theoretical analysis and practical implementation. The primary approach used is the comparative-analytical method, which enabled a systematic examination of existing annotation schemes such as LIAR, FakeNewsNet, MuMiN, NewsPolygraph, and others. This approach allowed us to identify their strengths and weaknesses, particularly in the context of multilingual data processing.
An important tool was content analysis, which was used to study the structure of fake news in Kazakh and Russian and to identify common linguistic and thematic features of disinformation.
The data classification method was applied in constructing the multi-level annotation scheme, allowing for information to be organized by layers: from basic categorization (e.g., REAL_NEWS) to detailed subtypes describing manipulative techniques, authorial intent, and target audience.
In addition, the study employed a practical annotation method using the Label Studio platform, which enabled the implementation of the scheme as a working annotation tool and supported the creation of an annotated text corpus.
The comparative analysis method was also used to benchmark the proposed scheme against existing solutions, demonstrating its advantages in terms of granularity, flexibility, and localization for Kazakh and Russian languages.
The main outcomes of this study include the development and implementation of a multi-level annotation scheme that captures not only the presence of fake content in a text, but also reveals its underlying nature, dissemination strategies, and targeting mechanisms.
Analysis of international and local datasets revealed that most existing annotation schemes are primarily English-centric and offer a limited set of labels, which makes them less suitable for comprehensive disinformation analysis in Kazakh–Russian bilingual contexts.
To assess the consistency and reproducibility of the annotation framework, we compared the percentage distribution of annotated entities between two independent experts and calculated Cohen’s Kappa coefficients for each major category.
Table 7 presents the results. High Kappa values for SOURCE (0.81) and CLAIM (0.78) confirm the formal definability and stability of these labels in news texts, making them reliable anchors for automating annotation processes. Labels with lower agreement values, such as AUTHOR_INTENT and FAKE_NEWS, reflect a higher interpretative burden, often requiring additional contextual and pragmatic analysis.
Figure 6 illustrates the inter-annotator agreement across core annotation categories using Cohen’s Kappa scores (see
Figure 6). The highest levels of agreement are observed for SOURCE and CLAIM, indicating their clarity and consistent interpretability. In contrast, categories such as FAKE_NEWS and AUTHOR_INTENT show moderate agreement, reflecting the increased subjectivity and contextual dependence involved in labeling them.
A progressive improvement in accuracy, recall, and F1-score is expected at each stage (see
Figure 7). These projections are based on preliminary baseline experiments using logistic regression on the top-level labels (REAL_NEWS, FAKE_NEWS) within the constructed corpus, where accuracy and F1 scores ranged between 72% and 75%. They are further supported by findings in the literature, where fine-tuning BERT and XLM-R models on similar tasks has led to performance improvements of 8–15% compared to simpler models [
6,
8,
9]. Additionally, analysis of transformer-based approaches in multilingual fake news detection shows that incorporating semantic relationship layers and transitioning from binary to multi-level classification enhances recall and F1 scores by enabling better differentiation of disinformation patterns. Therefore, the projected improvement presented in
Figure 5 is grounded in a synthesis of preliminary experimental outcomes, empirical findings from the literature, and the structural advantages of the proposed multi-level annotation scheme.
In addition to category and subtype levels, the proposed annotation scheme also incorporates formalized semantic relations between entities. The introduction of semantic relationships between annotated entities (e.g., CLAIM → EVIDENCE, CLAIM → DISINFORMATION_TECHNIQUE) is aimed not only at structuring the text but also at improving the quality of downstream fake news detection by enhancing contextual coherence and formalizing latent patterns. Traditional approaches often rely on isolated entities or features, without accounting for their interdependencies. This results in the loss of critical information, particularly in cases involving complex or covert forms of disinformation. Semantic links enable models to capture contextual dependencies between key components of a message. For instance, a CLAIM supported by weak or missing EVIDENCE and accompanied by a manipulative DISINFORMATION_TECHNIQUE is highly likely to be classified as fake. Such connections add an additional layer of formalized context, increasing the interpretability of the data and providing more precise rules for model training. As a result, semantic relationships contribute to greater detection accuracy by incorporating not only the content of the message but also the structural and relational dynamics between its elements (see
Figure 8).
These relations enable linking of claims to sources, evidence, timestamps, dissemination techniques, and authorial intent.
Table 8 below presents the complete list of such semantic relationships, which are essential for constructing logical-semantic graphs of disinformation (see
Table 8). These relations include both standard links, such as a claim supported by evidence and more specific ones that capture authorial motivation and manipulative strategies.
The semantic relations reflect the structural connections between entities within the annotation framework. They form the foundation for constructing logical-semantic graphs and enable the automated analysis of disinformation narratives.
An example of an annotated sentence in Kazakh, with full coverage of all entities and their interrelations, is presented in
Figure 9.
4. Conclusions
The scientific contribution of this study lies in the development of the first multi-level annotation scheme for disinformation detection in the Kazakh–Russian linguistic context, the creation of a corresponding annotated corpus, and the formalization of semantic relationships between key entities. From an applied perspective, the results of this work can be used to support the automation of fake news detection tasks and the development of intelligent systems for monitoring digital media. Future work will include the training and evaluation of automated models, the development of semi-automated annotation tools, and the integration of the proposed scheme into real-world analytical platforms. The proposed scheme enables the structured representation of key components of fake news, including claims (CLAIM), sources (SOURCE), evidence (EVIDENCE), disinformation techniques (DISINFORMATION_TECHNIQUE), authorial intent (AUTHOR_INTENT), target audience (TARGET_AUDIENCE), and temporal attributes (TIMESTAMP).
One of the key findings is the effectiveness of the multi-level approach compared to traditional binary classification. High inter-annotator agreement scores (Cohen’s Kappa up to 0.81) demonstrate the clarity of the guidelines and the reproducibility of the scheme. An additional advantage lies in the ability to formalize logical-semantic relationships between entities, paving the way for constructing graph-based representations of disinformation narratives.
The annotation results reveal consistent disinformation patterns, primarily targeting vulnerable audiences. Techniques used range from clickbait and exaggeration to fear-based manipulation and politically motivated content. Nevertheless, some limitations were identified, such as the interpretive complexity of subjective labels, limited availability of Kazakh-language content, and the absence of automated baseline models at the time of the study.
The observed differences in inter-annotator agreement across categories highlight the need for a differentiated approach to automation. Categories with high formal definability (e.g., CLAIM, SOURCE) can be effectively automated using standard sequential classification models. In contrast, more subjective categories (e.g., AUTHOR_INTENT, FAKE_NEWS) require the incorporation of contextual and discourse-level features, as well as additional annotation guidelines to enhance reproducibility and annotation consistency.
Future research should focus on the following areas:
Training and testing multilingual models using the annotated corpus;
Expanding the dataset with additional topics and sources;
Developing semi-automated annotation tools;
Investigating the impact of disinformation on different demographic groups.
Taken together, these findings confirm that the multi-level annotation scheme holds strong potential for both academic and applied domains. It can serve as a foundation for building more accurate and interpretable models for automated fake news detection, particularly in linguistically and culturally diverse media ecosystems.
Further research will focus on training and evaluating automated fake news detection models based on the proposed annotation scheme, as well as integrating these models into applied systems for monitoring and analyzing digital media.
In the longer term, prospective studies will explore the potential of using Large Language Models (LLMs) to automate or semi-automate the annotation process, including the preliminary generation of labels and the extraction of semantic relationships between entities. The application of such models could significantly enhance the speed and scalability of annotation. However, it also necessitates rigorous quality control and systematic validation, given the potential for automated errors, especially in socially sensitive tasks such as disinformation detection.