Next Article in Journal
Biofabricating Three-Dimensional Bacterial Cellulose Composites Using Waste-Derived Scaffolds
Previous Article in Journal
The Influence of the Spillover Punishment Mechanism Under P-MA Theory on the Balance of Perceived Value in the Intelligent Construction of Coal Mines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Graph-Based Maximum Independent Sets with Large Language Models for Extractive Text Summarization

Department of Computer Engineering, İnönü University, Malatya 44000, Turkey
Appl. Sci. 2025, 15(12), 6395; https://doi.org/10.3390/app15126395
Submission received: 2 May 2025 / Revised: 4 June 2025 / Accepted: 5 June 2025 / Published: 6 June 2025

Abstract

Large Language Models (LLMs) have shown a strong performance across various tasks but still face challenges in automatic text summarization. While they are effective in capturing semantic patterns from large corpora, they typically lack mechanisms for encoding structural relationships between sentences or paragraphs. Their high hardware requirements and limited analysis as to processing efficiency further constrain their applicability. This paper proposes a framework employing the Graph Independent Set approach to extract the essence of textual graphs and address the limitations of LLMs. The framework encapsulates nodes and relations into structural graphs generated through Natural Language Processing (NLP) techniques based on the Maximum Independent Set (MIS) theory. The incorporation of graph-derived structural features enables more semantically cohesive and accurate summarization outcomes. Experiments on the Document Understanding Conference (DUC) and Cable News Network (CNN)/DailyMail datasets are conducted with different summary lengths to evaluate the performance of the framework. The proposed method provides up to a 41.05% (Recall-Oriented Understudy for Gisting Evaluation, ROUGE-2 F1) increase in summary quality and a 60.71% improvement in response times on models such as XLNet, Pegasus, and DistilBERT. The proposed framework enables more informative and concise summaries by embedding structural relationships into LLM-driven semantic representations, while reducing computational costs. In this study, we explore whether integrating MIS-based graph filtering with LLMs significantly enhances both the accuracy and efficiency of extractive text summarization.

1. Introduction

LLMs have significantly advanced NLP, particularly in tasks such as text comprehension, summarization, and generation [1]. Recent models, including GPT [2] by OpenAI, BERT, RoBERTa, XLNet, ELECTRA, ALBERT [3], Pegasus [4] by Google, and DistilBERT [5] by Hugging Face, have demonstrated a strong performance across a range of NLP applications.
Despite their success, LLMs face difficulties in capturing the structural aspects of language, particularly in tasks such as the generation of abstracts. This limitation often results in outputs that are either repetitive or lack contextual coherence. Their strong dependence on training data characteristics further limits their generalizability [6], making them sensitive to variations in data scope and training strategies [7,8]. Although they are capable of processing large volumes of data conceptually, their inability to model internal structures between text components can lead to inconsistencies, especially when summarizing long and complex texts [9,10]. Their high computational and memory requirements restrict real-time deployment and reduce overall efficiency [11]. These challenges highlight the necessity of integrating structural information into LLMs to improve their summarization performance and produce more contextually coherent summaries.
This paper proposes a novel framework based on the MIS theory to address several limitations of LLMs. These include hardware-related issues such as high computational costs, a limited capacity to process lengthy documents, and significant energy consumption. The framework also targets content-related challenges, including repetitive outputs, the omission of key information, and the lack of structural awareness due to a dependence on training strategies. By modeling textual data as simplified, independent graphs, the proposed method extracts structural features using the MIS algorithm as its core component. Integrating these features with semantic content allows the generation of more coherent and informative summaries. The framework also facilitates a better representation of cause–effect relationships, contributing to an improved summary quality and reliability of output.
The framework begins with preprocessing to prepare documents for structural analysis. It then constructs graph representations that capture relationships between sentences based on shared phrases and semantic similarity. Using the MIS algorithm, the framework identifies independent structures within these graphs to reduce redundancy and simplify the input. This allows LLMs to work with a more structured and concise dataset, improving both the summarization speed and efficiency. Rather than processing entire documents, the model focuses on key sentence connections, which enhances energy efficiency and reduces computational demands. By incorporating structural information, the framework also helps to optimize the input data and mitigates misleading guidance during inference. Experiments conducted on the CNN/DailyMail and DUC-2002 datasets demonstrate that the framework consistently produces accurate and coherent summaries. The approach improves summarization performance by reducing redundancy, minimizing information loss, and lowering the reliance on specific training data and strategies.

Motivation and Objective

LLMs have demonstrated a strong performance not only in NLP but also in fields such as image processing and time-series analysis. However, their large size and high computational and training costs limit their deployment on resource-constrained devices. Another important limitation is their reliance on contextual information while overlooking the structural aspects of language. Although existing studies process large-scale content at the conceptual level, they often fail to effectively capture the internal structure of text. As a result, these models struggle to maintain contextual coherence, particularly when summarizing complex documents. This highlights the need for integrating structural information to improve the summarization capabilities of LLMs.
In this study, a new framework based on the Graph Independent Set is proposed that enables LLMs to be used in inferential text summarization tasks with a high performance and minimum cost. In this study, the MIS algorithm was chosen because it provides a structural filtering mechanism capable of systematically filtering out contextually irrelevant sentences before summarization. While node centrality-based methods are effective in determining the importance of nodes, they fall short in filtering out nodes with a low content contribution. In contrast, MIS prevents redundancy and semantic overlap, resulting in more compact, focused, and diverse summaries. This approach strengthens the semantic integrity by identifying mutually independent nodes, in contrast to centrality-focused methods that may cause repetition by prioritizing highly connected nodes. Our aim within the scope of this study is to eliminate the limitations of LLMs, including their high computational cost, limited scalability, redundancy in outputs, and their inability to utilize structural relationships within text. Collectively, these research questions aim to determine whether integrating MIS-based graph filtering with LLMs can significantly enhance both their summarization accuracy and computational efficiency. Our Graph Independent Set-based study on the effective and efficient use of LLMs seeks answers to the following fundamental research questions;
  • Does the proposed MIS-based framework improve the summarization accuracy compared to existing popular LLMs?
  • Can the framework reduce computational costs and increase efficiency in extractive summarization tasks?
  • Does integrating structured information into LLMs enable a more precise modeling of cause–effect relationships within texts?
  • Can the use of structured information storage mitigate the context loss inherent in transformer architectures?
  • Does structured and canonical data storage reduce the energy consumption of LLMs by optimizing data loads, and does this integration strengthen their performance?
  • Can the proposed method achieve a high summarization accuracy at a low cost across different LLMs and datasets?
The remainder of this paper is structured as follows: Section 2 reviews previous studies related to the topic. Section 3 describes the materials and methods used in the study. Section 4 presents the experimental setup and the obtained results. Section 5 discusses the findings. Finally, Section 6 summarizes the paper by highlighting its contributions and offers suggestions for future research.

2. Related Works

This section reviews academic studies that are directly or indirectly related to the proposed framework. The analysis aims to identify prevailing trends in the literature and to assess potential research gaps from a comprehensive perspective. Understanding these foundations is essential for positioning the current study within the broader scientific context. For clarity and thematic coherence, the literature is categorized into four main groups: (1) limitations in integrating graph structures with LLMs, (2) graph-based approaches to text summarization, (3) hybrid architectures combining LLMs with Graph Neural Networks (GNNs), and (4) graph-based models developed specifically for scientific text summarization.
One major limitation of LLMs in handling graph-based data is their inability to process graph structures directly. Nevertheless, integrating graph learning models with LLMs has shown promise for enhancing their performance. For instance, GraphLLM [12] incorporates structural information from graphs into the input of LLMs, achieving a 96.45% reduction in context length. However, its effectiveness is highly dependent on the choice of graph model and the capacity of the LLM, necessitating adaptation for different graph types. In [13], the authors explored node classification using graphs with textual node features. LLMs were used to enrich node representations, which were then processed by a GNN. While this approach leveraged LLMs for improved feature encoding, it also introduced a significant computational overhead. Moreover, over-reliance on the graph structure led to the omission of relational information between nodes. In another study [14], various graph-to-text transformation methods were evaluated to determine their impact on extractive summarization performance. With appropriate representations, performance improvements of up to 61.8% were achieved. However, the results did not generalize well to large-scale graphs, indicating limitations in scalability.
In [15], a benchmark was introduced to assess the structural and semantic understanding of LLMs of graphs. The input design and context structure were identified as key factors affecting performance. The study found that LLMs still perform worse than graph-specialized models.
Studies integrating graph-based text summarization typically utilize relationships at the document and sentence levels. GoSum [16] converts each document into a heterogeneous graph, modeling sentence-level and hierarchical contexts with GNNs before applying LLMs. The model achieved high ROUGE scores on the PubMed and arXiv datasets, but its performance depends heavily on the document structure. MultiBART-GAT (Graph Attention Network) [17] is another abstractive summarization model that combines knowledge graphs with BART. In this study, the limitations of LLMs in checking the accuracy of information were addressed. It leverages knowledge graphs and transformers for abstractive summarization. The researchers proposed a BART-based model called MultiBART-GAT. This model integrated knowledge graphs and multi-source transformers. LLMs were effective in controlling the information accuracy. The model’s knowledge graph creation and multi-source data processing are computationally expensive and complex. The HeterMDS model [18] constructs a heterogeneous graph structure consisting of document, paragraph, and sentence nodes. It employs a Graph Attention Network (GAT) to learn, while maintaining structural and content integrity, aiming to produce high-quality summaries.
The combination of GNNs and LLMs shows a strong potential for summarization and other NLP tasks. In [19], GNNs were used to predict protein interactions, while LLMs validated these predictions. This approach achieved a high accuracy and provided meaningful explanations. The study in [20] focused on modeling rich conversational structures, addressing the challenges caused by the unstructured nature of human interactions. Two graph types, Discourse Relationship Graph and Action Graph, were employed. The model performed well in both automatic and human evaluations. In [21], document-level information was incorporated alongside sentence-level relationships using a graph-structured representation. This method improved the summarization quality, especially for long scientific articles, and worked well across different document lengths and types. However, the model’s performance depended on dataset characteristics and required high integration costs.
Some studies apply graph-based summarization techniques to specific domains. In this study, an innovative model is proposed by integrating neural networks and graph-based methods for a multi-document summarization task [22]. This model aims to create effective and meaningful summaries by calculating the relationships between sentences. The model has been tested on the DUC-2004 dataset. Document section hierarchies are represented in the graph’s edges. Offline reinforcement learning is employed, and evaluations on PubMed and arXiv datasets yield high ROUGE scores. However, since the model selects sentences independently, context integrity is not provided. The desired level of semantic inference cannot be achieved due to repetitive sentences and heavy processing loads. In [23], the ability of LLMs to make sense of graph data and optimize network performance is focused on. The proposed framework aims to improve the route optimization and communication resource allocation of unmanned aerial vehicles. With LLM-supported graph analysis, a great potential has been achieved in network optimization and automatic decision-making processes. Plaza et al. [24] used graph structures for summarizing biomedical texts. In the generated graphs, semantic relationships between concepts were utilized to summarize important content. Additionally, nodes were classified using UMLS semantic types. This foundation was later extended by Givchi [25], whose new approach considered not only the semantic similarity between concepts but also their frequency of occurrence within documents. This led to more meaningful node connections and thus, the creation of more robust graphs. In a study where GNN processes were combined with text-based guidance [26], language-based graph representations were used. The issue of out-of-vocabulary (OOV) tokens used in nodes hindering task transferability was addressed through a vocabulary approach. However, the prompt-based GNN processed each graph independently, neglecting semantic similarities across graphs.
Despite significant progress in integrating graph structures with LLMs, several key limitations remain. The most notable is the high computational cost, as LLMs require considerable processing power due to their complexity. Another major drawback is the frequent inability to effectively combine structural and semantic information. Many models focus primarily on either structural relationships or conceptual content, with few successfully integrating both aspects.
This study proposes a framework that simultaneously evaluates structural and semantic information, achieves a high performance with low computational cost, and combines the robust theoretical foundations of graph theory with LLMs. It aims to address the context loss and processing inefficiencies found in existing methods. The need for a general-purpose framework constitutes the primary motivation for this study.
Table 1 presents the main features, advantages, and limitations of the reviewed studies, including the proposed method.

3. Datasets

To evaluate the performance of the proposed summarization model, two text summarization datasets, DUC-2002 [27] and CNN/DailyMail [28], were used in the experimental process. These datasets were selected to compare the model’s performance across different document types and to quantitatively assess its generalization ability. Evaluating the model on diverse contexts is essential. Some information about the datasets is presented in Table 2.

3.1. Description of the DUC Dataset

The Document Understanding Conference-2002 (DUC-2002) dataset is used for multi-document text summarization tasks. It contains sets of news articles, along with human-prepared reference summaries for each set, totaling 567 news articles. The DUC-2002 dataset is valuable due to the complexity of information across document sets and the presence of overlapping content that requires summarization. It is well-suited for evaluating the information fusion capability of summarization models [27].

3.2. Description of the CNN/DailyMail Dataset

The CNN/DailyMail dataset consists of news articles paired with their summaries. It is a single-document summarization dataset featuring a wide range of topics. Created by Hermann et al. [29], this dataset is well-suited for testing linguistic diversity due to its broad topic coverage. It contains approximately 290,000 news articles [28,30].

4. Materials and Methods

The main stages of this work are illustrated in Figure 1. The summarization process begins with a preprocessing step applied to the documents. Following this, graphs are generated from the processed texts. In these graphs, relationships between nodes, each representing a sentence, are established based on shared phrases. The next step is the Graph Independent Set analysis, which forms the core of the proposed framework. The MIS is identified on these graphs, and the nodes comprising the independent sets are extracted. This step aims to remove sentence groups represented by the MIS from the graphs. By doing so, repetitive information in the text graphs is eliminated before the quantitative evaluation. This restriction effectively prevents the inclusion of redundant word groups in the summary and addresses the limitations of LLMs.
The revised graphs’ representative sentences are then passed to the LLMs. This creates a much smaller subset of the original text, enabling LLMs to select sentences more quickly and with better contextual clarity. As a result, LLMs process a considerably more concise dataset compared to the full documents. Sentence coverage scores from XLNet, RoBERTa, DistilBERT, ELECTRA, ALBERT, and Pegasus are used to generate summaries of varying lengths (3, 200, and 400 words). To demonstrate the framework’s reliability and robustness, experiments are conducted on two distinct datasets: CNN/DailyMail and DUC-2002. While CNN/DailyMail offers broad, up-to-date news content, DUC-2002 provides more academic and comparative documents. Including both datasets is important to reflect the framework’s comprehensiveness, despite structural differences. Significant improvements in summarization quality are observed across both datasets. Additionally, the summarization process becomes notably faster.
Figure 2 shows the general block diagram of the proposed framework. This framework combines LLMs with graph-based methods for automatic text summarization, enabling LLMs to utilize both conceptual and structural information. The input data consists of texts from the DUC and CNN/DailyMail datasets. To support generalization with quantitative evidence, these texts are preprocessed and simplified while preserving their structural and contextual integrity.
In the second stage, a graph representing sentence phrases and terms is generated from the preprocessed data. Analyses based on Graph Independent Set principles are then conducted to identify the most independent nodes. In the third stage, nodes within the MIS cluster representing sentences least related to the main text are removed, resulting in new graphs containing sentences more closely aligned with the main idea.
In the fourth stage, the most valuable nodes from these graphs are selected using the eigenvector centrality metric. This step increases the understanding of the framework.
Finally, sentences corresponding to nodes with the highest centrality values are identified and selected for further evaluation by LLMs (XLNet, RoBERTa, DistilBERT, ELECTRA, ALBERT, and Pegasus). These pre-trained LLMs perform an additional scoring process on the selected sentences to determine the most suitable candidates for the summary. The refined texts, enriched with structural information, are then input to the LLMs. In the final stage, candidate summaries of the desired lengths are generated from the selected sentences. This framework aims to improve LLMs’ summarization efficiency through sentence selection techniques grounded in Graph Independent Set theory. Overall, the proposed model achieves both a high accuracy and rich summarization quality while reducing computational costs.

4.1. Integrating Natural Language and Graph-Based Representations

In this subsection, textual graphs are constructed. Creating sentence-based graphs that represent the input texts and extracting structural information from these graphs constitute an important stage. The effectiveness of these graphs in storing structural information directly impacts the quality and efficiency of the candidate summaries. This representation approach is based on modeling each sentence as a node, thereby forming a graph structure. Prior to the graphs’ construction, the texts undergo preprocessing steps, including lemmatization. The proposed framework generates graphs following these preprocessing procedures. Formally, a textual graph is represented as G = ( V , E ) where V ( V 1 , V 2 , , V n ) denotes the set of nodes (sentences), and E E 1 , E 2 , , E n denotes the set of edges (relationships between sentences). Typically, full textual graphs are analyzed using dense matrix representations, which have a time complexity of O(N2). An example of a text-based graph is illustrated in Figure 3. This stage provides the foundational framework and representation for subsequent processing steps.

4.2. Independent Set

MIS is a fundamental concept applied across various fields, facilitating the analysis of relational structures between nodes in a graph. It is utilized to address practical problems in areas such as network planning [31], social network analysis [32], bioinformatics [33], and NLP [34]. An MIS is defined as a subset of nodes in which no two nodes are adjacent; that is, no edges exist between any pair of nodes within the set. Importantly, multiple independent sets may satisfy this condition in a given graph. The MIS problem is equivalent to the maximum clique problem and is classified as NP-hard for general graph types [35]. Consequently, identifying optimal or near-optimal solutions poses significant computational challenges. Various methods have been proposed in the literature to approximate or solve this problem [36]. Formally, given a graph G = ( V , E ) , an independent set S V ( G ) is a set of nodes with no connecting edges between them. Adding any other node to S would introduce an edge within the set, thus violating the independence property. Among all independent sets identified in G , the one with the largest cardinality is designated as the MIS. Additionally, S must not be a proper subset of any other independent set in G , ensuring the mutual exclusivity or “no conflict” principle among nodes [31,32]. The seminal work by [37] investigates the problem of determining the maximum number of independent sets possible in a graph with nnn nodes, providing solutions for various graph classes.
Within the scope of this study, the textual content targeted for summarization is represented as a graph structure, as described in the corresponding section. From this graph, independent sets are identified, and the sentences associated with the nodes in these sets are excluded from the summarization process. This exclusion is based on the assumption that such sentences contribute minimally to the core content and should not appear in the final summary. By eliminating these elements, the summarization process can prioritize more semantically relevant and informative content. The remaining sentences are evaluated using node-centric graph metrics to guide the summary generation. The full sequence of these steps is detailed and illustrated in Figure 4.
Algorithm 1 takes textual graphs as input and outputs an MIS. This recursive algorithm continues until only independent sets remain or until the graph is empty. In each iteration, the graph is simplified, and eventually all possibilities are scanned to find the MIS.
Algorithm 1. MIS Algorithm: Maximum_Independent_Set (G).
1function Maximum_Independent_Set ( G )
2 if all vertices in V have degree 0:
3   return S = V
4 if   V is empty:
5   return empty set
6 select vertex u V with minimum degree
7 if u S       then                        //(if u is not included in the set)
8 G 1 = G { u }
9 S 1 = Maximum_Independent_Set ( G 1 )
10 else   u S                                   //(if u is included in the set)
11 G 2 = G { u }   neigbors ( u )
12 S 2 = Maximum_Independent_Set ( G 2 ) { u }
13 if   S 1   > S 2 then
14   return S 1
15 else
16   return S 2
17end function
In order to perform basic condition checks, all node degrees are checked to be 0. If all node degrees are 0, it means that there are no edges in this graph, and all nodes are independent sets. If there are no nodes in the graph, the independent set is empty. (line 2–4)
The node with the minimum degree (u) is selected from the remaining nodes. Since the relevant node will have fewer neighbors, fewer nodes will be eliminated. Thus, the probability of a larger independent set tries to be increased. (line 6)
After this stage, two possibilities are evaluated. The first of these possibilities is that the relevant node (u) is excluded. In this case, (u) is removed from the graph and is not included in the independent set. The algorithm will find the largest independent set over the remaining nodes. (line 7–9)
The second possibility is that node (u) and all its neighbors are removed from the graph. In this case, u is added to the independent set. (line 10–12)
In the last stage, the larger of the clusters found is selected. (line 13–16)
Although the Maximum Independent Set (MIS) problem is theoretically classified as NP-hard, the proposed algorithm significantly narrows the search space by employing a “minimum degree first” branching strategy. Therefore, within the current experimental scope, the proposed method demonstrates practical scalability without the need for additional approximation techniques.
Figure 4 illustrates the process step by step. First, the text to be summarized is divided into sentences, with each sentence represented as a node. Similarities and relationships between these sentences are mapped as edges to form a graph, preserving the structural relationships in the text. Then, the MIS algorithm is applied to eliminate sentences with a low information value. After selection, the graph is reconstructed with the remaining nodes, and meaningful sentences for the summary are identified.

4.3. Formatting of Mathematical Components

LLMs represent a major step beyond earlier statistical language models, which formed the basis of early computational linguistics [38]. The emergence of transformers, along with extensive training corpora and pre-training techniques, significantly enhanced models’ capabilities [39]. These models often contain hundreds of billions of parameters that are optimized during training on vast corpora, enabling sophisticated language generation. As the volume of textual data continues to grow, LLMs have driven major advances across NLP tasks. They have been applied in areas such as code generation, medical text processing, information retrieval, classification, and summarization [40,41,42]. LLMs generate output by capturing contextual dependencies through multi-layer transformer architectures. User input is first tokenized, then embedded into vector form, and passed through stacked transformer blocks. The final layer produces task-specific predictions (as shown in Figure 5).
Models such as XLNet, RoBERTa, DistilBERT, ELECTRA, ALBERT, and Pegasus are basically based on the transformer architecture presented in Figure 5 [44,45,46]. XLNet employs permutation-based training over the transformer architecture, improving its ability to capture word dependencies in longer phrases [47,48]. RoBERTa provides performance gains by a longer and optimized training process based on BERT architecture but with more data [49,50]. DistilBERT is faster and uses less resources with its compact structure. It achieves this by reducing the model size while preserving performance [51,52,53,54]. ELECTRA differs from BERT by using a replaced token detection task, allowing for more efficient pre-training. With ELECTRA, a high performance can be achieved even with small models [55,56,57]. ALBERT employs factorized embeddings and cross-layer parameter sharing, improving the resource efficiency in deep models [58,59,60]. Similarly, the Pegasus model developed by Google [4] demonstrates significant performance improvements in text summarization compared to traditional models. Its pre-training strategy is particularly effective in handling long and complex input texts.
Despite their capabilities, LLMs are often criticized as black-box systems that lack access to explicit structural features, focusing primarily on conceptual content. Furthermore, their high computational demands and long processing times remain a barrier to widespread deployment. To address these limitations, this study introduces a MIS-based framework that extracts salient structural information from sentence graphs, enhancing both the efficiency and summary quality.

4.4. Performance Metrics

In this study, ROUGE is used to evaluate the summarization performance by comparing model-generated summaries with human-written references. ROUGE measures the overlap of textual units such as n-grams and word sequences to assess the quality of generated summaries. This paper considers several ROUGE variants, including ROUGE-N, ROUGE-L, ROUGE-W, and ROUGE-S. Widely adopted in NLP tasks like summarization, machine translation, and text generation, ROUGE offers a practical and comparative metric. Its primary goal is to assess how accurately and comprehensively the model captures the content of the reference text. While ROUGE only approximates semantic similarity, it provides a cost-effective alternative to manual evaluation. It reports a single performance score using Precision (P), Recall (R), and the F-measure (F). Precision measures the proportion of relevant units in the candidate summary, while Recall assesses how much of the reference is captured. The F-measure balances P and R using the β parameter. These metrics take values between 0 and 1 and are calculated as follows,
R = c o r r e c t   s e n t e n c e s   i n   t h e   c a n d i d a t e   s u m m a r y t o t a l   n u m b e r   o f   s e n t e n c e s   i n   t h e   m o d e l   s u m m a r y
P = c o r r e c t   s e n t e n c e s   i n   t h e   c a n d i d a t e   s u m m a r y t o t a l   s e n t e n c e s   i n   c a n d i d a t e   s u m m a r y
F m e a s u r e = 1 + β 2 R P R + β 2 P
ROUGE-N specifically measures n-gram overlap, where N represents the n-gram length. If the candidate summary is denoted by s, ROUGE-N is computed as
R O U G E - N = r M Φ N r , Φ N ( s ) r M Φ N r , Φ N ( r )
ΦN(d) is the binary vector representing the N-grams in document d. Here, the ith component of the vector takes the value 1 if the ith N-gram is in d, and it is 0 otherwise. Here, (,) are the inner products of the vectors. ROUGE-L calculates the longest substrings in common between the candidate and the reference summary. ROUGE-W uses a weighting factor for consecutively overlapping words. It searches for strings with consecutive matches. ROUGE-W is a variant of ROUGE-L. ROUGE-S is a skip-bigram co-occurrence. ROUGE-SU is its average [29,61].

5. Experiment Setup and Results

This section describes the experimental setup, including the environment, datasets, and results obtained from evaluating the proposed text summarization framework.

5.1. Experiment Setup

The experiments were conducted on a desktop computer with an AMD Ryzen 7 PRO 4750 U processor (1.70 GHz), Radeon Graphics, and 16 GB RAM. All analyses were performed using Python v3.10. The proposed framework was evaluated on two benchmark datasets: DUC and CNN/DailyMail. While DUC contains more formal and academic content, CNN/DailyMail includes contemporary news articles from various domains. The texts were preprocessed with care to preserve their structural and contextual integrity, then converted into graph representations for summarization.

5.2. Experimental Results

This section presents performance analyses based on the application of the proposed text summarization framework to the DUC and CNN/DailyMail.
To provide LLMs with both conceptual and structural information, the DUC and CNN/DailyMail datasets were used as input to the integrated framework. The texts were preprocessed to preserve their structural and contextual integrity and transformed into simplified, structured forms. Graphs were then constructed from sentence-level phrases and key terms extracted from the processed texts. Using graph-based MIS theory, nodes deemed less important were identified and removed. This yielded new graphs composed of structurally inclusive and semantically rich nodes. Node centrality measures were applied to select the most significant nodes, and their corresponding sentences were identified. These selected sentences were further scored by pre-trained LLMs (XLNet, RoBERTa, DistilBERT, ELECTRA, ALBERT, and Pegasus) to determine final summary candidates. This process resulted in refined summaries that integrate both structural and conceptual dimensions. By combining MIS-based graph filtering with LLM evaluation, the framework significantly improves the summarization performance in terms of both accuracy and computational efficiency. The framework’s effectiveness was evaluated using ROUGE-1, ROUGE-2, ROUGE-L, ROUGE-W, ROUGE-S, and ROUGE-SU metrics, along with sub-metrics such as Precision, Recall, and F-measure. Comparative analyses were conducted on both datasets to validate the results.
The framework achieved notable performance gains on the DUC dataset compared to the baseline LLMs used in the experiments. Its effectiveness in summarizing current events was further evaluated using the CNN/DailyMail dataset. To assess its summarization capabilities across different summary lengths, the model was tested on both short and long summaries. This evaluation aimed to measure the model’s ability to retain key information in shorter outputs and preserve a detailed context in longer ones. Table 3 presents the ROUGE scores for 400-word summaries, comparing the proposed framework with baseline LLMs.
The F1 score, which balances Precision and Recall, serves as a more reliable performance metric, especially in scenarios involving class imbalance. Therefore, model performance in this study is primarily reported using the F1 score, calculated as the harmonic mean of Precision and Recall.
Experiments conducted on the DUC dataset with 400-word summaries demonstrate that the proposed framework consistently outperforms baseline LLMs. The most significant gain was observed for ALBERT, while all models including ELECTRA showed statistically meaningful improvements. According to the ROUGE-1 metric, which measures unigram overlap, the model achieved notable F1 increases for ALBERT (+0.0568), ELECTRA (+0.0504), RoBERTa (+0.0445), DistilBERT (+0.0109), Pegasus (+0.0294), and XLNet (+0.0086). The ROUGE-2 results, which reflect bigram-level contextual alignment, indicate improved consistency in ALBERT (+0.0678), ELECTRA (+0.0462), RoBERTa (+0.0385), Pegasus (+0.0332), XLNet (+0.0233), and DistilBERT (+0.0001). With ROUGE-L, which emphasizes the preservation of long sentence structures, the proposed model again showed strong gains: ALBERT (+0.0612), ELECTRA (+0.0498), RoBERTa (+0.0331), DistilBERT (+0.0071), XLNet (+0.0094), and Pegasus (+0.0015). ROUGE-W, which evaluates weighted sequence overlap and thus indicates the semantic richness of generated summaries, showed performance gains across all models—highest in ALBERT (+0.0258) and ELECTRA (+0.0223). For phrase-level continuity, as measured by ROUGE-S and ROUGE-SU, the framework improved contextual fluency in ALBERT (+0.0537), RoBERTa (+0.0380), ELECTRA (+0.0329), Pegasus (+0.0260), XLNet (+0.0092), and DistilBERT (+0.0054). Figure 6 provides a visual representation of ROUGE-F1 score improvements across models for the DUC dataset.
Table 4 summarizes the 200-word ROUGE scores of the proposed model compared to the baseline LLMs.
Experimental results on the DUC dataset for 200-word summaries show that the proposed framework significantly outperforms previously evaluated LLMs, with the exception of ELECTRA. One of the key strengths of the model lies in its ability to generate more meaningful and contextually consistent candidate summaries. In terms of ROUGE-1, which measures unigram overlap, the framework yields notable F1 score improvements: +0.0097 for ALBERT, +0.0146 for RoBERTa, +0.0229 for DistilBERT, and +0.0359 for Pegasus. The model also demonstrates strong performance in preserving contextual semantic integrity, as reflected by ROUGE-2 scores. Improvements include +0.0383 for ALBERT, +0.0108 for RoBERTa, +0.0185 for DistilBERT, and +0.0425 for Pegasus. ROUGE-L results, which emphasize the preservation of longer sentence structures, further confirm the model’s effectiveness. The proposed approach reports F1 gains of +0.0192 for ALBERT, +0.0099 for RoBERTa, +0.0301 for DistilBERT, and +0.0372 for Pegasus.
The results also indicate an improvement in computational efficiency. According to the ROUGE-W metric, which captures weighted sequence overlap, the proposed model yields F1 score gains of +0.0067 for ALBERT, +0.0156 for XLNet, and +0.0198 for Pegasus. In terms of contextual ordering and sentence consistency, evaluated through the ROUGE-S and ROUGE-SU metrics, the model reports improvements of +0.0082 and +0.0081 for ALBERT and +0.0332 for Pegasus. Figure 7 provides a graphical summary of the ROUGE-F1 score variations for the DUC dataset with 200-word summaries.
Experiments on the CNN/DailyMail dataset with three-word summaries reveal that the proposed model achieves notable performance gains, particularly when used with XLNet, DistilBERT, and ALBERT. According to the ROUGE-1 metric, which measures word-level overlap, F1 score improvements were observed as follows: +0.0578 for XLNet, +0.1321 for DistilBERT, and +0.0361 for ALBERT. In contrast, the model showed limited or no improvement with RoBERTa and Pegasus. These results indicate that the framework performs especially well with models such as XLNet and ALBERT.
The ROUGE-2 metric measures the contextual accuracy of summaries over two-word sequences. The proposed model demonstrates notable improvements in F1 scores, with enhancements observed in XLNet (+0.0406), ALBERT (+0.0812), and DistilBERT (+0.0147), as shown in Table 5. The ROUGE-L metric, which evaluates the preservation of long sentence structures, revealed positive impacts on F1 scores for XLNet (+0.0557) and ALBERT (+0.0303). However, no improvements were observed in the other LLMs. RoBERTa was disadvantaged in selections lacking contextual awareness. However, improvements in the ROUGE-W metric were observed in the F1 scores of the proposed model, including XLNet (+0.0341) and ALBERT (+0.0196). Notably, the substantial improvement observed in the XLNet model indicates the effective integration of the proposed approach. In the ROUGE-S and ROUGE-SU metrics, which assess contextual structure and sentence sequencing, the proposed model shows gains of (+0.08261), (+0.0257) for XLNet and (+0.0087), (+0.0083) for ALBERT. Figure 8 presents a graphical analysis reflecting the variations in ROUGE F1 scores observed for the CNN/DailyMail dataset and three-word abstracts.
The results indicate that the proposed model yields notable performance improvements with XLNet and ALBERT on the CNN/DailyMail dataset. While DistilBERT also showed positive gains, slight performance drops were observed for RoBERTa and Pegasus.
Table 6 presents the processing times of various transformer-based LLM models across two datasets (DUC and CNN/DailyMail) and three summary lengths (200 words, 400 words, and 3 words). Each cell displays the model’s processing times before and after applying the proposed summarization approach, along with the percentage reduction (gain). Notably, time reductions were observed across all models. Among these, ELECTRA demonstrated the fastest processing speeds, while Pegasus exhibited the highest percentage gains. Due to its large size and computational complexity, Pegasus also had the longest initial processing times. Nevertheless, the proposed framework yielded a remarkable 63.1% reduction in processing time for three-word summaries. DistilBERT, as a compressed LLM model, initially displayed shorter processing times; however, it achieved a 57.69% time reduction for 200-word summaries with the proposed approach. These findings suggest that the framework is effective for both large-scale and compressed models. ELECTRA, recognized as one of the fastest baseline models, also demonstrated gains of up to 58.56%. The most striking improvement was observed with ALBERT, which achieved a 60.22% reduction in processing time for DUC-200 and a 51.85% reduction for short text summarization, making it the most efficient model under the proposed framework. Figure 9 illustrates the time improvements observed across the different transformer-based LLM models, datasets, and summarization lengths.
Time reductions ranged from 54% to 69% for 200-word summaries, indicating that longer texts enhance the framework’s efficiency. For 400-word summaries, the gains were between 48% and 62%. These results show that the proposed approach remains effective even as the text length increases. For short summaries, time reductions varied between 25% and 63%. Since processing times are already low for shorter texts, the relative impact of the framework is more limited. Overall, the proposed method offers substantial time savings for both large and lightweight LLMs. A t-test was conducted to evaluate the impact of the MIS framework across different summary lengths (400-word, 200-word, and 3-word) and ROUGE metrics.
As shown in Table 7, the p-values for all summary lengths and each ROUGE metric (Rouge-1, Rouge-2, Rouge-L, Rouge-W, Rouge-S, and Rouge-SU) are below 0.05. This finding demonstrates that the improvements in summarization performance are statistically significant and not due to chance. The t-statistics and p-values for each summary length and ROUGE metric further support the robustness of the MIS framework and the effectiveness of the method. Table 7 confirms that the proposed approach leads to statistically significant improvements in summary quality.

6. Discussion

This study introduces an MIS-based graph filtering approach, integrated with LLMs, to enhance summarization performance. The proposed framework shows clear improvements in accuracy, semantic coherence, and computational efficiency, especially for 200-word summaries. Significant improvements were observed when using models such as XLNet, RoBERTa, DistilBERT, ALBERT, and Pegasus. However, ELECTRA performed relatively poorly, likely due to its masking-based strategy, which may not align well with the framework’s independent node selection. This highlights the need for tailored sentence selection strategies for models like ELECTRA.
Analyses of 400-word summaries show that the framework remains effective with longer content. The results confirm that it improves LLMs’ summarization and supports strong word-level representations. The notable gains with ALBERT likely stem from its ability to maintain semantic coherence. The summaries are structurally aligned with the original texts and preserve long-range contextual links. This suggests that even smaller LLMs can benefit from the framework. Selecting more information-dense sentences also enhances computational efficiency. By removing out-of-context content, the method produces clearer and more coherent summaries. In 200-word summaries, a consistency in word overlap, order, and sentence structure indicates the successful retention of key textual features. The use of MIS theory further supports contextual preservation by filtering irrelevant information.
Evaluations of three-word summaries showed a stronger integration with models like XLNet and ALBERT, leading to a better contextual coherence and handling of long sentence structures. Sentence selection was more effective with these models. The MIS-based structure reduced computational loads and enabled shorter, yet meaningful and dense, summaries. Pegasus, however, underperformed in short summaries, possibly due to its large architecture being less suited to structural modifications in brief contexts. In contrast, its performance improved significantly in 200- and 400-word summaries, suggesting better alignment with long-text summarization tasks.
Overall, the MIS-based approach has improved sentence selection in LLMs, leading to more realistic summaries. Strong results with models like ALBERT, despite its low parameter count, highlight the framework’s versatility. The MIS-driven structural transformation reduced computational demands, allowing a faster and resource-efficient generation of content-rich summaries.
Integrating structural and conceptual information has enabled more reliable inferences and clearer cause–effect relationships within the text. The graph-based structure preserved contextual links, reducing the risk of information loss. In conclusion, the experimental results show that combining MIS-based graph algorithms with LLMs creates an effective synergy in summarization, improving both accuracy and efficiency. The proposed framework offers valuable contributions to the field, both in theory and application.
By extending the capabilities of LLMs, the framework also improves their accessibility and practical utility.

7. Conclusions

The primary goal of this study is to address structural limitations in LLM-based text summarization and improve the accuracy, coherence, and efficiency of the generated summaries. Traditional LLMs emphasize contextual representations while neglecting structural features, often resulting in repetition, loss of context, and semantic drift, especially in long or complex texts. To overcome these issues, this study proposes a method to help LLMs produce more informative and reliable summaries.
To achieve this, a framework is proposed that represents texts as graphs and uses the MIS algorithm to filter sentences for more information-dense content. The method converts texts into sentence-level graph structures and passes the selected sentences to LLMs for summarization. Experimental results show clear improvements in ROUGE metrics. Processing time and energy efficiency also improved significantly.
A potential risk of the MIS algorithm is its tendency to overlook contextual information during sentence selection. In the proposed framework, this is addressed by modeling semantic relationships between sentences using graph structures. The MIS algorithm selects central, information-dense, and mutually dissimilar sentences to preserve the summary’s semantic scope. Performance gains in ROUGE-L and ROUGE-W, both sensitive to sequence and semantic consistency, indicate that contextual integrity is maintained. These results warrant a closer examination of how the MIS-based approach differs from prior graph-based summarization methods. Its novelty lies in structurally filtering content through MIS theory rather than relying on node centrality or attention weights. Instead, it identifies non-adjacent and thus semantically independent sentence clusters. This process removes irrelevant or weakly connected content, improving coherence and consistency while reducing LLMs’ processing load. The model-agnostic design also allows its integration with a wide range of Large Language Models. The experimental results confirm the value of incorporating structural information in graph-based summarizations.
The proposed framework also has limitations. In very short summaries, MIS-based filtering may exclude important contextual expressions, narrowing the informational scope and affecting semantic integrity, especially in real-time applications such as chatbots. Although only two datasets (DUC-2002 and CNN/DailyMail) were used, CNN/DailyMail includes texts from diverse domains (politics, health, sports, and economics), offering a degree of topic generalizability. Future work will explore applications in more technical domains, such as legal and biomedical texts. To assess its consistency, fluency, and informativeness, further analyses supported by human evaluation are planned. This will enable a more robust validation of the framework from both quantitative and qualitative perspectives.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The author declares no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LLMLarge Language Model
NLPNatural Language Process
MISMaximum Independent Set
DUCDocument Understanding Conference
CNNCable News Network
ROUGERecall-Oriented Understudy for Gisting Evaluation
GATGraph Attention Network
GNNGraph Neural Networks
OOVOut-of-Vocabulary

References

  1. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  2. Floridi, L.; Chiriatti, M. GPT-3: Its Nature, Scope, Limits, and Consequences. Minds Mach. 2020, 30, 681–694. [Google Scholar] [CrossRef]
  3. Shetty, A.M.; DH., M.; Aljunid, M.F. Fine-tuning XLNet for Amazon review sentiment analysis: A comparative evaluation of transformer models. ETRI J. 2025, 1–18. [Google Scholar] [CrossRef]
  4. Zhang, J.; Zhao, Y.; Saleh, M.; Liu, P. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In Proceedings of the International Conference on Machine Learning, Vienna, Austria, 13–18 July 2020; pp. 11328–11339. [Google Scholar]
  5. Pourkeyvan, A.; Safa, R.; Sorourkhah, A. Harnessing the Power of Hugging Face Transformers for Predicting Mental Health Disorders in Social Networks. IEEE Access 2024, 12, 28025–28035. [Google Scholar] [CrossRef]
  6. Tsai, Y.-H.; Chang, C.-M.; Chen, K.-H.; Hwang, S.-Y. An Integration of TextGCN and Autoencoder into Aspect-Based Sentiment Analysis; Springer: Berlin/Heidelberg, Germany, 2022; pp. 3–16. [Google Scholar]
  7. Yang, S.; Duan, X.; Xiao, Z.; Li, Z.; Liu, Y.; Jie, Z.; Tang, D.; Du, H. Sentiment Classification of Chinese Tourism Reviews Based on ERNIE-Gram+GCN. Int. J. Environ. Res. Public Health 2022, 19, 13520. [Google Scholar] [CrossRef]
  8. Zeng, D.; Zha, E.; Kuang, J.; Shen, Y. Multi-label text classification based on semantic-sensitive graph convolutional network. Knowl. -Based Syst. 2024, 284, 111303. [Google Scholar] [CrossRef]
  9. Zuo, Y.; Wu, J.; Zhang, H.; Lin, H.; Wang, F.; Xu, K.; Xiong, H. Topic Modeling of Short Texts. In Proceedings of the Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM: New York, NY, USA, 2016; pp. 2105–2114. [Google Scholar]
  10. Zhao, K.; Huang, L.; Song, R.; Shen, Q.; Xu, H. A Sequential Graph Neural Network for Short Text Classification. Algorithms 2021, 14, 352. [Google Scholar] [CrossRef]
  11. Mouratidis, D.; Kermanidis, K.L. Ensemble and Deep Learning for Language-Independent Automatic Selection of Parallel Data. Algorithms 2019, 12, 26. [Google Scholar] [CrossRef]
  12. Chai, Z.; Zhang, T.; Wu, L.; Han, K.; Hu, X.; Huang, X.; Yang, Y. GraphLLM: Boosting Graph Reasoning Ability of Large Language Model. arXiv 2023, arXiv:2310.05845. Available online: https://arxiv.org/abs/2310.05845 (accessed on 1 March 2025).
  13. Chen, Z.; Mao, H.; Li, H.; Jin, W.; Wen, H.; Wei, X.; Wang, S.; Yin, D.; Fan, W.; Liu, H.; et al. Exploring the Potential of Large Language Models (LLMs) in Learning on Graphs. arXiv 2024, arXiv:2307.03393. Available online: https://arxiv.org/abs/2307.03393 (accessed on 1 March 2025). [CrossRef]
  14. Fatemi, B.; Halcrow, J.; Perozzi, B. Talk like a Graph: Encoding Graphs for Large Language Models. arXiv 2023, arXiv:2310.04560. Available online: https://arxiv.org/abs/2310.04560 (accessed on 1 March 2025).
  15. Guo, J.; Du, L.; Liu, H.; Zhou, M.; He, X.; Han, S. GPT4Graph: Can Large Language Models Understand Graph Structured Data ? An Empirical Evaluation and Benchmarking. arXiv 2023, arXiv:2305.15066. Available online: https://arxiv.org/abs/2305.15066 (accessed on 2 March 2025).
  16. Bian, J.; Huang, X.; Zhou, H.; Huang, T.; Zhu, S. GoSum: Extractive summarization of long documents by reinforcement learning and graph-organized discourse state. Knowl. Inf. Syst. 2024, 66, 7557–7580. [Google Scholar] [CrossRef]
  17. Chen, T.; Wang, X.; Yue, T.; Bai, X.; Le, C.X.; Wang, W. Enhancing Abstractive Summarization with Extracted Knowledge Graphs and Multi-Source Transformers. Appl. Sci. 2023, 13, 7753. [Google Scholar] [CrossRef]
  18. Zeng, G.-H.; Liu, Y.-Q.; Zhang, C.-Y.; Cai, H.-C.; Chen, C.L.P. Adaptive Multi-Document Summarization Via Graph Representation Learning. IEEE Trans. Cogn. Dev. Syst. 2024, 1–12. [Google Scholar] [CrossRef]
  19. Ivanisenko, T.V.; Demenkov, P.S.; Ivanisenko, V.A. An Accurate and Efficient Approach to Knowledge Extraction from Scientific Publications Using Structured Ontology Models, Graph Neural Networks, and Large Language Models. Int. J. Mol. Sci. 2024, 25, 11811. [Google Scholar] [CrossRef]
  20. Chen, J.; Yang, D. Structure-Aware Abstractive Conversation Summarization via Discourse and Action Graphs. arXiv 2021, arXiv:2104.08400. Available online: https://arxiv.org/abs/2104.08400 (accessed on 2 March 2025).
  21. Cui, P.; Hu, L.; Liu, Y. Enhancing Extractive Text Summarization with Topic-Aware Graph Neural Networks. arXiv 2020, arXiv:2010.06253. Available online: https://arxiv.org/abs/2010.06253 (accessed on 4 March 2025).
  22. Yasunaga, M.; Zhang, R.; Meelu, K.; Pareek, A.; Srinivasan, K.; Radev, D. Graph-based Neural Multi-Document Summarization. arXiv 2017, arXiv:1706.06681. Available online: https://arxiv.org/abs/1706.06681 (accessed on 3 March 2025).
  23. Sun, G.; Wang, Y.; Niyato, D.; Wang, J.; Wang, X.; Poor, H.V.; Letaief, K.B. Large Language Model (LLM)-enabled Graphs in Dynamic Networking. IEEE Netw. 2024, 1. [Google Scholar] [CrossRef]
  24. Plaza, L.; Díaz, A.; Gervás, P. A semantic graph-based approach to biomedical summarisation. Artif. Intell. Med. 2011, 53, 1–14. [Google Scholar] [CrossRef] [PubMed]
  25. Givchi, A.; Ramezani, R.; Baraani-Dastjerdi, A. Graph-based abstractive biomedical text summarization. J. Biomed. Inform. 2022, 132, 104099. [Google Scholar] [CrossRef] [PubMed]
  26. Zhu, X.; Xue, H.; Zhao, Z.; Xu, W.; Huang, J.; Guo, M.; Wang, Q.; Zhou, K.; Zhang, Y. LLM as GNN: Graph Vocabulary Learning for Text-Attributed Graph Foundation Models. arXiv 2025, arXiv:2503.03313. Available online: https://arxiv.org/abs/2503.03313 (accessed on 3 March 2025).
  27. Document Understanding Conferences. Available online: http://duc.nist.gov (accessed on 14 July 2021).
  28. See, A.; Liu Google Brain, P.J.; Manning, C.D. Get To The Point: Summarization with Pointer-Generator Networks. arXiv 2017, arXiv:1704.04368. [Google Scholar]
  29. Lin, C.Y. Rouge: A package for automatic evaluation of summaries. In Text Summarization Branches Out (WAS 2004); Association for Computational Linguistics: Barcelona, Spain, 2004; pp. 25–26. [Google Scholar]
  30. Hermann, K.M.; Kocisky, T.; Grefenstette, E.; Espeholt, L.; Kay, W.; Suleyman, M.; Blunsom, P. Teaching machines to read and comprehend. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 7–12 December 2015. [Google Scholar]
  31. Moscibroda, T.; Wattenhofer, R. Maximal independent sets in radio networks. In Proceedings of the Twenty-Fourth Annual ACM Symposium on Principles of Distributed Computing, Las Vegas, NV, USA, 17–20 July 2005; ACM: New York, NY, USA, 2005; pp. 148–157. [Google Scholar]
  32. McIlroy, D. The UNIX Time-Sharing System: Foreword. In Proceedings of the First ACM SIGPLAN Conference on History of Programming Languages (HOPL-I), Los Angeles, CA, USA, 1–3 June 1978; ACM: New York, NY, USA, 2005; pp. 207–212. [Google Scholar]
  33. Mrzic, A.; Meysman, P.; Bittremieux, W.; Moris, P.; Cule, B.; Goethals, B.; Laukens, K. Grasping frequent subgraph mining for bioinformatics applications. BioData Min. 2018, 11, 20. [Google Scholar] [CrossRef]
  34. Uçkan, T.; Karcı, A. Extractive multi-document text summarization based on graph independent sets. Egypt. Inform. J. 2020, 21, 145–157. [Google Scholar] [CrossRef]
  35. Öztemiz, F. A greedy approach to solve maximum independent set problem: Differential Malatya independent set algorithm. Eng. Sci. Technol. Int. J. 2025, 63, 101995. [Google Scholar] [CrossRef]
  36. YAKUT, S.; ÖZTEMİZ, F.; KARCİ, A. A New Approach Based on Centrality Value in Solving the Maximum Independent Set Problem: Malatya Centrality Algorithm. Comput. Sci. 2023, 8, 16–23. [Google Scholar] [CrossRef]
  37. Jou, M.-J.; Chang, G.J. The Number Of Maximum Independent Sets In Graphs. Taiwan J. Math. 2000, 4, 685–695. [Google Scholar] [CrossRef]
  38. Zhao, W.X.; Zhou, K.; Li, J.; Tang, T.; Wang, X.; Hou, Y.; Min, Y.; Zhang, B.; Zhang, J.; Dong, Z.; et al. A Survey of Large Language Models. arXiv 2025, arXiv:2303.18223. Available online: https://arxiv.org/abs/2303.18223 (accessed on 5 March 2025).
  39. Feng, S.Y.; Gangal, V.; Wei, J.; Chandar, S.; Vosoughi, S.; Mitamura, T.; Hovy, E. A Survey of Data Augmentation Approaches for NLP. arXiv 2021, arXiv:2105.03075. Available online: https://arxiv.org/abs/2105.03075 (accessed on 5 March 2025).
  40. Novelli, C.; Casolari, F.; Rotolo, A.; Taddeo, M.; Floridi, L. Taking AI risks seriously: A new assessment model for the AI Act. AI Soc. 2024, 39, 2493–2497. [Google Scholar] [CrossRef]
  41. Cai, Y.; Mao, S.; Wu, W.; Wang, Z.; Liang, Y.; Ge, T.; Wu, C.; You, W.; Song, T.; Xia, Y.; et al. Low-code LLM: Graphical User Interface over Large Language Models. arXiv 2024, arXiv:2304.08103. Available online: https://arxiv.org/abs/2304.08103 (accessed on 2 March 2025).
  42. Yao, Y.; Duan, J.; Xu, K.; Cai, Y.; Sun, Z.; Zhang, Y. A survey on large language model (LLM) security and privacy: The Good, The Bad, and The Ugly. High-Confid. Comput. 2024, 4, 100211. [Google Scholar] [CrossRef]
  43. Zhongyu, S.; Zhou, W.; Ding, C.; Xia, M. Multi-Resolution Transformer Network for Building and Road Segmentation of Remote Sensing Image. ISPRS Int. J. Geo-Inf. 2022, 11, 165. [Google Scholar] [CrossRef]
  44. Kalyan, K.S. A survey of GPT-3 family large language models including ChatGPT and GPT-4. Nat. Lang. Process. J. 2024, 6, 100048. [Google Scholar] [CrossRef]
  45. Liu, Z.; Roberts, R.A.; Lal-Nag, M.; Chen, X.; Huang, R.; Tong, W. AI-based language models powering drug discovery and development. Drug Discov. Today 2021, 26, 2593–2607. [Google Scholar] [CrossRef]
  46. Kour, H.; Gupta, M.K. AI Assisted Attention Mechanism for Hybrid Neural Model to Assess Online Attitudes About COVID-19. Neural Process. Lett. 2023, 55, 2265–2304. [Google Scholar] [CrossRef]
  47. Marrone, S.; Sansone, C. On the transferability of adversarial perturbation attacks against fingerprint based authentication systems. Pattern Recognit. Lett. 2021, 152, 253–259. [Google Scholar] [CrossRef]
  48. Potamias, R.A.; Siolas, G.; Stafylopatis, A.-G. A transformer-based approach to irony and sarcasm detection. Neural Comput. Appl. 2020, 32, 17309–17320. [Google Scholar] [CrossRef]
  49. Yang, T.; Li, F.; Ji, D.; Liang, X.; Xie, T.; Tian, S.; Li, B.; Liang, P. Fine-grained depression analysis based on Chinese micro-blog reviews. Inf. Process. Manag. 2021, 58, 102681. [Google Scholar] [CrossRef]
  50. Capris, T.; Takagi, Y.; Figueiredo, D.; Henriques, J.; Pires, I.M. A Convolutional Neural Network-enabled IoT framework to verify COVID-19 hygiene conditions and authorize access to facilities. Procedia Comput. Sci. 2022, 203, 727–732. [Google Scholar] [CrossRef] [PubMed]
  51. Kinger, S.; Kinger, D.; Thakkar, S.; Bhake, D. Towards smarter hiring: Resume parsing and ranking with YOLOv5 and DistilBERT. Multimed. Tools Appl. 2024, 83, 82069–82087. [Google Scholar] [CrossRef]
  52. Wang, H.; Kang, X.; Ren, F. Emotion-Sentence-DistilBERT: A Sentence-BERT-Based Distillation Model for Text Emotion Classification; Springer: Berlin/Heidelberg, Germany, 2022; pp. 313–322. [Google Scholar]
  53. Leteno, T.; Gourru, A.; Laclau, C.; Gravier, C. An Investigation of Structures Responsible for Gender Bias in BERT and DistilBERT; Springer: Berlin/Heidelberg, Germany, 2023; pp. 249–261. [Google Scholar]
  54. Jojoa, M.; Eftekhar, P.; Nowrouzi-Kia, B.; Garcia-Zapirain, B. Natural language processing analysis applied to COVID-19 open-text opinions using a distilBERT model for sentiment categorization. AI Soc. 2024, 39, 883–890. [Google Scholar] [CrossRef]
  55. Qiu, K.; Zhang, Y.; Feng, Y.; Chen, F. LogAnomEX: An Unsupervised Log Anomaly Detection Method Based on Electra-DP and Gated Bilinear Neural Networks. J. Netw. Syst. Manag. 2025, 33, 33. [Google Scholar] [CrossRef]
  56. Yu, J.-L.; Dai, Q.-Q.; Li, G.-B. Deep learning in target prediction and drug repositioning: Recent advances and challenges. Drug Discov. Today 2022, 27, 1796–1814. [Google Scholar] [CrossRef]
  57. Ding, M. The road from MLE to EM to VAE: A brief tutorial. AI Open 2022, 3, 29–34. [Google Scholar] [CrossRef]
  58. Ye, Z.; Zuo, T.; Chen, W.; Li, Y.; Lu, Z. Textual emotion recognition method based on ALBERT-BiLSTM model and SVM-NB classification. Soft Comput. 2023, 27, 5063–5075. [Google Scholar] [CrossRef]
  59. Heim, E.; Ramia, J.A.; Hana, R.A.; Burchert, S.; Carswell, K.; Cornelisz, I.; Cuijpers, P.; El Chammay, R.; Noun, P.; van Klaveren, C.; et al. Step-by-step: Feasibility randomised controlled trial of a mobile-based intervention for depression among populations affected by adversity in Lebanon. Internet Interv. 2021, 24, 100380. [Google Scholar] [CrossRef]
  60. Hunhevicz, J.J.; Hall, D.M. Do you need a blockchain in construction? Use case categories and decision framework for DLT design options. Adv. Eng. Inform. 2020, 45, 101094. [Google Scholar] [CrossRef]
  61. Azmi, A.M.; Altmami, N.I. An abstractive Arabic text summarizer with user controlled granularity. Inf. Process. Manag. 2018, 54, 903–921. [Google Scholar] [CrossRef]
Figure 1. Flowchart for improving extractive summarization using graph-based independent sets in LLMs.
Figure 1. Flowchart for improving extractive summarization using graph-based independent sets in LLMs.
Applsci 15 06395 g001
Figure 2. Overall block diagram of the proposed framework.
Figure 2. Overall block diagram of the proposed framework.
Applsci 15 06395 g002
Figure 3. Example of text-based graph.
Figure 3. Example of text-based graph.
Applsci 15 06395 g003
Figure 4. Demonstration of graph-based content filtering with MIS selection.
Figure 4. Demonstration of graph-based content filtering with MIS selection.
Applsci 15 06395 g004
Figure 5. Generalized flow diagram of LLMs. Adapted from [43].
Figure 5. Generalized flow diagram of LLMs. Adapted from [43].
Applsci 15 06395 g005
Figure 6. Graphical analysis reflecting the observed variations in ROUGE F1 scores for the DUC dataset and 400-word summaries.
Figure 6. Graphical analysis reflecting the observed variations in ROUGE F1 scores for the DUC dataset and 400-word summaries.
Applsci 15 06395 g006
Figure 7. Graphical analysis reflecting the observed variations in ROUGE F1 scores for the DUC dataset and 200-word summaries.
Figure 7. Graphical analysis reflecting the observed variations in ROUGE F1 scores for the DUC dataset and 200-word summaries.
Applsci 15 06395 g007
Figure 8. Graphical analysis reflecting the observed variations in ROUGE-F1 scores for the CNN/DailyMail dataset.
Figure 8. Graphical analysis reflecting the observed variations in ROUGE-F1 scores for the CNN/DailyMail dataset.
Applsci 15 06395 g008
Figure 9. Effect of the proposed framework on processing time.
Figure 9. Effect of the proposed framework on processing time.
Applsci 15 06395 g009
Table 1. Comparative analysis of the proposed method and state-of-the-art approaches.
Table 1. Comparative analysis of the proposed method and state-of-the-art approaches.
ModelMain ContributionAdvantagesDisadvantages
[12] Graph + LLMContext length reductionPerformance boostModel heavily dependent on graph and LLM type
[14] Graph to TextConverts graphs to text for LLMEffective with suitable representationLarge-scale graphs less effective
[15] Benchmark for LLM Graph UnderstandingHighlights critical factors affecting LLMGood insight into LLM limitsLLM lags behind graph-specialized models
[16] GoSumHigh ROUGE on PubMed/arXiv datasetsEffective summarizationPerformance depends on document structure
[17] MultiBART-GATAbstractive summarization and info accuracyControls info accuracyComputationally expensive
[18] HeterMDSMaintains structure and content integrityHigh-quality summariesHigh integration cost
[19] Protein Interaction Pred.High-accuracy interaction predictionsMeaningful explanationsDomain-specific
[20] Discourse, Action GraphsImproves summary correctnessSuccess in automatic and human evalsComplexity of graphs
[21] Document-Level GraphsSignificant success on scientific datasetsHandles long documents wellDataset dependency and high cost
[22] Multi-Document Summ.Effective multi-doc summariesHigh ROUGE on DUC-2004No context integrity, repetitive sentences
[24] Biomed. Text Summ.Uses UMLS semantic typesMeaningful biomedical summariesDomain-specific
[25] Extended Biomed.
Graphs
More robust graph creationImproved node connectionsDomain-specific
[26] Prompt-Based GNNTask transferability improvementsOvercomes OOV issueIgnores semantic similarity across graphs
Proposed frameworkMIS-based structural filtering with LLM integrationHigh summarization accuracy, lower costPotential information loss in short summaries
Table 2. Comparison of text summarization datasets and their usage purposes.
Table 2. Comparison of text summarization datasets and their usage purposes.
DatasetsSDMDData SizeHuman AnnotationPurpose of Use
Duc-2002noyes567 documentyesInformation fusion–Content synthesis
CNN/DailyMailyesno290,000 newsyesMain idea extraction–Summary of information
Table 3. Performance improvements of ROUGE metric on 400-Word abstracts on DUC dataset (comparison of baseline and improved models).
Table 3. Performance improvements of ROUGE metric on 400-Word abstracts on DUC dataset (comparison of baseline and improved models).
ROUGE
Metrics
ROUGE Performance Values (DUC—400-Word Abstracts)
XLNetRobertaDistilbertElectraAlbertPegasus
BaselineProp.BaselineProp.BaselineProp.BaselineProp.BaselineProp.BaselineProp.
Rouge 1R0.484890.49119
(+0.0063)
0.450740.50124
(+0.0505)
0.487250.49911
(+0.0118)
0.443500.49335
(+0.0498)
0.474110.53538
(+0.0612)
0.501530.52736
(+0.02583)
P0.478430.48946
(+0.0110)
0.450950.48953
(+0.0385)
0.484470.49456
(+0.0100)
0.441730.49230
(+0.0505)
0.475340.52823
(+0.0528)
0.493710.52692
(+0.03321)
F10.481550.49018
(+0.0086)
0.450670.49519
(+0.0445)
0.485650.49659
(+0.0109)
0.442290.49278
(+0.0504)
0.474680.53156
(+0.0568)
0.497510.52693
(+0.02942)
Rouge 2R0.157390.18030
(+0.0229)
0.146510.18695
(+0.0404)
0.194730.19485
(+0.0001)
0.123820.16971
(+0.0458)
0.164930.23507
(+0.0701)
0.190690.22390
(+0.03321)
P0.155640.17936
(+0.0237)
0.145850.18261
(+0.0367)
0.193150.191700.122590.16900
(+0.0464)
0.165660.23143
(+0.0657)
0.187670.22250
(+0.03483)
F10.156480.17978
(+0.0233)
0.146120.18471
(+0.0385)
0.193850.193170.123130.16935
(+0.0462)
0.165280.23314
(+0.0678)
0.189140.22313
(+0.03399)
Rouge LR0.446620.45390
(+0.0072)
0.419370.45791
(+0.0385)
0.449720.45778
(+0.0080)
0.404100.45331
(+0.0492)
0.434600.50005
(+0.0654)
0.468170.49639
(+0.02822)
P0.440770.45237
(+0.0116)
0.419510.44727
(+0.0277)
0.447080.45336
(+0.0062)
0.402120.45221
(+0.0500)
0.435770.49323
(+0.0574)
0.460880.49587
(+0.03499)
F10.443600.45300
(+0.0094)
0.419270.45241
(+0.0331)
0.448210.45534
(+0.0071)
0.402830.45272
(+0.0498)
0.435140.49641
(+0.0612)
0.464420.49594
(+0.03152)
Rouge WR0.138540.13958
(+0.0010)
0.125770.14008
(+0.0143)
0.136730.14140
(+0.0046)
0.120540.13855
(+0.0180)
0.133560.15578
(+0.0222)
0.147220.15464
(+0.00742)
P0.215470.21928
(+0.0038)
0.198200.21550
(+0.0173)
0.214390.22085
(+0.0064)
0.188770.21788
(+0.0291)
0.211290.24207
(+0.0307)
0.228460.24374
(+0.01528)
F10.168600.17051
(+0.0019)
0.153820.16972
(+0.0159)
0.166890.17231
(+0.0054)
0.147010.16936
(+0.0223)
0.163620.18944
(+0.0258)
0.179010.18916
(+0.01015)
Rouge SR0.214240.22180
(+0.0075)
0.184820.22801
(+0.0431)
0.221860.22857
(+0.0067)
0.185910.21787
(+0.0319)
0.203700.26192
(+0.0582)
0.229060.25236
(+0.0233)
P0.208710.21989
(+0.0111)
0.184170.21726
(+0.0330)
0.218880.22330
(+0.0044)
0.183480.21667
(+0.0331)
0.204670.25468
(+0.0500)
0.221890.25099
(+0.0291)
F10.211290.22057
(+0.0092)
0.184200.22228
(+0.0380)
0.219990.22544
(+0.0054)
0.184200.21719
(+0.0329)
0.204110.25782
(+0.0537)
0.225260.25130
(+0.02604)
Rouge SUR0.215560.22310
(+0.0105)
0.186110.22934
(+0.0432)
0.223150.22988
(+0.0067)
0.187170.21921
(+0.0320)
0.205010.26325
(+0.0582)
0.230380.25369
(+0.02331)
P0.210010.22120
(+0.1119)
0.185470.21856
(+0.0330)
0.220170.22460
(+0.0044)
0.184740.21800
(+0.0332)
0.205980.25599
(+0.0500)
0.223190.25233
(+0.02914)
F10.212590.22188
(+0.0092)
0.185500.22359
(+0.0380)
0.221280.22675
(+0.0054)
0.185460.21853
(+0.0330)
0.205410.25914
(+0.0537)
0.226570.25263
(+0.02606)
▲ Improvement from baseline.
Table 4. Performance improvements of ROUGE metric on 200-word abstracts on DUC dataset (comparison of baseline and improved models).
Table 4. Performance improvements of ROUGE metric on 200-word abstracts on DUC dataset (comparison of baseline and improved models).
ROUGE
Metrics
ROUGE Performance Values (DUC—200-Word Abstracts)
XLNetRobertaDistilbertElectraAlbertPegasus
BaselineProp.BaselineProp.BaselineProp.BaselineProp.BaselineProp.BaselineProp.
Rouge 1R0.386520.40223
(+0.0157)
0.362190.37387
(+0.0116)
0.365040.39377
(+0.0287)
0.399180.389350.412100.42355
(+0.0114)
0.428150.46058
(+0.03243)
P0.389910.39357
(+0.0036)
0.354620.37229
(+0.0176)
0.366540.38293
(+0.0163)
0.401080.390350.409840.41782
(+0.0079)
0.414720.45413
(+0.03941)
F10.387740.39757
(+0.0141)
0.358140.37280
(+0.0146)
0.365160.38811
(+0.0229)
0.399800.389360.410660.42032
(+0.0096)
0.421100.45704
(+0.03594)
Rouge 2R0.106750.12093
(+0.0141)
0.082370.09336
(+0.0109)
0.084050.10225
(+0.0182)
0.122580.09785
0.132950.17328
(+0.0403)
0.143640.18557
(+0.04193)
P0.108680.11744
(+0.0087)
0.080750.09154
(+0.0107)
0.081200.09978
(+0.0185)
0.121550.097860.131960.16860
(+0.0366)
0.138170.18131
(+0.04314)
F10.107620.11911
(+0.0114)
0.081510.09237
(+0.0108)
0.082510.10097
(+0.0184)
0.121470.097760.132400.17072
(+0.0383)
0.140790.18332
(+0.04253)
Rouge LR0.342850.36832
(+0.0254)
0.328000.4235
(+0.0955)
0.324690.36007
(+0.0353)
0.357150.351300.368130.38939
(+0.0212)
0.387140.42147
(+0.03433)
P0.346490.36020
(+0.0137)
0.321390.4235
(+0.0121)
0.325590.34981
(+0.0242)
0.359230.352560.365990.38244
(+0.0164)
0.374930.41512
(+0.04019)
F10.344260.36397
(+0.0197)
0.324460.4235
(+0.099)
0.324580.35472
(+0.0301)
0.357900.351480.366780.38607
(+0.0192)
0.380730.41801
(+0.03728)
Rouge WR0.114700.12901
(+0.0143)
0.113110.112370.111650.12231
(+0.0006)
0.123320.120670.126950.13287
(+0.0059)
0.134770.15021
(+0.01544)
P0.182440.19869
(+0.0162)
0.174390.17599
(+0.0016)
0.176290.18719
(+0.0109)
0.195140.190630.198680.20661
(+0.0079)
0.205550.23306
(+0.02751)
F10.140670.15635
(+0.0156)
0.137140.137080.136530.14789
(+0.0113)
0.151020.147620.154800.16159
(+0.0067)
0.162740.18259
(+0.01985)
Rouge SR0.136500.14431
(+0.0078)
0.121810.12721
(+0.0054)
0.123330.14294
(+0.0196)
0.151220.141470.155310.16666
(+0.0113)
0.155660.18696
(+0.0313)
P0.138860.137400.116900.12506
(+0.0081)
0.122420.13501
(+0.0125)
0.152100.141930.153350.15891
(+0.0055)
0.145430.18059
(+0.03516)
F10.137060.14045
(+0.0033)
0.119010.12573
(+0.0067)
0.122050.13870
(+0.0166)
0.151140.141030.153910.16211
(+0.0082)
0.150050.18325
(+0.0332)
Rouge SUR0.138870.14676
(+0.0078)
0.124090.12956
(+0.0054)
0.125620.14532
(+0.0197)
0.153580.143820.157770.16910
(+0.0113)
0.158250.18957
(+0.03132)
P0.141260.139790.119100.12741
(+0.0083)
0.124760.13742
(+0.0126)
0.154490.144300.155800.16136
(+0.0055)
0.147910.18317
(+0.03526)
F10.139440.14287
(+0.0034)
0.121250.12807
(+0.0068)
0.124360.14103
(+0.0166)
0.153520.143380.156360.16455
(+0.0081)
0.152580.18584
(+0.03326)
▲ Improvement from baseline.
Table 5. Performance improvements of ROUGE metric on 3-word summaries for CNN/DailyMail Dataset (comparison of baseline and improved models).
Table 5. Performance improvements of ROUGE metric on 3-word summaries for CNN/DailyMail Dataset (comparison of baseline and improved models).
ROUGE
Metrics
ROUGE Performance Values (CNN/DailyMail—3-Word Abstracts)
XLNetRobertaDistilbertElectraAlbertPegasus
BaselineProp.BaselineProp.BaselineProp.BaselineProp.BaselineProp.BaselineProp.
Rouge 1R0.202740.33916
(+0.1364)
0.319280.296870.349640.39377▲
(+0.0441)
0.399180.389350.250920.34894▲
(+0.0980)
0.341340.32971
P0.173160.18269
(+0.0095)
0.209290.178420.211460.38293▲
(+0.1714)
0.401080.390350.185160.18799▲
(+0.0028)
0.224040.17474
F10.174620.23245
(+0.0578)
0.240240.216430.255970.38811▲
(+0.1321)
0.399800.389360.203670.23983▲
(+0.0361)
0.257710.22518
Rouge 2R0.027200.09733
(+0.0701)
0.103910.064820.126660.102250.122580.09785
0.044760.09292
(+0.0481)
0.107470.08703
P0.029320.05161
(+0.0222)
0.059580.040970.067960.09978
(+0.0318)
0.121550.097860.031200.04988
(+0.0186)
0.059330.04602
F10.025250.06587
(+0.0406)
0.073030.049200.086280.10097
(+0.0146)
0.121470.097760.034840.06367
(+0.0812)
0.074600.05945
Rouge LR0.159710.28513
(+0.0125)
0.261220.241690.290660.275010.357150.351300.197180.27847
(+0.0812)
0.285690.26888
P0.140870.15279
(+0.0119)
0.163150.148270.173260.156420.359230.352560.145960.14947
(+0.0035)
0.188350.13953
F10.139010.19480
(+0.0557)
0.190730.178010.211010.194640.357900.351480.160370.19068
(+0.0303)
0.216240.18105
Rouge WR0.074750.12937
(+0.0546)
0.121720.114130.137210.125380.123320.120670.093310.12843
(+0.0351)
0.132700.12461
P0.110910.11239
(+0.0014)
0.126160.114530.134000.117260.195140.190630.111450.11186
(+0.0004)
0.143860.10589
F10.083020.11708
(+0.0340)
0.116700.110350.131650.117700.151020.147620.096790.11641
(+0.0196)
0.131380.11225
Rouge SR0.031060.10828
(+0.0772)
0.100530.074240.123660.091240.151220.141470.056100.10345
(+0.0473)
0.106970.09945
P0.028260.03471 (+0.0064)0.034160.025510.041740.029000.152100.141930.029860.027720.035480.02697
F10.022840.04899
(+0.0261)
0.044080.035070.057220.040940.151140.141030.032850.04162
(+0.0087)
0.046590.04090
Rouge SUR0.039600.12013
(+0.0805)
0.112300.085740.134990.104160.153580.143820.067010.11635
(+0.0493)
0.119150.11093
P0.035720.03878 (+0.0030)0.041040.030590.047190.033880.154490.144300.036320.032390.042620.03094
F10.029000.05471
(+0.0257)
0.051650.041520.064090.047540.153520.143380.039870.04817
(+0.0083)
0.054400.04657
▲ Improvement from baseline.
Table 6. Time savings of the proposed framework for varying abstract lengths in LLMs.
Table 6. Time savings of the proposed framework for varying abstract lengths in LLMs.
ModelsEffect of the Proposed Framework on Processing Time (sn)
DUCBBC
200 Words400 Words3 Words
XLNet196.54→77.23▲
            (−60.71%)
185.35→82.17▲
            (−55.67%)
43.26→24.27▲
            (−43.90%)
RoBERTa122.33→56.24▲
            (−54.03%)
125.03→64.49▲
            (−48.42%)
32.61→21.73▲
            (−33.36%)
DistilBERT64.14→27.14▲
            (−57.69%)
67.58→33.43▲
            (−50.53%)
21.57→13.33▲
            (−38.20%)
ELECTRA37.84→15.68▲
            (−58.56%)
39.62→19.82▲
            (−49.97%)
11.91→8.92▲
            (−25.09%)
ALBERT113.62→45.2▲
            (−60.22%)
114.09→55.52▲
            (−51.34%)
24.32→11.71▲
            (−51.85%)
Pegasus4478.08→1384.51▲
            (−69.08%)
4843.66→1794.01▲
            (−62.96%)
903.73→333.52▲
            (−63.10%)
▲ Improvement from baseline.
Table 7. Statistical validation of ROUGE improvements via paired t-tests across summary length variants.
Table 7. Statistical validation of ROUGE improvements via paired t-tests across summary length variants.
Rouge Metricst-Statisticp-Value
400-wordRouge-15.4200.0056
Rouge-26.2680.0033
Rouge-L7.4070.0018
Rouge-W6.9090.0023
Rouge-S8.1290.0012
Rouge-SU9.1860.0008
200-wordRouge-17.3010.0019
Rouge-26.3890.0031
Rouge-L9.2310.0008
Rouge-W7.7140.0015
Rouge-S10.1570.0005
Rouge-SU8.3370.0011
3-wordRouge-110.1560.0005
Rouge-28.5000.0011
Rouge-L12.5090.0002
Rouge-W10.9540.0004
Rouge-S9.4360.0007
Rouge-SU15.9220.0001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hark, C. Using Graph-Based Maximum Independent Sets with Large Language Models for Extractive Text Summarization. Appl. Sci. 2025, 15, 6395. https://doi.org/10.3390/app15126395

AMA Style

Hark C. Using Graph-Based Maximum Independent Sets with Large Language Models for Extractive Text Summarization. Applied Sciences. 2025; 15(12):6395. https://doi.org/10.3390/app15126395

Chicago/Turabian Style

Hark, Cengiz. 2025. "Using Graph-Based Maximum Independent Sets with Large Language Models for Extractive Text Summarization" Applied Sciences 15, no. 12: 6395. https://doi.org/10.3390/app15126395

APA Style

Hark, C. (2025). Using Graph-Based Maximum Independent Sets with Large Language Models for Extractive Text Summarization. Applied Sciences, 15(12), 6395. https://doi.org/10.3390/app15126395

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop