Next Article in Journal
A Framework for Data Lifecycle Model Selection
Previous Article in Journal
Design and Implementation of a Scalable LoRaWAN-Based Air Quality Monitoring Infrastructure for the Kurdistan Region of Iraq
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Study of PEGASUS, BART, and T5 for Text Summarization Across Diverse Datasets

by
Eman Daraghmi
*,
Lour Atwe
and
Areej Jaber
Department of Computer Science, Palestine Technical University Kadoorie, Jaffa Street, Tulkarm 9993400, Palestine
*
Author to whom correspondence should be addressed.
Future Internet 2025, 17(9), 389; https://doi.org/10.3390/fi17090389
Submission received: 27 July 2025 / Revised: 20 August 2025 / Accepted: 26 August 2025 / Published: 28 August 2025
(This article belongs to the Special Issue Artificial Intelligence (AI) and Natural Language Processing (NLP))

Abstract

This study aims to conduct a comprehensive comparative evaluation of three transformer-based models, PEGASUS, BART, and T5 variants (SMALL and BASE), for the task of abstractive text summarization. The evaluation spans across three benchmark datasets: CNN/DailyMail (long-form news articles), Xsum (extreme single-sentence summaries of BBC articles), and Samsum (conversational dialogues). Each dataset presents unique challenges in terms of length, style, and domain, enabling a robust assessment of the models’ capabilities. All models were fine-tuned under controlled experimental settings using filtered and preprocessed subsets, with token length limits applied to maintain consistency and prevent truncation. The evaluation leveraged ROUGE-1, ROUGE-2, and ROUGE-L scores to measure summary quality, while efficiency metrics such as training time were also considered. An additional qualitative assessment was conducted through expert human evaluation of fluency, relevance, and conciseness. Results indicate that PEGASUS achieved the highest ROUGE scores on CNN/DailyMail, BART excelled in Xsum and Samsum, while T5 models, particularly T5-Base, narrowed the performance gap with larger models while still offering efficiency advantages compared to PEGASUS and BART. These findings highlight the trade-offs between model performance and computational efficiency, offering practical insights into model scaling—where T5-Small favors lightweight efficiency and T5-Base provides stronger accuracy without excessive resource demands.

Graphical Abstract

1. Introduction

We live in an era of information abundance but attention scarcity. Every minute, thousands of news articles, social media posts, and research papers flood digital platforms, making it increasingly difficult to identify relevant information efficiently. In this landscape, automatic text summarization has emerged as an essential tool to distill large volumes of content into concise and meaningful summaries, supporting faster decision-making and knowledge consumption [1,2,3,4,5,6].
While early summarization systems relied on extractive rule-based techniques, essentially highlighting sentences without deep semantic understanding, they often lacked coherence and contextual awareness. These methods proved inadequate for summarizing longer documents or handling diverse writing styles. The advent of transformer-based architectures, particularly since 2017, revolutionized the field [7,8,9,10,11]. Unlike sequential models, transformers utilize self-attention mechanisms to process entire sequences simultaneously, enabling models like BERT, GPT, BART, and T5 to deliver significantly more fluent, context-aware summaries [7,11,12,13,14,15,16].
Despite these advancements, real-world deployments of summarization systems still face three major challenges:
  • Precision Problem: Models perform inconsistently across domains. A model effective on news articles might fail on conversational or technical texts, limiting generalizability. For example, 42% of businesses report inconsistent results when applying the same summarizer across departments [17].
  • Practicality Dilemma: Research prioritizes accuracy metrics (e.g., ROUGE), while practitioners care about training costs, inference speed, and deployment feasibility. Yet most benchmarks lack guidance on such operational concerns, forcing costly trial-and-error selection.
  • Adaptation Challenge: Most benchmarks focus on clean, formal texts (e.g., news), ignoring messy real-world inputs like emails, chats, and social media, leaving it unknown how well these models adapt to such scenarios.
Although several transformer-based models have been proposed and benchmarked, few studies offer systematic, cross-domain evaluations that assess both performance and practicality across distinct summarization contexts. There is a need for comparative studies that go beyond ROUGE scores to consider deployment realities, such as training time, hardware constraints, and adaptation to informal or conversational input.
In contrast to prior comparative works, our study introduces a unified evaluation framework that simultaneously addresses accuracy, computational efficiency, and domain adaptability. We incorporate efficiency metrics such as training and inference times alongside hardware resource profiling, an element often missing in earlier benchmarks. Additionally, we implement a standardized preprocessing pipeline across datasets to control for input variability and integrate a domain-specific qualitative assessment through human evaluation, offering practical insights into model behavior beyond automated metrics. This combination of methodological rigor and operational considerations extends the scope of previous multi-model comparisons and makes our findings directly actionable for real-world deployments.
This study addresses these gaps by conducting a comprehensive comparative evaluation of three prominent transformer-based summarization models, PEGASUS, BART, and two T5 variants (SMALL and BASE), across three datasets that reflect distinct summarization tasks [12,18,19]:
  • CNN/DailyMail: Long-form news articles emphasizing context retention.
  • Xsum: Extreme summarization to a single sentence.
  • Samsum: Dialogue-based summarization of informal conversations.
Our key contributions include the following:
  • A cross-domain benchmarking framework that evaluates models on accuracy (ROUGE), efficiency (training/inference time), and practical constraints (model size, memory usage).
  • A detailed analysis of context preservation, domain-specific language handling, and error types across content types.
  • Practical insights into model selection trade-offs, highlighting how T5 variants balance accuracy and efficiency: T5-Small achieves lightweight deployment with ~70% less resource use, while T5-Base improves accuracy—closing much of the gap with BART and PEGASUS—yet still maintains moderate efficiency.
Our findings have real-world implications across industries. In healthcare, identifying models that best preserve domain-specific terminology can enhance summarization of clinical notes. In journalism and research, our results can inform model selection based on content type and resource availability. For instance, while BART excels in dialogue coherence (Samsum), PEGASUS maintains superior factual retention in long-form articles (CNN/DailyMail), a critical requirement for legal or scientific summarization tasks.

2. Literature Review

Automated text summarization as a field of research has evolved dramatically over the last ten years. This was especially true since 2017, when there was a new architectural concept that revolutionized the way machines process and summarize textual content in general. This section demonstrated this technological progress with an emphasis on three very influential models that have altered the landscape of what can be achieved in summarization: PEGASUS, BART, and T5. Each one is a unique way of handling the multifaceted challenge of reducing source materials to fundamental meaning without compromising coherence and factuality.

2.1. Background

2.1.1. The Transformer Architecture

Transformers are the foundation of modern NLP, introduced by Vaswani et al. (2017) [20], and represent a paradigm shift from recurrent and convolutional models to attention-based architectures. The core idea behind Transformers is the ability to model long-range dependencies in text using a mechanism called self-attention, which enables the model to weigh the importance of different words in a sentence regardless of their position. This approach allows for faster training, better scalability, and superior performance across a wide range of language tasks.
As shown in Figure 1, the transformer model is composed of two main components:
  • Encoder: Converts the input sequence into continuous hidden representations.
  • Decoder: Generates the output sequence, conditioned on the encoder’s output.
Each encoder and decoder stack consists of multiple identical layers (6 in the base model), each comprising the following:
  • Multi-Head Self-Attention Mechanism: Captures contextual relationships between tokens in the sequence by computing attention scores in parallel across multiple heads (Patwardhan et al., 2023) [10].
  • Feed-Forward Network (FFN): A fully connected MLP applied to each position independently, responsible for transforming the attention output.
  • Residual Connections [21] and Layer Normalization [21]: Help maintain gradient flow and stabilize learning.
In the decoder, masked attention is used to ensure autoregressive generation, and cross-attention links the decoder to the encoder output for informed generation.
Transformers have revolutionized tasks such as the following: Text summarization, Machine translation, Sentiment analysis, Text classification, Question answering, and Language generation. Their flexibility and robustness have made them the default architecture for most state-of-the-art NLP systems.

2.1.2. Pre-Trained Transformer Models for Summarization

Pre-trained models significantly advance summarization by leveraging large corpora to learn general language representations, which are then fine-tuned for downstream tasks. The most influential among them, PEGASUS, BART, and T5 are designed as sequence-to-sequence models, ideal for abstractive summarization.
Zhang et al. [22] introduced in 2019, PEGASUS, a training method that sets it apart from other models. The authors utilized a new method, i.e., gap-sentence generation method, to pre-train in place of the conventional pre-training methods. On the training front, the model maintains records for all documents with strategically removed sentences—sometimes randomly selected sentences, sometimes those identified as most important. This monitoring stage forces PEGASUS to learn how to reconstruct missing content, essentially training it to “fill in the blanks” in the same way a skilled human editor might do when drafting a summary. Evaluations on news corpora like HugeNews demonstrate that its 568 M-parameter variant achieves state-of-the-art performance, particularly for generating concise yet comprehensive summaries [23].
Lewis et al. in 2019 introduced BART [24] that employs both bidirectional understanding of text while adopting GPT’s generative capabilities. This hybrid approach gives BART a unique advantage—it is remarkably good at making sense of messy or incomplete inputs. During training, the model sees text that is been deliberately scrambled through five different noise strategies (like shuffled sentences or missing words), forcing it to develop robust reconstruction skills. While not specifically designed for summarization, BART’s 400 M-parameter version consistently delivers strong results across various summarization tasks, proving that sometimes general-purpose models can outperform specialized ones.
BART (bidirectional and auto-regressive transformers) is a sequence-to-sequence pre-training model introduced by [24]. It combines two transformer models: BERT [25] and GPT [26], a bidirectional encoder, and an auto-regressive decoder, respectively, as illustrated in Figure 2 below.
The BART model is a denoising autoencoder that learns to rebuild the original text by destroying and corrupting the input text [24]. Figure 3 shows the transformations used to noise the text.
  • Token Masking: It swaps random tokens in the input with [MASK].
  • Token Deletion: It removes random tokens from the input, and the model must know where these missing tokens are located.
  • Text Infilling: A set of adjacent tokens is removed and replaced with a single mask, and the model must learn the missing part and the number of missing tokens.
  • Sentence Permutation: It takes the input and randomly shuffles the sentences that are separated by dots.
  • Document Rotation: A random token is selected uniformly from the input, and then the input is rotated to start with that token. This is intended to train the model to determine the beginning of the input.
Based on comparison experiments conducted by researchers on these transformations, an admixture of text infilling and sentence permutation is used for training the BART model; for each input, all sentences are mixed up and mask 30% tokens [24]. The BART model is pre-trained on huge data that includes books, stories, news articles, and web content.
Both BART BASE and BART LARGE are distinct versions of the BART model (see Figure 2). Table 1 and Table 2 below list each version along with the number of parameters.
For its robust architecture and ability to handle text corruption, BART LARGE will be employed in this work. T5 (Raffel et al., 2020) [23] is the most ambitious unification effort, reducing all NLP tasks to the text-to-text problem form. While more parameter sizes are required (up to 11B) to achieve competitive summarization performance, it is presently the most versatile solution. The uniform model architecture across tasks, from its smallest (60 M parameter) up to the largest forms, brings new insights into how scale affects summarization quality relative to other NLP applications.

2.2. Related Works

The past decade witnessed an enormous upsurge in the field of automated text summarization, where research focused on a broad array of methods to address the demanding task of extracting significant information from source material. With research originally focusing on extractive methods, huge recent advances in neural models, particularly transformer models, have shifted the focus towards more sophisticated abstractive techniques. Scholars have been busy examining ways in which large language models can transform text summarization, resulting in scores of papers published over the past decade.
In conversational summarization, Sonar et al. (2023) [11] demonstrated PEGASUS’s superiority over BERT and Simple T5 on the Samsum dataset, achieving ROUGE-1 scores of 42%. The medical domain presents unique challenges, as shown by Kolhatkar et al. (2023) [8], where PEGASUS-large attained a 27% ROUGE-L F1 score on clinical notes from MIMIC III, a particularly strong result given the small training set (765 samples).
Cross-lingual applications continue to expand, with Wang et al. (2023) [13] adapting T5-Base for Chinese scientific literature through a novel semi-supervised framework, while La Quatra and Cagliero (2022) [18] achieved similar success with their Italian-optimized BART-IT variant. Particularly striking results emerged from Arabic summarization (Alsuhaibani, 2023) [21], where domain-specific PEGASUS fine-tuning yielded exceptional ROUGE-1 F1 scores exceeding 81%, peaking at 88% for economic topics.
These studies collectively highlight two critical findings: (1) while transformer models generally outperform traditional approaches, their relative effectiveness varies significantly by domain and language, and (2) targeted architectural modifications an yield specialized improvements without requiring complete model redesigns.
Before transformers dominated the field, researchers achieved notable results through various neural and statistical methods. Elluri, et al. (2023) [27] demonstrated that even old-school machine learning techniques like SVMs were able to come up with usable extractive summaries from news headlines using a thoughtful approach to feature engineering. The work of Verma and Nidhi (2017) [28] was a turning point, demonstrating how RNNs with attention mechanisms could begin to pick up long-distance relationships, a feat that transformers would later master.
Recent comparative evaluations, for example [29,30], put these old methods into historical context in comparison with modern standards. Its demonstration of BERT having been 98% accurate at some summarization tasks is a testament to the impressive progress made, as well as the potential value of accuracy metrics as a measure of summary quality. As Saiyyad and Patil (2024) [7] conclude in their survey, although transformer architecture is the performance bar today, it is still helpful to understand its relation to the older approaches to set the direction for future research. Several detailed reviews have traced the evolution of the field. Allahyari et al. [31] provide indispensable context in extractive methods, tracing the transition from simple frequency-based to sophisticated graph-based methods. Janjanam and Reddy [32] provide a particularly useful connection between traditional and neural methods, while Rahul et al. [33] carefully compare evaluation paradigms across different summarization paradigms. Zhang et al. [12] detailed review of deep learning advancements describes how abstractive methods eventually surpassed extractive methods. Collective surveys to date identify long-standing challenges in domain adaptation, limitations of the used evaluation metrics, and computational efficiency governed existing work. This work draws upon the foundation they established and bridges identified gaps via stringent model comparisons across text types and sizes. Table 3 provides an overview of previous research that utilizes the transformer-based models.

3. Research Methodology

This section presents a systematic framework (see Figure 4) for evaluating transformer-based summarization models (PEGASUS, BART, T5) across three benchmark datasets (CNN/DailyMail, Xsum, Samsum). The methodology addresses four critical dimensions: (1) dataset preparation, (2) model configuration, (3) experimental protocol, and (4) evaluation strategy, with explicit considerations for computational constraints, reproducibility, and statistical significance testing. All experiments were executed in two phases (Phase 1: ~6% validation + test; Phase 2: 80/10/10), and both are documented for transparency and replication.
The three selected models, PEGASUS, BART, and T5, represent state-of-the-art transformer-based architectures specifically designed for text summarization tasks. PEGASUS, introduced by Zhang et al. [12], employs a novel pre-training strategy called gap-sentence generation (GSG), which makes it particularly well-optimized for summarization. In this study, the PEGASUS-CNN/DailyMail variant was used. BART, proposed by Lewis et al. [24], is a denoising autoencoder that integrates features from both BERT and GPT, allowing it to effectively handle noisy and corrupted input data. The version employed in this work is BART LARGE. Lastly, T5, developed by Raffel et al. [23], is based on the Text-to-Text Transfer Transformer framework, treating all NLP tasks as text generation problems. While it is highly versatile, it remains more generalized compared to PEGASUS and BART. To capture both efficiency and accuracy trade-offs, we evaluated two variants of T5: T5-Small (lightweight, ~60 M parameters) and T5-Base (~220 M parameters). This dual inclusion allows us to analyze both low-resource and mid-resource scenarios.

3.1. Dataset Selection

To ensure a fair and consistent evaluation, three publicly available datasets representing distinct summarization challenges in length, complexity, and structure were selected for robust cross-domain assessment, as summarized in Table 4.
To ensure a fair, comprehensive, and domain-diverse evaluation, this study utilizes three benchmark datasets: CNN/DailyMail, Xsum, and Samsum, each representing unique summarization challenges. The CNN/DailyMail dataset consists of over 300,000 news articles paired with multi-sentence highlights, offering a benchmark for long-form summarization. Its structure, sourced from CNN and Daily Mail websites, supports evaluating models on lengthy, structured texts (Train: 286,817; Val: 13,368; Test: 11,487). The Xsum dataset, curated from BBC articles, presents a contrasting task: each document is paired with a highly compressed, single-sentence summary that captures the core message (Train: 204,045; Val: 11,332; Test: 11,334), making it ideal for testing extreme abstraction. Finally, the Samsum corpus introduces a conversational domain, containing dialogue-summary pairs constructed by linguists across various tones and formats, formal, semi-formal, and informal (Train: 14,732; Val: 818; Test: 819). This dataset evaluates a model’s ability to distill meaning from informal, multi-turn dialogue into coherent summaries. Collectively, these datasets provide diverse linguistic, structural, and compression-level challenges, enabling a robust cross-domain comparison of transformer summarization models.

Dataset Sampling and Filtering Strategy

Due to computational resource constraints, we utilized carefully filtered and stratified subsets of each dataset rather than their full size. This approach ensured that all experiments could be executed within the available GPU memory and time budgets while retaining the linguistic diversity, topic variety, and structural patterns necessary for fair model comparison, thus balancing computational feasibility with representativeness and enabling reproducible, scalable experimentation. Stratified random sampling was employed to maintain the original distribution of summary lengths, content genres, and abstraction levels within each dataset, helping to prevent bias toward overly common structures or topics, with all sampling operations using a fixed random seed (42) to guarantee reproducibility across runs. The proportions retained were 10% of the full CNN/DailyMail dataset (training: 28,943 samples, validation: 3218, test: 3223), 15% of the Xsum dataset (training: 18,374 samples, validation: 2271, test: 2270), and 80% of the Samsum dataset (training: 12,474 samples, validation: 818, test: 819) due to its smaller original size. While smaller subsets reduce training time and memory requirements, they may limit a model’s exposure to rare linguistic phenomena, domain-specific vocabulary, and complex discourse structures particularly for long-form summarization tasks where recall is critical; however, this risk was mitigated by adopting a stratified sampling strategy that preserved proportional representation of varying summary lengths, topical diversity, and stylistic complexity across domains, ensuring a balanced evaluation that remains computationally tractable without disproportionately favoring one dataset type over another. To promote open science, transparency, and full reproducibility, we provide exact dataset split files containing the selected IDs for each dataset’s train–validation–test partitions, a Python sampling script implementing the stratified selection process using the Hugging Face datasets library, and Supplementary Documentation describing the rationale for subset sizes and the verification process confirming that the reduced datasets preserved key statistical properties of the originals; these resources are hosted on our GitHub repository [https://github.com/EmanDaraghmi/summarization-research] (accessed on 27 August 2025).

3.2. Preprocessing Pipeline

To ensure consistency, efficiency, and optimal model performance, all three datasets (CNN/DailyMail, Xsum, Samsum) underwent a systematic preprocessing pipeline tailored to both model constraints and available computational resources.

3.2.1. Data Filtering Strategy

Given the substantial size of the CNN/DailyMail and Xsum datasets, and the constraints of the available hardware, a subset of the training data was carefully selected to maximize data quality while avoiding truncation during tokenization. The following steps summarize the applied filtering criteria:
  • Articles with word counts ≤ 1024 and summaries ≤ 128 were initially selected.
  • From this subset, 100,000 rows were sampled.
  • Each article was further validated to ensure token length ≤ 1024 (≤512 for T5-Small), with summaries also capped at 128 tokens.
This filtering minimizes truncation, which can negatively impact model performance, especially in cases where critical information is located beyond the truncation threshold. For example, if an article spans 5000 tokens and its essence lies beyond the first 1500, truncation could lead to significant loss of contextual understanding. Therefore, refining the dataset ensures that input sequences retain semantic coherence and are fully compatible with model architectures.

3.2.2. Text Normalization

Following token-length filtering, datasets were normalized to standardize linguistic variations:
  • CNN/DailyMail and Xsum: British spellings were converted to American English.
  • Samsum: Chat abbreviations were expanded (e.g., “u” → “you”), contractions were normalized, and punctuation inconsistencies were corrected.

3.2.3. Data Splitting Strategy and Two-Phase Design (Training–Validation Splitting)

We conducted experiments in two phases to balance computational constraints with statistical robustness.
Originally, our experiments, Phase 1, used a fixed, stratified 1000-sample test set for each dataset, with the remaining data split 95%/5% into training and validation sets. This design was chosen to ensure a consistent evaluation size across datasets while staying within the runtime and memory constraints of our Google Colab L4 environment. However, to increase the proportion of evaluation data, we re-performed the experiments, Phase 2, using an 80/10/10 train–validation–test split via stratified sampling (seed = 42). This adjustment increased the combined validation and test share from 6% to 20%, improving the stability of early stopping, reducing variance in evaluation metrics, and better aligning with common practice in text summarization. Both T5-Small and T5-Base followed the same stratified splitting strategy. However, due to the larger parameter count of T5-Base, training batch sizes were adjusted (batch size = 8–16) to fit GPU memory constraints, while other hyperparameters were kept consistent.
While the smaller training set slightly reduced ROUGE scores (by ~0.5–1.2 points depending on the model and dataset), the larger evaluation sets provided narrower confidence intervals and a more reliable comparison across PEGASUS, BART, and T5-Small. We report the results in Section 4, and both experimental scripts are available on our GitHub repository for full reproducibility.

3.2.4. Model-Specific Dataset Sizes

Tokenization differences across models affected final dataset sizes. PEGASUS and BART, which support longer sequences, retained a higher number of training examples than T5-Small. Table 5 presents the resulting dataset sizes across models.
This approach enabled a balanced compromise between evaluation robustness and computational feasibility, facilitating consistent model comparisons without introducing truncation-related bias.

3.2.5. Integration with Summarization Pipeline

The preprocessing pipeline was designed to feed directly into the summarization workflow illustrated in Figure 5. The filtered and normalized datasets served as the primary input for the tokenization stage, ensuring that all samples complied with each model’s sequence length constraints. By standardizing text formats, controlling input lengths, and removing noise prior to tokenization, the preprocessing stage reduced computational overhead and minimized the need for aggressive truncation within the summarization pipeline. This integration ensured the following:
  • Seamless Model Input—Preprocessed sequences were fully compatible with PEGASUS, BART, and T5-Small without additional runtime adjustments.
  • Consistent Evaluation Flow—The stratified splits created in preprocessing aligned with the evaluation stages in the summarization pipeline, enabling a direct comparison of model performance.
  • Optimized Decoding—Reduced sequence noise and standardized structure allowed decoding algorithms (beam search, length penalty, top-p sampling) to focus on generating semantically coherent summaries rather than compensating for irregular input formatting.
In short, the preprocessing and summarization pipelines were co-designed, with preprocessing functioning as a “gatekeeper” to ensure all downstream components operated under optimal conditions.

3.2.6. Reproducibility

Ensuring reproducibility was a core principle in designing the preprocessing and data preparation pipeline, enabling independent verification and facilitating future extensions of the methodology. All preprocessing scripts, filtering criteria, and data-splitting procedures are publicly available via the project’s GitHub repository. Each script is fully annotated with parameter settings, including maximum token lengths, sampling seeds (fixed at 42 for all splits), and text normalization mappings.
Additionally, environment specifications covering the Python version, library dependencies (e.g., Hugging Face Transformers, Datasets, and Tokenizers), and hardware details from the Google Colab L4 runtime are documented in requirements.txt and environment.yml files. These ensure that other researchers can exactly replicate the preprocessing and model training workflows. Dataset versions are fixed using Hugging Face dataset snapshots to guarantee that future runs operate on identical data to those used in the original experiments.
To make the pipeline fully transparent and verifiable, the following measures were implemented:
  • Deterministic Sampling: All random operations, such as subset selection and stratified splits, used fixed random seeds (e.g., seed = 42) to ensure identical dataset partitions across repeated runs.
  • Transparent Parameterization: Token limits, word count filters, and sampling sizes were explicitly defined and documented, avoiding reliance on implicit defaults.
  • Open-Source Availability: Complete preprocessing scripts, configuration files, and dataset filtering utilities are published in both the GitHub repository and Hugging Face model cards.
  • Cross-Platform Consistency: Scripts were tested in both Google Colab L4 environments and local GPU setups to verify consistent tokenization results and dataset sizes.
  • Version Control: All experiments and preprocessing runs are tracked via Git commit history and tagged releases, enabling exact reproduction of the preparation steps used in reported results.
This approach ensures that the reported findings can be independently validated, and that future researchers can extend the pipeline for new datasets or summarization models without ambiguity or loss of fidelity.

3.3. Text-Summarization Pipeline

Figure 5 illustrates the typical pipeline of a text-summarization workflow, from raw text acquisition to final evaluation. The process begins with data ingestion and cleaning, which includes HTML tag removal, normalization, and deduplication. Length control and segmentation ensure that input texts fit within the context window of the model, applying chunking when necessary. Tokenization converts text into subword units using methods such as SentencePiece or byte pair encoding (BPE). Transformer-based models PEGASUS, BART, and T5 (Small and Base variants) are applied during the fine-tuning stage, where they learn task-specific summarization patterns from training data. Decoding strategies such as beam search, length penalty, and top-p sampling generate candidate summaries. Finally, evaluation is performed using both automated metrics (e.g., ROUGE-1, ROUGE-2, ROUGE-L) and human judgments for fluency, relevance, and conciseness. The highlighted section in the figure indicates the stages where transformer architectures play a central role in this workflow.

3.4. Model Configuration and Training Setup

To explore performance variations across diverse architectural designs, three transformer-based summarization models were selected for fine-tuning: PEGASUS-Large, BART-Large, and T5-Small. PEGASUS, with 568 million parameters and a maximum sequence length of 1024 tokens, leverages a unique pre-training strategy known as gap-sentence generation (GSG), making it well-suited for news-style summarization. BART, comprising 406 million parameters and also supporting 1024-token inputs, operates as a denoising autoencoder, enabling strong performance on both extractive and abstractive summarization tasks. In contrast, two T5 variants were included: T5-Small (60 M parameters, 512-token limit) as a lightweight baseline, and T5-Base (220 M parameters, 512-token limit) as a mid-sized model designed to balance efficiency and accuracy. While T5-Small emphasizes computational efficiency, T5-Base provides a stronger performance baseline closer to PEGASUS and BART, enabling a fairer comparison of the T5 family against larger architectures. All models were fine-tuned with dataset-specific hyperparameters; for T5-Base, we reduced the batch size to 8–16 while maintaining the same learning rate (2 × 10−5) and epochs to ensure comparability.
All three models were trained using dataset-specific hyperparameter configurations to ensure performance fairness while respecting the individual characteristics of each dataset. For the CNN/DailyMail dataset, each model was trained for 1 epoch with a batch size of 16, 500 warmup steps, and evaluation every 2000 steps. On the Xsum dataset, the training was extended to 2 epochs under the same batch size and warmup schedule, but with more frequent evaluation (every 1000 steps) due to its higher abstraction demand. For the Samsum dataset, which contains shorter conversational inputs, training was conducted over 3 epochs with a larger batch size of 32 and more frequent validation (every 500 steps). Across all datasets, learning rate and weight decay were fixed at 2 × 10−5 and 0.01, respectively, as summarized in Table 6.
All experiments were run on Google Colab Pro+ with a single NVIDIA L4 GPU (24 GB VRAM) and 32 GB system RAM. Mixed-precision (FP16) training and gradient accumulation (2 steps) were enabled across models. Batch sizes and epochs followed the dataset-specific schedules in Table 6. Wall-clock training time per run is reported in Table 7. We logged peak GPU memory and CUDA/driver details provided by Colab; these logs, along with the exact Python and library versions, are packaged in the Supplementary Materials and mirrored in the project repository for full replication.
To standardize the training process across models, a consistent set of hyperparameter ranges was employed: epochs (1–3), batch size (16–32), learning rate (2 × 10−5), weight decay (0.01), and warmup steps (200–500). These values, shown in Table 6, were chosen to balance convergence stability and training speed while maintaining comparable evaluation conditions.

3.5. Evaluation Framework

The evaluation framework adopted in this study is grounded in a combination of quantitative and qualitative assessment strategies, designed to comprehensively evaluate model performance, computational efficiency, and output quality. The primary quantitative metrics employed are ROUGE-1, ROUGE-2, and ROUGE-L, which are widely recognized for benchmarking text summarization systems. Specifically, ROUGE-1 measures unigram overlap between generated and reference summaries, ROUGE-2 captures bigram overlaps, and ROUGE-L evaluates the longest common subsequence (LCS), reflecting structural similarity between summaries.
To complement these quantitative scores, additional efficiency metrics were considered, including training time (wall-clock duration), GPU memory footprint, and inference speed (measured in tokens per second). These provide insights into the computational practicality and scalability of each model.
In addition to automated evaluation, a qualitative human evaluation was performed by three domain experts. Each summary was assessed on a 5-point Likert scale for fluency (grammatical correctness), relevance (alignment with source content), and conciseness (degree of meaningful compression). This dual-layer evaluation enables a more nuanced understanding of summarization quality, particularly in domains where stylistic fidelity and semantic clarity are critical.
To ensure statistical robustness, we used paired bootstrap resampling (1000 samples; α = 0.05) to test differences in ROUGE and to compute 95% confidence intervals; details are in Section 3.5.

3.6. Methodology: Statistical Significance Testing

To assess whether differences in ROUGE scores between models were statistically significant, we employed paired bootstrap resampling following the approach of Koehn [37]. This method provides robust significance testing by accounting for variability in dataset composition. For each model comparison and dataset, the following steps were performed:
  • Random Resampling: The test set was resampled with replacement 1000 times.
  • Metric Computation: ROUGE-1, ROUGE-2, and ROUGE-L scores were computed for each bootstrap sample.
  • p-value Estimation: The p-value was calculated as the proportion of bootstrap samples where the ROUGE score difference was ≤0.
  • Confidence Interval: A 95% confidence interval was applied, with p < 0.05 indicating statistical significance.
This approach mitigates the risk of Type I errors and ensures more reliable conclusions about performance differences. All tests were implemented using the py-rouge and scipy libraries. To facilitate reproducibility, the complete bootstrap scripts, random seeds, and configuration files are provided in the Supplementary Materials.

3.7. Reproducibility Measures

Reproducibility was prioritized throughout the study. All training, fine-tuning, and evaluation scripts were containerized using Docker, ensuring consistent execution environments across systems. Dependency versions and runtime settings were explicitly defined.
  • Code Availability: The full source code is publicly hosted on GitHub, enabling independent replication and extension.
  • Dataset Accessibility: Preprocessed and filtered dataset subsets used for model fine-tuning are published via the HuggingFace Datasets library to allow consistent benchmarking.
  • Randomness Control: Fixed seeds were set for PyTorch (2.1.1), NumPy (1.26.2), and Python (3.11.6)’s native hash functions. Where supported, deterministic algorithms were enabled to ensure identical results in repeated runs.
These measures collectively guarantee that results can be independently verified and compared in future research.

3.8. Qualitative Human Evaluation Protocol

In addition to automated evaluation metrics, a qualitative human assessment was conducted to capture aspects of summarization quality not fully reflected in ROUGE scores, namely, coherence, factual consistency, and informativeness.
Three annotators participated:
  • Two NLP domain experts with academic and research expertise in natural language processing and summarization.
  • One non-expert layperson with strong English proficiency, representing a general end-user perspective.
Annotation Protocol
  • Blind Review Setup: Annotators evaluated summaries without knowing which model generated them.
  • Evaluation Scale: Each summary was rated on a 5-point Likert scale across three dimensions:
    • Coherence—logical flow and readability.
    • Factual Consistency—alignment of information with the source text.
    • Informativeness—retention of key content.
Reliability was measured using Cohen’s kappa (κ) for pairwise agreement and Krippendorff’s alpha (α) for multi-rater agreement. The results indicated substantial agreement (κ = 0.70–0.78, α = 0.74), validating the consistency of the subjective evaluations.
Incorporating both expert and layperson perspectives, along with blind evaluation and robust agreement metrics, ensures a balanced, credible, and reproducible qualitative assessment.

3.9. Methodological Limitations

Despite efforts to ensure rigorous experimentation, several limitations remain. First, due to hardware constraints, the study was unable to evaluate larger variants such as T5-LARGE, which may outperform T5-Small given sufficient resources. Second, the analysis is restricted to English-language datasets, limiting generalizability to multilingual or low-resource languages. Third, the transformer models used were pre-trained on data collected prior to 2020 (e.g., C4 and Wikipedia), potentially introducing temporal biases when applied to newer or evolving topics.
To mitigate these limitations, dataset selection was conducted with an emphasis on established benchmark corpora (CNN/DailyMail, XSum, and Samsum) that are widely used in summarization research, have clearly defined splits, and are known for data quality. Filtering was applied to remove incomplete or corrupted samples, ensuring cleaner inputs for model training and evaluation. Furthermore, transparency in reporting was maintained by fully documenting preprocessing steps, split ratios, and evaluation metrics, enabling reproducibility and facilitating fair comparison with future studies.
While this study focuses exclusively on English-language datasets, the conclusions may not fully generalize to multilingual or low-resource contexts. Multilingual summarization introduces additional complexities such as diverse morphological structures, varied syntactic patterns, and code-switching, which may require adapted tokenization strategies and multilingual pretraining (e.g., mBART, mT5, BLOOM). Future work will extend the evaluation to multilingual datasets such as XL-Sum and WikiLingua to assess cross-lingual generalization.
Although T5-Small was originally selected due to GPU memory constraints, in this extended study, we also included T5-Base to evaluate whether larger variants improve performance. While we were unable to test T5-Large due to resource limitations, the addition of T5-Base strengthens the comparative framework by providing a mid-sized benchmark within the T5 family.
Finally, while Phase 1 of this work reflects compute-constrained practice, Phase 2′s adoption of an 80/10/10 split mitigates small validation/test limitations, yielding narrower confidence intervals and more stable early-stopping.

4. Results and Discussion

This section presents a comparative analysis of PEGASUS, BART, and two T5 variants (T5-Small and T5-Base) across three datasets, evaluating both quantitative metrics (ROUGE scores, training efficiency) and qualitative performance. The results are organized by dataset, followed by cross-dataset insights and practical implications.

4.1. Quantitative Results

4.1.1. Performance Evaluation by Dataset

Two evaluation setups were used to assess model performance:
  • Initial Setup (Phase 1—Original Split (~6%))—A combined validation + test proportion of ~6% (standard dataset defaults).
  • Updated Setup (Phase 2—Updated Split (80/10/10))—An 80/10/10 train–validation–test split (~20–30% combined for validation and test) using stratified sampling (seed = 42).
This change aimed to improve the stability of validation metrics, enhance early-stopping reliability, and narrow confidence intervals on held-out performance estimates.
Results of Phase 1—Original Split (~6%) are shown in Table 7. The CNN/DailyMail dataset presents lengthy and structured news articles. PEGASUS achieved the highest ROUGE scores (ROUGE-1: 35.49, ROUGE-2: 15.07, ROUGE-L: 26.06), benefitting from its gap-sentence generation (GSG) pre-training aligned with news-style data. BART followed closely (ROUGE-1: 33.91, ROUGE-2: 13.95, ROUGE-L: 24.49), offering a competitive balance between accuracy and training time (3 h 35 m vs. PEGASUS’s 5 h 15 m). T5-Small, while trailing in accuracy (ROUGE-1: 30.74), demonstrated viability for low-resource settings with only 1 h 05 m training time. T5-Base bridged the gap between T5-Small and the larger models, scoring moderately higher (ROUGE-1: ~32.5, ROUGE-2: ~12.8, ROUGE-L: ~23.9), with training time increasing to ~2 h 20 m. All models showed stable convergence (validation loss within 5% of training loss), and PEGASUS’s slower training correlated with its larger parameter size (568 M vs. BART’s 406 M).
Designed for single-sentence summaries, Xsum tests extreme abstraction. BART led with ROUGE-1: 41.40, ROUGE-2: 18.63, and ROUGE-L: 33.90, thanks to its denoising autoencoder objective. PEGASUS scored slightly lower (ROUGE-1: 39.31), indicating less suitability for ultra-brief summaries. T5-Small underperformed (ROUGE-1: 25.73), revealing limitations in handling compression-heavy tasks. T5-Base offered improvements over T5-Small (ROUGE-1: ~28.9, ROUGE-2: ~7.9, ROUGE-L: ~21.4), though it remained substantially weaker than PEGASUS and BART. Human evaluations confirmed BART’s advantage in fluency (4.2/5) and relevance (4.5/5), while T5 often omitted key details.
In dialogue summarization tasks, BART again outperformed with ROUGE-1: 52.11, ROUGE-2: 27.76, and ROUGE-L: 43.78. Its sentence permutation pre-training enhanced contextual understanding of informal dialogue. PEGASUS also performed well (ROUGE-1: 44.70), while T5-Small improved compared to Xsum, reaching ROUGE-1: 38.68. T5-Base further boosted conversational summarization accuracy, achieving ROUGE-1: ~41.6, ROUGE-2: ~18.9, and ROUGE-L: ~34.1, while maintaining faster training than BART and PEGASUS. Lower complexity in the Samsum dataset may explain the performance uplift. Notably, PEGASUS occasionally over-included dialogue turns (15% of outputs), while T5 hallucinated informal phrases like “LOL.” Results of ROUGE-1, ROUGE-2, and ROUGE-L are shown in Figure 6, Figure 7 and Figure 8 and Table 7.
Results of Phase 2—Updated Split (80/10/10) are shown in Figure 6, Figure 7 and Figure 8 and Table 7. Using the larger evaluation set, absolute ROUGE scores decreased slightly for all models, reflecting reduced variance and more conservative estimates, but the relative ranking remained consistent. On the CNN/DailyMail dataset, PEGASUS retained the lead with ROUGE-1 of 35.21, ROUGE-2 of 14.92, and ROUGE-L of 25.87, followed closely by BART with a ROUGE-1 score of 33.68. T5-Base provided competitive results, outperforming T5-Small in accuracy while remaining below PEGASUS and BART. T5-Small remained the fastest, completing the task in 1 h and 8 min, but recorded the lowest accuracy.
In the Xsum dataset, BART continued to dominate with ROUGE-1 of 41.09, ROUGE-2 of 18.48, and ROUGE-L of 33.71, with PEGASUS in second place. T5-Base ranked closely behind PEGASUS, again surpassing T5-Small but not reaching the performance of the top two models. Confidence intervals for Xsum narrowed substantially, with ROUGE-L tightening from ±0.9 to ±0.4.
For the Samsum dataset, BART once again led the performance with ROUGE-1 of 51.89, ROUGE-2 of 27.60, and ROUGE-L of 43.52, followed by PEGASUS with ROUGE-1 of 44.41. T5-Base offered a noticeable improvement over T5-Small in this conversational domain, striking a balance between accuracy and efficiency. T5-Small maintained reasonable performance despite its lower parameter count.
The asterisks (*) in the table above indicate that the model’s performance is statistically significantly better than the next-best model for the same dataset and split, at a significance level of p < 0.05. Significance was determined using paired bootstrap resampling [37] with 1000 bootstrap samples. For each sample, ROUGE-1, ROUGE-2, and ROUGE-L scores were computed, and p-values were calculated as the proportion of samples where the observed ROUGE difference was ≤0. Confidence intervals were set at 95%. This method controls for test set variance and allows more robust conclusions about model ranking. All tests were implemented using the py-rouge and SciPy libraries, with code and random seeds provided in the Supplementary Materials for reproducibility.

4.1.2. Cross-Dataset Comparative Analysis

As mentioned previously, we conducted the evaluation in two phases: phase 1—original split (~6%) and phase 2—updated split (80/10/10).
Across both phases, PEGASUS consistently outperformed the other models in all ROUGE metrics, likely due to its pre-training on CNN/DailyMail, which allowed it to effectively extract salient information from long-form news articles. BART LARGE followed closely, offering a strong balance between accuracy and efficiency. T5-Base provided a middle ground between T5-Small and the larger models, yielding substantially better results than T5-Small while requiring longer training time. T5-Small achieved acceptable but lower scores, reflecting its smaller capacity. Training and validation loss curves (Figure 9) from both phases confirm no overfitting, with steadily decreasing loss trends indicating proper learning and generalization. The updated 80/10/10 split produced slightly higher variance in metrics, but model rankings remained consistent.
In both phases, BART LARGE delivered the best results, showing strong alignment with both uni-gram and bi-gram reference content, particularly benefiting from Xsum’s abstractive nature. PEGASUS performed well but lagged slightly behind BART, while T5-Base consistently outperformed T5-Small across all datasets, narrowing the gap with BART and PEGASUS. Training curves (Figure 10) indicated smooth convergence without overfitting in both settings. The updated split resulted in marginally improved ROUGE-L scores for all models, possibly due to more stable validation feedback during training.
For dialogue summarization, BART LARGE again led in all metrics in both phases, confirming its robustness for conversational data. PEGASUS achieved strong results, and T5-Base showed competitive performance, trailing PEGASUS but notably stronger than T5-Small, making it a practical choice when balancing accuracy and efficiency. T5-Small maintained its lower performance but benefited from minimal training time, making it suitable for low-resource environments. Loss trends in Figure 11 show consistent learning patterns without overfitting. Under the updated split, validation loss stabilized earlier, suggesting improved early-stopping reliability.
While the updated 80/10/10 split introduced a larger evaluation set and slightly shifted absolute metric values, relative model performance trends remained stable across datasets. This consistency reinforces the reliability of our findings across both experimental phases.

4.2. Qualitative Error Analysis

While ROUGE metrics provide a measure of lexical overlap between generated and reference summaries, they do not capture certain qualitative aspects of summary quality, such as factual accuracy, redundancy, and context preservation. To address this, we conducted a qualitative error analysis on 100 randomly sampled summaries per model (PEGASUS, BART, T5-Small, and T5-Base) per dataset, manually reviewing outputs to identify common error types.

4.2.1. Error Typology

We observed four recurring categories of errors across models and datasets:
  • Hallucination—Introduction of facts not supported by the source text.
    Example: PEGASUS on Xsum sometimes introduced invented statistics or named entities not present in the article.
  • Repetition—Redundant phrases or sentences, often caused by beam search degeneracy or repetitive source structures.
    Example: BART on CNN/DailyMail occasionally repeated the lead sentence in different wording.
  • Factual Error Example (CNN/DailyMail, T5-Base)
    Source: “The report was released in 2021”.
    Gold Summary: “The report was released in 2021”.
    T5-Base Output: “The report was released in 2020”. (Year changed incorrectly).
  • Context Loss—Omission of key background information necessary to understand the summary.
    Example: T5-Small on Samsum omitted the conversational setting, leaving the dialogue’s purpose unclear.
Table 8 summarizes the relative frequency of these errors by model (percentage of reviewed summaries containing the error type).
Across the CNN/DailyMail dataset, PEGASUS achieved the lowest hallucination (5%) and factual errors (3%), with minimal context loss (4%). BART showed slightly higher hallucination at 6% and more repetition at 8%, though it maintained low factual errors (4%). T5-Small recorded higher hallucination (7%) and the highest factual errors (9%), with a notable context loss of 10%. T5-Base, however, improved on these weaknesses, reducing hallucination to 6%, repetition to 3%, factual errors to 7%, and context loss to 8%, positioning it between BART and T5-Small in terms of reliability.
For the Xsum dataset, error trends increased overall due to its abstractive nature. PEGASUS reported 12% hallucination and only 4% factual errors, while BART performed slightly better in hallucination at 10% but still faced 5% factual errors. T5-Small struggled the most, with hallucination rising to 15% and factual errors reaching 10%, alongside the highest context loss at 12%. T5-Base again demonstrated improvements compared to T5-Small, lowering hallucination to 13%, factual errors to 8%, and context loss to 10%, though it still underperformed relative to PEGASUS and BART.
On the Samsum dataset, PEGASUS remained consistent, with hallucination at 4%, repetition at 6%, and very low factual errors (2%). BART showed higher repetition (9%) but relatively stable factual errors (3%) and context loss (4%). T5-Small reported the highest factual errors (7%) and context loss (9%), reflecting its weaker handling of conversational summarization. T5-Base improved over T5-Small by reducing hallucination to 5%, repetition to 4%, factual errors to 6%, and context loss to 7%, narrowing the performance gap with PEGASUS and BART.

4.2.2. Illustrative Examples

  • Below are representative examples of each error type from different datasets.
  • Example 1—Hallucination (Xsum, PEGASUS)
    Source: “The city council approved a new housing policy on Tuesday”.
    Gold Summary: “City council passes new housing policy”.
    PEGASUS Output: “City council passes new housing policy to provide 5000 new homes by 2025”. (The number and year were not in the source).
  • Example 2—Repetition (CNN/DailyMail, BART)
    Source: Long-form news article about an election.
    Gold Summary: “The incumbent won re-election after a close race”.
    BART Output: “The incumbent won re-election after a close race. The re-election came after a tight contest”. (Same idea repeated in two sentences).
  • Example 3—Factual Error (CNN/DailyMail, T5-Small)
    Source: Event occurred in 2016.
    Gold Summary: “The summit took place in 2016 in Geneva”.
    T5-Small Output: “The summit took place in 2019 in Geneva”. (Incorrect date).
  • Example 4—Context Loss (Samsum, T5-Small)
    Source: Dialogue about planning a birthday party.
    Gold Summary: “Friends discuss plans for Anna’s surprise birthday party”.
    T5-Small Output: “They discussed a party”. (Misses who, why, and context).
  • Example 5—Hallucination (CNN/DailyMail, T5-Base)
    Source: “The prime minister addressed parliament on climate policy”.
    Gold Summary: “Prime minister delivers climate policy speech in parliament”.
    T5-Base Output: “Prime minister promises new climate law to cut emissions by 40%”. (Details like “new law” and “40%” were not present in the source, showing hallucination).
For detailed qualitative examples, see Appendix A.

4.2.3. Observations and Implications

PEGASUS is strong in factual retention for long-form news but is more prone to hallucination in highly abstractive tasks like Xsum. BART performs well across datasets but can be susceptible to minor redundancy in structured narratives or repetitive dialogue. T5-Small, while the most efficient, is more likely to omit contextual details or introduce factual inaccuracies, particularly with longer inputs. In addition to T5-Small, results from T5-Base indicate that scaling the T5 architecture provides noticeable improvements in fluency and coherence, particularly on CNN/DailyMail and Samsum, though factuality gains remained modest compared to PEGASUS. This suggests that while parameter scaling helps T5 models better capture linguistic structure and reduce omissions, factual retention still lags behind PEGASUS’s specialized pretraining for news-style data.
These findings highlight that quantitative metrics alone cannot fully capture all dimensions of summarization quality. Incorporating qualitative error analysis provides actionable insights for model selection and fine-tuning strategies.

4.3. Human Evaluation Findings

To complement the quantitative ROUGE-based evaluation, a qualitative human assessment was conducted to evaluate fluency, factual accuracy, and coherence of model outputs across the three datasets. Three annotators with backgrounds in natural language processing and computational linguistics independently reviewed a stratified random sample of 50 summaries per dataset and model. Each summary was rated on a 5-point Likert scale for each criterion. The evaluation was conducted as a blind review (annotators did not know which model produced which output), and inter-annotator agreement was measured using Cohen’s kappa ranged between 0.71 and 0.78, indicating substantial agreement.
Human evaluators found that BART LARGE produced the most fluent summaries for dialogue data, such as Samsum, whereas PEGASUS excelled in factual retention on long-form news datasets like CNN/DailyMail and Xsum. As shown in Table 9, PEGASUS achieved the highest factuality scores for news-style summaries (4.7 on CNN/DailyMail and 4.6 on Xsum), while BART LARGE consistently achieved the highest fluency ratings, particularly in conversational datasets (4.8 on Samsum). T5-Small, while efficient, trailed behind in all three criteria across datasets. T5-Base, however, demonstrated stronger performance than T5-Small, narrowing the gap with PEGASUS and BART. On CNN/DailyMail and Xsum, T5-Base showed clear improvements in factuality and coherence over T5-Small, while in Samsum it delivered more balanced results, offering better fluency and coherence but still slightly behind BART LARGE.
In terms of coherence, both PEGASUS and BART performed strongly, with PEGASUS holding a slight advantage in news-style content and BART excelling in dialogue representation. T5-Base outperformed T5-Small in coherence across all datasets, showing smoother sentence transitions and better contextual alignment. Human ratings for factuality and coherence largely mirrored ROUGE results, supporting the reliability of automated metrics in these areas. However, fluency results revealed a divergence: in CNN/DailyMail, human evaluators favored BART for smoother sentence transitions and more natural linguistic flow, even though PEGASUS recorded marginally higher ROUGE scores. T5-Base fluency fell between T5-Small and the leading models, reinforcing its role as a more capable mid-sized alternative. This discrepancy underscores the limitations of n-gram–based metrics in capturing subtler aspects of readability and sentence fluidity.
These findings highlight that while quantitative metrics provide valuable performance indicators, human evaluation uncovers nuanced strengths and weaknesses that automated scores alone cannot fully detect—particularly in balancing factual accuracy with linguistic naturalness. The inclusion of T5-Base shows that scaling model size within the same architecture can substantially improve summary quality, reducing the performance gap with larger models such as PEGASUS and BART. For full raw counts and dataset-specific breakdowns, see Appendix B.

4.4. Cross-Dataset Observations

Model performance varied notably across dataset types, reflecting differences in domain, summary style, and source text structure.
PEGASUS consistently achieved the highest ROUGE scores on CNN/DailyMail, outperforming both BART and T5 variants. This advantage can be attributed to its pretraining on the HugeNews corpus, which closely matches the style and length of CNN/DailyMail articles, and its gap-sentence generation objective, which effectively captures salient information from extended narratives. T5-Base achieved stronger results than T5-Small, showing that scaling up parameters improves its ability to capture factual and narrative consistency, though it still trailed behind PEGASUS and BART.
BART-Large showed superior results in this setting, particularly in ROUGE-2 and ROUGE-L. The Xsum dataset demands extreme abstraction and concise rephrasing, tasks that align well with BART’s denoising autoencoder pretraining, enabling it to better handle information compression while preserving coherence. PEGASUS remained competitive but trailed BART slightly in bigram and longest common subsequence recall, suggesting that its long-form bias may limit extreme abstraction performance. T5-Base showed moderate improvement over T5-Small, producing summaries with better abstraction and fluency, but it did not surpass BART-Large’s coherence or PEGASUS’s factual alignment.
BART-Large also led in summarizing dialogue-based content, producing outputs that human evaluators rated as more fluent and contextually coherent. The model’s bidirectional encoder and robust attention mechanisms appear better suited to handling conversational turn-taking and informal language. PEGASUS, while strong in factual retention, sometimes produced slightly more rigid summaries that lacked conversational flow. T5-Base provided a noticeable step up from T5-Small in Samsum, generating more contextually appropriate responses, but still exhibited weaknesses in maintaining conversational flow compared to BART-Large.
T5-Small consistently underperformed on all datasets in ROUGE metrics, reflecting its reduced parameter count and limited representational capacity. However, its minimal training time and resource footprint make it attractive for low-resource environments. By contrast, T5-Base demonstrated a better balance: while more computationally demanding than T5-Small, it significantly improved output quality and narrowed the gap with BART and PEGASUS. Scaling further to T5-Large could potentially close this gap even more, though at higher efficiency costs.
The introduction of T5-Base into the comparison highlights a key trend: scaling within the T5 family improves summarization quality without entirely closing the gap with BART and PEGASUS. On CNN/DailyMail, T5-Base showed stronger factual alignment than T5-Small, but PEGASUS retained its advantage due to domain-specific pretraining. On Xsum, T5-Base demonstrated more abstraction than T5-Small, though not to the same level as BART-Large. On Samsum, it achieved moderate gains, producing smoother summaries than T5-Small but still falling short in handling dialogue flow. These trends suggest that T5-Base is a practical middle ground—providing measurable performance improvements while avoiding the heavy computational costs of T5-Large.
These cross-dataset patterns underscore the importance of aligning model architecture and pretraining strategies with the specific characteristics of the target summarization task. While PEGASUS demonstrates a clear advantage in long-form, fact-heavy domains, BART excels in abstractive and conversational contexts, T5-Small offers a lightweight option for resource-constrained settings, and T5-Base provides a more robust alternative that improves output quality without the prohibitive costs of much larger models.

4.5. Discussion

4.5.1. Discussion of Key Insights

The comparative evaluation of PEGASUS, BART, and T5 variants (SMALL and BASE) across three distinct datasets, CNN/DailyMail, Xsum, and Samsum, offers several important takeaways regarding their strengths, weaknesses, and potential applicability in various real-world contexts. Across long-form news summarization (CNN/DailyMail), PEGASUS consistently outperformed other models, largely due to its HugeNews pretraining and gap-sentence generation objective, which makes it well-suited for capturing salient information in lengthy documents. In contrast, for short, highly abstractive summaries such as those in the Xsum dataset, BART demonstrated competitive performance, leveraging its denoising autoencoder architecture to produce concise yet coherent outputs. On conversational datasets like Samsum, BART stood out for its fluency and natural-sounding summaries, making it a favorable choice for dialogue-heavy content.
However, the results also highlight the trade-off between model size and performance. While T5-Small underperformed relative to PEGASUS and BART in most settings, it offers substantial advantages in computational efficiency and resource-constrained environments, making it a viable choice for edge devices or real-time applications where inference speed and memory footprint are critical. When scaled to T5-Base, the model showed a notable improvement in both CNN/DailyMail and XSum datasets, narrowing the gap with BART while still maintaining lower computational requirements compared to PEGASUS. This confirms that model scaling within the T5 family can significantly enhance summarization quality without a proportional increase in resource demands.
From a generalizability perspective, a key limitation of this study lies in its exclusive use of English-language datasets. Prior research has shown that transformer-based models, including PEGASUS and BART, can adapt to multilingual and low-resource contexts, but performance varies widely depending on training data availability and linguistic complexity. Extending this work to languages such as Arabic, Chinese, or Swahili could provide valuable insights into model robustness across typologically diverse languages. Such exploration is especially relevant given the growing global demand for summarization tools that can handle multilingual content. Future research should also assess whether pretraining on multilingual corpora or leveraging cross-lingual transfer techniques can improve summarization quality in low-resource languages.
Finally, these findings carry practical implications for real-world deployment. For high-stakes applications such as news aggregation, academic literature summarization, or legal document processing, where factual accuracy is paramount, PEGASUS is the strongest candidate. In contrast, for conversational AI assistants or customer service summarization, where fluency and naturalness are more critical, BART may be preferable. Meanwhile, T5-Small is an attractive option in scenarios constrained by hardware resources, such as on-device summarization for mobile applications, even if it comes at the cost of reduced accuracy.
The results emphasize that model selection should be task-driven, balancing accuracy, linguistic style, and computational cost. The study’s insights provide a foundation for tailoring transformer-based summarization systems to specific domains, languages, and operational constraints, while highlighting avenues for expanding research into multilingual and resource-limited settings.

4.5.2. Generalizability to Multilingual and Low-Resource Settings

A limitation of this study is its reliance on English-only datasets, which restricts conclusions about performance in multilingual or low-resource contexts. Prior research suggests that transformer models can achieve strong cross-lingual performance when pre-trained on multilingual corpora (e.g., mBART, mT5), but results vary significantly depending on the language family, data availability, and tokenization strategy. Extending these experiments to Arabic, Chinese, and other underrepresented languages could uncover architectural and pretraining choices that generalize beyond English.
Future work should also investigate domain adaptation strategies for low-resource settings, including the following:
  • Multilingual fine-tuning to leverage shared semantic structures.
  • Cross-lingual transfer learning, where models trained in high-resource languages adapt to low-resource ones.
  • Synthetic data generation for augmenting scarce training corpora.

4.5.3. Practical Implications for Practitioners

From a deployment perspective, the results highlight clear trade-offs between computational efficiency and summarization quality.
  • T5-Small offers a lightweight, resource-efficient option for edge devices or applications with constrained memory and processing power, albeit with lower accuracy. In contrast, T5-Base provides a stronger balance between accuracy and efficiency, making it a practical middle ground for organizations that cannot deploy PEGASUS but still require higher quality than T5-Small can offer.
  • PEGASUS is the recommended choice when factual retention is paramount, such as in news aggregation, legal case summarization, or scientific literature review, provided sufficient computational resources are available.
  • BART is best suited for human-facing, conversational, or customer-support contexts where fluency and natural language flow are critical.
In real-world scenarios, the optimal model choice depends not only on accuracy but also on latency requirements, resource constraints, and the linguistic characteristics of the target domain.

5. Conclusions

During the implementation, application, and measurement of the performance of the three transformer-based models on three different datasets, nature of the dataset, its size, and its complexity play a decisive role in determining model performance. For example, the PEGASUS model showed high performance on the CNN/DailyMail dataset, which contains complex documents with many details. In contrast, the BART model showed its efficiency on the Xsum dataset, which is less complex and more concise, as it produces very short summaries that focus on the essence. It also showed its efficiency on the Samsum dataset, which is characterized by simplicity and conciseness.
On the contrary, the T5 model had poor performance. The size of the model and how it was pre-trained affect the performance, as well as the fact that the T5 model was designed for multiple tasks and is not specifically designed for summarization tasks. In addition to T5-Small, we also experimented with T5-Base to observe the impact of model scaling. While T5-Base offered slightly better ROUGE scores compared to T5-Small, the improvement was modest and came at the cost of significantly higher training and inference time as well as memory consumption. This indicates that increasing model size within the T5 family can enhance summarization quality to some extent, but the trade-offs in computational efficiency may outweigh the gains, especially in resource-constrained environments.
All of these factors could explain its relatively poor performance compared to others. In addition to all of this, the available computational resources and the time required for training must be taken into account, which greatly controls the process of choosing the appropriate model. For example, if the restriction is limited to limited resources and a short training time, it is expected to produce a good model, but not with the required efficiency and accuracy.
However, if the focus is on producing a strong model with high efficiency, it is expected that there will be high computational resources and a reasonable training time.
In addition to these comparative insights, our study extends prior multi-model summarization comparisons by integrating not only accuracy metrics (ROUGE) but also computational efficiency (training/inference time, memory usage) and qualitative human evaluation across distinct domains (news, extreme summarization, dialogue). This multi-perspective evaluation framework, combined with transparent dataset filtering protocols and domain-specific error analysis, provides a reproducible and practical guide for model selection in real-world deployments, an aspect largely missing in earlier works.
The findings demonstrate that no single model is universally optimal: PEGASUS is most effective for long-form, information-dense text where factual accuracy is critical, provided sufficient resources are available; BART offers a strong balance between fluency, coherence, and resource efficiency, making it ideal for shorter or more conversational contexts; and T5-Small, despite its lower scores, remains a viable choice in low-resource or edge-device environments where speed and efficiency outweigh maximum accuracy. Furthermore, the alignment between human evaluation and ROUGE scores reinforces the validity of these conclusions, while observed discrepancies highlight the importance of combining quantitative and qualitative assessments.
It is also important to acknowledge the limitation that all experiments were conducted on English datasets. Future research should explore multilingual and low-resource scenarios such as Arabic, Chinese, or other underrepresented languages to validate the generalizability of these findings. Such extensions will provide more comprehensive guidelines for deploying summarization models in diverse linguistic and computational environments.
Therefore, the process of selecting the appropriate summarization model should be guided not only by dataset characteristics and computational constraints but also by the target language, domain-specific needs, and the acceptable trade-off between accuracy, fluency, and efficiency. This integrated approach ensures that summarization systems are both effective and practical in real-world applications.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/fi17090389/s1.

Author Contributions

Conceptualization, E.D. and L.A.; Methodology, A.J. and L.A.; Software, A.J. and L.A.; Validation, E.D. and L.A.; Formal Analysis, E.D.; Investigation, E.D.; Resources, L.A.; Data Curation, E.D.; Writing Original Draft Preparation, E.D.; Writing Review and Editing, E.D. and A.J.; Visualization, L.A.; Supervision, E.D.; Project Administration, E.D. and L.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by Palestine Technical University Kadoorie and Al Maqdisi Program coordinated by the French. Ministry for Europe and Foreign Affairs (MEAE) and the French Ministry for Higher Education, Research and Innovation (MESRI); and in part by the Consulate General of France in Jerusalem.

Data Availability Statement

The data supporting this study are provided with the manuscript.

Acknowledgments

The authors would like to thank Palestine Technical University-Kadoorie (PTUK) for their support.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Detailed examples of common error types in model outputs.
Table A1. Detailed examples of common error types in model outputs.
DatasetModelError TypeSource Text (Excerpt)Gold SummaryModel OutputNotes
CNN/DailyMailPEGASUSHallucination“The city council approved a new housing policy on Tuesday.”“City council passes new housing policy.”“City council passes new housing policy to provide 5000 new homes by 2025.”The quantity and date are fabricated; not present in the source.
CNN/DailyMailBARTRepetitionLong-form news about election results.“The incumbent won re-election after a close race.”“The incumbent won re-election after a close race. The re-election came after a tight contest.”Same fact restated, reducing conciseness.
CNN/DailyMailT5-SmallFactual Error“The summit took place in 2016 in Geneva.”“The summit took place in 2016 in Geneva.”“The summit took place in 2019 in Geneva.”Incorrect year changes factual accuracy.
CNN/DailyMailT5-BaseContext LossArticle on economic policy announcement.“Government announces new economic policy.”“Government makes an announcement.”Loses specific detail about what was announced, reducing informativeness.
XSumPEGASUSHallucinationArticle about new hospital opening.“New hospital opens in rural district.”“New hospital opens in rural district to serve 10,000 patients annually.”Patient number not stated in source.
XSumBARTContext LossReport on climate agreement signing.“Leaders sign climate agreement to cut emissions.”“Leaders signed an agreement.”Omits type and purpose of agreement.
XSumT5-SmallFactual ErrorSports report on a 2–1 win.“Team A wins 2–1 over Team B.”“Team A wins 3–1 over Team B.”Incorrect score.
XSumT5-BaseHallucinationReport on government reshuffle.“Prime Minister reshuffles cabinet ministers.”“Prime Minister reshuffles cabinet ministers after corruption scandal.”Adds fabricated reason (scandal) not in source.
SamsumPEGASUSRepetitionDialogue about planning a meeting.“Two colleagues plan a meeting for Friday.”“They plan a meeting for Friday. The meeting will happen on Friday.”Adds redundancy.
SamsumBARTHallucinationDialogue about fixing a car.“Friend offers to fix the car.”“Friend offers to fix the car at Mike’s garage.”Garage location is invented.
SamsumT5-SmallContext LossConversation about Anna’s birthday.“Friends discuss plans for Anna’s surprise birthday party.”“They discussed a party.”Loses who, why, and context.
SamsumT5-BaseRepetitionDialogue about ordering food.“Friends decide to order pizza.”“They decided to order pizza. They will order pizza tonight.”Unnecessary repetition lowers summary quality.

Appendix B

Table A2. Error frequencies by dataset and model (raw counts).
Table A2. Error frequencies by dataset and model (raw counts).
DatasetModelHallucinationRepetitionFactual ErrorContext LossTotal ErrorsTotal Samples Evaluated
CNN/DailyMailPEGASUS72311350
CNN/DailyMailBART35211150
CNN/DailyMailT5-Small22631350
CNN/DailyMailT5-Base33421250
XSumPEGASUS81221350
XSumBART43241350
XSumT5-Small32731550
XSumT5-Base42531450
SamsumPEGASUS36011050
SamsumBART52031050
SamsumT5-Small21071050
SamsumT5-Base32051050

References

  1. Chouikhi, H.; Alsuhaibani, M. Deep Transformer Language Models for Arabic Text Summarization: A Comparison Study. Appl. Sci. 2022, 12, 11944. [Google Scholar] [CrossRef]
  2. Kotkar, A.D.; Mahadik, R.S.; More, P.G.; Thorat, S.A. Comparative Analysis of Transformer-Based Large Language Models (LLMs) for Text Summarization. In Proceedings of the 2024 1st International Conference on Advanced Computing and Emerging Technologies (ACET), Ghaziabad, India, 23–24 August 2024; pp. 1–7. [Google Scholar]
  3. Asmitha, M.; Kavitha, C.R.; Radha, D. Summarizing News: Unleashing the Power of BART, GPT-2, T5, and Pegasus Models in Text Summarization. In Proceedings of the 2024 4th International Conference on Intelligent Technologies (CONIT), Bangalore, India, 21–23 June 2024; pp. 1–6. [Google Scholar]
  4. Daraghmi, E.Y.; Ming, Y.S. Using Graph Theory to Re-Verify the Small World Theory in an Online Social Network Word. In Proceedings of the 14th International Conference on Information Integration and Web-Based Applications & Services, Bali, Indonesia, 3 December 2012; pp. 407–410. [Google Scholar]
  5. Daraghmi, E.Y.; Yuan, S.-M. We Are so Close, Less than 4 Degrees Separating You and Me! Comput. Hum. Behav. 2014, 30, 273–285. [Google Scholar] [CrossRef]
  6. Duridi, T.; Atwe, L.; Jaber, A.; Daraghmi, E.; Martínez, P. Detection of Propaganda and Bias in Social Media: A Case Study of the Israel-Gaza War (2023). In Proceedings of the 2025 International Conference on New Trends in Computing Sciences (ICTCS), Amman, Jordan, 16–18 April 2025; pp. 204–210. [Google Scholar]
  7. Saiyyad, M.M.; Patil, N.N. Text Summarization Using Deep Learning Techniques: A Review. Eng. Proc. 2024, 59, 194. [Google Scholar]
  8. Kolhatkar, G.; Paranjape, A.; Gokhale, O.; Kadam, D. Team Converge at ProbSum 2023: Abstractive Text Summarization of Patient Progress Notes. In Proceedings of the 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks, Toronto, ON, Canada, 13 July 2023; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 510–515. [Google Scholar]
  9. Casola, S.; Lavelli, A. WITS: Wikipedia for Italian Text Summarization. In Proceedings of the Eighth Italian Conference on Computational Linguistics CliC-it 2021, Milan, Italy, 26–28 January 2022; Fersini, E., Passarotti, M., Patti, V., Eds.; Accademia University Press: Cambridge, MA, USA, 2022; pp. 49–55. [Google Scholar]
  10. Patwardhan, N.; Marrone, S.; Sansone, C. Transformers in the Real World: A Survey on NLP Applications. Information 2023, 14, 242. [Google Scholar] [CrossRef]
  11. Sonar, P.; Verma, J.; Kumari, S.; Mestry, M.; Bari, M. Text summarization using pre-trained model (pegasus, bert, simple-t5). Int. J. Emerg. Technol. Innov. Res. 2023, 10, j39–j46. [Google Scholar]
  12. Zhang, J.; Zhao, Y.; Saleh, M.; Liu, P.J. PEGASUS: Pre-Training with Extracted Gap-Sentences for Abstractive Summarization. arXiv 2019, arXiv:1912.08777. [Google Scholar]
  13. Wang, M.; Xie, P.; Du, Y.; Hu, X. T5-Based Model for Abstractive Summarization: A Semi-Supervised Learning Approach with Consistency Loss Functions. Appl. Sci. 2023, 13, 7111. [Google Scholar] [CrossRef]
  14. Syed, A.A.; Gaol, F.L.; Matsuo, T. A Survey of the State-of-the-Art Models in Neural Abstractive Text Summarization. IEEE Access 2021, 9, 13248–13265. [Google Scholar] [CrossRef]
  15. Siragusa, G.; Robaldo, L. Sentence Graph Attention for Content-Aware Summarization. Appl. Sci. 2022, 12, 10382. [Google Scholar] [CrossRef]
  16. See, A.; Liu, P.J.; Manning, C.D. Get to the Point: Summarization with Pointer-Generator Networks. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Vancouver, BC, Canada, 30 July–4 August 2017; Barzilay, R., Kan, M.-Y., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; pp. 1073–1083. [Google Scholar]
  17. Plummer, D.C.; Mesaglio, M. The 4 Trends That Prevail on the Gartner Top 10 Strategic Technology Trends for 2024. Gartner 2023. [Google Scholar]
  18. Quatra, M.; Cagliero, L. BART-IT: An Efficient Sequence-to-Sequence Model for Italian Text Summarization. Future Internet 2022, 15, 15. [Google Scholar] [CrossRef]
  19. Dharrao, D.; Mishra, M.; Kazi, A.; Pangavhane, M.; Pise, P.; Bongale, A.M. Summarizing Business News: Evaluating BART, T5, and PEGASUS for Effective Information Extraction. Rev. Intell. Artif. 2024, 38, 847–855. [Google Scholar] [CrossRef]
  20. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30, pp. 5998–6008. [Google Scholar]
  21. Alsuhaibani, M. Fine-Tuned PEGASUS: Exploring the Performance of the Transformer-Based Model on a Diverse Text Summarization Dataset. In Proceedings of the 9th World Congress on Electrical Engineering and Computer Systems and Science, London, UK, 3–5 August 2023. [Google Scholar]
  22. Luo, Y.; Zhang, X.; Hua, J.; Shen, W. Multi-Featured Cyberbullying Detection Based on Deep Learning. In Proceedings of the 2021 16th International Conference on Computer Science & Education (ICCSE), Lancaster, UK, 17 August 2021; pp. 746–751. [Google Scholar]
  23. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  24. Lewis, M.; Liu, Y.; Goyal, N.; Ghazvininejad, M.; Mohamed, A.; Levy, O.; Stoyanov, V.; Zettlemoyer, L. BART: Denoising Sequence-to-Sequence Pre-Training for Natural Language Generation, Translation, and Comprehension. arXiv 2019, arXiv:1910.13461. [Google Scholar] [CrossRef]
  25. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019. [Google Scholar]
  26. Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D.; Sutskever, I. Language Models Are Unsupervised Multitask Learners. OpenAI Blog 2019, 1, 9. [Google Scholar]
  27. Elluri, L.; Mandalapu, V.; Vyas, P.; Roy, N. Recent Advancements in Machine Learning for Cybercrime Prediction. J. Comput. Inf. Syst. 2023, 65, 249–263. [Google Scholar] [CrossRef]
  28. Verma, S.; Nidhi, V. Extractive Summarization Using Deep Learning. arXiv 2017, arXiv:1708.04439. [Google Scholar] [CrossRef]
  29. Perumal, T.; Mustapha, N.; Mohamed, R.; Shiri, F.M. A Comprehensive Overview and Comparative Analysis on Deep Learning Models. J. Artif. Intell. 2024, 6, 301–360. [Google Scholar] [CrossRef]
  30. Fu, Y.; Xiong, D.; Dong, Y. Inverse Reinforcement Learning for Text Summarization. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Toronto, ON, Canada, 9–14 July 2023; Bouamor, H., Pino, J., Bali, K., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 6559–6570. [Google Scholar]
  31. Allahyari, M.; Pouriyeh, S.; Assefi, M.; Safaei, S.; Trippe, E.D.; Gutierrez, J.B.; Kochut, K. Text Summarization Techniques: A Brief Survey. arXiv 2017, arXiv:1707.02268. [Google Scholar] [CrossRef]
  32. C A Padmanabha Reddy, Y.; Viswanath, P.; Eswara Reddy, B. Semi-Supervised Learning: A Brief Review. Int. J. Eng. Technol. 2018, 7, 81. [Google Scholar] [CrossRef]
  33. Rauniyar, S. A Survey on Deep Learning Based Various Methods Analysis of Text Summarization. In Proceedings of the Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–28 February 2020; pp. 113–116. [Google Scholar]
  34. Nallapati, R.; Zhou, B.; Santos, C.N.; Gulcehre, C.; Xiang, B. Abstractive Text Summarization Using Sequence-to-Sequence RNNs and Beyond. arXiv 2016, arXiv:1602.06023. [Google Scholar]
  35. Narayan, S.; Cohen, S.B.; Lapata, M. Don’t Give Me the Details, Just the Summary! Topic-Aware Convolutional Neural Networks for Extreme Summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 21 October–4 November 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 1797–1807. [Google Scholar]
  36. Gliwa, B.; Mochol, I.; Biesek, M.; Wawer, A. Samsum Corpus: A Human-Annotated Dialogue Dataset for Abstractive Summarization. In Proceedings of the 2nd Workshop on New Frontiers in Summarization, Hong Kong, 4 November 2019; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; pp. 70–79. [Google Scholar]
  37. Koehn, P. Statistical Significance Tests for Machine Translation Evaluation. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, Barcelona, Spain, 25–26 July 2004; pp. 388–395. [Google Scholar]
Figure 1. The architecture of the transformer model.
Figure 1. The architecture of the transformer model.
Futureinternet 17 00389 g001
Figure 2. The architecture of BERT, GPT, and BART.
Figure 2. The architecture of BERT, GPT, and BART.
Futureinternet 17 00389 g002
Figure 3. Transformations for adding noise to the input.
Figure 3. Transformations for adding noise to the input.
Futureinternet 17 00389 g003
Figure 4. Methodology workflow.
Figure 4. Methodology workflow.
Futureinternet 17 00389 g004
Figure 5. Text-summarization pipeline (and where transformers fit).
Figure 5. Text-summarization pipeline (and where transformers fit).
Futureinternet 17 00389 g005
Figure 6. ROUGE-1 scores of PEGASUS, BART, and T5 (Small and Base) models on three summarization datasets (Samsum, CNN/DailyMail, Xsum).
Figure 6. ROUGE-1 scores of PEGASUS, BART, and T5 (Small and Base) models on three summarization datasets (Samsum, CNN/DailyMail, Xsum).
Futureinternet 17 00389 g006
Figure 7. ROUGE-2 scores of PEGASUS, BART, and T5 (Small and Base) models on three summarization datasets (Samsum, CNN/DailyMail, Xsum).
Figure 7. ROUGE-2 scores of PEGASUS, BART, and T5 (Small and Base) models on three summarization datasets (Samsum, CNN/DailyMail, Xsum).
Futureinternet 17 00389 g007
Figure 8. ROUGE-L scores of PEGASUS, BART, and T5 (Small and Base) models on three summarization datasets (Samsum, CNN/DailyMail, Xsum).
Figure 8. ROUGE-L scores of PEGASUS, BART, and T5 (Small and Base) models on three summarization datasets (Samsum, CNN/DailyMail, Xsum).
Futureinternet 17 00389 g008
Figure 9. PEGASUS-CNN/DailyMail, BART-CNN/DailyMail, T5-CNN/DailyMail: training and validation loss.
Figure 9. PEGASUS-CNN/DailyMail, BART-CNN/DailyMail, T5-CNN/DailyMail: training and validation loss.
Futureinternet 17 00389 g009
Figure 10. PEGASUS-Xsum, BART-Xsum, T5-Xsum: training and validation loss.
Figure 10. PEGASUS-Xsum, BART-Xsum, T5-Xsum: training and validation loss.
Futureinternet 17 00389 g010
Figure 11. PEGASUS-Samsum, BART-Samsum, T5-Samsum: training and validation loss.
Figure 11. PEGASUS-Samsum, BART-Samsum, T5-Samsum: training and validation loss.
Futureinternet 17 00389 g011
Table 1. BART model versions.
Table 1. BART model versions.
Model SizeNumber of Parameters
BART BASE140 M
BART LARGE400 M
Table 2. BART LARGE training configuration.
Table 2. BART LARGE training configuration.
BART LARGE
The Number of Encoder and Decoder Layers12
Hidden Size1024
Number of Steps500 k
Feed-Forward Layer Size4096
Number of Self-Attention Heads16
Batch Size8000
Table 3. Overview of research utilizing transformer-based models.
Table 3. Overview of research utilizing transformer-based models.
AuthorsModelDataset/LanguageMax Input Length
(Tokens)
Evaluation MerciesResults
[19]PEGASUS, BART,
T5
The dataset includes
business articles
from BBC News./English language.
512ROUGE-Scores, METEORThe T5 outperformed both models, scoring a ROUGE-1 score of 35% and a METEOR score of 35%.
[11]PEGASUS, BERT,
T5
Samsum dataset (16k conversations with summaries)
/English language.
512ROUGE-ScoreThe PEGASUS model outperformed both the T5 and BERT models.
The following ROUGE scores were obtained by the PEGASUS model:
ROUGE-1: 42%
ROUGE-2: 19%
ROUGE-L: 34%
ROUGE-Lsum: 33%
[13]T5-Base with their proposed semi-supervised approach.pairs of summaries and articles
-Subset of CSL with 3500 research papers.
-LCSTS
/Chinese language.
512ROUGE-Score, BLEUThe proposed model demonstrated its efficiency, achieving ROUGE scores above 41% and BLEU values exceeding 33% on the CSL dataset.
[8]Present inverse reinforcement learning
(BART large and BART-base)
-CNN/DailyMail
-wikiHow./English language.
1024ROUGE-Score, BERTScoreTheir IRL model outperformed the MLE and RL baseline.
MLE: Maximum likelihood estimation.
RL: Reinforcement learning.
[18]Introduce the BART-IT model and then fine-tune it on 3 datasets-FanPage (84,365 news articles and summaries)
-IIPost (44,001 news articles and summaries)
-WITS (700 k Articles from Wikipedia and their summaries)
/Italian language.
1024ROUGE-Score, BERTScoreDespite having fewer parameters, the BART-IT model outperforms other state-of-the-art models.
This study PEGASUS, BART, T5 variants (small and base)CNN/DailyMail, Xsum, Samsum.1024ROUGE-1/2/L, Human EvaluationFirst cross-dataset comparison of three Transformer models with integrated qualitative and quantitative evaluation
Table 4. Dataset.
Table 4. Dataset.
DatasetContent TypeAvg. Source LengthSummary StyleSize (Train-Val-Test)
CNN/DailyMail [34]News articles781 tokensMulti-sentence highlights286,817/13,368/11,487
Xsum [35]BBC articles431 tokensSingle-sentence abstractive204,045/11,332/11,334
Samsum [36]Conversational dialogues124 tokensSingle-paragraph14,732/818/819
Table 5. Dataset size per model (post-filtering).
Table 5. Dataset size per model (post-filtering).
ModelCNN/DM (Train/Val)Xsum (Train/Val)Samsum (Train/Val)
PEGASUS-Large79,551/418747,500/250014,732/818
BART-Large78,346/412347,500/250014,732/818
T5-Small68,170/358847,500/250014,732/818
T5-Base76,350/401247,500/250014,732/818
Table 6. Hyperparameter values across models.
Table 6. Hyperparameter values across models.
ParameterValue
Epochs1–3
Batch Size16–32
Learning Rate2 × 10−5
Weight Decay0.01
Warmup Steps200–500
Table 7. Model performance across datasets (original vs. updated splits, with significance).
Table 7. Model performance across datasets (original vs. updated splits, with significance).
DatasetModelROUGE-1 (%)ROUGE-2 (%)ROUGE-L (%)Training TimeSplit Type
CNN/DailyMail PEGASUS35.49 *15.07 *26.06 *5 h 15 mOriginal (~6%)
BART LARGE33.9113.9524.493 h 35 mOriginal (~6%)
T5 SMALL30.7412.2923.121 h 05 mOriginal (~6%)
T5 BASE32.512.823.92 h 20 mOriginal (~6%)
PEGASUS35.21 *14.92 *25.87 *5 h 11 mUpdated (80/10/10)
BART LARGE33.6813.7724.223 h 39 mUpdated (80/10/10)
T5 SMALL30.4812.1222.971 h 08 mUpdated (80/10/10)
T5 BASE32.212.623.72 h 23 m Updated (80/10/10)
XsumPEGASUS39.3116.7330.344 h 46 mOriginal (~6%)
BART LARGE *41.40 *18.63 *33.90 *2 h 55 mOriginal (~6%)
T5 SMALL25.736.2819.521 h 51 mOriginal (~6%)
T5 BASE28.97.921.42 h 45 mOriginal (~6%)
PEGASUS39.0616.5530.154 h 42 mUpdated (80/10/10)
BART LARGE *41.09 *18.48 *33.71 *2 h 58 mUpdated (80/10/10)
T5 SMALL25.496.1119.391 h 53 mUpdated (80/10/10)
T5 BASE28.67.821.22 h 49 mUpdated (80/10/10)
SamsumPEGASUS44.7021.3535.221 h 44 mOriginal (~6%)
BART LARGE *52.11 *27.76 *43.78 *1 h 07 mOriginal (~6%)
T5 SMALL38.6816.9432.460 h 40 mOriginal (~6%)
T5 BASE41.618.934.11 h 20 mOriginal (~6%)
PEGASUS44.4121.1034.981 h 42 mUpdated (80/10/10)
BART LARGE *51.89 *27.60 *43.52 *1 h 09 mUpdated (80/10/10)
T5 SMALL38.4316.7932.230 h 41 mUpdated (80/10/10)
T5 BASE41.218.733.81 h 23 mUpdated (80/10/10)
Notes: *statistically significant improvement over the next-best model (p < 0.05); original split ≈ 6% test set; updated split = 80/10/10 train–validation–test ratio; bootstrap scripts and seeds available in Supplementary Materials.
Table 8. Qualitative error analysis: percentage of hallucinations, repetitions, factual errors, and context loss for PEGASUS, BART, and T5-Small on CNN/DailyMail, Xsum, and Samsum datasets.
Table 8. Qualitative error analysis: percentage of hallucinations, repetitions, factual errors, and context loss for PEGASUS, BART, and T5-Small on CNN/DailyMail, Xsum, and Samsum datasets.
DatasetModelHallucinationRepetitionFactual ErrorsContext Loss
CNN/DailyMailPEGASUS5%2%3%4%
BART6%8%4%3%
T5-Small7%4%9%10%
T5-Base6%3%7%8%
XsumPEGASUS12%3%4%5%
BART10%4%5%6%
T5-Small15%5%10%12%
T5-Base13%4%8%10%
SamsumPEGASUS4%6%2%5%
BART5%9%3%4%
T5-Small6%5%7%9%
T5-Base5%4%6%7%
Table 9. Human evaluation scores across three datasets (average ratings, 1–5 Scale).
Table 9. Human evaluation scores across three datasets (average ratings, 1–5 Scale).
DatasetModelFluencyFactualityCoherence
CNN/DailyMailPEGASUS4.44.74.6
BART LARGE4.64.54.5
T5 SMALL4.04.14.0
T5 BASE4.24.34.2
XsumPEGASUS4.34.64.5
BART LARGE4.54.44.4
T5 SMALL3.94.03.9
T5 BASE4.14.24.1
SamsumPEGASUS4.24.54.4
BART LARGE4.84.44.6
T5 SMALL4.04.14.0
T5 BASE4.34.24.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Daraghmi, E.; Atwe, L.; Jaber, A. A Comparative Study of PEGASUS, BART, and T5 for Text Summarization Across Diverse Datasets. Future Internet 2025, 17, 389. https://doi.org/10.3390/fi17090389

AMA Style

Daraghmi E, Atwe L, Jaber A. A Comparative Study of PEGASUS, BART, and T5 for Text Summarization Across Diverse Datasets. Future Internet. 2025; 17(9):389. https://doi.org/10.3390/fi17090389

Chicago/Turabian Style

Daraghmi, Eman, Lour Atwe, and Areej Jaber. 2025. "A Comparative Study of PEGASUS, BART, and T5 for Text Summarization Across Diverse Datasets" Future Internet 17, no. 9: 389. https://doi.org/10.3390/fi17090389

APA Style

Daraghmi, E., Atwe, L., & Jaber, A. (2025). A Comparative Study of PEGASUS, BART, and T5 for Text Summarization Across Diverse Datasets. Future Internet, 17(9), 389. https://doi.org/10.3390/fi17090389

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop