Next Article in Journal
Comparing Classification Algorithms to Recognize Selected Gestures Based on Microsoft Azure Kinect Joint Data
Previous Article in Journal
Leveraging Learning Analytics to Improve the User Experience of Learning Management Systems in Higher Education Institutions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PROMPTHEUS: A Human-Centered Pipeline to Streamline Systematic Literature Reviews with Large Language Models

by
Joao Torres
1,
Catherine Mulligan
2,
Joaquim Jorge
1 and
Catarina Moreira
3,*
1
INESC-ID, Instituto Superior Técnico, Universidade de Lisboa, 1649-004 Lisbon, Portugal
2
SST, Imperial College London, London SW7 2AZ, UK
3
Data Science Institute, University of Technology Sydney, Ultimo, NSW 2007, Australia
*
Author to whom correspondence should be addressed.
Information 2025, 16(5), 420; https://doi.org/10.3390/info16050420
Submission received: 4 April 2025 / Revised: 10 May 2025 / Accepted: 13 May 2025 / Published: 21 May 2025

Abstract

:
The growing volume of academic publications poses significant challenges for researchers conducting timely and accurate systematic literature reviews (SLR), particularly in fast-evolving fields like artificial intelligence. This growth of academic literature also makes it increasingly difficult for lay people to access scientific knowledge effectively, meaning academic literature is often misrepresented in the popular press and, more broadly, in society. Traditional SLRs are labor-intensive and error-prone, and they struggle to keep up with the rapid pace of new research. To address these issues, we developed PROMPTHEUS: an AI-driven pipeline solution that automates the systematic literature review (LR) process using large language models (LLMs). We aimed to enhance efficiency by reducing the manual workload while maintaining the precision and coherence required for comprehensive literature synthesis. PROMPTHEUS automates key SLR stages, including systematic searches, data extraction, and topic modeling using BERTopic and summarization with transformer models. Evaluations across five research domains demonstrated that PROMPTHEUS reduces review time, achieves high precision, and provides coherent topic organization, offering a scalable and effective solution for conducting literature reviews in an increasingly crowded research landscape.

Graphical Abstract

1. Introduction

The exponential growth of academic publications poses a significant challenge for researchers trying to keep up-to-date with developments in numerous fields. More than 2.5 million papers are published annually in approximately 30,000 accredited journals worldwide, making it increasingly challenging to filter relevant research efficiently. This is particularly evident in rapidly evolving domains such as artificial intelligence, virtual reality, and blockchain technologies [1]. As the research landscape becomes saturated, the visibility of new and impactful studies diminishes, complicating the synthesis of existing knowledge [2]. SLRs provide a structured approach to summarizing and synthesizing research, offering essential information on specific topics. However, traditional manual SLR approaches are labor-intensive, prone to human error, and increasingly unsustainable due to the overwhelming data volume.
Recent work has advanced automation in systematic literature reviews using transformer-based models. Nanggala et al. [3] review how transformers like BERT enable tasks such as document classification and semantic clustering in literature reviews. Scherbakov et al. [4] survey transformer models adapted for automated screening, summarization, and topic modeling, highlighting their growing role in streamlining review workflows.
Despite advances in AI-driven tools for literature searches and summarization, a significant gap remains in fully automating the systematic review process. Individual tools exist for searching, filtering, and summarizing content; however, they often operate in isolation, leaving researchers to handle other crucial stages manually, such as selecting relevant literature, extracting insights, and generating coherent reports. Additionally, practitioners face persistent challenges in SLRs, such as manual screening burden, reproducibility gaps, and reviewer fatigue. The scientific community lacks a fully integrated, scalable solution that automates the SLR pipeline. Such a solution must ensure accuracy and efficiency while reducing the manual burden on researchers.
In response to this need, we propose PROMPTHEUS: process optimization and management of papers using emerging technologies for high efficiency in updated systematic reviews. PROMPTHEUS is an automated framework that integrates large language models (LLMs) to automate key phases of the SLR process: systematic search and screening, data extraction, synthesis, and summarization. While the critical planning phase remains in the hands of researchers, PROMPTHEUS significantly reduces the manual workload, improving the final outputs’ precision, accuracy, and relevance, allowing researchers to focus more on the innovative aspects of their work.
The contributions of this work are the following:
  • Novel Integration of SLR Phases: We present a fully automated approach to SLRs, combining multiple stages, search, extraction, and synthesis, into an end-to-end process powered by advanced natural language processing (NLP) techniques.
  • Precision in Literature Retrieval: We leverage state-of-the-art language models to enhance the precision of literature searches. This ensures that researchers receive high-quality and relevant studies, addressing a critical need for accurate literature filtering.
  • Structured Topic Modeling: PROMPTHEUS employs BERTopic, a topic modeling technique that structures the extraction and organization of information, allowing for clear, well-organized reviews.
  • Comprehensive Evaluation: We present a robust evaluation using several metrics, including ROUGE scores, Flesch readability scores, cosine similarity, and topic coherence. These evaluations demonstrate the effectiveness of PROMPTHEUS in automating the SLR process while maintaining high accuracy and improving the readability of generated content.
By automating the most time-consuming aspects of SLRs, PROMPTHEUS aims to make SLRs more accessible, efficient, and comprehensive. This will ultimately enable researchers to devote more time to innovative, high-impact research while ensuring they remain up-to-date with critical developments. In addition, such tools may reduce the increasing mistrust in science by making summarization more accessible to laypeople.
By automating the most labor-intensive stages of systematic literature reviews, PROMPTHEUS enhances the accessibility, efficiency, and rigor of the review process. This enables researchers to focus more on creative and high-impact work while remaining current with field developments. Furthermore, by improving the clarity and accessibility of synthesized knowledge, PROMPTHEUS may help bridge the gap between academic research and public understanding, thereby addressing growing concerns about scientific mistrust.
The remainder of this paper is organized as follows: Section 2 reviews related work and outlines current limitations in automated SLR systems. Section 3 presents the PROMPTHEUS framework in detail, including its architectural components and underlying NLP techniques. Section 4 describes the experimental setup and evaluation metrics, followed by the results and analysis. Section 5 discusses the implications, limitations, and potential future enhancements. Section 6 presents the main technical limitations of our approach, while Section 7 discusses ethical implications. Section 8 presents the link with our open-source framework. Finally, Section 9 concludes the paper.

2. Background and Related Work

Systematic literature reviews are crucial for synthesizing research, identifying knowledge gaps, and shaping future directions across various domains. The traditional SLR process, defined by the PRISMA guidelines [5], consists of four phases: planning, selection, extraction, and execution [6]. Despite its rigor, this method faces increasing challenges due to the sheer volume of academic publications. The manual nature of SLRs makes them labor-intensive, prone to error, and difficult to scale, particularly as research outputs grow exponentially.

2.1. Advances in Automating Systematic Literature Reviews

Recently, machine learning (ML) and natural language processing have emerged as powerful tools that can assist with these challenges by automating various SLR process stages. Large language models such as T5 [7], GPT-3.5, GPT-4o, and GPT-o1 [8] have been integrated into the SLR workflow, particularly for tasks like literature search, data extraction, and summarization. These AI approaches, including technology-assisted review (TAR) systems, apply NLP and ML techniques to automate the search and screening phases, significantly reducing manual effort by iteratively refining models to prioritize relevant studies. This automation extends to data extraction, where NLP techniques ensure consistency and synthesis, where models such as T5 and GPT generate coherent summaries of research findings, enhancing accuracy and readability.
Moreno-Garcia et al. [9] propose a novel AI-based framework that leverages ensemble learning techniques to improve the accuracy and efficiency of study selection and data extraction processes. Their model demonstrates how ensemble techniques, when applied to AI-assisted systematic reviews, can enhance the precision and recall of study identification while reducing manual effort. This work contributes to AI-driven SLR tools by highlighting the potential for combining multiple AI models to tackle the inherent variability and challenges in automating complex tasks like data extraction and synthesis.
Bolanos et al. [10] conducted a comprehensive review of AI-integrated SLR tools, highlighting the efficiency improvements AI brings while emphasizing usability-related challenges. The authors highlight the need for user-friendly tools and strategies to manage LLM hallucinations, notably through knowledge injection techniques. Similarly, Saeidmehr et al. [11] proposed a spiral approach to systematic reviews, significantly improving screening efficiency in smaller datasets while also addressing gaps in handling unbalanced datasets and improving article acquisition.
Other AI-based models, such as the multi-agent AI system developed by Sami et al. [12], offer a promising approach by automating most steps in the SLR process. Their system uses LLMs to automate tasks like generating search strings and screening abstracts. However, while this approach reduces the manual workload, it still faces limitations in managing complex queries and ensuring the relevance of the selected studies.
Automation techniques are increasingly used in systematic reviews, reducing manual workloads by up to 7%, as noted by Tóth et al. [13]. However, challenges remain in recall consistency and real-world adoption. The study emphasizes the need for standardized evaluation metrics to better assess automation’s impact, showing that while promising, automation’s full potential is not yet realized due to technical and practical limitations.

2.2. Limitations of Current Automated SLR Systems

Despite the progress in AI-assisted SLRs, several limitations remain. Many systems struggle with handling complex queries, often relying on simple keyword searches that fail to capture the depth and specificity needed for comprehensive reviews. Additionally, the criteria for inclusion and exclusion are frequently poorly defined, leading to the retention of irrelevant or low-quality studies. Existing research highlights the need for more sophisticated search algorithms, improved Boolean logic integration, and better strategies for managing large datasets without sacrificing accuracy and relevance [14,15,16].
Machine learning tools, such as Abstrackr, have effectively automated SLRs’ title and abstract screening stages. In the study by Gates et al. [17], Abstrackr reduced manual effort by up to 35%, significantly saving time while maintaining a high accuracy level in identifying relevant studies. However, the tool missed some important studies during the screening process, underscoring a critical limitation of AI-assisted systems. Despite the efficiency gains, these tools still require human oversight to ensure that essential research is not inadvertently excluded. This highlights the balance between leveraging AI to reduce workload and ensuring that expert validation is in place to preserve the comprehensiveness and quality of the review process.
To support these findings, Cierco Jimenez et al. [18] reviewed a range of ML tools for automating the SLR process, noting that many tools lacked user-friendly interfaces for researchers without programming skills. Affengruber et al. [19] presented this finding, showing that while tools like Abstrackr and Rayyan can enhance efficiency, there is still a need for more comprehensive evaluations of their usability and impact in real-world scenarios. In addition, Perlman-Arrow et al. [20] evaluated an NLP tool for abstract screening during a SARS-CoV-2 review, reducing screening time by 33.7% while still requiring human oversight to ensure accuracy.
While AI has improved screening efficiency, challenges remain in automating tasks like data extraction and risk of bias assessment. Ofori-Boateng et al. [21] highlight that screening is the most automated phase, but more advanced AI techniques are needed for accurate data extraction and bias assessment. This shows that while automation reduces workloads, it still falls short in handling complex tasks in systematic reviews.
The role of human expertise remains critical in maintaining the rigor of systematic reviews, particularly when AI tools are not yet fully capable of handling the complexities of comprehensive research synthesis. Qureshi et al. [22] and Li et al. [23] both highlighted the limitations of ChatGPT and similar LLMs, showing that while these models excel in specific tasks like abstract screening, they often require expert validation to prevent the inclusion of irrelevant or erroneous studies. Rather than replacing human researchers, these tools are best viewed as assistants in keeping up to date with emerging new work in the field.

2.3. Advanced NLP and LLM Techniques

Recent studies demonstrate the potential of advanced NLP techniques in addressing some of the limitations of current automated SLR systems. For instance, Kharawala et al. [24] explored using zero-shot classification combined with ML algorithms to automate abstract screening, demonstrating high precision and recall. Similarly, Dennstädt et al. [25] tested LLMs for title and abstract screening in the biomedical domain, showing high sensitivity but noting challenges related to resource demands and biases.
To enhance AI’s role in systematic reviews, Hamel et al. [26] developed a framework for integrating AI into the title and abstract screening phases of SLRs, stressing the importance of robust training sets and transparent reporting. In parallel, Masoumi et al. [27] demonstrated the effectiveness of BioBERT, a variant of BERT (https://maartengr.github.io/BERTopic/ (accessed on 12 May 2025)), fine-tuned for biomedical texts, in automating the abstract review process in medical research, showing that such models can significantly reduce manual workloads while maintaining high accuracy.

2.4. Challenges of Rapid Reviews and Methodological Shortcuts

AI-based approaches have also been applied to rapid reviews (RRs), often employing methodological shortcuts to expedite the review process. Guo et al. [15] examined the impact of these shortcuts, showing that while they improve efficiency, they can introduce biases and reduce comprehensiveness. Speckemeier et al. [28] echoed these concerns, calling for more rigorous methodologies to balance the need for speed with the maintenance of review quality.
Moreover, O’Connor et al. [29] examined the cultural and practical challenges of adopting automation tools in systematic reviews, particularly in healthcare. Their study emphasized better collaboration between AI systems and human experts to ensure these tools are effectively integrated into existing workflows.

2.5. Addressing the Gap

Building on the limitations identified in existing automated systems, our work presents PROMPTHEUS. This fully automated SLR pipeline system enhances the review process by addressing critical limitations such as inadequate inclusion/exclusion criteria and complex query handling. PROMPTHEUS automates the selection, extraction, and synthesis phases, allowing researchers to manage the Planning phase while leveraging advanced NLP techniques like BERTopic for topic modeling and Sentence-BERT for sentence similarity. By incorporating LLMs like GPT and T5 for summarization and post-editing, PROMPTHEUS ensures that the generated summaries are accurate and coherent.

3. PROMPTHEUS: A Framework for AI-Driven SLRs

Despite significant advancements in AI-assisted SLRs, challenges remain in ensuring automated systems’ accuracy, scalability, and relevance. Our proposed framework, PROMPTHEUS, introduces an integrated and fully automated SLR framework that enhances SLRs’ selection, extraction, and synthesis phases while maintaining human oversight during the planning phase. PROMPTHEUS leverages advanced NLP techniques such as BERTopic for topic modeling and Sentence-BERT for sentence similarity to improve the precision and relevance of selected studies. Our system also integrates LLMs like GPT and T5 for summarization and post-editing, ensuring the generated summaries are accurate and coherent.
By introducing early stage inclusion and exclusion criteria, PROMPTHEUS improves the rigor of study selection and reduces the likelihood of including irrelevant papers. This approach addresses the shortcomings identified by O’Connor et al. [29] and de la Torre-López et al. [30], who emphasized the importance of integrating AI tools that improve efficiency without compromising the accuracy and comprehensiveness of systematic reviews, and also the challenges highlighted by Affengruber et al. [31] and Shaheen et al. [32] who stressed the importance of balancing efficiency with comprehensive, high-quality reviews.

3.1. General Overview

Our automated SLR pipeline architecture is organized into three interconnected phases: (1) systematic search and screening, which uses prompt-engineered GPT-4o and GPT-3.5 models along with TF–IDF filtering and named-entity recognition (spaCy) to identify and select relevant academic papers; (2) data extraction and topic modeling, which leverages Sentence-BERT embeddings for document clustering and BERTopic for unsupervised theme discovery; (3) synthesis and summarization, which employs hierarchical summarization via GPT-4o with chain-of-thought prompting and extractive–abstractive fusion to generate coherent summaries and integrate findings into a structured review document. Figure 1 presents the overall process.

3.2. Systematic Search and Screening Module

The systematic search and screening module is the foundation of the automated systematic literature review process, which automates retrieving and filtering academic papers based on a user-defined research question or topic. This module addresses the limitations of traditional literature search methods, which often require extensive manual effort, by using LLMs and advanced NLP techniques to enhance the efficiency and precision of the search process.
While PROMPTHEUS supports the automation of literature retrieval and initial filtering using semantic similarity and topic modeling techniques, it currently does not include explicit mechanisms for defining and enforcing inclusion or exclusion criteria. These criteria are foundational to systematic review methodology, ensuring the transparency and consistency of selection decisions. In its current version, PROMPTHEUS approximates relevance via document embeddings and topic relevance, which serve as proxies for human judgment. In future iterations, we plan to incorporate rule-based modules that allow users to define customizable inclusion/exclusion filters based on attributes such as publication type, methodological approach, or domain-specific keywords.
Research Topic Expansion: The module begins with the user providing a research question or topic as input. The system leverages an LLM (GPT-3.5, GPT-4, or GPT-4o) to expand the initial input into a more detailed and semantically rich set of keywords and phrases to ensure the search captures a comprehensive range of relevant studies. This expansion is guided by a carefully crafted prompt that instructs the model to retain the core focus of the research topic while adding appropriate keywords and terms to cover variations and related concepts. Part of the prompt used for this task is the following:
System: “You are a knowledgeable AI specializing in generating expanded titles for research topics. Your expanded titles should be concise and focus on capturing the core semantic meaning of a topic, suitable for creating informative embeddings for tasks like similarity comparisons.”

User: “Task: Generate a slightly expanded title for the following research topic, keeping the core focus while potentially adding 1–2 highly relevant terms for improved semantic representation.
Topic: title
Guidelines:
* Include essential keywords directly related to the topic.
* If necessary, add 1–2 closely related terms to capture topic variations.
* Avoid introducing new concepts or significantly altering the original title’s meaning.
* Keep the expanded title concise and focused on the core meaning.
Output format:
* Provide the expanded title only. Do not include any additional explanations or commentary.”
For instance, given the input topic “AI-based literature review”, the LLM might generate an expanded version such as “AI-based literature review, automated systematic reviews, natural language processing for academic research synthesis”. This expanded set of keywords ensures that the subsequent search covers a broader scope, capturing essential variations and closely related studies that might be overlooked.
Automated Query Generation: After expanding the research topic, the system constructs a structured search query tailored to the arXiv repository. This step is guided by another prompt instructing the LLM to craft a precise and targeted search query incorporating all relevant keywords and phrases. The model is asked to include fields such as title and abstract to refine the search further, ensuring that the retrieved literature aligns closely with the expanded topic. The prompt used for generating the search query is shown in the box below:
System: “You are a skilled research assistant specializing in crafting precise and effective search queries for the arXiv scientific paper repository.”

User: “Task: Craft an effective search query tailored for the arXiv database, specifically designed to retrieve research papers on the following topic:
Topic: ‘expanded_title’
Guidelines:
1. Concise and Precise: The query should be succinct yet accurately represent the core concept of the topic.
2. Key Terms: Incorporate the most relevant keywords or phrases directly associated with the topic.
3. Synonyms and Variants (Optional): If applicable, include synonyms or alternative terms to broaden the search scope and capture nuanced variations of the topic.
4. Specificity: Prioritize terms specific to the field or subfield to minimize irrelevant results.
5. arXiv Compatibility: Utilize operators like ‘ti:’ (title) and ‘abs:’ (abstract) to target specific fields within the arXiv entries.
Output format:
* Provide the ArXiv query only. Do not include any additional explanations or commentary."
The output of this prompt might produce a query such as (ti:“AI-based literature review” OR abs:(“AI-based literature review” OR “automated systematic reviews”)) AND (ti:“NLP” OR abs:“NLP”). This structured query is then used to search the arXiv database through its API, retrieving up to 3000 academic papers that match the specified criteria. Once the search results are obtained, the module pre-processes the retrieved papers by extracting essential details such as paper ID, title, and abstract. The text is cleaned to ensure consistency and readability by removing unnecessary symbols and normalizing the format. This clean text is then used in the next stage of the module, where relevance filtering is performed.
Relevance Filtering Using Sentence Similarity: The module employs a similarity-based mechanism using Sentence-BERT embeddings to filter the most pertinent papers from the initial search results. It computes vector embeddings for both the expanded research topic and the cleaned abstracts of the retrieved papers. The cosine similarity between these embeddings is then calculated to assess the relevance of each paper. The top 200 papers with the highest similarity scores are selected for further analysis, ensuring the final literature set is focused and comprehensive. This structured approach significantly reduces manual effort while improving the quality and relevance of the selected studies, providing a robust foundation for subsequent stages.
Once the relevant papers are identified through a systematic search and screening, the next step is to organize these documents into coherent themes using the data extraction and topic modeling module.

3.3. Data Extraction and Topic Modeling Module

The data extraction and topic modeling module automates organizing and categorizing selected academic papers into meaningful topics based on semantic content. The module leverages topic modeling and language generation techniques to create a structured literature representation, making identifying key research themes and subtopics easier. The module’s core components include topic modeling and document clustering, keyword extraction and title generation, and topic report generation.
PROMPTHEUS distinguishes between two modes of information retrieval in the automated SLR pipeline: quantitative and qualitative. Quantitative retrieval refers to the extraction of structured metadata such as publication years, venues, author affiliations, or keyword frequencies. These elements are reliably parsed from bibliographic entries and document metadata using regular expressions or semantic filters. In contrast, qualitative retrieval, such as identifying key contributions, theoretical framing, or methodological limitations, requires interpretive reasoning. PROMPTHEUS employs transformer-based summarization and topic modeling to approximate this interpretive synthesis. While these tools offer scalable generation of thematic overviews, we acknowledge their current limitations in replicating the nuanced judgment of domain experts. As such, PROMPTHEUS is best viewed as an augmentative tool that can expedite preliminary synthesis, rather than fully substitute human interpretation.
Topic Modeling and Document Clustering: Once the most relevant documents are selected from the initial screening phase, this module initiates by creating embeddings for the textual content of each document using a Sentence-BERT model. These embeddings capture the semantic information of the documents, allowing for an effective clustering of papers based on their conceptual similarities. Topic modeling uses the BERTopic algorithm, which groups documents into coherent clusters reflecting the selected literature’s primary themes. The number of topics and the minimum topic size are dynamically adjusted based on the size and content of the dataset to ensure that the generated topics are both meaningful and interpretable.
Keyword Extraction and Title Generation: After clustering the documents into distinct topics, the system extracts keywords for each topic, summarizing the main themes in that cluster. The keywords are input into a language model, such as GPT-3.5, GPT-4, or GPT-4o, to generate concise and descriptive titles for each topic. This process is guided by a structured prompt instructing the language model to create topic titles that accurately represent the essence of the keywords while maintaining clarity and relevance. The prompt used for this task is as follows:
System: “You are an experienced researcher specializing in literature reviews. You are adept at crafting concise, informative, and engaging topic names for subsections that accurately reflect the content and guide the reader.”

User: “Task: Create a clear and concise topic name for a subsection in a literature review. The subsection covers the following keywords: topic_keywords:”
Guidelines:
* Length: Aim for 1–5 words.
* Accuracy: Ensure the topic name precisely reflects the keywords’ meaning.
* Relevance: The name should fit within the broader context of a literature review.
* Informativeness: Clearly indicate the subsection’s focus to the reader.
* Engagement: Make the topic name interesting and inviting to read.
Optional: If the keywords are too broad or ambiguous, suggest a more specific or narrowed-down focus within the topic.
Output format:
* Provide the topic title only. Do not include any additional explanations or commentary.
For instance, if the extracted keywords for a topic are “deep learning, neural networks, image recognition”, the generated title might be “Deep Learning for Image Recognition”. This descriptive title provides an overview of the underlying theme of the clustered documents, making it easier for researchers to navigate through the literature.
Topic Report Generation: After generating the titles, the system compiles a comprehensive report that includes the list of documents under each topic, the topic keywords, and the generated titles. This hierarchical organization of literature enhances the comprehensiveness and accessibility of the review, as it delineates different research themes and subtopics, making it easier for researchers to identify key trends and gaps in the literature. The module’s process is further supported by a series of iterations and parameter adjustments to refine the topic modeling. If the initial number of topics is too few or too many, the system dynamically tunes the parameters, such as the number of issues or the minimum size of a topic, to achieve optimal clustering.
In general, this module significantly improves the efficiency and effectiveness of the systematic literature review process by automating the categorization of papers and generating meaningful insights into the core themes of the literature. Automates document categorization, offering researchers valuable insights into the core themes of the literature and simplifying the identification of key trends and gaps.

3.4. Synthesis and Summarization Module

The synthesis and summarization module generates concise and coherent summaries for each identified topic cluster, significantly reducing the manual effort typically required in literature review processes. This module utilizes transformer-based models, such as T5, to summarize abstracts and GPT-based models for post-editing, ensuring that the resulting content is well-structured and easy to understand.
Abstract Summarization with T5: The process begins by generating summaries for individual abstracts within each topic cluster using a transformer-based model like T5. This model is specifically configured to produce short yet comprehensive summaries that capture each document’s key contributions and findings. The generated summaries retain essential details while significantly reducing the length of the original abstracts, making it easier to synthesize large volumes of research.
Topic-Level Summarization and Aggregation: After individual summaries are generated, they are aggregated into a comprehensive summary for each identified topic. This step synthesizes the insights from multiple papers within the same topic, offering a holistic view of the research contributions, trends, and open questions. The aggregated summaries provide a structured narrative highlighting the most significant findings across multiple studies.
Our method proceeds in four distinct steps:
  • Cluster Identification: We first group documents into thematic clusters using BERTopic. This step ensures that related studies are analyzed together, forming the basis for coherent summaries.
  • Extractive Summary Generation: For each cluster, we apply GPT-4o in extractive mode to select representative sentences. By focusing on key sentences, we preserve the most salient findings with minimal noise.
  • Abstractive Refinement: The extracted sentences are then fed into GPT-4o with chain-of-thought prompting to produce concise, fluent summaries. This refinement enhances readability and integrates insights across papers.
  • Aggregation into Narrative: Finally, we combine the abstractive summaries into the review’s thematic narrative, adding signposting and context.
This aggregation step transforms discrete summaries into a cohesive section suitable for publication.
Post-Editing and Refinement with GPT: To enhance the clarity, coherence, and flow of the aggregated summaries, a GPT-based model is employed for post-editing. The refinement process involves using a predefined prompt instructing GPT to improve readability and structure while preserving critical information. This step ensures that the final summaries are well-organized and suitable for inclusion in a structured literature review document. The following prompt is used for post-editing:
System: “You are an expert researcher specializing in literature reviews in the field of title. Your task is to meticulously refine and enhance machine-generated summaries of multiple research papers.”
User: Refine the following machine-generated summary for the section “section_name” in a literature review titled “title”
The original summary is a compilation of various papers. Please focus on retaining the most relevant information for this literature review section.
Crucially, ensure the inclusion of in-text citations (e.g.,
citepkadir2024revealing) for all information directly sourced from the referenced documents. Feel free to shorten the section summary if it enhances clarity and conciseness, but prioritize keeping essential details and all relevant citations.
Original Summary: summary
Output format:
* Provide only the revised summary. Do not include any additional explanations or commentary.
This refinement results in a more precise and cohesive summary that better communicates the core literature of the topic.
Document Compilation and Report Generation: The final step is compiling the generated summaries and topics into a coherent literature review document. This module integrates all the synthesized content into a structured LaTeX document, which includes an introduction, background information, detailed literature synthesis for each topic, and a conclusion. The system also generates a BibTeX file with the references for all included papers, ensuring proper citation and academic integrity.
The document generation process uses GPT, ensuring the final output is professionally formatted and adheres to the desired layout and style. The module supports various formats for exporting the final report, including LaTeX and PDF, providing researchers with a polished, ready-to-use literature review.

4. Experimental Setup

The proposed automated SLR framework was evaluated using a comprehensive experimental setup to assess its performance across different stages of the review process. We used five distinct research topics for the experiments: “Explainable Artificial Intelligence (XAI)”, “Virtual Reality (VR)”, “Blockchain”, “Large Language Models (LLMs)”, and “Neural Machine Translation (NMT)”. Each experiment focused on a specific phase of the proposed SLR framework: systematic search and screening, data extraction and topic modeling, and synthesis and summarization.
Datasets: We conducted experiments using five different research topics, each representing a unique area of academic research: explainable artificial intelligence, virtual reality, blockchain, large language models, and neural machine translation. We collected the papers for each research topic from the arXiv database. We retrieved papers based on search queries generated by GPT-3.5 and GPT-4o models, with a maximum limit of 3000 papers per query.
Experiments: We designed four experiments to assess the system’s performance across different phases: systematic search and screening, data extraction and topic modeling, synthesis and summarization, and document compilation and report generation. We reported the results using various metrics, including topic coherence, ROUGE scores, readability scores, and cosine similarity.
Readability Analysis: We evaluated the readability of the generated summaries and final LaTeX documents using the Flesch Reading Ease Score (FRES). The Flesch Reading Ease Score [33] provides insight into how easily a text can be read and understood. Higher FRES scores indicate simpler reading material, while lower scores denote more complex and challenging passages. We computed FRES at different stages of the summarization and document generation process to assess how readability changes as the content is processed through T5 summarization, GPT post-editing, and final document generation.
Metrics: To evaluate the quality and robustness of the proposed framework, we used the following metrics:
  • Topic coherence measures the semantic similarity between words in a topic, indicating how well the generated topics represent coherent and interpretable concepts. A higher coherence score suggests that the words within each topic are more closely related, making the topics more useful and understandable for further analysis [34].
  • ROUGE stands for recall-oriented understudy for gisting evaluation. It compares an automatically produced summary or translation against a set of reference summaries (typically human-produced). ROUGE evaluates various aspects, such as the overlap of n-grams, word sequences, and word pairs between the machine-generated output and the reference.
  • ROUGE-1 measures the overlap of unigrams (single words) between the generated and reference abstracts. ROUGE-1 is particularly useful for evaluating summarization techniques because it captures the essential content and ensures that key information from the original text is retained in the summary.
  • Precision for ROUGE-1 measures the fraction of relevant instances among the retrieved cases, indicating how much of the generated summary is present in the reference text, which is the abstract in our case.
  • Recall for ROUGE-1 measures the fraction of instances retrieved over the total number of cases in the reference, indicating how much of the reference abstract is covered by the generated summary.
  • F1-score for ROUGE-1 is the harmonic mean of precision and recall, providing a balance between the two metrics.
  • Cosine similarity measures the similarity between two non-zero vectors of an inner product space, effectively capturing the semantic closeness between the generated text and the reference text. Cosine similarity was used to evaluate the semantic alignment of abstracts with expanded topics during the systematic search and screening phase.
  • Flesch Reading Ease Score (FRES) [33] provides insight into how easily a piece of text can be read and understood. The FRES formula considers sentence length and syllable count, with higher scores indicating simpler and more accessible text. We computed the FRES for three stages: T5-generated summaries, GPT post-edited sections, and the final LaTeX document. The formula is as follows:
    FRES = 206.835 1.015 total words total sentences 84.6 total syllables total words .
    This formula provides a measure of how easy a text is to read. Higher scores indicate easier-to-read material, while lower scores denote more difficult passages.
  • The number of papers retrieved indicates the coverage of the search query and its ability to find relevant literature.
  • The number of papers filtered reflects the number of papers that passed an initial relevance filter based on the research topic.
  • The total CPU time is the computational time required for generating queries, retrieving papers, and filtering results.
Hardware: Experiments were conducted in a Google Colab environment using an Intel Xeon CPU @ 2.20 GHz (2 cores, 56 MB cache), with 12.7 GB of RAM and 107.7 GB of disk space.

4.1. Experiment 1: Systematic Search and Screening

This experiment evaluated the effectiveness of GPT-3.5 and GPT-4o in generating queries for retrieving research papers from the arXiv repository. Given their capabilities in generating structured and contextually rich queries, we sought to compare the two models regarding their retrieval performance, efficiency, and computational cost. The experiment aimed to identify which model performs better across various research topics.
We selected five diverse research topics for this evaluation: explainable artificial intelligence, virtual reality, blockchain, large language models, and neural machine translation. For each topic, we measured three key performance indicators: the number of papers retrieved, the number of papers filtered, and the CPU time.
The five selected topics, explainable AI, virtual reality, blockchain, large language models, and neural machine translation, represent high-volume research areas within the AI and computing domains. Query generation was based on a consistent prompt template used with both GPT-3.5 and GPT-4o.
Results and Analysis: The results are summarized in Table 1. GPT-4o consistently retrieved more papers than GPT-3.5 across all topics, indicating that GPT-4o generates comprehensive and relevant queries more effectively. For instance, GPT-4o retrieved 2833 papers for “Virtual Reality” compared to 1986 papers retrieved by GPT-3.5. Similarly, for “Explainable Artificial Intelligence”, GPT-4o retrieved 1712 papers, surpassing the 1287 papers retrieved by GPT-3.5.
While GPT-4o demonstrated superior retrieval capability, it also required significantly more computational time than GPT-3.5. For instance, the “Explainable Artificial Intelligence” topic took 1555 seconds to process using GPT-4o, whereas GPT-3.5 completed the same task in 1213 seconds, a difference of nearly 6 minutes. Similarly, GPT-4o required 2115 seconds to process the “Large Language Models” topic, which is approximately 10 minutes longer than GPT-3.5.
These results suggest that GPT-4o is more effective at generating queries that yield a more extensive set of relevant papers, making it well-suited for scenarios where comprehensive literature coverage is a priority. However, this increased retrieval capability comes at the cost of longer computational time, making GPT-4o less ideal for scenarios where efficiency and speed are critical considerations.
In conclusion, GPT-4o is preferable for use cases prioritizing comprehensive retrieval over computational efficiency, while GPT-3.5 may be better for time-sensitive applications. This insight provides a basis for selecting the appropriate LLM based on the specific requirements of different phases in the systematic literature review process.

4.2. Experiment 2: Data Extraction and Topic Modeling

This experiment evaluated the quality of topics generated during the data extraction and topic modeling phase. The goal was to determine how well the BERTopic algorithm organized the retrieved literature into meaningful and coherent themes.
We used the topic coherence metric from Gensim to measure the quality of the generated topics. Topic coherence quantifies the semantic similarity between words within a topic, indicating how well the topics represent coherent and interpretable concepts. This measure has been validated as a reliable method for assessing topic models in previous work by Röder et al. [35]. Their study evaluated over 237,912 coherence measures across six benchmark datasets and demonstrated that specific combinations of coherence metrics correlate highly with human ratings, setting a standard for evaluating topic models.
Our experiment applied Gensim’s implementation of the coherence metric to assess the topics generated from documents retrieved using GPT-3.5 and GPT-4o queries. This metric ensures that the topics produced are statistically sound and interpretable to human evaluators.
Results and Analysis: This experiment assessed the semantic coherence of topics generated from the documents retrieved using GPT-3.5 and GPT-4o queries. As shown in Table 2, coherence scores for most topics fall between 0.4 and 0.5, indicating a moderate level of topic quality. This range suggests that the topics are generally coherent and interpretable but could be further refined.
For instance, the topic coherence score for “Explainable Artificial Intelligence” was 0.467 using GPT-3.5 queries and 0.422 using GPT-4o queries, indicating that both models produce moderately coherent topics. Similarly, for “Virtual Reality”, GPT-4o achieved a coherence score of 0.481 compared to 0.434 by GPT-3.5, showing that GPT-4o produced slightly better-organized topics for this research area.
Although these scores indicate that the generated topics are generally coherent, they are lower than previous benchmarks, such as the BERTopic model, which achieved scores of 0.681 and 0.432 on different datasets, as reported by Rahimi et al. [34]. This suggests that while our system can generate meaningful topics, there is potential for further improvements in topic coherence to match or exceed these higher benchmark scores.

4.3. Experiment 3: Synthesis and Summarization

This experiment assessed the performance of the synthesis and summarization phase of our automated literature review framework. We evaluated the quality of the generated summaries using ROUGE scores to determine their relevance and content retention. Additionally, the readability of each summary was analyzed using the Flesch Reading Ease metric. The primary goal was to determine how effectively the system condenses and synthesizes information from multiple research papers while maintaining coherence and relevance.
ROUGE Score Analysis: We used the ROUGE-1 metric to compare the content overlap between the machine-generated summaries and the abstracts of the selected research papers, which served as reference texts. ROUGE-1 measures the degree of overlap in unigrams (single words) between the generated summaries and reference texts, making it suitable for evaluating content retention and relevance.
The evaluation was conducted in three stages:
  • Abstract Generated Summaries using T5: These serve as the baseline summaries generated by the T5 model, which captures the core content of the abstracts.
  • Post-Edited Generate Summaries using GPT: GPT-based models refine these summaries to enhance readability, coherence, and overall structure.
  • Document Compilation and Report Generation using LaTeX: Comprehensive sections formatted as LaTeX documents that integrate information from multiple summaries, providing a cohesive and structured literature overview.
We computed ROUGE-1 precision, recall, and F1-scores for each stage. While all three metrics provide valuable insights, we focused primarily on precision. High precision indicates that the summaries retain the most pertinent information from the reference abstracts, minimizing irrelevant details.
The results in Table 3 indicate that both GPT-3.5 and GPT-4o models achieved high precision (P) scores across all inputs, demonstrating that the generated summaries contain a significant proportion of relevant content present in the reference abstracts. However, the recall scores are relatively lower, reflecting that not all content from the reference abstracts is captured in the summaries. This is expected since we want to capture only the relevant information in the abstracts. In this table, there is no significant benefit in choosing GPT-4o instead of GPT-3.5, as the T5 model computes the summary. GPT, at this stage, was only used to gather documents by creating the ArXiv query.
Post-Editing Stage Results: The post-editing phase is crucial in refining the machine-generated summaries produced by the T5 model. This stage utilizes GPT-based models to enhance the initial summaries’ clarity, coherence, and structure. The objective is to condense and reorganize the content while preserving the most relevant information. Post-editing is essential for transforming raw summaries into well-structured sections that align with the broader context of an SLR.
Table 4 presents the ROUGE-1 scores for the summaries after being refined by GPT-based models.
The results show a clear drop in recall after post-editing compared to the initial T5-generated summaries, reflecting the focus on refining and condensing content. This reduction is expected as the goal is to produce well-structured, concise sections for the systematic literature review (SLR). Despite the decrease in recall, precision scores remain high, ensuring the retained information is relevant and concise.
GPT-4o demonstrates higher recall than GPT-3.5, indicating its ability to retain more content during post-editing. However, the F1-scores, which balance precision and recall, show only a slight advantage for GPT-4o, suggesting that both models perform similarly in maintaining a good balance between content relevance and retention.
Post-editing with GPT significantly improved the precision of the summaries, ensuring that the most relevant information was retained, even though recall slightly decreased. GPT-4o showed a slight edge in content retention, making it more suitable for comprehensive literature reviews.
Final LaTeX Document Evaluation: The system’s final LaTeX documents were evaluated to determine their effectiveness in creating cohesive literature review sections that integrate information from multiple sources.
As presented in Table 5, both GPT-3.5 and GPT-4o achieved exceptionally high precision scores (e.g., 0.991 for GPT-3.5 and 0.989 for GPT-4o in the “Explainable Artificial Intelligence” topic), indicating that the final documents are highly aligned with the reference abstracts in terms of relevance. However, the recall scores for these documents were relatively low, which is expected given that the final LaTeX documents are designed to provide comprehensive literature review sections, not direct summaries of the abstracts. These documents incorporate additional background information, contextual insights, and synthesized content from various sources, which broadens the scope and naturally reduces recall scores.
Despite this, the F1-scores, which balance precision and recall, show that the final documents maintain a strong balance between relevance and content coverage. GPT-4o generally achieved higher recall scores than GPT-3.5, suggesting it is more effective at incorporating additional relevant content while maintaining overall coherence. This makes GPT-4o particularly useful in scenarios where comprehensive literature coverage is essential.
While GPT-4o offers advantages in retaining more comprehensive content, GPT-3.5 remains a competitive option for generating concise and highly relevant summaries. Future efforts could focus on improving recall without sacrificing precision, allowing for even more comprehensive and well-rounded literature review sections.

4.4. Experiment 4: Readability Score

We used the Flesch Reading Ease Score to evaluate the readability of the generated summaries and final documents at different stages of the document generation process. This metric provides insights into how accessible the text is to a general audience, with higher scores indicating easier-to-read content. Readability was evaluated for the T5-generated summaries, GPT post-edited summaries, and the final LaTeX documents.
Table 6 outlines the interpretation of Flesch Reading Ease scores, with lower scores indicating text that requires a higher level of education to comprehend. In our evaluation, we compared these scores across the stages of the automated SLR process to assess how the readability evolved from the initial summarization to the final document creation.
Table 7 presents the Flesch Reading Ease scores for each stage of the document generation process. These scores provide an overview of how readability changes as the content is transformed from initial summaries to refined, structured documents.
T5-generated summaries exhibit low readability scores, indicating that the content is quite challenging to read. This outcome is expected due to the highly condensed nature of the T5-generated summaries, which prioritize brevity over readability, often lacking the narrative structure required for easier comprehension.
The readability of the summaries significantly improves after the GPT post-edited sections. The post-editing process refines the content by enhancing clarity, improving sentence structure, and providing a more coherent flow. This results in more accessible and readable sections, reflected in the enhanced Flesch scores.
The final LaTeX generated documents show further improvements in readability. The additional structuring, formatting, and content synthesis contribute to easier-to-read documents than the earlier stages. However, while the readability has improved, it remains lower than the baseline.
Baseline summaries exhibit the highest readability scores, demonstrating that maintaining readability while summarizing and synthesizing content remains a challenge. The baseline scores highlight the gap between the generated summaries and the clarity of the original abstracts.
The Flesch Reading Ease Scores improved progressively from T5-generated summaries to GPT post-edited sections and final LaTeX documents. However, despite these improvements, the readability of the generated documents remains lower than that of the original abstracts. This outcome underscores the challenge of maintaining high readability while compressing and synthesizing content, particularly in automated systems.

4.5. Experiment 5: Sentence Similarity

To further evaluate the effectiveness of our automated systematic literature review (SLR) system, we computed cosine similarity scores at various stages of document generation. This analysis quantifies how closely the generated summaries and final documents align with the original input queries, providing a measure of content retention and relevance. For comparison, a baseline (Random) was included, representing a document generated with random words, to serve as a control.
Cosine similarity measures the cosine of the angle between two vectors in a multidimensional space; here, these vectors represent text embeddings derived from the documents. Higher cosine similarity scores indicate greater alignment between the generated texts and the original input queries.
Table 8 presents the cosine similarity scores for each stage of the document generation process, offering insights into how well the system preserves content relevance across different stages:
Abstract Filtering: The filtered abstracts generated by GPT-3.5 and GPT-4o exhibit high cosine similarity scores, demonstrating that the initial search and screening phase effectively identifies documents closely related to the input queries. Both models perform well at this stage, confirming the robustness of the search process.
T5-Generated Summaries: The cosine similarity scores decrease slightly for the T5-generated summaries, which is expected. Summarization inherently condenses content and may omit some details, leading to a lower similarity with the full abstracts. However, the core information relevant to the input query remains retained, ensuring that the generated summaries focus on the main topics.
GPT Post-Edited Summaries: The cosine similarity scores increase after the post-editing process by GPT. This improvement suggests that the GPT-based post-editing refines the structure and readability and enhances alignment with the original input. The post-editing process ensures that the key content is retained while improving the coherence of the generated sections. GPT-4o generally outperforms GPT-3.5 in maintaining content similarity, indicating that GPT-4o is more effective at preserving relevant information.
Final LaTeX Documents: The final documents generated by GPT continue to exhibit high similarity scores, indicating that the synthesis and summarization process effectively retains the relevance of the content. The structured nature of LaTeX documents ensures that core themes from the input queries are well-represented. GPT-4o again shows slightly better performance than GPT-3.5, further suggesting that it is better at understanding and incorporating relevant content throughout the document generation process.
Random Baseline: As expected, the cosine similarity scores for the randomly generated document are very low. This serves as a control, validating the significance of the similarity scores observed for the generated summaries and final documents.
Overall, the system demonstrates robustness in generating documents that remain closely aligned with the original input queries, ensuring that the synthesized literature reviews preserve essential information while improving readability and structure.

4.6. Experiment 6: Finding the Optimal Number of Papers for SLR

This section explores several key performance metrics to determine the optimal number of papers to include in the SLR process. The metrics analyzed include CPU time, number of topics identified, topic coherence, ROUGE scores, readability scores, and cosine similarity scores. These metrics are used to assess the impact of different document limits on the quality and efficiency of the SLR. We recommend the most effective document limit that balances performance and computational resources based on the analysis results. Figure 2 presents the results obtained.
CPU Time: Figure 2a shows the CPU time shows how computational requirements scale with the number of documents processed. As the document count increases, CPU time rises significantly, with GPT-4o consistently requiring more time than GPT-3.5. This indicates that although GPT-4o can potentially offer more accurate results, it demands more computational resources, which is a trade-off to consider when processing large volumes of documents.
Number of Topics Found: In Figure 2b, BERTopic identifies an increasing number of topics as more documents are processed. GPT-4o consistently identifies more topics than GPT-3.5 across all document limits. This suggests that GPT-4o is more adept at detailed clustering, potentially offering a more nuanced breakdown of the literature. However, after a certain threshold, the increase in topics may not necessarily translate to better quality but rather more fragmented groupings.
Topic Coherence: The topic coherence metric measures the semantic similarity within the topics identified, providing insight into the quality of the generated clusters. Figure 2c illustrates the quality of the topics generated based on the semantic similarity of words within them. GPT-3.5 and GPT-4o maintain relatively stable topic coherence scores of up to 200 documents. Beyond this point, coherence begins to drop slightly for both models, likely due to overfitting or noise introduced by an excessive number of documents. This reinforces that 200 documents strike an optimal balance between quality and quantity regarding topic coherence.
ROUGE Scores for Summarization Quality: The ROUGE scores measure how well the generated summaries align with reference abstracts, focusing on content retention. Figure 2d shows that as the number of documents increases, the ROUGE scores improve, peaking around 200. This suggests that the system becomes better at generating summaries that capture the core content of the papers as more documents are processed. However, beyond the 200-document threshold, the improvement in ROUGE scores plateaus, indicating that additional documents do not contribute significantly to better summarization. This implies that while increasing the document count improves the system’s ability to summarize effectively, there is little benefit to going beyond 200 documents regarding content retention and quality.
Cosine Similarity for Content Alignment: The cosine similarity scores measure how closely the generated documents align with the input queries, indicating relevance and focus. Figure 2e shows that GPT-3.5 and GPT-4o achieve high similarity scores across all document limits, stabilizing around 200 documents. This indicates that 200 documents provide sufficient information to produce outputs well-aligned with the original research query without overwhelming the system with excess data. The plateau in similarity scores beyond this threshold suggests that additional documents do not significantly enhance the relevance of the generated summaries. Therefore, 200 documents appear to be the most efficient choice for maintaining high alignment with the research objectives while minimizing computational overhead.
Readability Scores: We use the Flesch Reading Ease [33] metric to evaluate how accessible and easy to read the generated summaries are. Figure 2f indicates that the readability scores increase as more documents are processed, reaching their highest point, around 200. This suggests that the generated summaries become clearer and easier to read as the system processes more documents, possibly due to having a more comprehensive pool of content to draw from. However, readability scores decline slightly after 200 documents, indicating that the system might introduce more complex or fragmented language as the document count grows. This highlights that 200 documents offer the best balance for generating summaries that are both informative and easy to read.
Optimal Number of Papers: Based on the analysis of the above metrics, 200 documents emerge as the optimal document limit for the SLR process. At this threshold, the system provides high-quality summaries, maintains strong topic coherence, and produces readable and relevant outputs without excessive computational resources. Using over 200 documents leads to diminishing returns, particularly regarding topic coherence, readability, and cosine similarity. Thus, we recommend 200 documents as the ideal balance between performance and efficiency for conducting automated systematic literature reviews.

5. Discussion

The results presented in this study demonstrate the potential of the proposed automated SLR framework to streamline and enhance the process of conducting literature reviews. By integrating advanced NLP techniques and LLMs such as GPT-3.5 and GPT-4o, the framework automates systematic search, data extraction, topic modeling, and summarization stages. However, a critical analysis of the results reveals both strengths and areas for improvement.
GPT-4o retrieves more papers than GPT-3.5: The experiments revealed that GPT-4o consistently outperformed GPT-3.5 in retrieving a larger number of papers across all research topics. This suggests that GPT-4o is better at generating more comprehensive and contextually rich search queries. The ability of GPT-4o to retrieve more papers is beneficial in scenarios where exhaustive literature coverage is important, such as systematic reviews and meta-analyses, as it ensures that a wider array of relevant research is considered. However, this improved retrieval capacity may also introduce more irrelevant or low-quality papers, necessitating more robust filtering mechanisms.
High ROUGE-1 precision scores for both models: Both GPT-3.5 and GPT-4o demonstrated high ROUGE-1 precision scores during the summarization phase, indicating that the generated summaries retained a significant amount of relevant content from the reference abstracts. This suggests that the models effectively focus on the most important information when creating summaries, which is critical in systematic reviews where maintaining the relevance of summarized content is paramount. However, the relatively lower recall scores reflect that some content was omitted during summarization, which may be intentional to avoid overwhelming the reader with excessive detail. The high precision with lower recall suggests a bias toward conciseness, which can be advantageous in certain contexts but may require adjustment depending on the goals of the review.
Post-editing improved precision but reduced recall: The post-editing phase significantly improved the precision of the generated summaries but reduced recall, indicating that while the content became more concise and focused, some relevant details were omitted. This aligns with the goal of post-editing, which is to refine and streamline the summaries for clarity and coherence. GPT-4o demonstrated higher recall than GPT-3.5 in this phase, suggesting it more effectively retained content during post-editing. This slight advantage highlights GPT-4o’s ability to balance relevance and conciseness better, making it more suitable for generating comprehensive yet readable summaries in systematic reviews.
GPT-4o achieved higher recall in final LaTeX documents: The final LaTeX documents generated by GPT-4o achieved higher recall scores than those generated by GPT-3.5, indicating that GPT-4o was more successful in incorporating additional relevant content while maintaining coherence. This makes GPT-4o particularly advantageous for use cases that require comprehensive literature coverage, as the final documents generated by GPT-4o were better at synthesizing information from multiple sources. However, this increase in recall may come at the cost of readability, as the additional content could make the final documents more complex and challenging to navigate. Future work could explore optimizing the balance between recall and readability in final document generation.
Readability improved through post-editing and final document generation: The Flesch Reading Ease Scores demonstrated a clear improvement in readability from the initial T5-generated summaries to the GPT post-edited sections and final LaTeX documents. This suggests that the post-editing process significantly enhanced the clarity and coherence of the summaries, making them easier to read and understand. However, despite these improvements, the readability of the generated documents remained lower than the baseline abstracts. This outcome reflects the inherent difficulty in maintaining high readability while condensing and synthesizing technical content. Future work could explore more advanced techniques to improve readability, especially in the post-editing phase, to close the gap with the original abstracts.
Cosine similarity confirms robust content retention: Cosine similarity scores across all stages of document generation were high, confirming that the system retained key content from the original input queries. The post-editing and final document generation stages further improved content alignment, particularly with GPT-4o, which generally outperformed GPT-3.5 in maintaining content relevance. These results suggest that both models effectively ensure the generated summaries and documents stay focused on the core topics of the input queries, making them reliable tools for systematic literature reviews. The consistently high similarity scores also validate the robustness of retrieval, summarization, and synthesis, ensuring that essential information is not lost throughout the stages.
The findings from this study underscore the utility of combining GPT models and NLP techniques to automate key phases of systematic literature reviews, from retrieval to summarization. While GPT-4o demonstrates superior performance in content retrieval and recall, GPT-3.5 remains competitive for tasks prioritizing efficiency and conciseness. The framework shows promise in automating extensive literature reviews with relatively high precision and robust content retention. However, challenges remain, particularly in optimizing readability and balancing recall with document complexity. Future work should focus on refining the post-editing processes, improving the coherence and accessibility of generated documents, and ensuring that the system remains adaptable to diverse academic domains. Enhancing the framework’s ability to filter irrelevant or lower-quality content will strengthen its applicability in high-demand, resource-intensive reviews.

6. Limitations

A key limitation of this study is that it only analyzed proprietary models, specifically OpenAI’s GPT-3.5 and GPT-4o, and did not include open-source models like LLaMA or Falcon. This is important because open-source models are becoming increasingly popular for research and practical applications due to their accessibility and customization potential. By focusing only on proprietary models, this study misses the opportunity to evaluate the performance, bias mitigation strategies, and transparency advantages that open-source models may offer.
Bias in Literature Selection: AI models can introduce bias into the literature selection process. These models are trained on large, potentially unbalanced datasets, which can skew the selection towards more popular or well-represented topics, ignoring less-covered research areas. Future work should explore ways to address this bias, possibly through fairness-aware algorithms or more inclusive data sources.
Hallucination and Misinformation: Both GPT-3.5 and GPT-4o are prone to generating content that is not directly grounded in the input data, which can lead to inaccurate summaries. This is particularly risky in a systematic literature review where factual accuracy is critical. Adding validation steps, such as human review or factual grounding mechanisms, would help mitigate this issue.
Exclusion of Proprietary Models: By not including freely available and modifiable models like LLaMA or Falcon, this study does not address how open-source solutions could improve transparency, reproducibility, and control in the SLR process. These models offer the potential for better alignment with specific research needs and ethical considerations such as data privacy and bias control.
Scalability: Although the system performed well with up to 3000 papers per query, larger datasets may introduce computational challenges. Future iterations should improve scalability, possibly by adopting distributed computing methods.
Future studies should incorporate open-source models to compare their effectiveness and address the broader needs of the academic community, offering more flexibility, transparency, and cost control.

7. Ethical Implications

Despite the promise of efficiency and scalability, the fully automated approach outlined in this paper also raises significant ethical challenges that must be critically addressed in future work. The absence of human oversight introduces epistemic risks, including unrecognized bias, misclassification, or omitting foundational literature, especially when dealing with interdisciplinary or emergent topics. Moreover, while fluent and structurally sound, the synthesized outputs may lack the nuanced interpretation or methodological rigor expected by domain experts. Ensuring such reviews’ accuracy, completeness, and contextual sensitivity remains an open problem. Future developments should explore hybrid models that allow for configurable human-in-the-loop interaction, post-hoc validation, and model transparency. Additionally, a systematic evaluation framework involving domain experts is essential to assess the reliability, trustworthiness, and publication readiness of fully automated SLR outputs.
It is critical, therefore, to highlight that the availability of fully automated SLR systems does not diminish the need for proper training in literature review methodology, particularly for early-career researchers and PhD students. Mastery of the principles underpinning rigorous evidence synthesis is critical for understanding the strengths and limitations of such systems and maintaining academic standards. These tools should be viewed as augmentative technologies, which would be beneficial for keeping pace with the accelerating volume of academic publications rather than as replacements for expert judgment. A key element of responsible use must include a structured review stage in which researchers validate the completeness of the results, particularly by checking for omitted seminal works and ensuring that domain-specific knowledge has been appropriately captured. Embedding this review step into the workflow is essential for assuring quality, relevance, and scholarly integrity.

8. Reproducibility

To facilitate transparency and support future research, all code and resources for PROMPTHEUS are publicly available at the following GitHub repository: https://github.com/joaopftorres/PROMPTHEUS.git, accessed on 12 May 2025.

9. Conclusions

This study proposed PROMPTHEUS, an automated SLR framework that integrates advanced NLP techniques and large language models to streamline the literature review process. By automating systematic search, data extraction, topic modeling, and synthesis, the framework effectively manages the growing volume of academic literature. Our experiments across five research topics demonstrated the system’s strengths in selecting relevant papers, retaining key content, and improving readability in post-editing stages. However, areas like recall and topic coherence still require improvement.
Despite achieving high precision, lower recall scores, and moderate topic coherence, these metrics suggest that some relevant content was omitted and topics could be better organized. GPT-4o outperformed GPT-3.5 in recall and content retention. However, readability, though improved in post-editing, remained below the clarity of the original abstracts.
In conclusion, PROMPTHEUS contributes to automating systematic literature reviews by combining advanced NLP techniques and large language models. Our framework addresses the growing challenges of managing vast academic literature by streamlining critical processes while maintaining high precision and content relevance. However, the trade-offs between precision, recall, and readability underscore further refinement, particularly in improving topic coherence and ensuring that the most relevant content is consistently included. Looking ahead, the future potential of PROMPTHEUS is vast. We will continue to optimize these elements while incorporating open-source models to improve flexibility, transparency, and scalability, empowering the academic community to more effectively manage the ever-rapidly growing body of research with enhanced comprehensiveness and efficiency.

Author Contributions

Conceptualization, J.T.; Methodology, J.T.; Software, J.T.; Formal analysis, J.T.; Investigation, C.M. (Catherine Mulligan), J.J. and C.M. (Catarina Moreira); Resources, C.M. (Catherine Mulligan), J.J. and C.M. (Catarina Moreira); Writing—original draft, J.T. and C.M. (Catarina Moreira); Writing—review & editing, C.M. (Catherine Mulligan), J.J. and C.M. (Catarina Moreira); Visualization, J.T.; Supervision, C.M. (Catherine Mulligan), J.J. and C.M. (Catarina Moreira). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Fundacao para a Ciencia e Tecnologia (FCT).

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ware, M.; Mabe, M. The STM Report: An Overview of Scientific and Scholarly Journal Publishing; International Association of Scientific, Technical and Medical Publishers: The Hague, The Netherlands, 2015. [Google Scholar]
  2. Chu, J.S.G.; Evans, J.A. Slowed canonical progress in large fields of science. Proc. Natl. Acad. Sci. USA 2021, 118, e2021636118. [Google Scholar] [CrossRef]
  3. Nanggala, K.; Elwirehardja, G.N.; Pardamean, B. Systematic Literature Review of Transformer Model Implementations in Detecting Depression. In Proceedings of the 2023 6th International Conference of Computer and Informatics Engineering (IC2IE), IEEE, Lombok, Indonesia, 14–15 September 2023; pp. 203–208. [Google Scholar]
  4. Scherbakov, D.; Hubig, N.; Jansari, V.; Bakumenko, A.; Lenert, L.A. The emergence of large language models as tools in literature reviews: A large language model-assisted systematic review. J. Am. Med. Inform. Assoc. 2025, ocaf063. [Google Scholar] [CrossRef] [PubMed]
  5. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
  6. Okoli, C. A Guide to Conducting a Standalone Systematic Literature Review. Commun. Assoc. Inf. Syst. 2015, 37, 43. [Google Scholar] [CrossRef]
  7. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. arXiv 2023, arXiv:1910.10683. [Google Scholar]
  8. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Few-Shot Learners. arXiv 2020, arXiv:2005.14165. [Google Scholar]
  9. Moreno-Garcia, C.F.; Jayne, C.; Elyan, E.; Aceves-Martins, M. A novel application of machine learning and zero-shot classification methods for automated abstract screening in systematic reviews. Decis. Anal. J. 2023, 6, 100162. [Google Scholar] [CrossRef]
  10. Bolanos, F.; Salatino, A.; Osborne, F.; Motta, E. Artificial Intelligence for Literature Reviews: Opportunities and Challenges. Artif. Intell. Rev. 2024, 57, 259. [Google Scholar] [CrossRef]
  11. Saeidmehr, A.; Steel, P.D.G.; Samavati, F.F. Systematic review using a spiral approach with machine learning. Syst. Rev. 2024, 13, 32. [Google Scholar] [CrossRef]
  12. Sami, A.M.; Rasheed, Z.; Kemell, K.K.; Waseem, M.; Kilamo, T.; Saari, M.; Duc, A.N.; Systä, K.; Abrahamsson, P. System for systematic literature review using multiple AI agents: Concept and an empirical evaluation. arXiv 2024, arXiv:2403.08399. [Google Scholar]
  13. Tóth, B.; Berek, L.; Gulácsi, L.; Péntek, M.; Zrubka, Z. Automation of systematic reviews of biomedical literature: A scoping review of studies indexed in PubMed. Syst. Rev. 2024, 13, 174. [Google Scholar] [CrossRef]
  14. Chappell, M.; Edwards, M.; Watkins, D.; Marshall, C.; Graziadio, S. Machine learning for accelerating screening in evidence reviews. Cochrane Evid. Synth. Methods 2023, 1, e12021. [Google Scholar] [CrossRef]
  15. Guo, Q.; Jiang, G.; Zhao, Q.; Long, Y.; Feng, K.; Gu, X.; Xu, Y.; Li, Z.; Huang, J.; Du, L. Rapid review: A review of methods and recommendations based on current evidence. J. Evid. Based Med. 2024, 17, 434–453. [Google Scholar] [CrossRef] [PubMed]
  16. Robledo, S.; Grisales Aguirre, A.M.; Hughes, M.; Eggers, F. “Hasta la vista, baby”—Will machine learning terminate human literature reviews in entrepreneurship? J. Small Bus. Manag. 2023, 61, 1314–1343. [Google Scholar] [CrossRef]
  17. Gates, A.; Gates, M.; DaRosa, D.; Elliott, S.A.; Pillay, J.; Rahman, S.; Vandermeer, B.; Hartling, L. Decoding semi-automated title-abstract screening: Findings from a convenience sample of reviews. Syst. Rev. 2020, 9, 272. [Google Scholar] [CrossRef] [PubMed]
  18. Cierco Jimenez, R.; Lee, T.; Rosillo, N.; Cordova, R.; Cree, I.A.; Gonzalez, A.; Indave Ruiz, B.I. Machine learning computational tools to assist the performance of systematic reviews: A mapping review. BMC Med. Res. Methodol. 2022, 22, 322. [Google Scholar] [CrossRef]
  19. Affengruber, L.; van der Maten, M.M.; Spiero, I.; Nussbaumer-Streit, B.; Mahmić-Kaknjo, M.; Ellen, M.E.; Goossen, K.; Kantorova, L.; Hooft, L.; Riva, N.; et al. An exploration of available methods and tools to improve the efficiency of systematic review production: A scoping review. BMC Med. Res. Methodol. 2024, 24, 210. [Google Scholar] [CrossRef]
  20. Perlman-Arrow, S.; Loo, N.; Bobrovitz, N.; Yan, T.; Arora, R.K. A real-world evaluation of the implementation of NLP technology in abstract screening of a systematic review. Res. Synth. Methods 2023, 14, 608–621. [Google Scholar] [CrossRef]
  21. Ofori-Boateng, R.; Aceves-Martins, M.; Wiratunga, N.; Moreno-Garcia, C.F. Towards the automation of systematic reviews using natural language processing, machine learning, and deep learning: A comprehensive review. Artif. Intell. Rev. 2024, 57, 200. [Google Scholar] [CrossRef]
  22. Qureshi, R.; Shaughnessy, D.; Gill, K.A.; Robinson, K.A.; Li, T.; Agai, E. Are ChatGPT and large language models “the answer” to bringing us closer to systematic review automation? Syst. Rev. 2023, 12, 72. [Google Scholar] [CrossRef]
  23. Li, M.; Sun, J.; Tan, X. Evaluating the effectiveness of large language models in abstract screening: A comparative analysis. Syst. Rev. 2024, 13, 219. [Google Scholar] [CrossRef] [PubMed]
  24. Kharawala, S.; Mahajan, A.; Gandhi, P. Artificial intelligence in systematic literature reviews: A case for cautious optimism. J. Clin. Epidemiol. 2021, 138, 243. [Google Scholar] [CrossRef]
  25. Dennstädt, F.; Zink, J.; Putora, P.M.; Hastings, J.; Cihoric, N. Title and abstract screening for literature reviews using large language models: An exploratory study in the biomedical domain. Syst. Rev. 2024, 13, 158. [Google Scholar] [CrossRef]
  26. Hamel, C.; Hersi, M.; Kelly, S.E.; Tricco, A.C.; Straus, S.; Wells, G.; Pham, B.; Hutton, B. Guidance for using artificial intelligence for title and abstract screening while conducting knowledge syntheses. BMC Med. Res. Methodol. 2021, 21, 285. [Google Scholar] [CrossRef]
  27. Masoumi, S.; Amirkhani, H.; Sadeghian, N.; Shahraz, S. Natural language processing (NLP) to facilitate abstract review in medical research: The application of BioBERT to exploring the 20-year use of NLP in medical research. Syst. Rev. 2024, 13, 107. [Google Scholar] [CrossRef] [PubMed]
  28. Speckemeier, C.; Niemann, A.; Wasem, J.; Buchberger, B.; Neusser, S. Methodological guidance for rapid reviews in healthcare: A scoping review. Res. Synth. Methods 2022, 13, 394–404. [Google Scholar] [CrossRef] [PubMed]
  29. O’Connor, A.M.; Tsafnat, G.; Thomas, J.; Glasziou, P.; Gilbert, S.B.; Hutton, B. A question of trust: Can we build an evidence base to gain trust in systematic review automation technologies? Syst. Rev. 2019, 8, 143. [Google Scholar] [CrossRef]
  30. de la Torre-López, J.; Ramírez, A.; Romero, J.R. Artificial intelligence to automate the systematic review of scientific literature. Computing 2023, 105, 2171–2194. [Google Scholar] [CrossRef]
  31. Affengruber, L.; Nussbaumer-Streit, B.; Hamel, C.; Van der Maten, M.; Thomas, J.; Mavergames, C.; Spijker, R.; Gartlehner, G. Rapid review methods series: Guidance on the use of supportive software. BMJ Evid. Based Med. 2024, 29, 264–271. [Google Scholar] [CrossRef]
  32. Shaheen, N.; Shaheen, A.; Ramadan, A.; Hefnawy, M.T.; Ramadan, A.; Ibrahim, I.A.; Hassanein, M.E.; Ashour, M.E.; Flouty, O. Appraising systematic reviews: A comprehensive guide to ensuring validity and reliability. Front. Res. Metrics Anal. 2023, 8, 1268045. [Google Scholar] [CrossRef]
  33. Kincaid, J. Derivation of New Readability Formulas (Automated Readability Index, Fog Count and Flesch Reading Ease Formula) for Navy Enlisted Personnel; Chief of Naval Technical Training: Millington, TN, USA, 1975. [Google Scholar]
  34. Rahimi, H.; Mimno, D.; Hoover, J.; Naacke, H.; Constantin, C.; Amann, B. Contextualized Topic Coherence Metrics. In Proceedings of the Findings of the Association for Computational Linguistics: EACL 2024; Graham, Y., Purver, M., Eds.; EACL: St. Julian’s, Malta, 2024; pp. 1760–1773. [Google Scholar]
  35. Röder, M.; Both, A.; Hinneburg, A. Exploring the Space of Topic Coherence Measures. In Proceedings of the Eight International Conference on Web Search and Data Mining, Shanghai, China, 2–6 February 2015. [Google Scholar]
  36. Wikipedia. Flesch–Kincaid Readability Tests—Wikipedia, The Free Encyclopedia. 2024. Available online: http://en.wikipedia.org/w/index.php?title=Flesch%E2%80%93Kincaid%20readability%20tests&oldid=1214583693 (accessed on 12 May 2025).
Figure 1. The PROMPTHEUS framework consists of three phases: (1) systematic search and screening using GPT and Sentence-BERT for paper selection, (2) data extraction and topic modeling with BERTopic and GPT for organizing and generating section titles, and (3) synthesis and summarization with T5 and GPT to refine and compile the findings into an SLR LaTeX document. This framework leverages NLP techniques and LLMs for an efficient and scalable SLR process.
Figure 1. The PROMPTHEUS framework consists of three phases: (1) systematic search and screening using GPT and Sentence-BERT for paper selection, (2) data extraction and topic modeling with BERTopic and GPT for organizing and generating section titles, and (3) synthesis and summarization with T5 and GPT to refine and compile the findings into an SLR LaTeX document. This framework leverages NLP techniques and LLMs for an efficient and scalable SLR process.
Information 16 00420 g001
Figure 2. Performance metrics across different document limits for GPT-3.5 and GPT-4o in the SLR process. (a) CPU Time: GPT-4o consistently requires more time than GPT-3.5 as the number of documents increases, reflecting its computational complexity. (b) Number of Topics: GPT-4o identifies more topics, indicating a finer level of clustering. (c) Topic Coherence: Coherence is stable up to 200 documents for both models, but it declines as more documents are added, suggesting overfitting or noise. (d) ROUGE Scores: Summarization quality improves and plateaus around 200 documents. (e) Cosine Similarity: Both models show stable alignment with input queries, with diminishing returns beyond 200 documents. (f) Readability Scores: Readability peaks around 200 documents before declining, suggesting this as the optimal limit for accessible summaries.
Figure 2. Performance metrics across different document limits for GPT-3.5 and GPT-4o in the SLR process. (a) CPU Time: GPT-4o consistently requires more time than GPT-3.5 as the number of documents increases, reflecting its computational complexity. (b) Number of Topics: GPT-4o identifies more topics, indicating a finer level of clustering. (c) Topic Coherence: Coherence is stable up to 200 documents for both models, but it declines as more documents are added, suggesting overfitting or noise. (d) ROUGE Scores: Summarization quality improves and plateaus around 200 documents. (e) Cosine Similarity: Both models show stable alignment with input queries, with diminishing returns beyond 200 documents. (f) Readability Scores: Readability peaks around 200 documents before declining, suggesting this as the optimal limit for accessible summaries.
Information 16 00420 g002
Table 1. Comparison of GPT-3.5 and GPT-4o in finding papers for the SLR process. CPU time indicates the total time for the entire automated SLR process.
Table 1. Comparison of GPT-3.5 and GPT-4o in finding papers for the SLR process. CPU time indicates the total time for the entire automated SLR process.
InputModelCPU Time (s)Papers Found
Explainable Artificial IntelligenceGPT-3.512131287
GPT-4o15551712
Virtual RealityGPT-3.513191986
GPT-4o14962833
BlockchainGPT-3.514763000
GPT-4o15633000
Large Language ModelsGPT-3.515051400
GPT-4o21153000
Neural Machine TranslationGPT-3.516482018
GPT-4o16732073
Table 2. Topic coherence analysis using Gensim’s topic coherence metric for GPT-3.5 and GPT-4o generated queries.
Table 2. Topic coherence analysis using Gensim’s topic coherence metric for GPT-3.5 and GPT-4o generated queries.
InputModelTopic CoherenceNumber of Topics
Explainable Artificial IntelligenceGPT-3.50.4675
GPT-4o0.4226
Virtual RealityGPT-3.50.4348
GPT-4o0.4818
BlockchainGPT-3.50.4118
GPT-4o0.4755
Large Language ModelsGPT-3.50.4285
GPT-4o0.4735
Neural Machine TranslationGPT-3.50.4776
GPT-4o0.4707
Table 3. ROUGE-1 scores for the T5-generated summaries (P = precision, R = recall, and F1 = F-measure), with the selected abstracts as a reference.
Table 3. ROUGE-1 scores for the T5-generated summaries (P = precision, R = recall, and F1 = F-measure), with the selected abstracts as a reference.
InputModelROUGE
PRF1
Explainable Artificial IntelligenceGPT-3.50.9630.4050.570
GPT-4o0.9640.3870.552
Virtual RealityGPT-3.50.9670.4180.583
GPT-4o0.9690.4250.591
BlockchainGPT-3.50.9670.4010.567
GPT-4o0.9680.4000.567
Large Language ModelsGPT-3.50.9660.3760.540
GPT-4o0.9650.3810.546
Neural Machine TranslationGPT-3.50.9650.4620.625
GPT-4o0.9650.4600.623
Table 4. ROUGE-1 scores for GPT post-edited sections (P = precision, R = recall, and F1 = F-measure).
Table 4. ROUGE-1 scores for GPT post-edited sections (P = precision, R = recall, and F1 = F-measure).
InputModelROUGE
PRF1
Explainable Artificial IntelligenceGPT-3.50.9220.0290.055
GPT-4o0.8840.0630.118
Virtual RealityGPT-3.50.9200.0290.057
GPT-4o0.9060.0630.118
BlockchainGPT-3.50.9240.0380.072
GPT-4o0.8970.0290.056
Large Language ModelsGPT-3.50.9190.0280.055
GPT-4o0.9010.0420.080
Neural Machine TranslationGPT-3.50.9110.0340.066
GPT-4o0.8830.0750.138
Table 5. ROUGE-1 scores for the GPT-generated final LaTeX document (P = precision, R = recall, and F1 = F-measure).
Table 5. ROUGE-1 scores for the GPT-generated final LaTeX document (P = precision, R = recall, and F1 = F-measure).
InputModelROUGE
PRF1
Explainable Artificial IntelligenceGPT-3.50.9910.0180.036
GPT-4o0.9890.0260.049
Virtual RealityGPT-3.50.9880.0150.029
GPT-4o0.9720.0340.065
BlockchainGPT-3.50.9870.0230.045
GPT-4o0.9950.0280.054
Large Language ModelsGPT-3.50.9950.0210.042
GPT-4o0.9900.0390.076
Neural Machine TranslationGPT-3.50.9760.0210.041
GPT-4o0.9690.0380.074
Table 6. Interpretation of Flesch Reading Ease Scores [33,36].
Table 6. Interpretation of Flesch Reading Ease Scores [33,36].
ScoreSchool Level (US)Notes
100–905thVery easy to read.
Easily understood by
an average 11-year-old student.
90–806thEasy to read.
Conversational English
for consumers.
80–707thFairly easy to read.
70–608th–9thPlain English.
Easily understood by
13- to 15-year-old students.
60–5010th–12thFairly difficult to read.
50–30CollegeDifficult to read.
30–10GraduateVery difficult to read.
Best understood by
university graduates.
10–0ProfessionalExtremely difficult to read.
Best understood by
university graduates.
Table 7. Readability scores for T5-generated summaries, GPT post-edited sections, and the GPT-generated final LaTeX document.
Table 7. Readability scores for T5-generated summaries, GPT post-edited sections, and the GPT-generated final LaTeX document.
InputModelT5SumGPTSecGPTSLRBaseline
Explainable Artificial IntelligenceGPT-3.58.45030.94229.94550.827
GPT-4o9.19628.05334.542
Virtual RealityGPT-3.516.12633.97426.92654.297
GPT-4o14.48132.56026.237
BlockchainGPT-3.56.87422.15819.02447.461
GPT-4o9.95424.33920.314
Large Language ModelsGPT-3.514.39934.50535.04153.834
GPT-4o21.73430.46633.798
Neural Machine TranslationGPT-3.520.81532.19129.24259.556
GPT-4o22.50827.10631.754
Table 8. Cosine similarity between the embedding of the input and the embeddings of generated documents (Abs = abstracts, T5Sum = T5-generated summaries, GPTSec = GPT post-edited sections, GPTSLR = GPT-generated final LaTeX document, and Random = document created with randomly generated words).
Table 8. Cosine similarity between the embedding of the input and the embeddings of generated documents (Abs = abstracts, T5Sum = T5-generated summaries, GPTSec = GPT post-edited sections, GPTSLR = GPT-generated final LaTeX document, and Random = document created with randomly generated words).
InputModelCosine Similarity
AbsT5SumGPTSecGPTSLRRandom
Explainable Artificial IntelligenceGPT-3.50.6440.5870.7180.7630.118
GPT-4o0.4460.5390.6870.753 
Virtual RealityGPT-3.50.4820.5390.5970.5520.082
GPT-4o0.6130.6240.6020.689 
BlockchainGPT-3.50.5830.5110.6650.6810.071
GPT-4o0.6460.5870.5500.612 
Large Language ModelsGPT-3.50.5510.4460.5980.6100.127
GPT-4o0.4360.6190.5910.691 
Neural Machine TranslationGPT-3.50.6220.6910.6920.6940.108
GPT-4o0.6220.6640.6690.743 
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Torres, J.; Mulligan, C.; Jorge, J.; Moreira, C. PROMPTHEUS: A Human-Centered Pipeline to Streamline Systematic Literature Reviews with Large Language Models. Information 2025, 16, 420. https://doi.org/10.3390/info16050420

AMA Style

Torres J, Mulligan C, Jorge J, Moreira C. PROMPTHEUS: A Human-Centered Pipeline to Streamline Systematic Literature Reviews with Large Language Models. Information. 2025; 16(5):420. https://doi.org/10.3390/info16050420

Chicago/Turabian Style

Torres, Joao, Catherine Mulligan, Joaquim Jorge, and Catarina Moreira. 2025. "PROMPTHEUS: A Human-Centered Pipeline to Streamline Systematic Literature Reviews with Large Language Models" Information 16, no. 5: 420. https://doi.org/10.3390/info16050420

APA Style

Torres, J., Mulligan, C., Jorge, J., & Moreira, C. (2025). PROMPTHEUS: A Human-Centered Pipeline to Streamline Systematic Literature Reviews with Large Language Models. Information, 16(5), 420. https://doi.org/10.3390/info16050420

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop