Next Article in Journal
Optimizing a Double Stage Heat Transformer Performance by Levenberg–Marquardt Artificial Neural Network
Previous Article in Journal
On Classification of the Human Emotions from Facial Thermal Images: A Case Study Based on Machine Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accelerating Disease Model Parameter Extraction: An LLM-Based Ranking Approach to Select Initial Studies for Literature Review Automation

1
School of Veterinary Science, Massey University, Palmerston North 4442, New Zealand
2
School of Mathematical and Computational Sciences, Massey University, Auckland 0632, New Zealand
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2025, 7(2), 28; https://doi.org/10.3390/make7020028
Submission received: 15 January 2025 / Revised: 4 March 2025 / Accepted: 21 March 2025 / Published: 26 March 2025
(This article belongs to the Section Learning)

Abstract

:
As climate change transforms our environment and human intrusion into natural ecosystems escalates, there is a growing demand for disease spread models to forecast and plan for the next zoonotic disease outbreak. Accurate parametrization of these models requires data from diverse sources, including the scientific literature. Despite the abundance of scientific publications, the manual extraction of these data via systematic literature reviews remains a significant bottleneck, requiring extensive time and resources, and is susceptible to human error. This study examines the application of a large language model (LLM) as an assessor for screening prioritisation in climate-sensitive zoonotic disease research. By framing the selection criteria of articles as a question–answer task and utilising zero-shot chain-of-thought prompting, the proposed method achieves a saving of at least 70% work effort compared to manual screening at a recall level of 95% (NWSS@95%). This was validated across four datasets containing four distinct zoonotic diseases and a critical climate variable (rainfall). The approach additionally produces explainable AI rationales for each ranked article. The effectiveness of the approach across multiple diseases demonstrates the potential for broad application in systematic literature reviews. The substantial reduction in screening effort, along with the provision of explainable AI rationales, marks an important step toward automated parameter extraction from the scientific literature.

1. Introduction

Zoonotic diseases are becoming increasingly more prevalent due to increased interactions among, humans, livestock, wildlife, disease vectors, and pathogens, exacerbated by climate change and rapid human expansion into natural habitats [1,2]. Globally, zoonotic diseases have disproportionately impacted impoverished livestock workers in low- and middle-income countries, and are responsible for millions of human deaths every year [3]. To address this threat, there is a mounting need for systems to forecast and model the spread of diseases, thereby aiding public health planning and supporting early warning systems [1].
Constructing effective disease models depends on precise parametrisation, which requires data from multiple sources, including clinical records, environmental datasets, grey literature, and scientific publications [4,5]. A systematic literature review (SLR), considered the gold standard [6], is used to reliably extract and synthesise this information. The key stages of the SLR process consist of planning and developing protocols, searching, screening articles for relevance, extracting information, assessing the quality of studies, and, finally, synthesising and reporting [7,8].
Despite the reliability of SLRs, the manual process demands significant time and resource investment [9], it incurs substantial financial costs [10], and is difficult to scale given the exponential growth of publications [11]. This inefficiency creates a major bottleneck, resulting in outdated reviews, unnecessary waste, and potential harm in decision making [12].
Conducting SLRs in multidisciplinary fields, such as climate-sensitive zoonotic diseases, introduces further challenges. Research in this area spans diverse domains including epidemiology, ecology, veterinary sciences, and public health [13,14]. Identifying the relevant literature requires broad search terms and careful screening, as studies may focus on indirect indicators to assess disease risk, such as disease vector distribution or environmental factors like rainfall, temperature, humidity, and extreme weather events [15,16,17], rather than direct pathogen detection. Furthermore, confounding factors such as vector control, antimicrobial treatment, and infrastructure changes obscure the assessment of climate impact [18], rendering the selection of primary studies a cognitively intensive task.
Automation offers a cost-efficient, faster, and more reliable solution, capable of meeting the demands for scale and quality [19,20,21]. Among the most promising advancements in this domain are large language models (LLMs) based on the transformer architecture [22], which have gained significant popularity [23,24]. These models leverage vast amounts of text data and advanced contextual understanding to handle complex language tasks, including screening citations [25,26,27,28,29], extracting data [29,30,31], and synthesising information [29,32,33]. However, integrating LLMs into SLR workflows poses major challenges [24]. They can produce plausible-sounding erroneous responses known as hallucinations [34], promote negative biases seen in training data [35], lack transparency, and are considered “black-box” systems [36]. Furthermore, comprehensive evaluation of LLMs must consider diverse factors including language tasks, reasoning, robustness, trustworthiness, and ethics [37]. While these issues are significant, researchers are actively exploring various options to mitigate risks [36,37,38,39]. Targeted applications of LLMs in specific SLR tasks offer promising opportunities to enhance efficiency while minimising risks.
One effective way to leverage LLMs in the SLR workflow is through relevancy ranking for screening prioritisation, the process of ranking the most relevant studies to streamline the screening stage. Title–abstract screening is the most time-consuming aspect of the manual SLR process [40]. Best-practice guidelines recommend that primary studies are sorted by relevance before being screened [41]. This can improve reviewer efficiency, increase productivity, and facilitate faster decision making, especially for highly skewed datasets, such as those in the medical domain, which constitute only 3 % to 6 % [42] of relevant documents. The output of a relevancy ranking system can be directly utilised by selecting the top-k ranked articles based on an approximation of relevant document coverage [43,44] or, alternatively, assist users in selecting seed or example articles for input into a further specialised binary classifier tool.
Current methods for automated screening prioritisation can be categorised into query-based and model-based approaches [45]. Query-based approaches determine relevancy ranking via a query generated from attributes of the review protocol, including title, Boolean queries, or selection criteria. These methods are often grounded in traditional information retrieval techniques, such as BM25 [46]. Model-based approaches utilise classification models trained to distinguish between relevant and irrelevant documents, often employing techniques such as active learning and relevance feedback [45].
Both query-based and model-based approaches have proven effective in studies of encoder-only BERT-based large language models [43,45,47,48]. However, these models require computationally intensive fine-tuning within the operational domain [43,45,47] and present challenges in understanding the rationale behind their ranking decisions. In contrast, generative large language models, like the GPT series, which utilises only the decoder layer of the transformer architecture [49], exhibit advanced natural language understanding and can perform zero-shot task solving while providing explanations through their language generation capabilities, as demonstrated in recommendation systems [50]. Despite these advantages, their application has mostly been explored in SLR binary classification research [25,51,52,53] and remains under-explored for SLR screening prioritisation [54,55], particularly in domain-specific applications.
While query-based methods have demonstrated effectiveness in automated screening prioritisation, the success of these approaches heavily depends on the values used in the query, which are derived from the SLR eligibility criteria. These criteria constitute a series of rules and specific requirements that a document must satisfy for inclusion in the SLR. Typically, these criteria are evaluated based on research question or inclusion/exclusion criteria, but could also be framed as a question-answering problem [56]. The Question-Answering (QA) framework, a best-practice screening tool in manual SLR workflow [41], provides a fine-grained, consistent, and targeted approach to determining eligibility. The integration of a QA framework has demonstrated promising results in previous research when applied to screening prioritisation [55], while also offering a mechanism to capture model reasoning, thereby enhancing transparency and interpretability in the decision-making process. The capacity to capture model reasoning and improve transparency has frequently been overlooked in prior research.
To address these challenges, this research proposes leveraging a generative LLM for relevancy ranking for screening prioritisation in climate-sensitive zoonotic diseases through a QA framework. This approach aims to enhance the SLR workflow while capturing model responses and investigating its utility for improving transparency and interpretability for human reviewers.

1.1. Contribution

This paper contributes to enhancing information retrieval techniques using generative LLMs, advancing the automation of SLRs and investigating explainable AI methods for ranking and reviewing the literature on climate-sensitive zoonotic diseases. The specific contributions can be summarised as follows:
  • Validating the use of a generative LLM for relevancy ranking of primary studies by utilising a QA framework in the area of climate-sensitive zoonotic diseases.
  • Evaluating how well the solution can generalise across the climate-sensitive zoonotic disease literature.
  • Evaluating the utility of the LLM-generated reasoning text for human reviewers to enhance transparency and trust.

1.1.1. Problem Definition

The focus of this study is to investigate the utility of using an LLM to rank primary studies by relevance utilizing a QA-framework. Given a set of documents D and a set of eligibility criteria C , the task of the relevancy ranker is to assign each document d D a relevance score r d [ 0 , 1 ] , where r d indicates how well document d satisfies the criteria C . A score closer to 1 indicates higher relevance, while a score closer to 0 indicates lower relevance.
Using the QA framework approach, Kusa et al. describes the eligibility criteria C being transformed into a set of questions Q = { q 1 , , q | C | } , where q k corresponds to a specific criteria C. A set of predicated answers A ^ d = { a ^ k d | meets ( q k , a ^ k d ) } can be obtained for each document d D , where meets ( q k , a ^ k d ) denotes that the document d should meet the criterion expressed by q k . The relevancy score r ^ d of a document can be computed on the predicate answers A ^ d using an aggregation method such as linear weighted sum (LWS).

1.1.2. Background

There are multiple ways to use LLMs for ranking documents; rankers [43,54,57,58] and assessors are [59] popular choices. Rankers determine the order of items based on their perceived value, while assessors provide an evaluation or judgement of the quality or suitability of a single item.
Zero-shot rankers [50], have no need for prior training or examples and can be categorised into point-wise [58], which scores one query and document at a time, ranking the documents based on a score; pair-wise, where the model assesses a pair of documents against a query; or list-wise [57], which involves presenting both query and the complete list of items in the prompt. Pair-wise and list-wise methods do not scale for long lists, as is the case with SLRs where the initial search could yield a substantial list of results. Wang et al. examined the performance of zero-shot and fine-tuned point-wise neural rankers in the context of SLR document ranking and discovered that zero-shot neural rankers performed worse than traditional methods like BM25 in the absence of fine-tuning. Moreover, research utilising point-wise rankers with generative LLMs indicated that obtaining ranking scores from log-likelihood values resulted in superior ranking performance relative to employing LLM-generated labels [58]. The potential for additional output from an LLM, such as reasoning text, is constrained in this approach, as each output token, the smallest text unit a model can process, must be individually parsed, ultimately increasing complexity and limiting the interpretability of the ranker. Such constraints and the reliance on fine-tuning for improved performance highlights the need for approaches that balance scalability and transparency in zero-shot settings.
Zero-shot LLM assessors [59] perform relevancy judgement on a query-document pair when provided with a set of relevancy labels. When ground truth labels are available, these assessors are frequently employed to generate annotated datasets [60]. In addition to issuing judgement, LLM assessors can offer natural language explanations to support their conclusions; they provide scalability, consistency, and the potential to complement human assessors in judgement tasks [59].
Several studies investigating the application of LLMs in SLR workflow have highlighted the importance for human oversight [25,32,51,53] or the use of LLMs as assistant reviewers [25,61]. Adopting an LLM as an assessor can effectively realise these strategies. While there are no ground truth data available at the start of a review, the assessor’s approach can still assign labels based on predefined eligibility criteria, establishing an initial framework to guide the ranking process and provide transparency. This study examines the effectiveness of an LLM as an assessor only, utilising its capacity to provide answer labels to questions.
Both ranker and assessor approaches can be enhanced using various strategies ranging from resource intensive options, such as pre-training and fine-tuning of models [43,62], to prompt engineering [38]. Domain-specific pre-training and fine-tuning have demonstrated substantial performance gains [33,43,63]. However, techniques such as few-shot and chain-of-thought (CoT) prompting, leveraging a persona, and adopting fine-grained labels offer a more economical, less complex initial approach prior to pursuing advanced optimisations. While zero-shot prompting does not utilise task-specific examples, few-shot prompting incorporates both positive and negative examples, demonstrating promising results in title–abstract screening automation [27,62,64,65]. Considering that ranking occurs at the very start of the screening process, users typically lack domain-specific examples at this early stage, making few-shot prompts less practical for an initial ranking task. Chain-of-thought (CoT) prompting instructs the model to adopt a grounded, “step-by-step” approach to task resolution, reducing the likelihood of hallucinations, improving performance and accuracy [38]. To increase transparency, this study captures the CoT reasoning so users can gain insights into the model reasoning. Additionally, integrating a persona can enhance a model’s capacity to deliver more tailored and consistent responses across multiple interactions, adapting to a wide range of scenarios [51,66]. Finally, recent experiments with zero-shot LLM rankers indicate that fine-grained relevancy labels help guide the model to differentiate documents more effectively [58]. This study specifically examines the impact of fine-grained labels on the ranking performance using an LLM as an assessor within a zero-shot setting for prioritising primary studies.

1.2. Research Questions

This study aims to evaluate the effectiveness of using a generative LLM in the role of an assessor to assist in the relevancy ranking of primary studies in climate-sensitive zoonotic diseases. To guide the investigation, the following research questions are posed:
RQ1 
How does an LLM-based assessor utilising a QA framework compared to baseline models utilising review title and selection criteria?
RQ2 
Does the label granularity effect the ranking performance of an LLM-based assessor utilising a QA framework for climate-sensitive zoonotic disease?
RQ3 
Does the ranking performance of an LLM-based assessor generalise across climate-sensitive zoonotic disease datasets with varying relevance rate?
RQ4 
Does CoT rationale provided by an LLM assist a human reviewer’s ability to detect misclassifications in SLR?

2. Methodology

A series of experiments was conducted to investigate the research questions outlined in Section 1.2. This section provides details of the dataset, models, the evaluation metrics, and the experimental design.

2.1. Dataset

The dataset used in this study consists of 2905 title–abstracts related to four zoonotic diseases: Crimean–Congo haemorrhagic fever (CCHF), Ebola virus, Leptospirosis (Lepto), and Rift Valley fever virus (RVF), together with one climate variable, rainfall. Each combination of disease and climate variable constitutes a distinct SLR. An SLR protocol was established, outlining the selection criteria for each combination of disease and climate variable. A search was conducted in the PubMed and Scopus online journal repositories and the results were imported into a reference manager, along with titles and abstracts followed by de-duplication. A team of 5 researchers (early and mid-career researchers, postgraduate student, and undergraduate student) were trained to screen abstracts using the criteria outlined in Table 1. Abstracts that met the inclusion criteria were assigned a score of 1; otherwise a score of 0. When a reviewer was unsure, abstracts were discussed with one of the mid-career researchers who had experience in reviewing infectious diseases climate sensitivity. Abstracts that were unclear, or did not precisely meet the study selection criteria, were scored 1 to be further examined in the full-text inspection stage.
Table 2 illustrates the composition of the dataset. Ebola publications exhibited a particularly pronounced skew, with only 1.5% (14/915) deemed relevant, while the remaining diseases ranged from 10.8% to 12.6%. The titles–abstracts were exported to a CSV file along with the manually assigned scores, and labels were assigned per record to indicate the target disease and climate variable (rainfall).

2.2. QA Framework

Following the abstract screening best-practice guideline [41], a set of eligibility questions was developed to evaluate the relevance of title–abstracts based on the selection criteria. Distinct sets of eligibility questions were formulated for each disease and the associated climate variable (rainfall), ensuring their relevance to the SLR topic, as detailed in Table 3. It should be noted that reviewers did not use these questions during their screening, as they were developed retrospectively for the evaluation process. The guidelines stipulate that the questions should be clear and concise, and must be (1) objective, (2) “single-barrelled” or focused on a specific aspect of the citation, (3) use a consistent sentence structure, and (4) ensure responses are limited to yes, no, and unsure only. Furthermore, the questions should be organised hierarchically, starting with the easiest and progressing to difficult. In this study, the question text and the number of questions assigned to each disease remained constant across experiments, while the answer labels were varied. The specific variations are detailed in Section 2.4.

2.3. Prompts

The design and choice of prompting has implications on the model response [61,67]. The prompts deployed in this study were inspired by those from previous research [25,27,38,51,61,68] and received iterative refinement. The prompt performance was evaluate using recall@k and mean average precision MAP metrics detailed in Section 2.7.
A persona description stating “You are a world leading expert veterinary epidemiologist screening abstracts of scientific papers for the systematic literature review of ’$topic”’ was established prior to the main instruction prompt. All experimental runs used the same persona, with the $topic placeholder dynamically replaced by the topics listed in Table 3.
Two main instruction prompts were designed for this study. The first prompt, TSC prompt, serves as a baseline, utilising the review title and selection criteria. This prompt is an adaptation of the Zero-shot Framework CoT prompt [27], directing the model to analyse the title and apply inclusion and exclusion criteria to classify a given abstract as “Definitely Include”, “Probably Include”, “Probably Exclude”, “Definitely Exclude”, or “Unsure” (see Listing 1). The placeholders $title, $inclusion, $exclusion, and $abstract were populated dynamically according to the disease context of the experiment.
Listing 1. TSC prompt, based on the review title and selection criteria.
Make 07 00028 i001
The second prompt, QA prompt in Listing A1 (see Appendix A), utilises the QA framework discussed in Section 2.2. This prompt instructs the model to follow a structured approach to answering the eligibility questions with respect to a given abstract. The answer labels available to the model are discussed in Section 2.4. Similar to the first prompt, placeholders for $abstract and $question are dynamically populated with title–abstract and eligibility questions based on disease context.
Both prompts employ predefined answer labels and explicit reasoning, adhering to the principles of CoT prompting [38], and align with the instructive template category, as described in [65]. Given that GPT models are autoregressive [62], where previous tokens influence the final answer, the prompts were designed to explicitly request reasoning prior to delivering a final answer label. Finally, the prompts instruct the model to format the response as a JSON object; fields capture the reasoning text, answer, and, in the case of the QA framework, the question number. This structure helps ensure consistent and reliable output formatting and eliminates the need for parsing natural text while facilitating easy integration with downstream components.

2.4. Answer Labels

Table 4 lists the predefined answer schemas designed to be used with the QA and TSC prompt, restricting the model to selecting only from these options. The schema QA-3 was adapted from the best-practice guidelines [41], while QA-4 and Qa-5 were inspired by [58].
Additionally, a variant of the QA-2 schema, termed QA-2-C, a multi-level answer schema, was explored by instructing the model to generate a separate confidence score (Table 5). This method aims to more accurately capture the uncertainty that the original QA-2 fails to account for, yet appears in the other schemas. Each answer label was assigned a predefined answer score to quantify its contribution to the relevancy ranking calculation (described in Section 2.5); in the case of QA-2-C, the confidence score was used in place of the answer score. The QA prompt was adjusted to accommodate the multi-level scoring, as shown in Listing A2 (see Appendix A).

2.5. Relevancy Ranking

A relevancy score was computed for each title–abstract record by employing a linear weighted sum (LWS) of the answer scores derived from all questions within the QA framework. The records are subsequently sorted in descending order based on the relevancy score. In this study, it was assumed that all questions carried equal weight. Formally, let
  • Q = { q 1 , , q | C | } be the eligibility questions derived from the selection criteria.
  • A ^ d = { a ^ k d } be the set of predicated answers for each question q k in document d.
  • w k be a predefined weight reflecting the importance of each question q k .
Then, the relevancy score r ^ d of document d is expressed as
r ^ d = k = 1 | C | w q a ^ k d
where, for equal-weighted questions, w k = 1 k for all q. This naturally simplifies to
r ^ d = 1 k k = 1 | C | a ^ k d
In cases where ties occured in the relevancy score, title–abstract length was used as a tiebreaker by sorting the title–abstracts in descending order of length, assuming that longer articles were more relevant.

2.6. Models

2.6.1. Baseline Models

The BM25 and MiniLM v2 [69] models were utilised to establish a zero-shot baseline. BM25 is a widely used ranking algorithm used in information retrieval; it evaluates a document’s relevance with respect to a query. The query in this instance is a concatenation of the review title with the selection criteria. BM25 scores were computed for each title–abstract and query pair following a preprocessing step that included lower casing, punctuation elimination, stop word removal, and stemming. The results were ranked according to the BM25 scores, with the text length serving as a tie-breaker.
MiniLM v2 is a BERT-based, pre-trained sentence-transformer model optimised for embedding text, and it is trained on a massive and diverse dataset [70]. This study used the all-MiniLM-L6-v2 model without fine-tuning. Embedding vectors were generated for both title–abstracts and queries. Cosine similarity was computed as per Equation (3) for each vector pair, and subsequently normalised and sorted, with the text length serving as a tie-breaker.
cosine _ similarity ( d , q ) = d · q d q
This methodology for establishing a baseline is aligned with the approach implemented by CSMed, a meta-dataset comprising 325 systematic literature reviews from the medical and computer scientific fields [56].

2.6.2. Large Language Models

For this experiment, GPT-4o-mini-2024-07-18, a smaller and optimised version of GPT-4, was utilised to achieve a balance between cost, performance, and computational efficiency [71]. The model was used in its pre-trained form without any fine-tuning and accessed via the OpenAI API. The sampling temperature value was set to 0 to improve reproducibility and ensure deterministic responses between invocations but also to ensure that it did not affect the model’s ability to self-correct [67]. The max_token parameter was set to 512 to provide sufficient context for processing the prompts and to generate concise reasoning responses. The response_format parameter was used to ensure consistent and reliable output formatting. A JSON Schema was assigned to this parameter that describes the structure of the expected output, preventing the need for parsing of natural text and facilitating integration with downstream code.

2.7. Evaluation Metrics

The performance of ranking tasks in information retrieval challenges is evaluated using rank-based metrics and metrics at predefined cut-offs, such as the top k % of retrieved documents [56,72]. The evaluation metrics used to measure the performance of the approach are listed below:
  • Recall @ k%
Recall, also known as sensitivity, provides a measure of how many relevant documents are retrieved within the top-k% of ranked results.
  • nWSS @ r%
Normalized work saved over sampling (nWSS), equivalent to true negative rate (or Specificity), quantifies the effort or work saved by the automated system when compared to random sampling, assuming a fixed recall level, and can be utilized to evaluate outcomes across models and datasets [73]. Here nWSS is evaluated at r = 95 % and r = 100 % .
  • AP
Average precision (AP) represents the average of precision computed at each relevant document’s position, considering all documents retrieved up to a specific rank. It combines both precision and recall, evaluating how effective documents are ranked by relevance. Unlike precision @ k%, which is influenced by the total number of relevant documents, AP addresses this limitation by providing a more balanced assessment [72].
  • MAP
Mean average precision (MAP) represents the average AP across individual information needs and provides a single robust metric for evaluating the ranking quality across recall levels [72].

2.8. Experimental Setup

The experiments were automated using Python 3.9.21, with each experimental run organised into separate Python scripts. The baseline experiments utilised title and selection criteria, abbreviated as TSC in conjunctions with BM25, MiniLM, and ChatGPT-4o-mini models. Meanwhile, the QA framework was exclusively tested with the ChatGPT-4o-mini model using all QA answer schemas list in Table 4 and Table 5.
Each execution script requires two input files: a CSV of title–abstracts and a prompt template file, which is dynamically populated with experimental context-specific values for each zoonotic disease: CCHF, Ebola, Lepto, and RVF (Figure 1). The code then calls the OpenAI API completion endpoint using predefined model parameters and the adapted prompts. By utilizing the API, the study ensures reproducibility. All responses from the API are captured and subsequently processed to calculate a relevancy ranking.

3. Results

This section presents the results of the eight models evaluated across the four diseases: CCHF, Ebola, Lepto, and RVF, evaluated against the study metric (Table 6).
The following sections will examine the results by each research question.

3.1. RQ1. LLM-Based QA Assessor vs. Baseline

For RQ1, the comparison is between QA-based models and the TSC baseline generative (TSC-5) and non-generative (TSC-BM25, TSC-MiniLM) models.
Recall@k%: The results illustrated in Figure 2 exhibit that QA-based models consistently performed better than the TSC baseline models. At k = 50%, all QA-based models reached recall levels above 0.95. They also attained higher recall scores than the baseline in the early ranks k < 20%, particularly for Ebola.
nWSS@r%: In Figure 3, QA-4 showed the highest efficiency, especially for CCHF (0.78 at 95%), Ebola (0.86 at 95%), Lepto (0.72 at 95%), and RVF (0.70 at 95%), surpassing baseline models across all diseases. While the nWSS of QA-4 dropped when targeting r = 100%, it still remained superior to the baseline models. In contrast, the TSC models displayed lower overall work savings, with particularly high variability in the LLM-based TSC model: it ranged from negative savings with Ebola to 46% nWSS@95% with Lepto.
AP: This metric (Figure 4) evaluates ranking quality by measuring the number of relevant documents and how early they appear. The QA models, particularly QA-4, showed superior performance over the baseline models, with QA-4 achieving the highest AP values for all diseases except Ebola, where the performance was comparable to QA-5.
Ranking: The box plot in Figure 5 shows the distribution of the ranking positions for each model across the diseases. The QA models have narrower inter-quartile ranges (IQRs) and shorter whiskers, while the baseline models display higher medians and wider variability in rank positions.
MAP: The MAP metric in Table 7 offers a single performance measure for each model. The QA-4 model attained the highest MAP score of 0.691, while the baseline TSC-BM25 model had the lowest of 0.229. Notably, the TSC-5 model achieved a 0.489, beating the QA-2 model. A precision–recall curve (Figure 6) visualises overall performance by illustrating the effect of varying thresholds on the precision–recall balance. The QA models exhibit a larger area under the PR curve compared to the baseline models, although the TSC-5 model is positioned closer to the QA model curve in the precision–recall space than those of TSC-BM25 and TSC-MiniLM.

3.2. RQ2. Effect of Label Granularity

This section analyses performance across the QA-based models. QA-2 and QA-3 are categorised as single-level coarse-grained label models, while QA-4 and QA-5 are categorised as single-level fine-grained label models. QA-2-C is identified as a multi-level label model.
Recall@k%: In Figure 2, generally, recall improves with granularity (QA-2 to QA-4), with diminishing returns observed from QA-4 to QA-5. Single-level fine-grained labels (QA-4, QA-5) exhibit high recall at early ranks (k < 20%). By k = 50%, there is not much difference as they reach near complete recall ( 0.98 ) across all disease datasets.
nWSS@r%: Results for nWSS@r% indicate that QA-2-c, QA-4, and QA-5 exhibit significant work savings at 95% recall target over coarse-grained labels (QA-2, QA-3), achieving a minimum of 61%, 70%, and 67%, respectively. In particular, QA-4 and QA-5 exhibit exceptional savings on highly skewed datasets such as Ebola, realising 83% to 88% savings at recall levels of 95% and 100%.
AP: Fine-grained labels tend to lead to higher precision, with QA-4 consistently achieving the highest AP scores across diseases, particularly for Lepto and RVF, with QA-5 not far behind.
Ranking: Fine-grained label models (QA-4, QA-5) are positioned higher in the plot, indicating lower ranking, and consistently have lower medians across all diseases when compared to the other models.
MAP: The MAP scores in Table 7 demonstrate a clear trend, indicating improved ranking performance with increased label granularity but peaking at QA-4 (0.691). This is also evident in the precision–recall curve, where QA-4 maintains higher precision across most recall values compared to QA-5, which trails slightly below it in the precision–recall space.

3.3. RQ3. Performance Across Zoonotic Diseases

This subsection examines performance variations across all models (QA and TSC) and diseases (Ebola, Lepto, CCHF, RVF).
Recall@k%: When looking across zoonotic disease datasets, QA-based models maintain a recall level between 0.95 and 1.00 at k = 50%. In the Ebola dataset, QA-4 and QA-5 models demonstrate a significantly high recall, achieving complete recall at k = 15 % . The baseline models did not achieve a comparable level of recall, with BM25 exhibiting a consistent but significantly lower recall across all diseases. In contrast, TSC-5 and TSC-MiniLM performed well with Lepto but severely unperformed with other diseases.
nWSS@r%: For all diseases, the QA models retain a minimum of 54% work saving at nWSS@95%; however, performance drops when transitioning to nWSS@100% for all disease except Ebola, where savings slightly increase. Notably, TSC-MiniLM achieves the same work savings as QA-4 and QA-2-C (nWSS@95% = 0.72); however, when striving for full recall, this value reduces dramatically, proving high variability.
AP: A consistently high AP across diseases indicates models that can handle disease-specific terminology and generalise well independently of relevancy rates. The leading QA models (QA-4, QA-5) achieved a minimum of 0.5 AP across all diseases while the baseline models exhibit lower and greater variability in AP.
Ranking: QA models displayed the best rank distribution with the Ebola dataset (Figure 5), demonstrated by the lower median rank, tighter IQR, and few outliers. Similar patterns were seen in Lepto, CCHF, and RVF; however, it is important to note that the Lepto dataset has the highest number of outliers.

3.4. RQ4. Utility of Generated CoT Rationale

The CoT reasoning logic proved to be consistent, in line with human expectations, helpful in detecting edge cases, and capable of revealing variations in misclassifications across disease datasets. Listing 2 is an example RVF response from the QA-4 model that demonstrates the logic that the model provides for each title–abstract record. A senior reviewer on the team was presented with the results to manually assess the validity of the responses. For each disease, title–abstracts with relevancy scores exceeding a predetermined threshold and classified as irrelevant, as well as those below the threshold classified as relevant, were displayed in MS Excel alongside the answer scores and reasoning text for each question. The reviewer did report any significant issues in the model’s reasoning and considered the explanatory text highly beneficial for comprehending the logic underlying the answer scores. Using the sort and filter facilities provided in MS Excel, the reviewer was able to efficiently review and identify several misclassified items, as shown in Table 8. Lepto (16), RVF (13), and CCHF(10) had the most misclassifications, highlighting the inherent variability in manual SLR screening quality but also opportunities to refine the selection criteria and eligibility questions. These findings not only highlight the potential of CoT reasoning to ensure transparency in the SLR process but also highlight a means to iteratively enhance the ranking process.
Listing 2. Example JSON response for an RVF abstract.
Make 07 00028 i002

4. Discussion

This study explored the use of a generative LLM as an assessor for screening prioritisation in climate-sensitive zoonotic disease research. Findings revealed that a zero-shot LLM-based QA assessor, leveraging fine-grained labels, can effectively rank primary studies by relevance and demonstrates broad generalisability across four climate-sensitive zoonotic disease datasets with differing relevancy rates. Additionally, the CoT reasoning text generated by this approach provided valuable insight to human reviewers and aided in the identification of misclassified records in the disease datasets.
Before further interpretation of the results, the following important limitations are noteworthy. Most importantly, the study depended on a single reviewer for the manual evaluation of the model’s reasoning text. No metrics were established to measure the quality or potential bias of the text generated by the LLM, which are prone to “hallucination” [34,35]. Nevertheless, the results from the study demonstrated a potential use case for such data and the possibility for creating a unique dataset to enable the team to conduct further research into reasoning and bias within LLM-based approaches.
This study used a single closed-source commercial LLM from OpenAI to conduct all experiments. Restricting the evaluation solely to ChatGPT-4-o-mini restricts the generalisation to other LLMs. Additionally, while a fixed model version was used with the temperature parameter set to 0 to ensure deterministic behaviour, and prior research confirming consistent outcomes on repeated invocation [61], ongoing optimisations by OpenAI may have led to performance changes that could have influenced the results. In the future, the team intends to explore additional models, including open-source alternatives applying the same methodology.
Moreover, the eligibility questions used in the experiments were generated in retrospect following manual screening; hence, there is a risk of overfitting and misalignment with broader applicability. Polanin et al. advocates for conducting a pilot screening session with a subset of abstracts to refine and validate eligibility questions prior to screening. This approach could ensure clarity and consistency of the questions while improving the effectiveness of the rankings. Such an exercise may prove beneficial for future screening efforts, as it could mitigate bias and facilitate the development of a more comprehensive dataset. Additionally, the collection of ground truth data at the question level would facilitate a more precise and detailed evaluation of the model and support further research.
Lastly, although the real-world setting of this study’s dataset demonstrated the effectiveness of the approach, it has not been thoroughly evaluated on datasets from other domains. CSMed is an initiative to create a standardised dataset to evaluate the performance of automated SLR screening models [56], a gap highlighted in recent research [24]. It provides access to several hundred SLRs in the medical and computer science domain and presents an opportunity to benchmark the approach proposed in this study and gain a broader understanding of its performance, generalisability, and adaptability across domains.
Despite the limitations, this study demonstrates the effectiveness of employing a QA framework with an LLM-based assessor to robustly rank the literature on four climate-sensitive zoonotic diseases by relevance. The QA models consistently outperformed all the baseline models, achieving high recall@k% and MAP scores across all disease datasets in this study. High recall is crucial in SLR automation systems [74] to prevent bias and ensure that all relevant articles are identified, while a high MAP score reflects strong discrimination and resilience to imbalanced datasets [72], ensuring that reviewers are presented with both manageable and highly relevant articles. Additionally, the high nWSS@r% values (Figure 3) obtained by the top QA models demonstrate their ability to detect irrelevant documents early in the screening process, thereby reducing the number of items requiring manual review. The leading QA models (QA-4) demonstrate a minimum savings of 70% at r = 95% across all diseases; however, performance declines when aiming for complete recall. Lastly, the QA models exhibit less variability and more consistency across the datasets compared to the baseline models, as evidenced in Figure 5 by the narrower inter-quartile ranges.
While the study results are promising, it is important to contextualise the system performance. Direct comparisons to other systems must be made with caution due to the challengingly unique climate-sensitive zoonotic disease dataset employed in this study. However, the baseline TSC-BM25 model provides a simple yet informative reference point. In the study by Wang et al. examining the performance of neural rankers on the CLEF [75] datasets, a baseline BM25 model recorded a recall@20% ranging from 0.52 to 0.64, alongside an MAP score of 0.16, using title as the input query to the model. By comparison, TSC-BM25 in this study recorded a recall@20% ranging from 0.37 to 0.57 and a slightly higher MAP score of 0.23 using title and selection criteria as the input query. Their most effective fine-tuned BioBERT neural ranker demonstrated a recall@20% ranging from 0.82 to 0.89, and an MAP score of 0.381. In contrast, the QA-4 model in this study attained a recall@20% ranging from 0.83 to 1.0, and a significantly higher MAP score of 0.670. Although the highly skewed Ebola dataset may have contributed to the elevated recall score, the findings highlight the potential of the zero-shot LLM-based QA framework for enhancing ranking performance without the need for fine-tuning. However, any conclusions should be deferred until a comprehensive evaluation with a standardised benchmarking dataset is conducted.
The adoption of a QA framework has been shown to be highly effective in the current study, contrary to the findings of prior research such as Kohandel Gargari et al., where a screening tool approach yielded poor performance. Upon examining their methodology, several factors may have contributed to the discrepancy, including the complexity of the prompt, ambiguity in managing “unclear” labels, and the assumption that the model will adhere to the embedded logic flow in the prompt without errors. Additionally, their research employed GPT 3.5, an older model that is less advanced and less accurate. In contrast, the findings from Akinseloyin et al. align more closely with this study, utilising a similar QA framework methodology. However, their eligibility questions were derived from the inclusion/exclusion criteria using an LLM and observed that the generated questions lacked complete independence and recommended that such questions be created by humans for better reliability, a recommendation followed in this study.
In addition to the QA framework, the use of fine-grained labels appears to have a consistent positive impact on ranking tasks, whether in an assessor or a ranker context. The single-level fine-grained models, QA-4 and QA-5, demonstrated superior recall, precision, and work savings, with minimal variability, compared to the coarser-grained models, QA-2 and QA-3. Fine-grained labels appear to enhance the system performance by supplying the model with broader answer options, allowing it to more effectively represent uncertainty. However, the benefits diminished as granularity increased beyond four levels, as evidenced by the reduced MAP score for QA-5. Zhuang et al. found that fine-grained relevancy labels improved performance, with no advantage in exceeding four levels of granularity using a point-wise LLM-based ranker. While the tasks in these studies differ (assessor vs. ranker), this suggests that fine-grained labels may be broadly beneficial for LLM-based ranking tasks.
Interestingly, the shift from a single-level to a multi-level labelling approach (QA-2-c), which pairs binary labels with confidence scores and demonstrates a middle ground between fine-grained (QA-4, QA-5) and coarse-grained (QA-2, QA-3) labels, resulted in a reduced performance. Several factors could account for this outcome. Firstly, the multi-level model employs a more intricate prompt; simple prompt structures are more effective than complex ones [76]. Furthermore, a single-level classification inherently captures both the relevance and uncertainty in one consolidated label (e.g., “Probably Yes”), whereas a two-step classification introduces uncertainty post-decision, thereby splitting the context. Finally, the model’s autoregressive characteristic [62], combined with single versus multiple decision-making points, may account for the observed performance differences.
The study approach also shows promising generalisability across climate-sensitive zoonotic disease datasets, despite their varied characteristics. The Ebola dataset reveals a significant skew, with only 1.5% of relevant records, whereas the CCHF, Lepto, and RVF datasets demonstrate moderate skews, as outlined in Table 2. This variation is apparent when evaluating model performance across the different diseases. Yet, QA-based models, specifically QA-4 and QA-5, retained higher recall@k%, nWSS@95%, and AP scores while maintaining consistent ranking across the datasets in comparison to the baseline models. The QA models also maintained the best rank distribution across all diseases (Figure 5), demonstrated by the lower median rank and tighter IQR, suggesting that QA models are superior to baseline models at prioritising relevant articles in the earlier ranks. Nevertheless, the Lepto dataset has the greatest number of outliers, which may be a sign of ambiguous abstracts, misaligned classification criteria, or model limitations.
Nicholson Thomas et al. used a selection-criteria-based prompt, similar to the TSC-5 baseline model, to screen articles for ecosystem condition indicators. They report that to handle multidimensional topics with high precision, iterative refinement of the selection criteria was essential. While further enhancements of the selection criteria could enhance the performance of the TSC-5 model for diseases where it underperformed (other than Lepto), the QA framework offers an additional layer of flexibility. It decouples the decision making from the LLM’s internal logic by decomposing the assessment into granular questions. Weak signals in the form of uncertainty labels (e.g., “Probably No” or “Unsure”) allow studies that marginally fail to still contribute to overall ranking through the LWS. This flexibility ensures that relevant studies are not prematurely excluded, and provides adaptability to the varied disease datasets.
Complementing the flexibility of the QA framework is the utility of reasoning text generated through CoT prompting, which provides a detailed explanation for each relevancy assessment, exposing the models’ decision-making processes to the reviewers. For example, the reasoning text in Listing 2 illustrates the model’s ability to justifying uncertain responses like “Probably No” in question 2, where it explains that although the study addresses disease vectors, it does not directly assess virus prevalence. On the spectrum of human–machine collaboration, this approach of automatically generating a first-pass judgement with rationale falls into the human verification category, or human-in-the-loop approach [59]. While CoT prompting does not ensure the accuracy of the reasoning path, it does enable the model to more effectively access relevant data learned during pre-training [38], which may explain the QA models’ performances over the baseline models. Further investigation is required to assess the validity and reliability of the generated rationale.
Finally, the approach proposed in this study can currently be used as a standalone decision aid to reduce the risk of human error and bias during the initial screening phase. The main prompt provided in the supplementary code is generic and applicable across domains. All that is needed is to formulate research-specific eligibility questions, define the review topic, and modify the persona text accordingly. Data are output as a CSV file containing ranking scores, answers, and reasoning text, and can be reviewed in a tool such as MS Excel, providing an additional layer of adaptability. By prioritising and streamlining the review process, this approach allows human reviewers to focus on the edge cases, making it a practical and efficient solution for SLRs.

5. Conclusions

This study reports the empirical results of a relevancy ranking approach that leverages an LLM as an assessor, guided by a QA framework, to rank the literature on four climate-sensitive zoonotic diseases and one climate variable (rainfall). The findings demonstrate that an LLM-based QA assessor using zero-shot CoT prompting can effectively and reliably rank primary studies by relevance across four zoonotic disease datasets with varying relevance rates. Notably, fine-grained QA models significantly outperform the baseline models, achieving strong recall at multiple thresholds, high MAP scores, and substantial work savings of at least 70% (NWSS@95%). Additionally, the CoT reasoning text generated by this approach provides valuable insight, assisting researchers to identify several misclassified items and enhancing transparency and confidence in the screening process.
Although further empirical research is necessary to validate the approach against standardised benchmark datasets, the substantial reduction in screening effort, combined with the provision of explainable AI rationales, represents an important step toward automated parameter extraction from the scientific literature.

Author Contributions

Conceptualization, M.S., M.W., E.V. and T.S.; Methodology, M.S. and E.V.; Software, M.S.; Validation, M.S., M.W., E.V. and T.S.; Formal analysis, M.S.; Investigation, M.S.; Resources, E.V.; Data curation, M.S., M.W., E.V. and N.H.; Writing—original draft, M.S.; Writing—review & editing, E.V. and T.S.; Visualization, M.S.; Supervision, M.W., E.V. and T.S.; Project administration, E.V.; Funding acquisition, M.W. and E.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Wellcome Trust grant 226044/Z/22/Z, “a Digital Technology Development Award in Climate Sensitive Infectious Disease Modelling”.

Data Availability Statement

All of the code used in this study is accessible in a github repository (link provided upon publishing).

Acknowledgments

We gratefully acknowledge funding from the Wellcome grant.

Conflicts of Interest

This work is funded by a Wellcome grant for the development of the CliZod system. The funding body had no role in the study design, data collection, analysis, or interpretation. We have no conflicts of interest to declare.

Abbreviations

The following abbreviations are used in this manuscript:
MDPIMultidisciplinary Digital Publishing Institute
LLMLarge language model
QAQuestion and answer
CCHFCrimean–Congo haemorrhagic fever
RVFRift valley fever

Appendix A. Additional Prompts and Disease Information

Listing A1. QA single-level prompt, based on QA framework.
Make 07 00028 i003
Listing A2. QA multi-level prompt, based on QA framework and supporting an answer and a confidence level.
Make 07 00028 i004

References

  1. Ryan, S.J.; Lippi, C.A.; Caplan, T.; Diaz, A.; Dunbar, W.; Grover, S.; Johnson, S.; Knowles, R.; Lowe, R.; Mateen, B.A.; et al. The Current Landscape of Software Tools for the Climate-Sensitive Infectious Disease Modelling Community. Lancet Planet. Health 2023, 7, e527–e536. [Google Scholar] [CrossRef]
  2. Allen, T.; Murray, K.A.; Zambrana-Torrelio, C.; Morse, S.S.; Rondinini, C.; Di Marco, M.; Breit, N.; Olival, K.J.; Daszak, P. Global Hotspots and Correlates of Emerging Zoonotic Diseases. Nat. Commun. 2017, 8, 1124. [Google Scholar] [CrossRef]
  3. Grace, D.; Mutua, F.K.; Ochungo, P.; Kruska, R.L.; Jones, K.; Brierley, L.; Lapar, M.L.; Said, M.Y.; Herrero, M.T.; Phuc, P.M.; et al. Mapping of Poverty and Likely Zoonoses Hotspots; Technical Report; International Livestock Research Institute: Nairobi, Kenya, 2012. [Google Scholar]
  4. Gubbins, S.; Carpenter, S.; Mellor, P.; Baylis, M.; Wood, J. Assessing the Risk of Bluetongue to UK Livestock: Uncertainty and Sensitivity Analyses of a Temperature-Dependent Model for the Basic Reproduction Number. J. R. Soc. Interface 2008, 5, 363–371. [Google Scholar] [CrossRef]
  5. Guis, H.; Caminade, C.; Calvete, C.; Morse, A.P.; Tran, A.; Baylis, M. Modelling the Effects of Past and Future Climate on the Risk of Bluetongue Emergence in Europe. J. R. Soc. Interface 2011, 9, 339–350. [Google Scholar] [CrossRef]
  6. Dekkers, R.; Carey, L.D.; Langhorne, P. Making Literature Reviews Work: A Multidisciplinary Guide to Systematic Approaches; Springer: Cham, Swizerland, 2023. [Google Scholar]
  7. Chandler, J.; Cumpston, M.; Li, T.; Page, M.J.; Welch, V.J.H.W. Cochrane Handbook for Systematic Reviews of Interventions, 2nd ed.; Higgins, J.P.T., Thomas, J., Chandler, J., Cumpston, M., Li, T., Page, M.J., Welch, V.A., Eds.; Cochrane Book Series; Wiley-Blackwell: Hoboken, NJ, USA, 2019. [Google Scholar]
  8. Kitchenham, B.; Brereton, P. A systematic review of systematic review process research in software engineering. Inf. Softw. Technol. 2013, 55, 2049–2075. [Google Scholar] [CrossRef]
  9. Tricco, A.C.; Brehaut, J.; Chen, M.H.; Moher, D. Following 411 Cochrane Protocols to Completion: A Retrospective Cohort Study. PLoS ONE 2008, 3, e3684. [Google Scholar] [CrossRef]
  10. Michelson, M.; Reuter, K. The Significant Cost of Systematic Reviews and Meta-Analyses: A Call for Greater Involvement of Machine Learning to Assess the Promise of Clinical Trials. Contemp. Clin. Trials Commun. 2019, 16, 100443. [Google Scholar] [CrossRef]
  11. Bornmann, L.; Haunschild, R.; Mutz, R. Growth Rates of Modern Science: A Latent Piecewise Growth Curve Approach to Model Publication Numbers from Established and New Literature Databases. Humanit. Soc. Sci. Commun. 2021, 8, 224. [Google Scholar] [CrossRef]
  12. Bashir, R.; Surian, D.; Dunn, A.G. Time-to-Update of Systematic Reviews Relative to the Availability of New Evidence. Syst. Rev. 2018, 7, 195. [Google Scholar] [CrossRef]
  13. Jones, B.A.; Grace, D.; Kock, R.; Alonso, S.; Rushton, J.; Said, M.Y.; McKeever, D.; Mutua, F.; Young, J.; McDermott, J.; et al. Zoonosis Emergence Linked to Agricultural Intensification and Environmental Change. Proc. Natl. Acad. Sci. USA 2013, 110, 8399–8404. [Google Scholar] [CrossRef]
  14. Shaheen, M.N.F. The Concept of One Health Applied to the Problem of Zoonotic Diseases. Rev. Med. Virol. 2022, 32, e2326. [Google Scholar] [CrossRef]
  15. Jones, K.E.; Patel, N.G.; Levy, M.A.; Storeygard, A.; Balk, D.; Gittleman, J.L.; Daszak, P. Global Trends in Emerging Infectious Diseases. Nature 2008, 451, 990–993. [Google Scholar] [CrossRef]
  16. Liao, H.; Lyon, C.J.; Ying, B.; Hu, T. Climate Change, Its Impact on Emerging Infectious Diseases and New Technologies to Combat the Challenge. Emerg. Microbes Infect. 2025, 13, 2356143. [Google Scholar] [CrossRef]
  17. Van de Vuurst, P.; Escobar, L.E. Climate Change and Infectious Disease: A Review of Evidence and Research Trends. Infect. Dis. Poverty 2023, 12, 51. [Google Scholar] [CrossRef]
  18. Altizer, S.; Ostfeld, R.S.; Johnson, P.T.J.; Kutz, S.; Harvell, C.D. Climate Change and Infectious Diseases: From Evidence to a Predictive Framework. Science 2013, 341, 514–519. [Google Scholar] [CrossRef]
  19. Clark, J.; McFarlane, C.; Cleo, G.; Ishikawa Ramos, C.; Marshall, S. The Impact of Systematic Review Automation Tools on Methodological Quality and Time Taken to Complete Systematic Review Tasks: Case Study. JMIR Med. Educ. 2021, 7, e24418. [Google Scholar] [CrossRef]
  20. Thomas, J.; McDonald, S.; Noel-Storr, A.; Shemilt, I.; Elliott, J.; Mavergames, C.; Marshall, I.J. Machine Learning Reduced Workload with Minimal Risk of Missing Studies: Development and Evaluation of a Randomized Controlled Trial Classifier for Cochrane Reviews. J. Clin. Epidemiol. 2021, 133, 140–151. [Google Scholar] [CrossRef]
  21. Tsafnat, G.; Glasziou, P.; Choong, M.K.; Dunn, A.; Galgani, F.; Coiera, E. Systematic Review Automation Technologies. Syst. Rev. 2014, 3, 74. [Google Scholar] [CrossRef]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2017; Volume 30. [Google Scholar]
  23. Santos, Á.O.D.; Da Silva, E.S.; Couto, L.M.; Reis, G.V.L.; Belo, V.S. The Use of Artificial Intelligence for Automating or Semi-Automating Biomedical Literature Analyses: A Scoping Review. J. Biomed. Inform. 2023, 142, 104389. [Google Scholar] [CrossRef]
  24. Bolanos, F.; Salatino, A.; Osborne, F.; Motta, E. Artificial Intelligence for Literature Reviews: Opportunities and Challenges. Artif. Intell. Rev. 2024, 57, 259. [Google Scholar] [CrossRef]
  25. Guo, E.; Gupta, M.; Deng, J.; Park, Y.J.; Paget, M.; Naugler, C. Automated Paper Screening for Clinical Reviews Using Large Language Models: Data Analysis Study. J. Med. Internet Res. 2024, 26, e48996. [Google Scholar] [CrossRef]
  26. Issaiy, M.; Ghanaati, H.; Kolahi, S.; Shakiba, M.; Jalali, A.; Zarei, D.; Kazemian, S.; Avanaki, M.; Firouznia, K. Methodological Insights into ChatGPT’s Screening Performance in Systematic Reviews. BMC Med. Res. Methodol. 2024, 24, 78. [Google Scholar] [CrossRef]
  27. Cao, C.; Sang, J.; Arora, R.; Kloosterman, R.; Cecere, M.; Gorla, J.; Saleh, R.; Chen, D.; Drennan, I.; Teja, B.; et al. Prompting Is All You Need: LLMs for Systematic Review Screening. medRxiv 2024. [Google Scholar] [CrossRef]
  28. Alshami, A.; Elsayed, M.; Ali, E.; Eltoukhy, A.E.E.; Zayed, T. Harnessing the Power of ChatGPT for Automating Systematic Review Process: Methodology, Case Study, Limitations, and Future Directions. Systems 2023, 11, 351. [Google Scholar] [CrossRef]
  29. Fernandes Torres, J.P.; Mulligan, C.; Jorge, J.; Moreira, C. PROMPTHEUS: A Human-Centered Pipeline to Streamline Slrs with Llms. arXiv 2024, arXiv:2410.15978. [Google Scholar]
  30. Schmidt, L.; Finnerty Mutlu, A.N.; Elmore, R.; Olorisade, B.K.; Thomas, J.; Higgins, J.P.T. Data Extraction Methods for Systematic Review (Semi)Automation: Update of a Living Systematic Review. F1000Research 2023, 10, 401. [Google Scholar] [CrossRef]
  31. Polak, M.P.; Morgan, D. Extracting Accurate Materials Data from Research Papers with Conversational Language Models and Prompt Engineering. Nat. Commun. 2024, 15, 1569. [Google Scholar] [CrossRef]
  32. Nicholson Thomas, I.; Roche, P.; Grêt-Regamey, A. Harnessing Artificial Intelligence for Efficient Systematic Reviews: A Case Study in Ecosystem Condition Indicators. Ecol. Inform. 2024, 83, 102819. [Google Scholar] [CrossRef]
  33. Susnjak, T.; Hwang, P.; Reyes, N.H.; Barczak, A.L.C.; McIntosh, T.R.; Ranathunga, S. Automating Research Synthesis with Domain-Specific Large Language Model Fine-Tuning. arXiv 2024, arXiv:2404.08680. [Google Scholar] [CrossRef]
  34. Ji, Z.; Yu, T.; Xu, Y.; Lee, N.; Ishii, E.; Fung, P. Towards Mitigating Hallucination in Large Language Models via Self-Reflection. arXiv 2023, arXiv:2310.06271. [Google Scholar] [CrossRef]
  35. Zack, T.; Lehman, E.; Suzgun, M.; Rodriguez, J.A.; Celi, L.A.; Gichoya, J.; Jurafsky, D.; Szolovits, P.; Bates, D.W.; Abdulnour, R.E.E.; et al. Assessing the Potential of GPT-4 to Perpetuate Racial and Gender Biases in Health Care: A Model Evaluation Study. Lancet Digit. Health 2024, 6, e12–e22. [Google Scholar] [CrossRef] [PubMed]
  36. Zhao, H.; Chen, H.; Yang, F.; Liu, N.; Deng, H.; Cai, H.; Wang, S.; Yin, D.; Du, M. Explainability for Large Language Models: A Survey. ACM Trans. Intell. Syst. Technol. 2024, 15, 20:1–20:38. [Google Scholar] [CrossRef]
  37. Chang, Y.; Wang, X.; Wang, J.; Wu, Y.; Yang, L.; Zhu, K.; Chen, H.; Yi, X.; Wang, C.; Wang, Y.; et al. A Survey on Evaluation of Large Language Models. ACM Trans. Intell. Syst. Technol. 2024, 15, 39:1–39:45. [Google Scholar] [CrossRef]
  38. Wei, J.; Wang, X.; Schuurmans, D.; Bosma, M.; Xia, F.; Chi, E.; Le, Q.V.; Zhou, D. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. In Advances in Neural Information Processing Systems 35 (NeurIPS 2022); Neural Information Processing Systems Foundation, Inc. (NeurIPS): New Orleans, LA, USA, 2022; Volume 35, pp. 24824–24837. [Google Scholar]
  39. Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.t.; Rocktäschel, T.; et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2020; Volume 33, pp. 9459–9474. [Google Scholar]
  40. Scott, A.M.; Forbes, C.; Clark, J.; Carter, M.; Glasziou, P.; Munn, Z. Systematic Review Automation Tools Improve Efficiency but Lack of Knowledge Impedes Their Adoption: A Survey. J. Clin. Epidemiol. 2021, 138, 80–94. [Google Scholar] [CrossRef]
  41. Polanin, J.R.; Pigott, T.D.; Espelage, D.L.; Grotpeter, J.K. Best Practice Guidelines for Abstract Screening Large-Evidence Systematic Reviews and Meta-Analyses. Res. Synth. Methods 2019, 10, 330–342. [Google Scholar] [CrossRef]
  42. Sampson, M.; Tetzlaff, J.; Urquhart, C. Precision of Healthcare Systematic Review Searches in a Cross-sectional Sample. Res. Synth. Methods 2011, 2, 119–125. [Google Scholar] [CrossRef]
  43. Wang, S.; Scells, H.; Koopman, B.; Zuccon, G. Neural Rankers for Effective Screening Prioritisation in Medical Systematic Review Literature Search. In Proceedings of the26th Australasian Document Computing Symposium, ADCS’22, Adelaide, SA, Australia, 15–16 December 2022; pp. 1–10. [Google Scholar] [CrossRef]
  44. Mitrov, G.; Stanoev, B.; Gievska, S.; Mirceva, G.; Zdravevski, E. Combining Semantic Matching, Word Embeddings, Transformers, and LLMs for Enhanced Document Ranking: Application in Systematic Reviews. Big Data Cogn. Comput. 2024, 8, 110. [Google Scholar] [CrossRef]
  45. Mao, X.; Zhuang, S.; Koopman, B.; Zuccon, G. Dense Retrieval with Continuous Explicit Feedback for Systematic Review Screening Prioritisation. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, Washington, DC, USA, 14– 18 July 2024; pp. 2357–2362. [Google Scholar] [CrossRef]
  46. Robertson, S.; Zaragoza, H. The Probabilistic Relevance Framework: BM25 and Beyond. Found. Trends® Inf. Retr. 2009, 3, 333–389. [Google Scholar] [CrossRef]
  47. Yang, E.; MacAvaney, S.; Lewis, D.D.; Frieder, O. Goldilocks: Just-Right Tuning of BERT for Technology-Assisted Review. In Proceedings of the Advances in Information Retrieval; Springer: Cham, Switzerland, 2022; pp. 502–517. [Google Scholar] [CrossRef]
  48. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers); Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Minneapolis, MN, USA, 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
  49. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving language understanding by generative pre-training. OpenAI Tech Report; OpenAI Research: San Francisco, CA, USA, 2018. [Google Scholar]
  50. Wu, L.; Zheng, Z.; Qiu, Z.; Wang, H.; Gu, H.; Shen, T.; Qin, C.; Zhu, C.; Zhu, H.; Liu, Q.; et al. A Survey on Large Language Models for Recommendation. World Wide Web 2024, 27, 60. [Google Scholar] [CrossRef]
  51. Kohandel Gargari, O.; Mahmoudi, M.H.; Hajisafarali, M.; Samiee, R. Enhancing Title and Abstract Screening for Systematic Reviews with GPT-3.5 Turbo. BMJ Evid.-Based Med. 2024, 29, 69–70. [Google Scholar] [CrossRef] [PubMed]
  52. Matsui, K.; Utsumi, T.; Aoki, Y.; Maruki, T.; Takeshima, M.; Takaesu, Y. Human-Comparable Sensitivity of Large Language Models in Identifying Eligible Studies Through Title and Abstract Screening: 3-Layer Strategy Using GPT-3.5 and GPT-4 for Systematic Reviews. J. Med. Internet Res. 2024, 26, e52758. [Google Scholar] [CrossRef]
  53. Sanghera, R.; Thirunavukarasu, A.J.; Khoury, M.E.; O’Logbon, J.; Chen, Y.; Watt, A.; Mahmood, M.; Butt, H.; Nishimura, G.; Soltan, A. High-Performance Automated Abstract Screening with Large Language Model Ensembles. arXiv 2024, arXiv:2411.02451. [Google Scholar] [CrossRef]
  54. Wang, S.; Scells, H.; Koopman, B.; Potthast, M.; Zuccon, G. Generating Natural Language Queries for More Effective Systematic Review Screening Prioritisation. In Proceedings of the SIGIR-AP 2023—Annual International ACM SIGIR Conference on Research and Development in Information Retrieval in the Asia Pacific Region, Beijing, China, 26–28 November 2023; pp. 73–83. [Google Scholar] [CrossRef]
  55. Akinseloyin, O.; Jiang, X.; Palade, V. A Question-Answering Framework for Automated Abstract Screening Using Large Language Models. J. Am. Med. Inform. Assoc. 2024, 31, 1939–1952. [Google Scholar] [CrossRef] [PubMed]
  56. Kusa, W.; Mendoza, O.E.; Samwald, M.; Knoth, P.; Hanbury, A. CSMeD: Bridging the Dataset Gap in Automated Citation Screening for Systematic Literature Reviews. Adv. Neural Inf. Process. Syst. 2023, 36, 23468–23484. [Google Scholar]
  57. Hou, Y.; Zhang, J.; Lin, Z.; Lu, H.; Xie, R.; McAuley, J.; Zhao, W.X. Large Language Models Are Zero-Shot Rankers for Recommender Systems. In Proceedings of the Advances in Information Retrieval; Goharian, N., Tonellotto, N., He, Y., Lipani, A., McDonald, G., Macdonald, C., Ounis, I., Eds.; Springer: Cham, Switzerland, 2024; pp. 364–381. [Google Scholar] [CrossRef]
  58. Zhuang, H.; Qin, Z.; Hui, K.; Wu, J.; Yan, L.; Wang, X.; Bendersky, M. Beyond Yes and No: Improving Zero-Shot LLM Rankers via Scoring Fine-Grained Relevance Labels. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers); Duh, K., Gomez, H., Bethard, S., Eds.; Association for Computational Linguistics: Mexico City, Mexico, 2024; pp. 358–370. [Google Scholar] [CrossRef]
  59. Faggioli, G.; Dietz, L.; Clarke, C.L.A.; Demartini, G.; Hagen, M.; Hauff, C.; Kando, N.; Kanoulas, E.; Potthast, M.; Stein, B.; et al. Perspectives on Large Language Models for Relevance Judgment. In Proceedings of the 2023 ACM SIGIR International Conference on Theory of Information Retrieval, ICTIR ’23, Taipei, Taiwan, 23 July 2023; pp. 39–50. [Google Scholar] [CrossRef]
  60. Thomas, P.; Spielman, S.; Craswell, N.; Mitra, B. Large Language Models Can Accurately Predict Searcher Preferences. In Proceedings of the 47th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’24, Washington, DC, USA, 14–18 July 2024; pp. 1930–1940. [Google Scholar] [CrossRef]
  61. Syriani, E.; David, I.; Kumar, G. Screening Articles for Systematic Reviews with ChatGPT. J. Comput. Lang. 2024, 80, 101287. [Google Scholar]
  62. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models Are Few-Shot Learners. arXiv 2020, arXiv:2005.14165. [Google Scholar] [CrossRef]
  63. Gu, Y.; Tinn, R.; Cheng, H.; Lucas, M.; Usuyama, N.; Liu, X.; Naumann, T.; Gao, J.; Poon, H. Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing. ACM Trans. Comput. Healthc. 2021, 3, 2:1–2:23. [Google Scholar] [CrossRef]
  64. Huotala, A.; Kuutila, M.; Ralph, P.; Mäntylä, M. The Promise and Challenges of Using LLMs to Accelerate the Screening Process of Systematic Reviews. In Proceedings of the 28th International Conference on Evaluation and Assessment in Software Engineering, EASE ’24, Salerno, Italy, 18–21 June 2024; pp. 262–271. [Google Scholar] [CrossRef]
  65. Kojima, T.; Gu, S.S.; Reid, M.; Matsuo, Y.; Iwasawa, Y. Large Language Models Are Zero-Shot Reasoners. In Proceedings of the 36th International Conference on Neural Information Processing Systems, NIPS ’22, New Orleans, LA, USA, 28 November–9 December 2022; pp. 22199–22213. [Google Scholar]
  66. Tseng, Y.M.; Huang, Y.C.; Hsiao, T.Y.; Chen, W.L.; Huang, C.W.; Meng, Y.; Chen, Y.N. Two Tales of Persona in LLMs: A Survey of Role-Playing and Personalization. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024; Al-Onaizan, Y., Bansal, M., Chen, Y.N., Eds.; Association for Computational Linguistics: Miami, FL, USA, 2024; pp. 16612–16631. [Google Scholar] [CrossRef]
  67. Liu, D.; Nassereldine, A.; Yang, Z.; Xu, C.; Hu, Y.; Li, J.; Kumar, U.; Lee, C.; Qin, R.; Shi, Y.; et al. Large Language Models Have Intrinsic Self-Correction Ability. arXiv 2024, arXiv:2406.15673. [Google Scholar] [CrossRef]
  68. Spillias, S.; Tuohy, P.; Andreotta, M.; Annand-Jones, R.; Boschetti, F.; Cvitanovic, C.; Duggan, J.; Fulton, E.; Karcher, D.; Paris, C.; et al. Human-AI Collaboration to Identify Literature for Evidence Synthesis. Cell Rep. Sustain. 2024, 1, 100132. [Google Scholar] [CrossRef]
  69. Wang, W.; Wei, F.; Dong, L.; Bao, H.; Yang, N.; Zhou, M. MINILM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers. In Advances in Neural Information Processing Systems 33 (NeurIPS 2020); Neural Information Processing Systems Foundation, Inc. (NeurIPS): Vancouver, BC, Canada, 2020; Volume 2020. [Google Scholar]
  70. SentenceTransformers Documentation—Sentence Transformers Documentation. Available online: https://www.sbert.net/ (accessed on 29 December 2024).
  71. OpenAI Platform. Available online: https://platform.openai.com (accessed on 26 December 2024).
  72. Manning, C.D.; Raghavan, P.; Schütze, H. Introduction to Information Retrieval; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  73. Kusa, W.; Lipani, A.; Knoth, P.; Hanbury, A. An Analysis of Work Saved over Sampling in the Evaluation of Automated Citation Screening in Systematic Literature Reviews. Intell. Syst. Appl. 2023, 18, 200193. [Google Scholar] [CrossRef]
  74. Feng, Y.; Liang, S.; Zhang, Y.; Chen, S.; Wang, Q.; Huang, T.; Sun, F.; Liu, X.; Zhu, H.; Pan, H. Automated Medical Literature Screening Using Artificial Intelligence: A Systematic Review and Meta-Analysis. J. Am. Med. Inform. Assoc. 2022, 29, 1425–1432. [Google Scholar] [CrossRef]
  75. Kanoulas, E.; Li, D.; Azzopardi, L.; Spijker, R. CLEF 2017 Technologically Assisted Reviews in Empirical Medicine Overview. In Proceedings of the 8th International Conference of the CLEF Association, CLEF 2017, Dublin, Ireland, 11–14 September 2017. [Google Scholar]
  76. Linzbach, S.; Tressel, T.; Kallmeyer, L.; Dietze, S.; Jabeen, H. Decoding Prompt Syntax: Analysing Its Impact on Knowledge Retrieval in Large Language Models. In Proceedings of the Companion Proceedings of the ACM Web Conference 2023, WWW ’23 Companion, Austin, TX, USA, 30 April 2023–4 May 2023; pp. 1145–1149. [Google Scholar] [CrossRef]
Figure 1. Overview of the experimental setup for title–abstract ranking using LLM-based QA framework.
Figure 1. Overview of the experimental setup for title–abstract ranking using LLM-based QA framework.
Make 07 00028 g001
Figure 2. Plot of recall at varying levels of k% across four zoonotic diseases.
Figure 2. Plot of recall at varying levels of k% across four zoonotic diseases.
Make 07 00028 g002
Figure 3. Plot of nWSS across four zoonotic diseases datasets at recall threshold of 95% and 100%.
Figure 3. Plot of nWSS across four zoonotic diseases datasets at recall threshold of 95% and 100%.
Make 07 00028 g003
Figure 4. Plot of average precision across four zoonotic diseases.
Figure 4. Plot of average precision across four zoonotic diseases.
Make 07 00028 g004
Figure 5. Plot of the distribution of rank positions of relevant title–abstract across four zoonotic diseases datasets.
Figure 5. Plot of the distribution of rank positions of relevant title–abstract across four zoonotic diseases datasets.
Make 07 00028 g005
Figure 6. Plot of precision and recall across four zoonotic disease datasets.
Figure 6. Plot of precision and recall across four zoonotic disease datasets.
Make 07 00028 g006
Table 1. Selection criteria used in the systematic review for the abstract screening process.
Table 1. Selection criteria used in the systematic review for the abstract screening process.
Inclusion CriteriaExclusion Criteria
  • Primary research or meta-analysis
  • Assesses the relationship between the selected climate variable and either:
    -
    Disease incidence or prevalence.
    -
    Pathogen survival.
    -
    Transmission.
    -
    Virulence.
    -
    Demonstrated vector or maintenance host survival, development or distribution.
  • Reviews, opinions, books, editorials.
  • Laboratory-focused studies, e.g., studies to develop an appropriate culture method.
Table 2. Overview of dataset composition by disease, climate variable, and relevance proportions.
Table 2. Overview of dataset composition by disease, climate variable, and relevance proportions.
DiseaseClimate VariableRelevantIrrelevantTotal% Relevant
CCHFRainfall5739745412.6%
EbolaRainfall149019151.5%
LeptoRainfall10889199910.8%
RVFRainfall6347453711.7%
Table 3. Disease-specific topics and question-based eligibility criteria for climate-sensitive zoonotic disease studies.
Table 3. Disease-specific topics and question-based eligibility criteria for climate-sensitive zoonotic disease studies.
DiseaseTopic and Eligibility Questions
CCHFTopic:
Impact of Climate Change on CCHF: A Focus on Rainfall
Eligibility Questions:
Q1. Does the study report on primary research or a meta-analysis rather than a review, opinion, or book?
Q2. Does the study measure the incidence or prevalence or virulence or survival or transmission of Crimean-Congo haemorrhagic fever or a relevant vector (such as ticks) without specifically measuring the incidence of the pathogens?
Q3. Does the research examine environmental factors such as rainfall, seasonality (e.g., wet vs. dry season) or regional comparisons impacting disease prevalence or vector distribution?
Q4. Is the study focused on field-based or epidemiological research rather than laboratory method validation?
EbolaTopic:
Impact of Climate Change on Ebola: A Focus on Rainfall
Eligibility Questions:
Q1. Does the study report on primary research or a meta-analysis rather than a review, opinion, or book?
Q2. Does the study measure the incidence or prevalence or virulence or survival or transmission of Ebola or Marburg, a relevant vector, or reservoir hosts abundance or distribution (such as bats or primates) without specifically measuring the incidence of the pathogens?
Q3. Does the research examine environmental factors such as rainfall, seasonality (e.g., wet vs. dry season) or regional comparisons impacting disease prevalence or vector distribution?
Q4. Is the study focused on field-based or epidemiological research rather than laboratory method validation?
LeptoTopic:
Impact of Climate Change on Leptospirosis: A Focus on Rainfall
Eligibility Questions:
Q1. Does the study report on primary research or a meta-analysis rather than a review, opinion, or book?
Q2. Does the study measure the incidence or prevalence or virulence or survival or transmission of Leptospirosis, a relevant arthropod vector, or reservoir hosts (such as rodents) without specifically measuring the incidence of the pathogens?
Q3. Does the research examine environmental factors such as rainfall, seasonality (e.g., wet vs. dry season) or regional comparisons impacting disease prevalence or vector distribution?
Q4. Is the study focused on field-based or epidemiological research rather than laboratory method validation?
RVFTopic:
Impact of Climate Change on Rift Valley Fever Virus: A Focus on Rainfall
Eligibility Questions:
Q1. Does the study report on primary research or a meta-analysis rather than a review, opinion, or book?
Q2. Does the study measure the incidence or prevalence or virulence or survival or transmission of Rift Valley fever or other vector-borne diseases (such as malaria) that share similar vectors (e.g., mosquitoes) without specifically measuring the incidence of the pathogen?
Q3. Does the research examine environmental factors such as rainfall, seasonality (e.g., wet vs. dry season) or regional comparisons impacting disease prevalence or vector distribution?
Q4. Is the study focused on field-based or epidemiological research rather than laboratory method validation?
Table 4. Answer labels and scoring scales for single-level QA and TSC models.
Table 4. Answer labels and scoring scales for single-level QA and TSC models.
Answer SchemaAnswer LabelsAnswer Score
QA-2Yes1.0
No0.0
QA-3Yes1.00
Unsure0.50
No0.00
QA-4Definitely Yes0.95
Probably Yes0.75
Probably No0.25
Definitely No0.05
QA-5Definitely Yes1.00
Probably Yes0.75
Unsure0.50
Probably No0.25
Definitely No0.00
TSC-5Definitely Include1.00
Probably Include0.75
Unsure0.50
Probably Exclude0.25
Definitely Exclude0.00
Table 5. Answer labels and scoring scales for multi-level QA models.
Table 5. Answer labels and scoring scales for multi-level QA models.
Answer SchemaAnswer LabelsConfidence LabelsConfidence Score
QA-2-CYesHigh1.00
Medium0.75
Low0.50
NoHigh0.00
Medium0.25
Low0.50
Table 6. Performance metrics for models by zoonotic diseases.
Table 6. Performance metrics for models by zoonotic diseases.
DiseaseModelRecall@k%nWSS@r%AP
51020305095100
CCHFQA-20.330.470.810.881.000.690.630.70
QA-2-C0.390.540.820.930.980.640.530.77
QA-30.390.530.790.930.980.670.510.76
QA-40.350.630.860.961.000.780.610.80
QA-50.330.650.840.951.000.740.620.78
TSC-50.370.460.580.630.820.160.120.55
TSC-BM250.230.320.460.580.790.220.170.40
TSC-MiniLM0.250.330.400.540.740.150.180.38
EbolaQA-20.500.710.791.001.000.670.720.18
QA-2-C0.360.790.930.931.000.610.660.33
QA-30.430.860.861.001.000.660.710.23
QA-40.710.931.001.001.000.860.910.51
QA-50.640.931.001.001.000.830.880.50
TSC-50.430.430.500.500.71 0.01 0.040.37
TSC-BM250.140.360.570.570.860.020.070.05
TSC-MiniLM0.210.290.570.710.860.320.370.04
LeptoQA-20.320.540.820.920.950.570.410.58
QA-2-C0.360.650.810.940.960.720.280.69
QA-30.320.520.770.890.970.560.250.57
QA-40.390.750.920.940.990.720.420.80
QA-50.400.720.920.940.980.700.400.78
TSC-50.380.780.870.900.940.460.100.75
TSC-BM250.150.280.480.600.850.260.140.26
TSC-MiniLM0.320.560.810.940.980.720.260.62
RVFQA-20.210.370.760.840.950.540.310.44
QA-2-C0.290.410.760.900.970.730.360.53
QA-30.240.410.760.900.980.610.290.51
QA-40.320.560.830.890.980.700.270.66
QA-50.290.520.830.890.980.670.300.61
TSC-50.210.270.400.480.730.230.080.28
TSC-BM250.110.220.370.490.730.190.060.21
TSC-MiniLM0.140.240.400.560.900.350.260.26
Table 7. Results for mean average precision (MAP) and area under the precision–recall curve (PR-AUC) by model.
Table 7. Results for mean average precision (MAP) and area under the precision–recall curve (PR-AUC) by model.
ModelMAPPR-AUC
QA-20.4760.621
QA-2-C0.5790.669
QA-30.5190.636
QA-40.6910.761
QA-50.6700.740
TSC-50.4890.639
TSC-BM250.2290.206
TSC-MiniLM0.3250.271
Table 8. Misclassifications identified by the system.
Table 8. Misclassifications identified by the system.
DiseaseRevised to IncludeRevised to ExcludeTotal
CCHF7310
Ebola101
Lepto13316
RVF8513
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sujau, M.; Wada, M.; Vallée, E.; Hillis, N.; Sušnjak, T. Accelerating Disease Model Parameter Extraction: An LLM-Based Ranking Approach to Select Initial Studies for Literature Review Automation. Mach. Learn. Knowl. Extr. 2025, 7, 28. https://doi.org/10.3390/make7020028

AMA Style

Sujau M, Wada M, Vallée E, Hillis N, Sušnjak T. Accelerating Disease Model Parameter Extraction: An LLM-Based Ranking Approach to Select Initial Studies for Literature Review Automation. Machine Learning and Knowledge Extraction. 2025; 7(2):28. https://doi.org/10.3390/make7020028

Chicago/Turabian Style

Sujau, Masood, Masako Wada, Emilie Vallée, Natalie Hillis, and Teo Sušnjak. 2025. "Accelerating Disease Model Parameter Extraction: An LLM-Based Ranking Approach to Select Initial Studies for Literature Review Automation" Machine Learning and Knowledge Extraction 7, no. 2: 28. https://doi.org/10.3390/make7020028

APA Style

Sujau, M., Wada, M., Vallée, E., Hillis, N., & Sušnjak, T. (2025). Accelerating Disease Model Parameter Extraction: An LLM-Based Ranking Approach to Select Initial Studies for Literature Review Automation. Machine Learning and Knowledge Extraction, 7(2), 28. https://doi.org/10.3390/make7020028

Article Metrics

Back to TopTop