Next Article in Journal
Customer Churn Prediction: A Systematic Review of Recent Advances, Trends, and Challenges in Machine Learning and Deep Learning
Previous Article in Journal
Leveraging LLMs for Automated Extraction and Structuring of Educational Concepts and Relationships
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Screening Smarter, Not Harder: Budget Allocation Strategies for Technology-Assisted Reviews (TARs) in Empirical Medicine

by
Giorgio Maria Di Nunzio
Department of Information Engineering, University of Padova, 35131 Padova, Italy
Mach. Learn. Knowl. Extr. 2025, 7(3), 104; https://doi.org/10.3390/make7030104
Submission received: 2 May 2025 / Revised: 11 September 2025 / Accepted: 16 September 2025 / Published: 20 September 2025

Abstract

In the technology-assisted review (TAR) area, most research has focused on ranking effectiveness and active learning strategies within individual topics, often assuming unconstrained review effort. However, real-world applications such as legal discovery or medical systematic reviews are frequently subject to global screening budgets. In this paper, we revisit the CLEF eHealth TAR shared tasks (2017–2019) through the lens of budget-aware evaluation. We first reproduce and verify the official participant results, organizing them into a unified dataset for comparative analysis. Then, we introduce and assess four intuitive budget allocation strategies—even, proportional, inverse proportional, and threshold-capped greedy—to explore how review effort can be efficiently distributed across topics. To evaluate systems under resource constraints, we propose two cost-aware metrics: relevant found per cost unit (RFCU) and utility gain at budget (UG@B). These complement traditional recall by explicitly modeling efficiency and trade-offs between true and false positives. Our results show that different allocation strategies optimize different metrics: even and inverse proportional allocation favor recall, while proportional and capped strategies better maximize RFCU. UG@B remains relatively stable across strategies, reflecting its balanced formulation. A correlation analysis reveals that RFCU and UG@B offer distinct perspectives from recall, with varying alignment across years. Together, these findings underscore the importance of aligning evaluation metrics and allocation strategies with screening goals. We release all data and code to support reproducibility and future research on cost-sensitive TAR.

Graphical Abstract

1. Introduction

Systematic reviews are a cornerstone of evidence-based decision making across a range of disciplines. Unlike traditional literature reviews, systematic reviews follow a rigorous methodology to identify, evaluate, and synthesize all available evidence relevant to a specific research question. This process typically involves a comprehensive search of the literature, predefined inclusion and exclusion criteria, and systematic data extraction and analysis. While essential in building cumulative knowledge, systematic reviews are also time-consuming and labor-intensive, often requiring months of work by domain experts [1].
In medicine, systematic reviews play a critical role in synthesizing clinical evidence to inform guidelines, policy decisions, and individual patient care. These reviews are particularly demanding due to the volume and heterogeneity of biomedical literature, the need for high precision and recall, and the importance of reproducibility and transparency [2]. Given these challenges, technology-assisted review (TAR) systems have emerged as crucial tools in the healthcare domain, particularly in supporting the growing demand for timely and comprehensive systematic reviews. As the volume of biomedical literature continues to rise, traditional manual review processes are becoming increasingly insufficient to ensure that healthcare decisions are informed by the best available evidence. TAR systems leverage intelligent algorithms—ranging from classical machine learning to neural ranking models and active learning frameworks—to enhance the efficiency, accuracy, and scalability of the systematic review pipeline [3]. These systems not only reduce the cognitive and temporal burden on human reviewers but also contribute to reproducibility, transparency, and adaptability in evidence-based medicine [4].
Evaluating TAR systems presents distinct challenges that go beyond those of traditional information retrieval settings. Standard metrics used in text classification and information retrieval, such as recall, precision, and the area under the curve, primarily measure the system’s ability to retrieve relevant documents [5]. However, these metrics often assume an idealized environment where resources—time, money, and human effort—are unlimited. In practice, systematic reviews, particularly in medicine, operate under strict resource constraints: reviewers must balance the goal of achieving high recall with the reality of limited budgets and tight timelines [6,7]. Furthermore, existing metrics rarely account for the actual cost of reviewing documents, which includes the cognitive burden and time demands placed on human experts. Another complication arises when the true recall is unknown or difficult to estimate, especially in live systematic review projects where the entire set of relevant documents is not available [8]. As a result, there is growing recognition that the evaluation of TAR systems should incorporate cost awareness, risk management, and practical stopping criteria, rather than relying solely on traditional retrieval performance indicators. In real-world settings—like in medical domains—reviewers often operate under limited time, budgets, and cognitive resources. As such, cost-aware perspectives are critical in assessing the true utility of TAR systems and guiding their design [9].
The CLEF eHealth (https://clefehealth.imag.fr/ (accessed on 1 May 2025)) initiative has been a key venue in evaluating the performance of TAR systems in the context of medical systematic reviews. Between 2017 and 2019 [10,11,12], the CLEF eHealth TAR tasks provided a benchmark for comparing automated and semi-automated approaches using realistic, large-scale datasets derived from Cochrane reviews (https://www.cochranelibrary.com/ (accessed on 1 May 2025)). These tasks aimed to simulate real-world conditions by offering a continuous active learning setup, assessing systems based on their ability to retrieve relevant documents efficiently and effectively. This paper provides a critical overview of the experiments conducted in the CLEF eHealth TAR tasks over these three years, highlighting methodological trends, performance outcomes, and implications for future development in medical TAR systems.
This paper presents a comprehensive re-examination of the CLEF eHealth TAR tasks conducted from 2017 to 2019. In particular, we propose the following key contributions from this study.
  • Reproduction of Official Results: We collect, organize, and verify the official participant runs from the CLEF eHealth TAR tasks (2017–2019), ensuring reproducibility and providing a unified view of system performance across editions.
  • Exploration of Budget Allocation Strategies: We investigate alternative strategies for allocating limited review budgets across multiple topics. By analyzing budget-aware screening performance, we provide practical insights for the optimization of systematic reviews under resource constraints.
  • Introduction of Cost-Aware Evaluation Metrics: We propose two novel measures—relevant found per cost unit (RFCU@k) and utility gain at budget (UG@B)—designed to capture the trade-offs between retrieval effectiveness and resource consumption in TAR systems.
This study makes a distinct contribution to the TAR literature by shifting the focus from single-topic ranking performance to the broader problem of budget allocation across multiple topics under realistic constraints. Whereas most prior evaluations have assumed unconstrained or uniform review budgets, we formalize allocation as a policy optimization problem and introduce two novel budget-aware metrics, RFCU and UG@B, that capture the complementary dimensions of efficiency and utility. To the best of our knowledge, no prior study has systematically investigated budget allocation strategies in evaluating TAR systems. By reanalyzing three years of CLEF eHealth TAR tasks under standardized budget assumptions, we show that allocation policies, whether simple heuristics or adaptive approaches, can materially influence review outcomes. These results provide both methodological advances and practical insights, offering the community new tools and benchmarks to better align TAR evaluation with the realities of systematic review practice.
The remainder of this paper is organized as follows. In Section 2, we introduce budget-aware perspectives on technology-assisted reviews and discuss their implications for screening strategies. Section 3 presents the CLEF eHealth TAR lab, describing its setting, tasks, and relevance for empirical medicine. Section 4 reports on the experimental analysis, evaluating different budget allocation strategies. Finally, Section 5 concludes the paper and outlines directions for future work.

2. Beyond Recall: Budget-Aware Perspectives in Technology- Assisted Reviews

Although metrics such as recall and workload reduction are standard in evaluating TAR systems, they only partially capture the practical constraints under which systematic reviews are conducted. In real-world settings—especially in medical domains—reviewers often operate under limited time, budgets, and cognitive resources. As such, cost-aware perspectives are critical in assessing the true utility of TAR systems and guiding their design. Several dimensions of cost remain underexplored in current TAR research.
First, most evaluation frameworks assume unlimited resources for document screening, with few efforts to simulate budget-constrained or time-sensitive review scenarios. Integrating explicit constraints—such as fixed review time or monetary limits—could yield insights into the robustness of different systems under realistic operational pressures. Second, human-in-the-loop efficiency remains largely nonquantified. The time per decision, fatigue effects, and the interpretability of ranking outputs are rarely measured but have substantial impacts on reviewer performance and user acceptance. Third, the computational cost of TAR systems, especially neural architectures, is seldom considered relative to their performance gains. Lightweight, non-neural models may offer similar effectiveness at a significantly lower cost in terms of infrastructure, energy consumption, and deployment complexity—factors that are particularly relevant for smaller institutions or global health contexts. Fourth, few studies account for the opportunity costs of screening inefficiencies or misclassifications. The delayed identification of relevant studies can affect the timeliness of clinical guideline implementation or policy decisions, which carries downstream consequences that are not reflected in retrieval metrics alone.
To address these limitations, research on budget allocation should adopt broader experimental protocols that model cost explicitly. This includes collecting timing data, incorporating economic analyses, and conducting mixed-method user studies. Comparative benchmarks should evaluate TAR systems not only by their retrieval effectiveness but also by their efficiency, transparency, and long-term sustainability.

2.1. Examples of Budget Allocation in Information Retrieval

While considerable research in the technology-assisted review (TAR) field has focused on stopping strategies and recall-oriented optimization, the problem of how to distribute review effort across multiple topics or tasks has received far less attention. In particular, budget allocation—how limited annotation or screening resources should be apportioned across topics—remains an underexplored area. To the best of our knowledge, budget allocation in TAR systems has not yet been systematically investigated, despite its growing relevance in scenarios involving large-scale, multitopic systematic reviews.
In this section, we examine related work on budget allocation in adjacent areas, highlighting strategies and frameworks that could inform future developments in TAR.
The TREC 2017 Common Core track explored how bandit techniques could be used not only to simulate but also to construct a test collection in real time. This requires addressing practical challenges beyond simulations, such as allocating annotation effort across topics and enabling assessors to familiarize themselves with each topic, while also developing infrastructure for real-time document selection and judgment acquisition [13].
Constructing reliable and low-cost IR test collections often depends on how annotation budgets are distributed across topics. Traditional strategies, as used in TREC, typically assign a uniform budget per topic, often tied to a fixed top-k pool size. However, more recent work has highlighted the limitations of such static allocation and has proposed dynamic approaches based on topic-specific needs [14].
Flexible budget models have also been studied in the context of sequential decision making. A generalization of restless multiarmed bandits (RMABs), known as flexible RMABs (F-RMABs), has been proposed to allow resource redistribution across decision rounds. This framework supports more realistic planning scenarios and provides a theoretically grounded algorithm for optimal budget allocation over time [15].
In multichannel advertising, reinforcement learning techniques have been applied to optimize long-term outcomes under budget constraints. A hybrid Q-learning and differential evolution (DE) approach has been introduced to dynamically allocate resources across channels, taking into account delayed effects and interdependencies between actions and outcomes [16].
Budget-sensitive approaches have also emerged in data annotation settings such as multilabel classification. A reverse auction framework has been proposed to select crowd workers based on cost and confidence, combined with systematic budget selection strategies (e.g., greedy and multicover selection) to address domain coverage limitations within fixed budgets [17].
Finally, contextual bandits, widely used in online recommendation systems, have been extended to handle biased user feedback caused by herding effects. This line of research, while not directly focused on budget, provides useful insights into sequential decision making under partial observability and feedback bias, which are relevant for adaptive resource allocation [18].

2.2. A Proposal of Budget Allocation Strategies in TAR Systems

In cost-constrained learning scenarios such as active learning, annotation campaigns, or document screening workflows, a central challenge is how to allocate a fixed labeling or review budget in order to maximize utility [16,19,20]. The budget may represent time, a monetary cost, or human effort and must be judiciously spent across a sequence of uncertain decisions. While most traditional active learning approaches focus on selecting the next most informative item, they often assume a static or uniform cost per query and overlook the global allocation question: how much effort should be spent on which data, tasks, or domains?
This budget allocation problem has been relatively understudied in TAR systems, despite its crucial role in real-world deployments. For example, in multiclass classification, multidomain retrieval, or document triage tasks, allocating too much effort to low-yield areas can waste resources, while underinvesting in promising areas reduces overall effectiveness. In [21], the authors mention that “it has been estimated that systematic reviews take, on average, 1139 h (range 216–2518 h) to complete and usually require a budget of at least $100,000”. Therefore, a robust allocation strategy must consider both the expected gain from labeling additional examples and the associated cost, potentially adapting over time as feedback is observed. This requires balancing the exploitation of known high-utility areas with the exploration of uncertain regions and motivates strategies that incorporate cost awareness, utility estimation, or uncertainty modeling into the allocation process.
Despite being a critical aspect, most TAR systems, when evaluated, focus on optimizing the within-topic ranking performance, assuming unlimited or per-topic isolated review budgets. However, in real-world scenarios such as legal reviews, systematic medical literature reviews, or multitopic classification, practitioners operate under a global screening budget that must be distributed across heterogeneous topics of varying size, relevance prevalence, and ranking quality [9,22,23,24,25,26,27,28].
In this study, we present a set of straightforward yet informative strategies for distributing review effort across topics. Rather than proposing complex adaptive policies, our objective is to analyze how different allocation schemes affect system performance and evaluation metrics under constrained review conditions. To this end, we wish to re-evaluate the CLEF eHealth TAR experiments from the 2017 to 2019 editions by applying a budget allocation strategy across all runs. This approach ensures that each TAR system is assessed under the same labeling cost constraints, allowing for a fair and consistent comparison of their effectiveness. By standardizing the annotation budget, we eliminate variability caused by differing cost assumptions in the original submissions, thereby focusing the evaluation on the intrinsic capabilities of the TAR methods themselves, rather than on disparities in resource expenditure.
In Section 2.2.1, we present four baseline strategies for budget allocation:
  • even allocation;
  • proportional allocation;
  • inversely proportional allocation; and
  • threshold-capped greedy allocation.
We selected these four strategies to reflect a range of simple, interpretable policies with different assumptions about topic relevance distribution. These strategies provide a useful contrast between uniform effort (even), topic prior-based effort (proportional/inverse), and a greedy baseline driven by estimated gains.
We assume that we have a budget, B, available, and we need to allocate it across the set of topics T = { t 1 , t 2 , , t i , } , with a total number of topics equal to | T | . B i will be the budget allocated for the i-th topic t i . The total number of documents available for the i-th topic is D i .

2.2.1. Even Allocation

The even allocation strategy divides the total review budget equally across all topics:
B i = B | T |
This approach ensures the uniform treatment of all topics, regardless of their size or difficulty. It is simple, fair, and effective in scenarios where topics are equally important or we lack prior estimations of relevance prevalence or difficulty. It may underperform when topic sizes are highly imbalanced or when some topics are significantly more relevant than others.

2.2.2. Proportional Allocation

Proportional allocation distributes the budget in proportion to the number of documents in each topic:
B i = D i j = 1 | T | D j · B
This strategy assumes that larger topics deserve more review effort. This approach reflects the topic scale and avoids undersampling large collections, but it can overcommit to large topics with low relevance prevalence, reducing the overall retrieval efficiency.

2.2.3. Inversely Proportional Allocation

Inversely proportional allocation gives more effort to smaller topics, based on the notion that they can be reviewed more thoroughly and may yield higher accuracy averaged across the topics:
B i = 1 / D i j = 1 | T | 1 / D j · B
It prioritizes small topics that may otherwise be overlooked, and it should improve the recall by ensuring better topic coverage under tight budgets. The limitation of this approach is that it may underallocate effort to large but relevance-rich topics.

2.2.4. Threshold-Capped Greedy Allocation

This strategy allocates budget in a greedy, depth-first manner, prioritizing smaller topics. Topics are sorted in ascending order regarding the available number of documents D i , and the budget is allocated sequentially until exhaustion. For each topic, allocation is capped at either a fraction τ ( 0 , 1 ] of its total size or the remaining budget, B remaining :
B i = min ( τ · D i , B remaining )
The idea of this approach is to quickly screen small topics, enabling broader coverage when only partial review is feasible. The drawback is that it may lead to the neglect of later (often larger) topics unless combined with a secondary reallocation strategy.
These strategies provide foundational tools for managing review effort in multitopic TAR settings. While simple, they offer contrasting trade-offs between fairness, coverage, and efficiency. In Section 4, we analyze their empirical performance and explore adaptive alternatives based on relevance prevalence and utility gain.
One final remark concerns the fact that the budget allocation strategies may operate in parallel with a stopping strategy at the topic level. While a global screening budget constrains the total effort across all topics, individual topics may consume less than their allocated share if an early stopping condition is met—for example, when an estimated recall threshold is achieved or the marginal utility drops below a predefined threshold. In practice, this allows the reallocation of unused budget to other topics that remain uncertain or underexplored. However, the problem of recall estimation and stopping decision making introduces additional complexity, requiring reliable recall estimators or predictive models. In this work, we focus exclusively on the budget allocation problem under the assumption of fixed topic-level effort, and we leave the investigation of stopping strategies to future work.

2.3. Budget-Aware Evaluation Metrics

Traditional evaluations of TAR systems often prioritize effectiveness metrics, such as recall, while focusing on reducing the workload [10,11,12]. However, these metrics only offer a partial perspective on system utility, particularly when reviews must be performed under real-world resource constraints. To address this gap, we propose a budget-aware framework for evaluating TAR systems where we treat the systematic review process as a decision-making problem under budget constraints. Each document-screening action is associated with a cost—whether in terms of time, cognitive load, or financial expenditure—while the benefit is measured by how effectively the system retrieves relevant studies. To quantify this balance, we introduce budget-sensitive metrics like the relevant found per cost unit (RFCU) and utility gain at budget (UG@B), which measure system effectiveness while factoring in resource limitations.
These budget-aware metrics are intended to complement, rather than replace, traditional evaluation measures such as recall, precision, or F1. While standard metrics provide essential insights into the overall effectiveness of TAR systems, they do not capture how efficiently this performance is achieved in terms of the annotation cost. Budget-aware metrics fill this gap by explicitly linking performance to resource usage, offering a more refined view of system behavior under realistic constraints. By considering both sets of metrics, we gain a more comprehensive understanding of each system’s strengths—balancing absolute effectiveness with cost-efficiency—which is crucial for informed deployment in high-cost, time-sensitive review settings.
The proposed budget-aware metrics used in our evaluation operate under the realistic assumption that the total number of relevant documents is unknown—a typical scenario in technology-assisted review tasks. Rather than relying on oracle knowledge of how many relevant items remain undiscovered, these measures focus exclusively on the performance achieved within the constraints of the allocated budget. This means that we evaluate systems based on the actual annotation cost incurred and the utility of the results retrieved and not based on hypothetical outcomes that assume full awareness of the relevance distribution. This approach ensures that our evaluation remains grounded in the practical limitations of real-world applications, where exhaustive relevance judgments are rarely available.
We will begin our presentation of the definitions of the new metrics by reviewing Recall@k, which measures the proportion of relevant documents retrieved within the first k screened documents:
Recall @ k = TP k P total
where TP k is the number of relevant documents (true positives—TP) found up to rank k, and P total is the total number of relevant documents (positives—P) in the collection. While Recall@k effectively captures early retrieval performance, it assumes that screening cost is uniform and unlimited. In real-world TAR scenarios, reviewers operate under strict resource constraints. Consequently, Recall@k may fail to reflect the efficiency with which relevant documents are identified relative to the effort expended.
To address these limitations, we propose two complementary baseline budget-aware evaluation metrics: relevant found per cost unit (RFCU) and utility gain at budget (UG@B). Let B denote the reviewing budget, expressed as the number of documents that experts can examine for a systematic review. Given a ranked list of documents and their binary relevance labels, we wish to measure how effectively a system identifies relevant documents while minimizing wasted effort on nonrelevant ones.

2.3.1. Relevant Found per Cost Unit@k

RFCU@k measures how many relevant documents are retrieved per unit of cost within the first k documents. It is defined as
RFCU @ k = TP k k · c
where
  • TP k is the number of relevant documents found within the first k screened documents;
  • c is the cost of reviewing a single document (e.g., in units of time, money, or effort).
RFCU@k captures the cost-efficiency of the screening process. Unlike Recall@k, it directly accounts for the screening cost, making it particularly appropriate when the objective is to maximize the yield of relevant documents per dollar or hour spent.
To better illustrate the difference between Recall@k and RFCU@k, consider the following hypothetical scenario: suppose that a document collection contains P total = 100 relevant documents. A reviewer screens the top k = 50 documents, finding TP k = 20 relevant documents. The cost per document assessed is c = 2 cost units (e.g., due to document length or review complexity).
Recall@k measures the proportion of relevant documents retrieved within the top k documents:
Recall @ k = TP k P total = 20 100 = 0.2
Thus, 20% of all relevant documents are found after screening 50 documents.
RFCU@k measures the number of relevant documents retrieved in the top-k documents per unit of cost spent:
RFCU @ k = TP k k × b = 20 50 × 2 = 20 100 = 0.2
Thus, the system achieves efficiency of 0.2 relevant documents per budget unit.
Although, in this example, the numerical values of Recall@k and RFCU@k coincide, they measure fundamentally different aspects:
  • Recall@k evaluates how much of the total relevant information has been retrieved;
  • RFCU@k evaluates how efficiently the review effort has been translated into relevant findings, accounting for the cost per document.
If the cost per document c were higher (e.g., b = 4 ), RFCU@k would drop:
RFCU @ k = 20 50 × 4 = 20 200 = 0.1
whereas Recall@k would remain unchanged.
Another important aspect to highlight is the fact that RFCU@k is similar to Precision@k, but the interpretation and the objectives are quite different also in this case. Precision@k measures the proportion of relevant documents among the top-k retrieved documents:
Precision @ k = TP k k
Precision@k reflects the purity of the top-k results and is bounded between 0 and 1. It assumes that the cost of assessing each document is constant and does not influence the evaluation. RFCU@k accounts explicitly for the screening cost. When c = 1 , RFCU@k reduces to Precision@k. However, when the document assessment costs vary (e.g., due to the document length or reviewer expertise requirements), RFCU@k would provide an evaluation of system efficiency when screening costs need to be taken into account.

2.3.2. Utility Gain at Budget (UG@B)

UG@B formalizes the balance between the gain in finding relevant documents and the cost of reviewing nonrelevant ones. It is defined as
UG @ B = j = 1 B i u j where u j = + g , if the j - th document is relevant c , if the j - th document is nonrelevant
where
  • B i is the total number of documents reviewed for the i-th topic;
  • g is the gain associated with reviewing a relevant document;
  • c is the penalty associated with reviewing a nonrelevant document.
While Recall@k, and all other traditional measures, remains a valuable metric in assessing early retrieval effectiveness, it does not account for the practical costs associated with reviewing documents. RFCU@k and UG@B extend the traditional evaluation by incorporating budget awareness and cost-efficiency, providing a more actionable framework for the assessment of TAR systems under real-world constraints.

3. The CLEF eHealth Technology-Assisted Review Lab

CLEF eHealth was a large and extensive evaluation challenge in the medical and biomedical domain within the CLEF initiative. The goal was to provide researchers with datasets, evaluation frameworks, and events. The CLEF eHealth TAR lab, conducted from 2017 to 2019 [10,11,12], aimed to advance the development of intelligent systems to support the systematic review process in evidence-based medicine. Systematic reviews are vital in synthesizing research findings and guiding clinical and policy decisions, yet they are often labor-intensive and time-consuming due to the exponential growth of the medical literature. The TAR lab addressed this challenge by providing a benchmark platform for evaluating automated and semi-automated methods capable of assisting reviewers in identifying relevant scientific articles.
The main objectives of the TAR lab were threefold:
  • To develop and evaluate technology-assisted methods that support different stages of systematic reviews, including document retrieval, title and abstract screening, and ranking;
  • To assess how well such systems could replicate or improve upon expert-generated results, particularly in the context of diagnostic test accuracy (DTA) reviews;
  • To create reusable test collections and protocols that could foster reproducible and comparative experimentation across the research community.
Each year, the lab introduced realistic tasks simulating distinct phases of the systematic review process. In 2017, the lab focused on prioritizing citations retrieved by expert Boolean queries. In 2018 and 2019, the lab expanded to two subtasks: (1) retrieving relevant documents using only structured review protocols (no Boolean search) and (2) ranking documents retrieved by expert searches (title and abstract screening). In particular, some of the TAR tasks challenges were as follows:
  • Data sparsity and imbalance: Relevant studies were a small minority within large retrieval sets, mimicking the real-world screening burden.
  • Domain specificity: The DTA reviewed the required understanding of complex medical terminology and concepts.
  • Limited supervision: Especially in the no Boolean search subtask, systems had to operate with minimal training data and no expert-crafted queries.
Across the three editions, the TAR lab demonstrated that automatic and semi-automatic methods could substantially reduce the human effort required to complete systematic reviews without compromising their comprehensiveness. Participants employed a wide range of techniques, including classic machine learning, neural ranking models, and query expansion strategies. Some of the general findings were as follows:
  • High recall achievable with limited review effort: Several systems reached near-complete recall for relevant studies after reviewing only a fraction of the dataset.
  • Protocol-based retrieval remains challenging: The no Boolean search subtask proved to be the most difficult, with mixed results across teams and significant room for improvement.
  • Iterative strategies outperform static ones: Approaches that leveraged feedback loops, such as active learning or reranking, were generally more effective than static ranking methods.

3.1. CLEF eHealth 2017 Technology-Assisted Review Task

The CLEF eHealth 2017 evaluation lab introduced a new pilot task focusing on technology-assisted reviews (TAR) in empirical medicine [12,29]. The primary objective of this task was to support health scientists and policymakers by enhancing information access during the systematic review process. The TAR task specifically addressed the abstract and title screening phase of conducting diagnostic test accuracy (DTA) systematic reviews. Participants were challenged to develop and evaluate search algorithms aimed at efficiently and effectively ranking studies during this screening phase. The task involved constructing a benchmark collection of fifty systematic reviews, along with corresponding relevant and irrelevant articles identified through original Boolean queries. Fourteen teams participated, submitting a total of 68 automatic and semi-automatic runs, utilizing various information retrieval and machine learning algorithms over diverse text representations, in both batch and iterative modes.

3.2. CLEF eHealth 2018 Technology-Assisted Review Task

The CLEF eHealth 2018 evaluation lab continued its exploration into technology-assisted reviews (TAR) within empirical medicine, aiming to enhance the efficiency and effectiveness of systematic reviews [11,30]. Building upon the foundation laid in 2017, the 2018 lab introduced refinements to its tasks and methodologies. The 2018 TAR task was structured into two subtasks, each designed to simulate different stages of the systematic review process.
  • Subtask 1: No Boolean Search—Participants were provided with elements of a systematic review protocol, such as the objective, study type, participant criteria, index tests, target conditions, and reference standards. Using this information, they were to retrieve relevant studies from the PubMed database without constructing explicit Boolean queries. The goal was to emulate the initial search phase of a systematic review, relying solely on protocol information to identify pertinent literature.
  • Subtask 2: Title and Abstract Screening—In this subtask, participants received the results of a Boolean search executed by Cochrane experts, including the set of PubMed Document Identifiers (PMIDs) retrieved. The task involved ranking these abstracts to prioritize the most relevant studies, effectively simulating the screening phase, when reviewers assess titles and abstracts to determine inclusion in the systematic review.

3.3. CLEF eHealth 2019 Technology-Assisted Review Task

The CLEF eHealth 2019 evaluation lab, building upon the foundations established in previous years, introduced refinements to its tasks and methodologies [10,31]. The 2019 TAR task comprised the same two subtasks, each designed to simulate different stages of the systematic review process: subtask 1—no Boolean search; subtask 2—title and abstract screening.

4. Experimental Analysis

In this section, we present the experimental framework adopted to analyze the results of CLEF eHealth TAR tasks from 2017 to 2019. Our primary focus is on reproducing and systematically studying the official runs submitted by participants, with particular attention to the title and abstract screening task. We restrict our analysis to this task because it is the one with sufficient evaluation data across all three editions, whereas the other subtask (full-text screening) was less consistently populated by participants and thus less suitable for comprehensive comparison. The goal of this experimental analysis is to reproduce the retrieval effectiveness of TAR systems under a recall-oriented perspective, reflecting the critical importance of high recall in systematic reviews. Accordingly, we adopt a set of well-established evaluation metrics that were used in the evaluation of the CLEF TAR labs, each capturing complementary aspects of system performance.
  • Recall@k: This measures the proportion of relevant documents retrieved within the top k screened documents. This metric provides a direct view of how quickly systems identify relevant studies. We will evaluate the runs when k is equal to the total number of relevant documents.
  • Area Under the Curve (AUC): Specifically, we compute the AUC of the recall versus documents screened curve, offering a global assessment of the retrieval efficiency across the screening process.
  • Mean Average Precision (MAP): This summarizes the precision achieved for each relevant document retrieved, providing an aggregate measure of the ranking quality. Although it is traditionally precision-biased, we have seen that this measure is strongly correlated with RFCU@k.
  • Work Saved over Sampling at Recall (WSS@r): This quantifies the reduction in screening effort relative to random sampling, at a given level of recall, thus highlighting the practical workload benefits offered by TAR systems.
These metrics collectively allow us to capture both the effectiveness (in terms of the retrieval and ranking of relevant documents) and the efficiency (in terms of effort saved) of the systems evaluated during the CLEF eHealth TAR challenges. This analysis serves as the foundation for comparing the results on new cost- and budget-aware evaluation measures proposed Section 2.3.
We implemented all measures in R and did not rely on the official scripts in order to be consistent with the R programming framework that we established for this analysis.

4.1. Adaptive Baseline: Epsilon-Greedy Multiarmed Bandit

While our main analysis focuses on fixed, hand-crafted allocation strategies, we also include in this work a simple adaptive baseline to illustrate how learning-based methods can dynamically adjust budget allocation. We adopt an epsilon-greedy multiarmed bandit approach [32,33], where each review topic is treated as an “arm” that yields a reward when relevant documents are discovered. At each allocation step, the policy selects the topic with the highest estimated reward (exploitation) with probability 1 ϵ or a random topic (exploration) with probability ϵ . In this work, the reward is defined as the number of relevant documents identified in the screened set, consistent with prior work in active learning [33,34].
This adaptive baseline highlights two key properties absent in static heuristics: (1) the ability to incorporate observed feedback into future allocation decisions and (2) the trade-off between exploiting high-yield topics and exploring uncertain ones. While epsilon-greedy is intentionally simple and does not model topic heterogeneity or review costs, it provides a principled adaptive comparison point and a foundation for more sophisticated strategies.

4.2. Datasets and Data Availability

All data used in this study were obtained from the official CLEF TAR GitHub repository (https://github.com/CLEF-TAR/tar/ (accessed on 1 May 2025)). This repository contains the complete set of topics, document collections, and submitted runs for the CLEF eHealth TAR tasks from 2017 to 2019, as well as associated relevance judgments and evaluation scripts. By relying on this openly accessible and curated resource, we ensure the transparency, reproducibility, and consistency of our experimental setup and results. We will also make available all the code that we implemented to reproduce the study online (https://github.com/gmdn (accessed on 1 May 2025)).
In Table 1, we provide a summary of the number of topics, documents, and relevant documents for each year of the CLEF eHealth TAR lab.

4.3. Reproducing the Original Results

In this section, we revisit and reproduce the results of the CLEF eHealth TAR lab from the 2017 to 2019 editions. By systematically reproducing the original runs, we aim to create a consistent evaluation framework that allows for direct comparison across years, systems, and strategies. This reproduction effort not only reinforces the transparency and reliability of previous findings but also lays the groundwork for introducing new, budget-aware evaluation methods that reflect the practical constraints of real-world TAR deployments in a better way.
To ensure completeness, we have included not only the runs that were officially evaluated during the original campaigns but also those that were submitted but not formally assessed at the time. This comprehensive approach allows for a more thorough comparison across systems and years and supports the introduction of new budget-aware evaluation metrics that reflect the practical constraints and priorities of real-world screening workflows.
In Table 2, Table 3 and Table 4, we show the reproduced results of the three editions, CLEF 2017, CLEF 2018, and CLEF 2019, respectively. We observed that our replicated evaluation closely matched the original scores. In fewer than 10% of the evaluated system–topic combinations, we observed minor deviations, with absolute differences typically ranging from 2% to 3%, and all of them occurred in the work saved over sampling (WSS) metric. This is due to the behavior of trec_eval (https://github.com/usnistgov/trec_eval (accessed on 1 May 2025)), the software used by CLEF lab organizers to compute results, which applies a truncation mechanism when a run fails to retrieve all relevant documents, leading to slight variations in WSS calculation. These isolated cases were limited in scope and did not exhibit any systematic bias in favor of a particular system. For each system–topic–year tuple, we computed both absolute and relative differences, confirming that the observed discrepancies were negligible and did not compromise the integrity or fairness of the comparative evaluation—a key objective of the CLEF benchmarking initiative. This high level of agreement confirms the reliability of the reproduction process and provides a solid foundation for further analysis. With this strongly reproduced dataset, we can confidently proceed to study the impact of budget allocation strategies, ensuring that any observed differences in performance are due to the evaluation framework rather than inconsistencies in the data or original runs.

4.4. Analysis of Budget Allocation Strategies

To explore the impact of budget-aware evaluation, we designed an experimental setting in which each TAR system is allocated a fixed annotation budget equivalent to 10% of the total number of documents in the collection and a value of 1 for both cost and gain. This budget simulates a realistic constraint on the number of documents that can be manually reviewed or labeled, reflecting common limitations in resources. All systems are evaluated under this uniform budget threshold to ensure comparability, and selection strategies are assumed to proceed until the budget is exhausted. This controlled setting enables us to assess not only how effectively systems retrieve relevant documents but also how efficiently they operate within predefined resource limits.
For better readability, we show the tables of all results in Appendix A. Our comparative evaluation across multiple years reveals that the effectiveness of budget allocation strategies varies substantially depending on the evaluation metric used. In particular, when optimizing for recall—the proportion of relevant documents retrieved—the best-performing allocation strategies were even allocation and inverse proportional allocation. These strategies prioritize topic coverage and allocate resources in a way that avoids overcommitting to large, potentially low-yield topics. In particular, inverse proportional allocation favors smaller topics, which may contain concentrated pockets of relevant documents and benefit from deeper review. This behavior aligns with recall’s sensitivity to missing relevant items, especially in underrepresented or niche topics.
In contrast, when optimizing for RFCU, the highest scores were achieved by the proportional and threshold-capped greedy strategies. These methods allocate budget relative to the topic size or limit effort given to large topics and thus avoid overspending review effort where relevant documents are sparse. RFCU emphasizes efficiency—finding the most relevant documents per unit cost—rather than coverage and is therefore better served by strategies that exploit expected high-yield regions while avoiding low-utility areas.
For UG@B, we observed no significant performance differences among the four allocation strategies. This suggests that UG@B is less sensitive to the specific distribution of the budget, possibly due to its balancing of gains (true positives) and penalties (false positives). UG@B captures the net benefit rather than pure coverage or efficiency and may tolerate a wider range of allocation behaviors without large variance in utility.
These results are better visualized in Figure 1, Figure 2 and Figure 3. For each year, we show a boxplot of the metrics of all experiments according to the allocation strategy. Strategies like even allocation and inverse proportional allocation promote fairness and topic-wide coverage, which benefits recall. Conversely, strategies such as proportional allocation and capped greedy allocation optimize for cost-effectiveness, which aligns with RFCU. UG@B’s neutrality reflects its more balanced formulation, integrating both gain and cost into a single score and, in this specific case, the fact that gains and costs are equal.
The epsilon-greedy bandit adaptive baseline achieved performance that was broadly comparable to that of the static policies, although it did not consistently outperform them. In particular, its ability to dynamically reallocate effort allowed it to avoid severe underinvestment in small but relevant-rich topics, yielding recall scores close to those of even and inverse proportional allocation. At the same time, the exploration component prevented the systematic neglect of large topics, producing RFCU values that were again similar to those of even allocation and inverse proportional allocation. However, under our uniform cost/gain assumptions, epsilon-greedy did show an advantage in terms of UG@B for all years. This behavior is similar to what we observed when using UG@B as the reward, rather than the number of relevant documents, for the adaptive baseline [35]. This reinforces the view that, in budget-constrained TAR settings, adaptive strategies offer flexibility, but their advantages depend strongly on the chosen rewards, task characteristics, and evaluation criteria. Overall, these results confirm that even the simplest adaptive strategy can provide a reasonable middle ground between coverage- and efficiency-oriented heuristics, and they highlight the potential of more advanced learning-based allocation methods.

4.5. Analysis of Correlations Between Evaluation Metrics

To better understand the relationships between evaluation metrics, we conducted a year-by-year correlation analysis comparing recall with RFCU and UG@B. The result, plotted in Figure 4 and Figure 5, shows that RFCU is, in some cases, only weakly correlated with recall—demonstrating that systems that achieve high recall may do so inefficiently, while others can yield high RFCU with relatively low recall. This divergence reinforces the view that RFCU captures a distinct, efficiency-oriented dimension of performance. In contrast, UG@B shows a more consistent and moderate correlation with recall across years, suggesting that it partially aligns with traditional effectiveness goals while also incorporating cost sensitivity. Importantly, the strength of these correlations varies year by year, reflecting differences in system behavior, task design, and relevance prevalence across CLEF eHealth editions. These findings support the use of multiple, complementary metrics in TAR evaluation to provide a broader understanding of system trade-offs under budget constraints.

4.6. On the Effect of Gain/Loss Ratios in UG@B

In our main experiments, the budget B was defined as the number of documents that could be reviewed. Under this interpretation, every reviewed document is either a true positive ( T P ) or a false positive ( F P ) and thus T P + F P = B . Let us rewrite Equation (8) in the following manner:
U G @ B = g · T P c · F P
where g > 0 is the gain per relevant document found within the budget B, and c > 0 is the cost per nonrelevant document reviewed.
Substituting F P = B T P , we obtain
U G @ B = ( g + c ) · T P c B .
Since B is constant for all systems, the ranking induced by UG@B is strictly increasing in T P . This implies that, when the budget is defined in terms of the number of documents reviewed, the relative weighting of gain and loss (i.e., the ratio g / c ) does not affect the ranking between systems. A system that retrieves more relevant documents within the fixed number of screened documents will always achieve a higher UG@B, regardless of how much losses are penalized. In this setting, no crossover between systems (a crossover occurs when a system performs better or worse than another system according to the parameters of the metric—in this case, g and c) can occur unless a stopping strategy is implemented for one of the systems (more details are given in the Conclusions).
As an alternative, the budget can be defined in terms of time rather than the number of screened documents. In this case, B represents the total annotation time available, where each reviewed document consumes a variable amount of time depending on whether it is relevant or not, and possibly also on the system’s workflow. For example, a system might route likely negatives through a faster triage process, while relevant items require more careful reading. Under this interpretation, the number of documents reviewed within the same time budget may differ across systems, and the mix of T P and F P achieved depends on both the effectiveness and review cost. Let us consider this version of UG@B, which highlights the gain/loss ratio up to a multiplicative constant c:
U G @ B g c · T P F P .
This scenario renders UG@B more sensitive to the gain/loss ratio, and crossovers between systems become possible. To illustrate this, consider the following example. Suppose that the time budget is fixed at B = 100 min. We compare two systems with different review dynamics:
  • System A: True positives take 2.0 min to review, and false positives take 1.0 min. Within 100 min, this system can review ( T P A ,   F P A ) = ( 33 ,   34 ) documents, since 2 · 33 + 1 · 34 = 100 .
  • System B: Both true and false positives take 2.0 and 1.33 min to review, respectively. Within 100 min, this system can review ( T P B ,   F P B ) = ( 30 ,   30 ) documents, since 2 · 30 + 1.33 · 30 100 .
In this situation, the evaluation of the two systems would be
U G A = 2 1 33 34 = 32 U G B = 2 1.33 30 30 = 15.1
and system A would be preferred in terms of UG@B. The question now is the following: is there a gain/cost ratio value for which the preference of one system over the other would switch? Let us rewrite the utilities as a function of the gain/loss ratio r = g / c (with c = 1 without loss of generality):
U G A ( r ) = 33 r 34 , U G B ( r ) = 30 r 30 .
The crossover point occurs when U G A ( r ) = U G B ( r ) :
33 r 34 = 30 r 30 3 r = 4 r 1.33 .
This means that
  • For r < 1.33 (loss-dominant setting), system B is preferred because it minimizes wasted effort on nonrelevant items;
  • For r > 1.33 (gain-dominant setting), system A is preferred because its higher number of true positives outweighs the additional cost of false positives.
This example, displayed in Figure 6, shows that when the budget is defined in terms of time, the choice of gain/loss ratio directly influences the ranking of systems. UG@B thus provides a tunable framework to reflect different operational scenarios, such as prioritizing recall (high g / c ) versus minimizing annotation effort (low g / c ).

5. Conclusions and Future Work

This study presents a comprehensive, budget-aware analysis of document screening strategies in technology-assisted review using real-world evaluation data from the CLEF eHealth shared tasks (2017–2019). Our investigation emphasizes the importance of modeling both retrieval performance and resource constraints when designing and evaluating TAR systems.
In particular, our main contributions are the following.
  • Reproduction of Official Results: We collect, organize, and verify the official participant runs from the CLEF eHealth TAR tasks, providing a unified and reproducible dataset spanning three years of evaluations.
  • Exploration of Budget Allocation Strategies: We investigate how limited screening budgets can be distributed across topics using several allocation strategies, including even, proportional, inverse proportional, and capped greedy approaches. Our analysis reveals that different strategies optimize different objectives: even and inverse proportional strategies favor recall, while proportional and capped strategies enhance efficiency.
  • Introduction of Cost-Aware Evaluation Metrics: We propose two novel measures—relevant found per cost unit (RFCU) and utility gain at budget (UG@B)—that directly account for resource expenditure. These metrics complement traditional recall by quantifying screening efficiency and net utility, and they help to expose trade-offs not visible through recall alone.
Our findings show that no single allocation strategy dominates across all evaluation criteria, underscoring the need to align strategy selection with screening goals—whether completeness, efficiency, or balanced utility. Furthermore, our correlation analysis demonstrates that cost-aware metrics like RFCU and UG@B provide complementary insights alongside recall, rather than duplicating it. The choice of evaluation metric in TARs should reflect the goals and constraints of the review task. In resource-unconstrained environments—such as legal e-discovery with completeness mandates—recall remains the gold standard. However, in practice, many reviews operate under strict cost or time budgets. In such settings, traditional recall may fail to capture a system’s efficiency or utility under constraints. To address this, we propose the following guidelines for selecting evaluation metrics that align with practical review objectives. For a comprehensive performance analysis, we recommend using a composite evaluation approach: recall should be used to ensure topic coverage, while RFCU and UG@B provide complementary perspectives on efficiency and cost sensitivity. This multimetric framework offers a more robust and actionable view of TAR performance, especially in real-world screening scenarios where budgets are finite and reviewers must prioritize effort.
This work opens up several promising directions. We envision the development of cost–risk frontiers, which plot the relationship between resource expenditure and the risk of missing relevant documents. These frontiers would enable a nuanced comparison of TAR strategies, not only by their performance but also by how well they handle the degradation of results under limited resources. Evaluating systems along cost–risk frontiers will provide a deeper understanding of their robustness when constrained by resource availability.
Incorporating human decision making into TAR system evaluation is a crucial avenue for future work, as the effectiveness of TAR workflows is ultimately shaped by human-in-the-loop dynamics. Reviewer behavior can vary widely—in terms of speed, risk tolerance, consistency, fatigue, and stopping strategies—introducing variance in the effective cost and duration of screening. These behavioral factors mean that the performance and utility of a policy can diverge from the outcomes predicted under the assumptions of this paper. To capture this complexity, research in this area will require access to real-world data on reviewer behavior. While simulations can approximate variations in speed, risk tolerance, or stopping strategies, they inevitably rely on simplifying assumptions that may not capture the full complexity of human decision making. In this sense, synthetic data can therefore be misleading, particularly when subtle patterns of fatigue, inconsistency, or adaptation to system guidance play a decisive role. To obtain reliable insights, future work should complement simulations with empirical studies involving actual users, ensuring that evaluations reflect the realities of human interaction with TAR systems.
Automatic early stopping is a crucial aspect of TAR systems, as it directly affects the trade-off between recall and cost. Stopping too early risks missing a substantial portion of relevant documents, while stopping too late expends resources without meaningful gains. Reviewer behavior further complicates this picture: a conservative reviewer may continue screening well beyond the optimal point, consuming unnecessary resources, while a more aggressive one may stop prematurely, jeopardizing recall. These behavioral and policy-driven factors highlight the importance of jointly modeling human decision making and system-level stopping strategies, ensuring that evaluations reflect both theoretical performance and the realities of practice. We propose the exploration of adaptive stopping mechanisms that rely on marginal utility or predictive confidence thresholds, allowing the review process to adjust dynamically to budget constraints. In future work, we aim to estimate the opportunity cost of early stopping by quantifying the risk of missing relevant documents when resources are limited, thereby identifying the most effective strategies for cost-sensitive review settings.
Looking ahead, we aim to move beyond static budget allocation by enabling each TAR system to dynamically determine its own optimal budget usage based on performance signals and uncertainty estimates. Rather than imposing a fixed budget ceiling, future evaluations will explore adaptive budget strategies where systems decide how much annotation effort is necessary to achieve a given level of recall or confidence.

Funding

The APC was funded by the “National Biodiversity Future Center—NBFC” project funded under the National Recovery and Resilience Plan (NRRP), Mission 4 Component 2 Investment 1.4—Call for tender No. 3138 of 16 December 2021, rectified by Decree n. 3175 of 18 December 2021 of Italian Ministry of University and Research funded by the European Union—NextGenerationEU. Project code CN_00000033.

Acknowledgments

This work is partially supported by the HEREDITARY Project, as part of the European Union’s Horizon Europe research and innovation programme under grant agreement No GA 101137074. During the preparation of this manuscript, the author used ChatGPT, OpenAI, GPT-5 exclusively for language refinement (e.g., grammar, style, and clarity). The author have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A

In this appendix, we provide the detailed results of the different allocation budget approaches for each year of the CLEF TAR lab.
Table A1. CLEF TAR 2017 lab. Even allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A1. CLEF TAR 2017 lab. Even allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@95RFCU@kUG@B
1auth.simple.run6_NO_OVERVIEW0.980.980.980.980.820.15−274.67
2auth.simple.run8_NO_OVERVIEW0.980.980.980.980.820.15−274.67
3auth.simple.run7_NO_OVERVIEW0.980.980.980.980.820.14−274.93
4auth.simple.run5_NO_OVERVIEW0.950.950.950.950.720.13−284.53
5auth.simple.run30.890.280.840.270.330.12−301.73
6auth.simple.run10.890.300.840.280.340.11−302.40
7auth.simple.run20.890.300.830.280.340.11−302.67
8auth.simple.run40.890.300.830.280.340.11−302.60
9waterloo.B-rank-cost_NO_OVERVIEW0.890.320.840.310.350.12−299.87
10waterloo.B-rank-normal0.890.320.840.310.350.12−299.87
11waterloo.B-thresh0.890.320.840.310.350.12−299.93
12waterloo.B-thresh-cost_NO_OVERVIEW0.890.320.840.310.350.12−299.93
13waterloo.A-rank-cost_NO_OVERVIEW0.880.290.830.270.310.12−300.20
14waterloo.A-rank-normal0.880.290.830.270.310.12−300.20
15waterloo.A-thresh0.880.290.830.270.310.12−300.20
16waterloo.A-thresh-cost_NO_OVERVIEW0.880.290.830.270.310.12−300.20
17padua.iafap_m10p5f0m100.840.290.780.260.260.11−309.33
18padua.iafapc_m10p20f0t300p2m10.NO_OVERVIEW0.840.300.800.260.290.12−290.03
19padua.iafas_m10k50f0m100.840.280.780.250.260.11−310.80
20padua.iafap_m10p2f0m100.830.280.760.240.210.11−310.93
21padua.iafa_m10k150f0m100.830.300.780.270.280.10−311.93
22uos.sis.AL30Q0.830.210.740.200.170.11−310.50
23padua.iafapc_m10p20f0t150p2m10.NO_OVERVIEW0.820.300.770.260.210.12−292.93
24padua.iafapc_m10p10f0t150p2m10.NO_OVERVIEW0.790.280.730.240.140.12−268.60
25ncsu.abs0.730.080.600.110.120.09−322.67
26padua.iafapc_m10p5f0t0p2m10.NO_OVERVIEW0.710.260.690.220.080.16−213.27
27uos.bm25_10.710.180.620.160.110.09−328.63
28uos.bm25_20.710.180.620.160.110.09−328.63
29uos.bm25_2.50.710.180.620.160.110.09−328.63
30uos.sis.BM25_NO_OVERVIEW0.710.180.620.160.110.09−328.63
31uos.sis.bm25_1.5_NO_OVERVIEW0.710.180.620.160.110.09−328.63
32uos.sis.TMAL30Q_BM250.700.150.610.150.160.09−323.43
33cnrs.noaffull.all0.700.180.610.160.110.09−325.87
34sheffield.run30.700.210.620.180.150.08−329.53
35eth.m10.700.220.630.200.150.09−324.67
36eth.m20.700.220.630.200.150.09−324.67
37eth.m40.700.220.630.200.150.09−324.67
38sheffield.run10.690.180.610.150.190.08−334.47
39sheffield.run20.690.220.610.200.160.09−328.87
40sheffield.run40.690.220.610.200.160.09−328.80
41cnrs.noaf.all0.680.160.570.130.110.08−335.67
42ncsu.simple0.660.070.530.100.060.10−316.80
43amc.run0.610.130.500.110.090.07−338.47
44cnrs.abrupt.all0.610.160.510.130.040.07−340.67
45ecnu.run30.590.200.540.17−0.010.10−280.47
46qut.ca_bool_ltr0.570.130.460.100.030.08−337.00
47ecnu.run20.570.200.520.16−0.020.11−265.93
48qut.ca_pico_ltr0.570.160.480.140.060.07−339.53
49uos.sis.TMBEST_BM250.570.120.460.110.140.07−343.57
50cnrs.gradual.all0.560.150.460.130.050.08−337.53
51qut.rf_bool_ltr0.550.110.420.090.000.08−338.20
52iiit.run10.550.160.510.140.050.12−196.37
53iiit.run20.550.160.510.140.050.12−196.37
54iiit.run30.550.160.510.140.050.12−196.37
55iiit.run40.550.160.510.140.050.12−196.37
56iiit.run5_NO_OVERVIEW0.550.160.510.140.050.12−196.37
57iiit.run6_NO_OVERVIEW0.550.160.510.140.050.12−196.37
58iiit.run7_NO_OVERVIEW0.550.160.510.140.050.12−196.37
59iiit.run8_NO_OVERVIEW0.550.160.510.140.050.12−196.37
60qut.rf_pico_ltr0.540.130.430.100.050.07−344.40
61qut.pico_es0.500.160.450.110.050.08−290.47
62ntu.run10.490.070.350.06−0.020.06−348.80
63qut.bool_es0.490.150.440.120.030.07−305.33
64ecnu.run10.470.090.350.08−0.000.06−356.13
65ntu.run20.420.050.290.05−0.010.05−358.20
66ntu.run30.420.040.280.040.020.04−369.47
67ucl.run_abstract0.400.050.260.05−0.000.05−362.47
68ucl.run_fulltext0.380.050.250.04−0.000.04−371.27
69uos.sis.pubmed.random_NO_OVERVIEW0.360.040.210.03−0.040.04−369.33
Table A2. CLEF TAR 2017 lab. Proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A2. CLEF TAR 2017 lab. Proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1auth.simple.run6_NO_OVERVIEW0.940.940.940.940.720.33−276.73
2auth.simple.run8_NO_OVERVIEW0.940.940.940.940.720.33−276.73
3auth.simple.run7_NO_OVERVIEW0.940.940.940.940.720.33−277.00
4auth.simple.run5_NO_OVERVIEW0.910.910.910.910.620.32−285.80
5waterloo.B-rank-cost_NO_OVERVIEW0.740.290.720.250.110.24−300.20
6waterloo.B-rank-normal0.740.290.720.250.110.24−300.20
7waterloo.B-thresh0.730.290.710.250.110.24−298.23
8waterloo.B-thresh-cost_NO_OVERVIEW0.730.290.710.250.110.24−298.23
9auth.simple.run10.730.270.710.220.080.22−302.53
10auth.simple.run30.710.250.700.210.050.21−302.67
11waterloo.A-rank-cost_NO_OVERVIEW0.710.260.690.200.110.22−302.20
12waterloo.A-rank-normal0.710.260.690.200.110.22−302.20
13waterloo.A-thresh0.710.260.690.200.110.22−302.20
14waterloo.A-thresh-cost_NO_OVERVIEW0.710.260.690.200.110.22−302.20
15auth.simple.run20.700.270.680.220.110.20−303.20
16auth.simple.run40.690.270.680.220.080.20−303.27
17padua.iafap_m10p5f0m100.680.270.660.210.080.20−309.13
18padua.iafa_m10k150f0m100.670.270.660.210.080.20−310.80
19padua.iafas_m10k50f0m100.670.260.650.200.080.19−310.60
20padua.iafapc_m10p20f0t300p2m10.NO_OVERVIEW0.660.270.650.200.080.19−283.07
21padua.iafapc_m10p5f0t0p2m10.NO_OVERVIEW0.640.240.630.190.080.16−312.07
22padua.iafap_m10p2f0m100.640.260.620.190.050.18−312.13
23padua.iafapc_m10p10f0t150p2m10.NO_OVERVIEW0.630.270.620.200.010.19−215.57
24padua.iafapc_m10p20f0t150p2m10.NO_OVERVIEW0.630.270.620.200.010.19−215.57
25uos.sis.AL30Q0.580.190.560.150.010.15−312.87
26uos.sis.TMAL30Q_BM250.510.140.490.10−0.020.14−323.40
27cnrs.noaffull.all0.510.170.490.12−0.020.15−326.67
28eth.m10.510.200.490.150.010.17−327.07
29eth.m20.510.200.490.150.010.17−327.07
30eth.m40.510.200.490.150.010.17−327.07
31sheffield.run40.500.200.490.15−0.020.16−330.20
32sheffield.run20.500.200.480.15−0.020.16−330.00
33sheffield.run30.490.200.480.130.010.15−330.87
34sheffield.run10.470.160.450.100.010.15−338.13
35uos.bm25_10.450.160.440.11−0.020.12−334.00
36uos.bm25_20.450.160.440.11−0.020.12−334.00
37uos.bm25_2.50.450.160.440.11−0.020.12−334.00
38uos.sis.BM25_NO_OVERVIEW0.450.160.440.11−0.020.12−334.00
39uos.sis.bm25_1.5_NO_OVERVIEW0.450.160.440.11−0.020.12−334.00
40cnrs.noaf.all0.440.150.430.09−0.020.13−341.40
41ncsu.abs0.430.060.420.06−0.050.10−334.47
42ecnu.run30.430.170.420.12−0.020.15−302.13
43iiit.run10.430.150.410.10−0.020.13−225.37
44iiit.run20.430.150.410.10−0.020.13−225.37
45iiit.run30.430.150.410.10−0.020.13−225.37
46iiit.run40.430.150.410.10−0.020.13−225.37
47iiit.run5_NO_OVERVIEW0.430.150.410.10−0.020.13−225.37
48iiit.run6_NO_OVERVIEW0.430.150.410.10−0.020.13−225.37
49iiit.run7_NO_OVERVIEW0.430.150.410.10−0.020.13−225.37
50iiit.run8_NO_OVERVIEW0.430.150.410.10−0.020.13−225.37
51ecnu.run20.420.180.410.12−0.020.16−291.37
52ncsu.simple0.420.070.400.06−0.020.11−317.60
53qut.bool_es0.370.140.360.09−0.050.13−338.53
54qut.pico_es0.370.150.360.08−0.020.12−340.87
55cnrs.abrupt.all0.370.140.360.08−0.050.12−347.40
56amc.run0.340.110.330.07−0.050.12−341.67
57qut.ca_pico_ltr0.340.150.330.10−0.050.13−343.20
58cnrs.gradual.all0.330.130.320.08−0.050.14−342.53
59qut.ca_bool_ltr0.310.120.300.06−0.050.09−345.47
60uos.sis.TMBEST_BM250.310.110.290.07−0.050.12−351.07
61qut.rf_bool_ltr0.280.100.270.06−0.050.09−347.27
62qut.rf_pico_ltr0.260.110.250.06−0.050.10−351.07
63ecnu.run10.200.080.190.04−0.050.08−364.33
64ntu.run10.200.060.190.03−0.050.07−358.47
65ntu.run20.150.050.150.01−0.050.05−367.13
66ucl.run_abstract0.110.040.100.02−0.050.05−374.80
67ucl.run_fulltext0.100.030.100.01−0.050.04−383.07
68ntu.run30.090.030.090.01−0.050.03−381.53
69uos.sis.pubmed.random_NO_OVERVIEW0.090.030.090.01−0.050.04−379.67
Table A3. CLEF TAR 2017 lab. Inverse proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A3. CLEF TAR 2017 lab. Inverse proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1auth.simple.run6_NO_OVERVIEW0.980.980.980.980.820.15−275.37
2auth.simple.run8_NO_OVERVIEW0.980.980.980.980.820.15−275.37
3auth.simple.run7_NO_OVERVIEW0.980.980.980.980.820.15−275.63
4auth.simple.run5_NO_OVERVIEW0.950.950.950.950.720.14−284.50
5auth.simple.run30.880.280.830.270.330.12−305.23
6auth.simple.run10.880.300.830.280.360.12−305.63
7auth.simple.run20.880.300.820.280.370.12−305.03
8auth.simple.run40.880.300.820.280.340.12−305.43
9waterloo.A-rank-cost_NO_OVERVIEW0.880.280.820.260.280.12−303.00
10waterloo.A-rank-normal0.880.280.820.260.280.12−303.00
11waterloo.A-thresh0.880.280.820.260.280.12−303.00
12waterloo.A-thresh-cost_NO_OVERVIEW0.880.280.820.260.280.12−303.00
13waterloo.B-rank-cost_NO_OVERVIEW0.870.320.830.300.290.12−303.07
14waterloo.B-rank-normal0.870.320.830.300.290.12−303.07
15waterloo.B-thresh0.870.320.830.300.290.12−303.07
16waterloo.B-thresh-cost_NO_OVERVIEW0.870.320.830.300.290.12−303.07
17padua.iafas_m10k50f0m100.830.280.780.250.250.11−313.30
18padua.iafa_m10k150f0m100.830.300.780.260.250.11−314.37
19padua.iafapc_m10p20f0t300p2m10.NO_OVERVIEW0.820.300.770.260.270.12−286.80
20padua.iafap_m10p5f0m100.820.290.760.250.230.11−312.10
21uos.sis.AL30Q0.820.210.730.200.200.11−313.37
22padua.iafap_m10p2f0m100.820.270.740.240.180.11−313.50
23padua.iafapc_m10p20f0t150p2m10.NO_OVERVIEW0.810.300.760.250.220.12−281.70
24padua.iafapc_m10p10f0t150p2m10.NO_OVERVIEW0.780.280.730.230.140.13−250.77
25ncsu.abs0.750.080.610.110.180.09−324.10
26uos.bm25_10.710.170.610.160.110.09−329.90
27uos.bm25_20.710.170.610.160.110.09−329.90
28uos.bm25_2.50.710.170.610.160.110.09−329.90
29uos.sis.BM25_NO_OVERVIEW0.710.170.610.160.110.09−329.90
30uos.sis.bm25_1.5_NO_OVERVIEW0.710.170.610.160.110.09−329.90
31cnrs.noaffull.all0.700.180.600.160.180.09−329.03
32uos.sis.TMAL30Q_BM250.700.150.600.140.200.09−326.03
33eth.m10.690.220.620.200.170.09−326.43
34eth.m20.690.220.620.200.170.09−326.43
35eth.m40.690.220.620.200.170.09−326.43
36sheffield.run30.690.210.610.180.150.08−331.13
37padua.iafapc_m10p5f0t0p2m10.NO_OVERVIEW0.690.260.670.220.040.16−194.37
38sheffield.run20.690.220.610.200.160.08−330.67
39sheffield.run10.680.180.590.150.190.08−336.87
40sheffield.run40.680.220.600.200.160.08−331.13
41ncsu.simple0.670.070.520.100.060.10−320.03
42cnrs.noaf.all0.660.160.550.130.130.08−337.57
43amc.run0.600.120.490.110.100.07−340.77
44cnrs.gradual.all0.600.150.470.130.080.08−338.17
45cnrs.abrupt.all0.590.160.490.130.050.07−343.57
46ecnu.run30.580.190.520.160.000.10−272.90
47qut.ca_pico_ltr0.570.160.480.140.060.07−340.70
48qut.ca_bool_ltr0.570.120.450.100.030.08−338.37
49ecnu.run20.560.200.520.160.000.11−258.07
50uos.sis.TMBEST_BM250.560.120.450.110.140.07−345.10
51qut.rf_bool_ltr0.550.110.420.090.000.07−339.30
52iiit.run10.550.160.500.140.050.12−190.10
53iiit.run20.550.160.500.140.050.12−190.10
54iiit.run30.550.160.500.140.050.12−190.10
55iiit.run40.550.160.500.140.050.12−190.10
56iiit.run5_NO_OVERVIEW0.550.160.500.140.050.12−190.10
57iiit.run6_NO_OVERVIEW0.550.160.500.140.050.12−190.10
58iiit.run7_NO_OVERVIEW0.550.160.500.140.050.12−190.10
59iiit.run8_NO_OVERVIEW0.550.160.500.140.050.12−190.10
60qut.rf_pico_ltr0.540.130.430.100.050.07−345.23
61ntu.run10.520.070.360.06−0.020.06−349.50
62qut.bool_es0.500.150.440.120.030.07−297.00
63ecnu.run10.490.090.360.08−0.000.05−356.37
64qut.pico_es0.490.160.440.110.050.08−277.13
65ntu.run20.440.050.290.05−0.000.05−358.43
66ntu.run30.430.040.290.040.020.04−367.80
67ucl.run_abstract0.420.050.260.05−0.000.05−361.53
68uos.sis.pubmed.random_NO_OVERVIEW0.400.040.220.03−0.040.04−367.93
69ucl.run_fulltext0.400.050.260.04−0.000.04−369.60
Table A4. CLEF TAR 2017 lab. Capped allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A4. CLEF TAR 2017 lab. Capped allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1auth.simple.run6_NO_OVERVIEW0.980.980.980.980.820.23−274.67
2auth.simple.run8_NO_OVERVIEW0.980.980.980.980.820.23−274.67
3auth.simple.run7_NO_OVERVIEW0.980.980.980.980.820.23−274.93
4auth.simple.run5_NO_OVERVIEW0.950.950.950.950.720.22−284.20
5auth.simple.run10.820.300.790.270.210.17−300.13
6auth.simple.run30.810.280.780.260.140.17−300.00
7waterloo.B-rank-cost_NO_OVERVIEW0.810.320.780.290.210.17−298.47
8waterloo.B-rank-normal0.810.320.780.290.210.17−298.47
9auth.simple.run20.810.300.770.260.220.17−300.87
10auth.simple.run40.800.300.770.260.190.17−300.87
11waterloo.B-thresh0.800.320.780.290.210.17−296.50
12waterloo.B-thresh-cost_NO_OVERVIEW0.800.320.780.290.210.17−296.50
13waterloo.A-rank-cost_NO_OVERVIEW0.790.290.760.250.200.16−300.20
14waterloo.A-rank-normal0.790.290.760.250.200.16−300.20
15waterloo.A-thresh0.790.290.760.250.200.16−300.20
16waterloo.A-thresh-cost_NO_OVERVIEW0.790.290.760.250.200.16−300.20
17padua.iafa_m10k150f0m100.780.300.750.250.210.16−308.13
18padua.iafapc_m10p20f0t300p2m10.NO_OVERVIEW0.770.300.740.240.200.16−289.23
19padua.iafas_m10k50f0m100.770.280.740.240.210.16−308.40
20padua.iafap_m10p5f0m100.760.290.730.240.170.16−307.00
21padua.iafapc_m10p20f0t150p2m10.NO_OVERVIEW0.740.300.710.240.130.16−221.73
22padua.iafap_m10p2f0m100.730.280.690.220.090.15−309.80
23padua.iafapc_m10p10f0t150p2m10.NO_OVERVIEW0.710.280.690.220.070.16−221.50
24uos.sis.AL30Q0.710.210.660.190.070.15−309.67
25padua.iafapc_m10p5f0t0p2m10.NO_OVERVIEW0.690.260.670.210.110.16−304.23
26uos.sis.TMAL30Q_BM250.640.150.580.130.060.15−319.97
27eth.m10.610.220.570.190.040.13−323.87
28eth.m20.610.220.570.190.040.13−323.87
29eth.m40.610.220.570.190.040.13−323.87
30sheffield.run40.600.220.560.180.030.13−327.40
31sheffield.run20.600.220.560.190.030.13−327.13
32sheffield.run30.600.210.560.160.030.13−328.40
33cnrs.noaffull.all0.600.180.560.150.010.13−324.00
34uos.bm25_10.570.180.530.140.010.12−331.07
35uos.bm25_20.570.180.530.140.010.12−331.07
36uos.bm25_2.50.570.180.530.140.010.12−331.07
37uos.sis.BM25_NO_OVERVIEW0.570.180.530.140.010.12−331.07
38uos.sis.bm25_1.5_NO_OVERVIEW0.570.180.530.140.010.12−331.07
39sheffield.run10.570.180.520.140.050.12−335.40
40ncsu.abs0.550.080.490.08−0.020.12−330.47
41cnrs.noaf.all0.530.160.490.110.030.10−339.47
42ncsu.simple0.520.070.460.080.030.13−314.53
43iiit.run10.520.160.490.130.020.12−218.67
44iiit.run20.520.160.490.130.020.12−218.67
45iiit.run30.520.160.490.130.020.12−218.67
46iiit.run40.520.160.490.130.020.12−218.67
47iiit.run5_NO_OVERVIEW0.520.160.490.130.020.12−218.67
48iiit.run6_NO_OVERVIEW0.520.160.490.130.020.12−218.67
49iiit.run7_NO_OVERVIEW0.520.160.490.130.020.12−218.67
50iiit.run8_NO_OVERVIEW0.520.160.490.130.020.12−218.67
51ecnu.run30.510.200.480.15−0.020.14−304.47
52ecnu.run20.490.200.470.15−0.020.13−293.90
53cnrs.abrupt.all0.450.160.420.11−0.010.09−345.27
54amc.run0.440.130.400.09−0.020.10−339.53
55qut.ca_pico_ltr0.430.160.390.12−0.000.10−341.00
56qut.bool_es0.420.150.400.11−0.050.10−333.73
57cnrs.gradual.all0.420.150.390.110.010.11−340.33
58uos.sis.TMBEST_BM250.420.120.360.09−0.030.11−348.30
59qut.pico_es0.410.160.390.10−0.000.10−335.67
60qut.ca_bool_ltr0.400.130.360.08−0.040.08−343.67
61qut.rf_bool_ltr0.360.110.330.07−0.040.08−344.87
62qut.rf_pico_ltr0.350.130.320.09−0.000.09−348.80
63ecnu.run10.300.090.270.06−0.050.07−362.13
64ntu.run10.290.070.250.05−0.040.07−356.20
65ntu.run20.250.050.210.03−0.040.06−364.93
66ntu.run30.210.040.160.03−0.050.05−378.33
67ucl.run_abstract0.190.050.150.03−0.040.05−372.47
68ucl.run_fulltext0.180.050.150.02−0.050.03−381.00
69uos.sis.pubmed.random_NO_OVERVIEW0.160.040.120.02−0.050.04−378.07
Table A5. CLEF TAR 2017 lab. Epsilon-greedy multiarmed bandit. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A5. CLEF TAR 2017 lab. Epsilon-greedy multiarmed bandit. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1waterloo.B-thresh-cost_NO_OVERVIEW0.830.310.780.300.400.12−282.67
2waterloo.B-thresh0.820.310.770.300.400.12−279.07
3waterloo.A-rank-normal0.780.280.740.250.350.14−286.73
4waterloo.A-thresh-cost_NO_OVERVIEW0.780.280.730.250.350.14−287.37
5waterloo.A-thresh0.780.280.730.250.350.13−286.00
6waterloo.A-rank-cost_NO_OVERVIEW0.770.280.730.250.320.14−285.50
7padua.iafa_m10k150f0m100.770.300.720.260.290.13−292.50
8padua.iafapc_m10p20f0t150p2m10.NO_OVERVIEW0.760.290.710.250.180.14−221.47
9padua.iafapc_m10p20f0t300p2m10.NO_OVERVIEW0.750.290.710.250.240.14−278.77
10padua.iafas_m10k50f0m100.750.280.710.240.240.13−292.63
11waterloo.B-rank-cost_NO_OVERVIEW0.750.310.720.290.360.14−286.70
12waterloo.B-rank-normal0.740.310.710.290.360.13−287.17
13padua.iafap_m10p5f0m100.740.280.690.250.220.13−290.97
14padua.iafap_m10p2f0m100.700.270.660.220.140.14−289.90
15padua.iafapc_m10p10f0t150p2m10.NO_OVERVIEW0.680.270.650.220.100.15−199.30
16sheffield.run40.650.220.590.200.120.10−306.60
17padua.iafapc_m10p5f0t0p2m10.NO_OVERVIEW0.630.260.610.210.080.18−171.23
18sheffield.run30.630.210.570.180.090.10−308.17
19sheffield.run20.620.210.570.200.080.10−307.67
20sheffield.run10.610.180.540.150.100.08−310.57
21uos.sis.AL30Q0.600.200.540.180.110.13−289.87
22uos.sis.TMAL30Q_BM250.560.150.490.130.130.11−301.03
23ncsu.abs0.560.070.490.090.060.09−283.47
24uos.bm25_10.550.170.480.140.030.11−306.23
25iiit.run5_NO_OVERVIEW0.550.160.500.140.020.12−242.03
26uos.bm25_2.50.540.170.480.140.020.11−304.93
27iiit.run40.540.160.490.140.020.12−241.63
28iiit.run7_NO_OVERVIEW0.540.160.490.140.020.12−243.47
29uos.bm25_20.540.170.480.140.020.11−307.77
30iiit.run20.540.160.490.140.020.12−239.20
31iiit.run6_NO_OVERVIEW0.540.160.490.140.020.12−239.20
32uos.sis.bm25_1.5_NO_OVERVIEW0.540.170.470.140.020.11−303.43
33iiit.run8_NO_OVERVIEW0.540.160.480.140.020.12−243.43
34iiit.run30.530.160.480.140.020.12−241.17
35qut.ca_pico_ltr0.530.160.460.13−0.010.09−314.43
36uos.sis.BM25_NO_OVERVIEW0.530.170.470.14−0.000.11−307.50
37amc.run0.470.120.410.100.020.09−308.30
38qut.bool_es0.450.150.410.12−0.050.09−276.53
39uos.sis.TMBEST_BM250.440.120.370.100.020.09−316.33
40qut.pico_es0.430.150.390.11−0.030.10−257.17
41qut.rf_pico_ltr0.430.120.370.09−0.010.08−319.87
42qut.ca_bool_ltr0.430.120.360.08−0.030.07−314.70
43qut.rf_bool_ltr0.420.100.350.08−0.040.07−317.93
44ntu.run30.380.040.260.04−0.020.04−339.77
45ntu.run10.370.070.310.05−0.030.06−314.77
46ncsu.simple0.340.050.280.050.090.06−265.93
47ucl.run_fulltext0.320.050.210.04−0.040.04−323.73
48ntu.run20.320.050.240.03−0.030.05−329.80
49uos.sis.pubmed.random_NO_OVERVIEW0.290.040.180.03−0.040.04−336.97
50ucl.run_abstract0.240.050.190.03−0.040.04−339.50
Table A6. CLEF TAR 2018 lab. Even allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A6. CLEF TAR 2018 lab. Even allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1auth_run30.930.390.890.380.510.14−512.50
2auth_run10.920.400.890.390.510.14−512.77
3auth_run20.920.400.890.390.510.14−512.77
4UWB0.920.390.880.370.450.14−510.90
5UWA0.920.370.880.350.440.14−512.37
6cnrs_RF_bi0.840.330.800.290.270.12−550.03
7cnrs_comb0.840.340.800.320.300.12−553.50
8shef-feed0.840.560.800.590.350.12−543.83
9cnrs_RF_uni0.830.350.790.300.230.12−551.70
10unipd_t5000.830.360.770.300.260.12−551.43
11unipd_t10000.820.360.770.300.170.12−553.23
12unipd_t15000.820.360.770.300.170.12−553.23
13shef-general0.750.280.690.240.200.10−579.97
14shef-query0.720.240.660.210.190.10−586.83
15uic_model70.640.210.550.16−0.020.08−608.57
16uic_model80.630.200.540.16−0.020.08−610.30
17ECNU_RUN30.610.200.510.13−0.040.08−605.83
18ECNU_RUN10.600.190.490.12−0.040.08−612.70
19ECNU_RUN20.460.120.400.07−0.050.07−606.10
Table A7. CLEF TAR 2018 lab. Proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A7. CLEF TAR 2018 lab. Proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1auth_run10.790.380.770.330.170.28−535.03
2auth_run20.790.380.770.330.170.28−535.03
3auth_run30.780.370.760.320.170.28−535.57
4UWB0.770.360.760.300.140.27−535.30
5UWA0.770.350.750.290.110.27−536.57
6cnrs_comb0.720.310.700.260.110.26−567.30
7shef-feed0.720.540.710.550.110.29−563.50
8cnrs_RF_bi0.700.320.680.240.080.24−567.23
9unipd_t5000.690.330.670.250.110.24−568.97
10cnrs_RF_uni0.690.330.670.250.080.24−568.97
11unipd_t15000.680.330.670.250.080.24−572.37
12unipd_t10000.680.330.660.250.080.24−572.83
13shef-general0.550.260.530.18−0.050.22−614.90
14shef-query0.500.220.480.14−0.050.20−625.43
15uic_model70.420.190.410.11−0.050.17−640.43
16uic_model80.410.180.400.11−0.050.16−642.17
17ECNU_RUN30.410.180.400.09−0.050.16−634.43
18ECNU_RUN10.380.170.370.09−0.050.14−646.03
19ECNU_RUN20.260.100.250.04−0.050.11−657.90
Table A8. CLEF TAR 2018 lab. Inverse proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A8. CLEF TAR 2018 lab. Inverse proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1UWB0.910.390.870.360.470.14−520.83
2auth_run10.910.400.870.380.480.14−523.87
3auth_run20.910.400.870.380.480.14−523.87
4auth_run30.910.390.870.380.480.14−523.07
5UWA0.900.370.860.340.470.14−523.10
6cnrs_RF_bi0.840.330.790.290.320.12−554.50
7cnrs_comb0.830.340.790.320.370.11−557.77
8shef-feed0.830.560.800.590.400.12−547.77
9cnrs_RF_uni0.820.350.770.290.250.12−558.17
10unipd_t5000.820.360.760.300.270.12−556.77
11unipd_t10000.820.360.760.300.190.12−558.23
12unipd_t15000.820.360.760.300.190.12−558.23
13shef-general0.730.280.680.240.200.09−585.43
14shef-query0.710.240.650.210.210.09−590.63
15uic_model70.650.210.550.17−0.000.08−607.63
16uic_model80.640.200.540.16−0.000.08−609.97
17ECNU_RUN30.620.200.500.13−0.040.08−604.70
18ECNU_RUN10.600.190.480.13−0.040.08−610.57
19ECNU_RUN20.450.120.390.07−0.050.07−589.57
Table A9. CLEF TAR 2018 lab. Capped allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A9. CLEF TAR 2018 lab. Capped allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1auth_run30.850.380.830.350.260.24−532.37
2auth_run10.850.380.830.350.260.24−532.23
3auth_run20.850.380.830.350.260.24−532.23
4UWB0.850.370.820.330.230.24−531.83
5UWA0.840.360.810.310.200.24−533.23
6cnrs_comb0.780.320.760.290.180.22−564.83
7cnrs_RF_bi0.770.320.740.270.150.22−563.30
8shef-feed0.770.550.750.560.190.24−561.10
9cnrs_RF_uni0.750.340.730.270.150.22−565.37
10unipd_t5000.750.340.730.270.130.20−566.30
11unipd_t15000.740.340.720.270.100.20−569.77
12unipd_t10000.740.340.720.270.100.20−570.30
13shef-general0.610.260.590.20−0.000.19−611.70
14shef-query0.570.220.550.170.000.17−622.23
15uic_model70.470.190.450.13−0.040.15−638.10
16uic_model80.460.190.440.13−0.030.14−639.97
17ECNU_RUN30.440.190.420.10−0.050.13−633.23
18ECNU_RUN10.410.170.390.10−0.050.11−644.90
19ECNU_RUN20.320.110.290.05−0.050.11−655.80
Table A10. CLEF TAR 2018 lab. Epsilon-greedy multiarmed bandit. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A10. CLEF TAR 2018 lab. Epsilon-greedy multiarmed bandit. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall_at_kAUCMAPWSS_at_r95rfbuug_at_budget
1auth_run30.890.390.850.370.510.15−483.23
2auth_run20.880.400.850.370.510.15−490.10
3auth_run10.880.400.840.370.510.15−490.03
4UWB0.870.380.830.350.490.15−489.37
5UWA0.840.350.810.330.490.15−499.77
6shef-feed0.830.560.800.590.330.13−487.33
7unipd_t5000.810.350.750.290.230.13−512.27
8cnrs_RF_bi0.800.330.750.290.320.13−514.90
9unipd_t15000.790.350.740.290.170.13−511.53
10cnrs_comb0.790.330.750.310.350.13−525.73
11unipd_t10000.790.350.730.290.170.13−512.63
12cnrs_RF_uni0.790.340.740.290.240.13−519.07
13shef-general0.720.270.670.240.270.10−544.50
14shef-query0.670.230.620.200.260.10−546.57
15ECNU_RUN30.570.190.480.12−0.050.09−565.67
16uic_model70.550.200.490.160.020.08−550.17
17ECNU_RUN10.540.180.460.12−0.040.08−571.93
18uic_model80.540.200.470.150.030.08−551.33
19ECNU_RUN20.430.110.380.07−0.050.07−559.73
Table A11. CLEF TAR 2019 lab. Even allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A11. CLEF TAR 2019 lab. Even allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
12018_stem_original_p10_t10000.800.260.720.240.240.14−193.39
22018_stem_original_p10_t15000.800.260.720.240.240.14−193.39
32018_stem_original_p10_t3000.800.260.720.240.240.14−193.39
42018_stem_original_p10_t4000.800.260.720.240.240.14−193.39
52018_stem_original_p10_t5000.800.260.720.240.240.14−193.39
6distributed_effort_p10_t10000.800.260.720.230.230.14−193.39
7distributed_effort_p10_t15000.800.260.730.240.240.14−193.39
82018_stem_original_p50_t10000.800.260.730.240.190.14−193.42
92018_stem_original_p50_t15000.800.260.730.240.190.14−193.42
102018_stem_original_p50_t3000.800.260.730.240.190.14−193.42
112018_stem_original_p50_t4000.800.260.730.240.190.14−193.42
122018_stem_original_p50_t5000.800.260.730.240.190.14−193.42
132018_stem_original_p10_t2000.800.260.720.230.240.14−193.58
14distributed_effort_p10_t5000.790.260.720.230.190.14−194.16
152018_stem_original_p50_t2000.790.260.730.240.210.14−193.42
16distributed_effort_p10_t4000.790.260.710.230.190.14−194.81
17distributed_effort_p10_t3000.790.260.710.230.190.14−195.32
18distributed_effort_p10_t2000.780.250.700.230.180.14−195.84
19abs-hh-ratio-ilps0.730.460.690.510.180.15−186.71
20distributed_effort_p10_t1000.730.240.620.210.120.13−198.35
212018_stem_original_p50_t1000.720.260.660.230.210.13−197.10
222018_stem_original_p10_t1000.710.260.640.230.180.13−197.65
23abs-th-ratio-ilps0.710.430.670.480.210.14−193.90
24sheffield-baseline0.630.210.560.190.120.10−218.65
25baseline_bm25_t1000.570.180.500.160.110.09−223.10
26baseline_bm25_t2000.560.180.500.160.140.09−223.94
27baseline_bm25_t10000.560.180.490.160.140.09−224.00
28baseline_bm25_t15000.560.180.490.160.140.09−224.00
29baseline_bm25_t3000.560.180.490.160.140.09−224.00
30baseline_bm25_t4000.560.180.490.160.140.09−224.00
31baseline_bm25_t5000.560.180.490.160.140.09−224.00
Table A12. CLEF TAR 2019 lab. Proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A12. CLEF TAR 2019 lab. Proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
1abs-hh-ratio-ilps0.630.440.620.460.140.29−194.81
22018_stem_original_p10_t5000.620.230.610.180.040.21−206.06
32018_stem_original_p50_t5000.620.230.600.180.040.21−206.13
42018_stem_original_p10_t4000.620.230.600.180.040.21−206.26
52018_stem_original_p50_t4000.620.230.600.180.040.21−206.32
62018_stem_original_p10_t10000.620.230.600.180.040.21−206.32
72018_stem_original_p10_t15000.620.230.600.180.040.21−206.32
8distributed_effort_p10_t10000.620.230.600.180.040.21−206.32
9distributed_effort_p10_t15000.620.230.600.180.040.21−206.32
10distributed_effort_p10_t3000.620.230.600.180.040.21−206.32
11distributed_effort_p10_t4000.620.230.600.180.040.21−206.32
12distributed_effort_p10_t5000.620.230.600.180.040.21−206.32
132018_stem_original_p50_t10000.620.230.600.180.040.21−206.39
142018_stem_original_p50_t15000.620.230.600.180.040.21−206.39
152018_stem_original_p10_t3000.620.230.600.180.010.21−206.45
162018_stem_original_p50_t3000.620.230.600.180.010.21−206.52
172018_stem_original_p10_t2000.610.230.590.180.010.21−207.10
182018_stem_original_p50_t2000.610.230.590.180.010.21−207.16
19distributed_effort_p10_t2000.600.230.580.170.100.20−207.13
20abs-th-ratio-ilps0.590.400.580.420.140.27−202.61
212018_stem_original_p10_t1000.540.240.520.17−0.020.21−209.29
222018_stem_original_p50_t1000.540.240.520.17−0.020.21−209.29
23distributed_effort_p10_t1000.520.220.500.160.010.17−211.13
24sheffield-baseline0.450.190.440.140.010.15−229.68
25baseline_bm25_t1000.410.160.400.110.010.14−232.06
26baseline_bm25_t4000.410.160.390.110.010.14−232.32
27baseline_bm25_t3000.410.160.390.110.010.14−232.77
28baseline_bm25_t2000.410.160.390.110.010.14−232.90
29baseline_bm25_t5000.400.160.390.110.010.14−232.71
30baseline_bm25_t10000.400.160.390.110.010.14−233.10
31baseline_bm25_t15000.400.160.390.110.010.14−233.10
Table A13. CLEF TAR 2019 lab. Inverse proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A13. CLEF TAR 2019 lab. Inverse proportional allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
12018_stem_original_p10_t10000.780.260.700.230.250.14−196.16
22018_stem_original_p10_t15000.780.260.700.230.250.14−196.16
32018_stem_original_p10_t4000.780.260.700.230.250.14−196.16
42018_stem_original_p10_t5000.780.260.700.230.250.14−196.16
5distributed_effort_p10_t10000.780.260.700.230.240.14−196.16
62018_stem_original_p10_t3000.780.260.700.230.240.14−196.23
7distributed_effort_p10_t15000.780.260.700.230.240.14−196.29
82018_stem_original_p10_t2000.780.260.700.230.250.14−195.71
92018_stem_original_p50_t2000.780.260.710.240.250.14−195.74
10distributed_effort_p10_t5000.770.260.690.230.220.14−196.42
11distributed_effort_p10_t4000.770.260.690.230.210.14−196.48
12distributed_effort_p10_t3000.770.260.690.230.210.14−196.61
132018_stem_original_p50_t10000.770.250.710.240.220.14−196.19
142018_stem_original_p50_t15000.770.250.710.240.220.14−196.19
152018_stem_original_p50_t3000.770.250.710.240.220.14−196.19
162018_stem_original_p50_t4000.770.250.710.240.220.14−196.19
172018_stem_original_p50_t5000.770.250.710.240.220.14−196.19
18distributed_effort_p10_t2000.770.250.680.220.180.14−197.13
19abs-hh-ratio-ilps0.730.460.690.510.190.16−187.23
20distributed_effort_p10_t1000.710.240.600.200.130.13−199.00
212018_stem_original_p50_t1000.710.260.650.230.220.13−198.19
222018_stem_original_p10_t1000.710.260.640.230.200.13−198.42
23abs-th-ratio-ilps0.710.430.660.480.210.14−194.16
24sheffield-baseline0.640.210.560.190.140.10−217.68
25baseline_bm25_t10000.590.180.500.160.150.09−220.90
26baseline_bm25_t15000.590.180.500.160.150.09−220.90
27baseline_bm25_t4000.590.180.500.160.150.09−220.90
28baseline_bm25_t5000.590.180.500.160.150.09−220.90
29baseline_bm25_t3000.590.180.500.160.150.09−220.90
30baseline_bm25_t2000.590.180.500.160.160.09−221.10
31baseline_bm25_t1000.590.180.500.160.130.09−220.58
Table A14. CLEF TAR 2019 lab. Capped allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A14. CLEF TAR 2019 lab. Capped allocation of budget. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall@kAUCMAPWSS@r95RFCU@kUG@B
12018_stem_original_p50_t5000.700.250.670.210.110.18−202.19
22018_stem_original_p10_t5000.700.260.660.200.110.18−202.42
32018_stem_original_p50_t4000.700.250.670.210.110.18−202.32
42018_stem_original_p10_t4000.700.260.660.200.110.18−202.55
52018_stem_original_p50_t10000.700.250.670.210.110.18−202.45
62018_stem_original_p50_t15000.700.250.670.210.110.18−202.45
72018_stem_original_p10_t10000.700.260.660.200.110.18−202.68
82018_stem_original_p10_t15000.700.260.660.200.110.18−202.68
9distributed_effort_p10_t10000.700.260.660.200.100.18−202.68
10distributed_effort_p10_t15000.700.260.660.200.110.18−202.68
11distributed_effort_p10_t3000.700.260.660.200.100.18−202.42
12distributed_effort_p10_t4000.700.260.660.200.100.18−202.48
13distributed_effort_p10_t5000.700.250.660.200.100.18−202.68
142018_stem_original_p50_t3000.700.250.660.210.080.18−202.52
152018_stem_original_p10_t3000.700.260.650.200.080.18−202.74
162018_stem_original_p50_t2000.690.250.660.210.080.18−203.23
172018_stem_original_p10_t2000.690.260.650.200.080.18−203.45
18abs-hh-ratio-ilps0.680.450.660.490.170.21−191.87
19distributed_effort_p10_t2000.680.250.630.200.070.17−203.26
20abs-th-ratio-ilps0.650.420.640.450.150.20−199.19
212018_stem_original_p50_t1000.620.260.590.200.050.18−205.55
222018_stem_original_p10_t1000.620.260.580.200.050.17−205.77
23distributed_effort_p10_t1000.600.230.560.180.040.16−207.39
24sheffield-baseline0.530.210.500.160.060.12−226.26
25baseline_bm25_t1000.480.180.460.140.030.11−229.23
26baseline_bm25_t3000.480.180.450.140.030.11−229.87
27baseline_bm25_t4000.480.180.450.140.030.11−229.55
28baseline_bm25_t2000.480.180.450.140.030.11−230.06
29baseline_bm25_t5000.480.180.450.140.030.11−230.00
30baseline_bm25_t10000.470.180.450.140.030.11−230.06
31baseline_bm25_t15000.470.180.450.140.030.11−230.06
Table A15. CLEF TAR 2019 lab. Epsilon-greedy multiarmed bandit. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
Table A15. CLEF TAR 2019 lab. Epsilon-greedy multiarmed bandit. The results are ordered decreasingly by recall. Runs that were not included in the official overview are named accordingly.
RunRecallRecall_at_kAUCMAPWSS_at_r95rfbuug_at_budget
12018_stem_original_p10_t5000.710.250.640.220.190.13−187.00
22018_stem_original_p10_t10000.700.250.640.220.190.13−185.29
3distributed_effort_p10_t5000.700.250.630.220.180.13−185.94
4distributed_effort_p10_t3000.700.250.630.220.170.13−187.65
5abs-hh-ratio-ilps0.700.460.670.500.210.16−173.35
62018_stem_original_p10_t3000.690.250.630.220.190.13−184.77
72018_stem_original_p50_t4000.690.250.640.230.180.13−184.77
8distributed_effort_p10_t10000.690.250.630.220.180.13−185.52
92018_stem_original_p10_t15000.690.250.630.220.190.13−186.23
102018_stem_original_p10_t4000.690.250.630.220.190.13−185.00
11distributed_effort_p10_t15000.680.250.630.220.190.13−184.10
12abs-th-ratio-ilps0.680.430.650.470.200.14−179.97
132018_stem_original_p10_t2000.680.250.620.220.190.13−188.13
142018_stem_original_p50_t2000.680.250.630.230.180.13−187.42
152018_stem_original_p50_t10000.680.240.630.220.160.13−186.52
162018_stem_original_p50_t3000.660.240.620.220.180.12−184.68
172018_stem_original_p50_t5000.660.230.620.210.160.12−185.87
18distributed_effort_p10_t2000.660.240.590.200.140.12−187.87
19distributed_effort_p10_t4000.650.240.600.210.160.12−188.42
202018_stem_original_p50_t15000.650.220.610.210.140.13−186.52
212018_stem_original_p10_t1000.620.240.570.200.180.12−190.77
22sheffield-baseline0.610.200.540.180.140.10−205.58
232018_stem_original_p50_t1000.610.230.570.200.170.13−192.81
24distributed_effort_p10_t1000.590.220.530.180.110.11−191.06
25baseline_bm25_t1000.520.180.460.160.120.08−207.29
26baseline_bm25_t10000.520.180.460.150.110.08−210.45
27baseline_bm25_t5000.520.180.450.150.120.08−210.58
28baseline_bm25_t2000.510.180.450.150.120.08−206.77
29baseline_bm25_t15000.510.180.450.150.120.08−207.97
30baseline_bm25_t3000.510.180.450.150.120.08−208.65
31baseline_bm25_t4000.510.180.450.150.120.08−207.00

References

  1. Needleman, I.G. A guide to systematic reviews. J. Clin. Periodontol. 2002, 29, 6–9. [Google Scholar] [CrossRef] [PubMed]
  2. Timsina, P.; Liu, J.; El-Gayar, O. Advanced analytics for the automation of medical systematic reviews. Inf. Syst. Front. 2016, 18, 237–252. [Google Scholar] [CrossRef]
  3. Di Nunzio, G.M. Technology Assisted Review Systems: Current and Future Directions. In Proceedings of the 3rd Workshop on Augmented Intelligence for Technology-Assisted Reviews Systems (ALTARS 2024), Glasgow, UK, 28 March 2024; Di Nunzio, G.M., Kanoulas, E., Majumder, P., Eds.; CEUR Workshop Proceedings. ACM: New York, NY, USA, 2024; Volume 3832, ISSN 1613-0073. [Google Scholar]
  4. Cormack, G.V.; Grossman, M.R. Engineering Quality and Reliability in Technology-Assisted Review. In Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, New York, NY, USA, 17–21 July 2016; SIGIR ’16. pp. 75–84. [Google Scholar] [CrossRef]
  5. Manning, C.; Schutze, H. Foundations of Statistical Natural Language Processing; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  6. Jørgensen, L.; Paludan-Müller, A.S.; Laursen, D.R.T.; Savović, J.; Boutron, I.; Sterne, J.A.C.; Higgins, J.P.T.; Hróbjartsson, A. Evaluation of the Cochrane tool for assessing risk of bias in randomized clinical trials: Overview of published comments and analysis of user practice in Cochrane and non-Cochrane reviews. Syst. Rev. 2016, 5, 80. [Google Scholar] [CrossRef] [PubMed]
  7. Koffel, J.B. Use of Recommended Search Strategies in Systematic Reviews and the Impact of Librarian Involvement: A Cross-Sectional Survey of Recent Authors. PLoS ONE 2015, 10, e0125931. [Google Scholar] [CrossRef] [PubMed]
  8. Elliott, J.H.; Synnot, A.; Turner, T.; Simmonds, M.; Akl, E.A.; McDonald, S.; Salanti, G.; Meerpohl, J.; MacLehose, H.; Hilton, J.; et al. Living systematic review: 1. Introduction—The why, what, when, and how. J. Clin. Epidemiol. 2017, 91, 23–30. [Google Scholar] [CrossRef] [PubMed]
  9. Yang, E.; Lewis, D.D.; Frieder, O. On minimizing cost in legal document review workflows. In Proceedings of the 21st ACM Symposium on Document Engineering, New York, NY, USA, 24–27 August 2021; DocEng ’21. pp. 1–10. [Google Scholar] [CrossRef]
  10. Kelly, L.; Suominen, H.; Goeuriot, L.; Neves, M.; Kanoulas, E.; Li, D.; Azzopardi, L.; Spijker, R.; Zuccon, G.; Scells, H.; et al. Overview of the CLEF eHealth Evaluation Lab 2019. In Proceedings of the Experimental IR Meets Multilinguality, Multimodality, and Interaction, Lugano, Switzerland, 9–12 September 2019; Crestani, F., Braschler, M., Savoy, J., Rauber, A., Müller, H., Losada, D.E., Heinatz Bürki, G., Cappellato, L., Ferro, N., Eds.; Springer: Cham, Switzerland, 2019; pp. 322–339. [Google Scholar] [CrossRef]
  11. Suominen, H.; Kelly, L.; Goeuriot, L.; Névéol, A.; Ramadier, L.; Robert, A.; Kanoulas, E.; Spijker, R.; Azzopardi, L.; Li, D.; et al. Overview of the CLEF eHealth Evaluation Lab 2018. In Proceedings of the Experimental IR Meets Multilinguality, Multimodality, and Interaction, Avignon, France, 10–14 September 2018; Bellot, P., Trabelsi, C., Mothe, J., Murtagh, F., Nie, J.Y., Soulier, L., SanJuan, E., Cappellato, L., Ferro, N., Eds.; Springer: Cham, Switzerland, 2018; pp. 286–301. [Google Scholar] [CrossRef]
  12. Goeuriot, L.; Kelly, L.; Suominen, H.; Névéol, A.; Robert, A.; Kanoulas, E.; Spijker, R.; Palotti, J.; Zuccon, G. CLEF 2017 eHealth Evaluation Lab Overview. In Proceedings of the Experimental IR Meets Multilinguality, Multimodality, and Interaction, Dublin, Ireland, 11–14 September 2017; Jones, G.J., Lawless, S., Gonzalo, J., Kelly, L., Goeuriot, L., Mandl, T., Cappellato, L., Ferro, N., Eds.; Springer: Cham, Switzerland, 2017; pp. 291–303. [Google Scholar] [CrossRef]
  13. Voorhees, E.M. On Building Fair and Reusable Test Collections using Bandit Techniques. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, New York, NY, USA, 22–26 October 2018; CIKM ’18. pp. 407–416. [Google Scholar] [CrossRef]
  14. Rahman, M.M.; Kutlu, M.; Lease, M. Constructing Test Collections using Multi-armed Bandits and Active Learning. In Proceedings of the The World Wide Web Conference, New York, NY, USA, 13–17 May 2019; WWW ’19. pp. 3158–3164. [Google Scholar] [CrossRef]
  15. Rodriguez-Diaz, P.; Killian, J.A.; Xu, L.; Suggala, A.S.; Taneja, A.; Tambe, M. Flexible budgets in restless bandits: A primal-dual algorithm for efficient budget allocation. In Proceedings of the Thirty–Seventh AAAI Conference on Artificial Intelligence and Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence and Thirteenth Symposium on Educational Advances in Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; AAAI’23/IAAI’23/EAAI’23. AAAI Press: Washington, DC, USA, 2023; Volume 37, pp. 12103–12111. [Google Scholar] [CrossRef]
  16. Li, M.; Zhang, J.; Alizadehsani, R.; Pławiak, P. A Multi-Channel Advertising Budget Allocation Using Reinforcement Learning and an Improved Differential Evolution Algorithm. IEEE Access 2024, 12, 100559–100580. [Google Scholar] [CrossRef]
  17. Suyal, H.; Singh, A. Multilabel classification using crowdsourcing under budget constraints. Knowl. Inf. Syst. 2024, 66, 841–877. [Google Scholar] [CrossRef]
  18. Xu, L.; Wang, L.; Xie, H.; Zhou, M. Contextual Bandit with Herding Effects: Algorithms and Recommendation Applications. In Proceedings of the PRICAI 2024: Trends in Artificial Intelligence, Kyoto, Japan, 19 November 2024; Hadfi, R., Anthony, P., Sharma, A., Ito, T., Bai, Q., Eds.; Springer: Singapore, 2025; pp. 132–144. [Google Scholar] [CrossRef]
  19. Stradiotti, L.; Perini, L.; Davis, J. Combining Active Learning and Learning to Reject for Anomaly Detection. In ECAI 2024; IOS Press: Amsterdam, The Netherlands, 2024; pp. 2266–2273. [Google Scholar] [CrossRef]
  20. Yang, Y.; Zhang, J.; Qin, R.; Li, J.; Wang, F.Y.; Qi, W. A Budget Optimization Framework for Search Advertisements Across Markets. IEEE Trans. Syst. Man-Cybern.-Part Syst. Humans 2012, 42, 1141–1151. [Google Scholar] [CrossRef]
  21. Tricco, A.C.; Antony, J.; Zarin, W.; Strifler, L.; Ghassemi, M.; Ivory, J.; Perrier, L.; Hutton, B.; Moher, D.; Straus, S.E. A scoping review of rapid review methods. BMC Med. 2015, 13, 224. [Google Scholar] [CrossRef] [PubMed]
  22. O’Halloran, T.; McManus, B.; Harbison, A.; Grossman, M.R.; Cormack, G.V. Comparison of Tools and Methods for Technology-Assisted Review. In Proceedings of the Information Management, Suva, Fiji, 23–25 January 2024; Li, S., Ed.; Springer Nature: Cham, Switzerland, 2024; pp. 106–126. [Google Scholar] [CrossRef]
  23. Cormack, G.V.; Grossman, M.R. Evaluation of machine-learning protocols for technology-assisted review in electronic discovery. In Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval, New York, NY, USA, 6–11 July 2014; SIGIR ’14. pp. 153–162. [Google Scholar] [CrossRef]
  24. Wallace, B.C.; Dahabreh, I.J.; Schmid, C.H.; Lau, J.; Trikalinos, T.A. Modernizing the systematic review process to inform comparative effectiveness: Tools and methods. J. Comp. Eff. Res. 2013, 2, 273–282. [Google Scholar] [CrossRef] [PubMed]
  25. Kusa, W.; Lipani, A.; Knoth, P.; Hanbury, A. An analysis of work saved over sampling in the evaluation of automated citation screening in systematic literature reviews. Intell. Syst. Appl. 2023, 18, 200193. [Google Scholar] [CrossRef]
  26. Molinari, A.; Esuli, A.; Sebastiani, F. Improved risk minimization algorithms for technology-assisted review. Intell. Syst. Appl. 2023, 18, 200209. [Google Scholar] [CrossRef]
  27. Di Nunzio, G.M. A Study on a Stopping Strategy for Systematic Reviews Based on a Distributed Effort Approach. In Proceedings of the Experimental IR Meets Multilinguality, Multimodality, and Interaction, Thessaloniki, Greece, 22–25 September 2020; Arampatzis, A., Kanoulas, E., Tsikrika, T., Vrochidis, S., Joho, H., Lioma, C., Eickhoff, C., Névéol, A., Cappellato, L., Ferro, N., Eds.; Springer: Cham, Switzerland, 2020; pp. 112–123. [Google Scholar] [CrossRef]
  28. Lewis, D.D.; Yang, E.; Frieder, O. Certifying One-Phase Technology-Assisted Reviews. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, New York, NY, USA, 1–5 November 2021; CIKM ’21. pp. 893–902. [Google Scholar] [CrossRef]
  29. Névéol, A.; Robert, A.; Anderson, R.; Cohen, K.B.; Grouin, C.; Lavergne, T.; Rey, G.; Rondet, C.; Zweigenbaum, P. CLEF eHealth 2017 Multilingual Information Extraction task Overview: ICD10 Coding of Death Certificates in English and French. In Proceedings of the Working Notes of CLEF 2017—Conference and Labs of the Evaluation Forum, Dublin, Ireland, 11–14 September 2017; Cappellato, L., Ferro, N., Goeuriot, L., Mandl, T., Eds.; CEUR Workshop Proceedings. RWTH: Aachen, Germany, 2017; Volume 1866, ISSN 1613-0073. [Google Scholar]
  30. Kanoulas, E.; Li, D.; Azzopardi, L.; Spijker, R. CLEF 2018 Technologically Assisted Reviews in Empirical Medicine Overview. In Proceedings of the Working Notes of CLEF 2018—Conference and Labs of the Evaluation Forum, Avignon, France, 10–14 September 2018; Cappellato, L., Ferro, N., Nie, J.Y., Soulier, L., Eds.; CEUR Workshop Proceedings. RWTH: Aachen, Germany, 2018; Volume 2125, ISSN 1613-0073. [Google Scholar]
  31. Kanoulas, E.; Li, D.; Azzopardi, L.; Spijker, R. CLEF 2019 Technology Assisted Reviews in Empirical Medicine Overview. In Proceedings of the Working Notes of CLEF 2019—Conference and Labs of the Evaluation Forum, Lugano, Switzerland, 9–12 September 2019; Cappellato, L., Ferro, N., Losada, D.E., Müller, H., Eds.; CEUR Workshop Proceedings. RWTH: Aachen, Germany, 2019; Volume 2380, ISSN 1613-0073. [Google Scholar]
  32. Tran-Thanh, L.; Chapman, A.; de Cote, E.M.; Rogers, A.; Jennings, N.R. Epsilon–First Policies for Budget–Limited Multi-Armed Bandits. In Proceedings of the AAAI Conference on Artificial Intelligence, Atlanta, GA, USA, 11–15 July 2010; Volume 24, pp. 1211–1216. [Google Scholar] [CrossRef]
  33. Auer, P.; Cesa-Bianchi, N.; Fischer, P. Finite-Time Analysis of the Multiarmed Bandit Problem. Mach. Learn. 2002, 47, 235–256. [Google Scholar] [CrossRef]
  34. Li, L.; Chu, W.; Langford, J.; Schapire, R.E. A Contextual-Bandit Approach to Personalized News Article Recommendation. In Proceedings of the 19th International Conference on World Wide Web, New York, NY, USA, 26–30 April 2010; WWW ’10. pp. 661–670. [Google Scholar] [CrossRef]
  35. Di Nunzio, G. POLAR: Policy Optimization for Literature Analysis under Review Constraints. In Proceedings of the 34th ACM International Conference on Information and Knowledge Management, CIKM 2025, Seoul, Republic of Korea, 10–14 November 2025; ACM: New York, NY, USA, 2025. [Google Scholar] [CrossRef]
Figure 1. Boxplots of recall by budget allocation strategy across years.
Figure 1. Boxplots of recall by budget allocation strategy across years.
Make 07 00104 g001
Figure 2. Boxplots of RFCU by budget allocation strategy across years.
Figure 2. Boxplots of RFCU by budget allocation strategy across years.
Make 07 00104 g002
Figure 3. Boxplots of UG@B by budget allocation strategy across years.
Figure 3. Boxplots of UG@B by budget allocation strategy across years.
Make 07 00104 g003
Figure 4. Correlation between RFCU and recall across years.
Figure 4. Correlation between RFCU and recall across years.
Make 07 00104 g004
Figure 5. Correlation between UF@B and recall across years.
Figure 5. Correlation between UF@B and recall across years.
Make 07 00104 g005
Figure 6. UG@B for systems A and B under a fixed time budget of B = 100 min. System A reviews ( 33 ,   34 ) true and false positives, while system B reviews ( 30 ,   30 ) . The curves show the utility as a function of the gain/loss ratio r. The vertical dashed line marks the crossover point ( r 1.33 ), where both systems achieve an equal UG@B.
Figure 6. UG@B for systems A and B under a fixed time budget of B = 100 min. System A reviews ( 33 ,   34 ) true and false positives, while system B reviews ( 30 ,   30 ) . The curves show the utility as a function of the gain/loss ratio r. The vertical dashed line marks the crossover point ( r 1.33 ), where both systems achieve an equal UG@B.
Make 07 00104 g006
Table 1. Summary of CLEF eHealth TAR collections from 2017 to 2019.
Table 1. Summary of CLEF eHealth TAR collections from 2017 to 2019.
YearTopics (Test)DocumentsRelevant Documents
201730117,5621857
201830218,4963964
20193182,4211682
Table 2. CLEF TAR 2017 lab. The results shown are the ones recalculated in this experimental work, while the last three columns display the differences between our results and the official results.
Table 2. CLEF TAR 2017 lab. The results shown are the ones recalculated in this experimental work, while the last three columns display the differences between our results and the official results.
RunAUCMAPWSS@95AUC_diffMAP_diffWSS@95_diff
1amc.run0.750.130.320.010.000.01
2auth.simple.run10.930.300.700.000.00−0.00
3auth.simple.run20.920.290.700.00−0.00−0.00
4auth.simple.run30.920.290.670.00−0.000.00
5auth.simple.run40.920.290.690.000.00−0.00
6cnrs.abrupt.all0.730.140.240.010.000.01
7cnrs.gradual.all0.700.150.260.01−0.000.03
8cnrs.noaf.all0.770.140.350.010.000.01
9cnrs.noaffull.all0.830.180.490.00−0.000.00
10ecnu.run10.620.090.110.010.000.01
11ntu.run10.600.080.100.01−0.000.01
12ntu.run20.590.060.120.01−0.000.01
13ntu.run30.530.050.070.01−0.000.00
14padua.iafa_m10k150f0m100.890.280.500.000.000.01
15padua.iafap_m10p2f0m100.870.250.470.00−0.000.01
16padua.iafap_m10p5f0m100.880.270.480.00−0.000.01
17padua.iafas_m10k50f0m100.890.270.510.000.000.01
18qut.ca_bool_ltr0.730.110.270.01−0.000.02
19qut.ca_pico_ltr0.740.150.290.01−0.000.01
20qut.rf_bool_ltr0.700.110.240.010.000.03
21qut.rf_pico_ltr0.720.120.290.010.000.01
22sheffield.run10.810.170.420.00−0.00−0.00
23sheffield.run20.840.220.500.00−0.00−0.01
24sheffield.run30.840.200.480.000.00−0.00
25sheffield.run40.840.220.490.000.00−0.01
26ucl.run_abstract0.500.060.060.010.000.01
27ucl.run_fulltext0.510.050.060.010.000.02
28uos.sis.TMAL30Q_BM250.830.160.520.00−0.000.01
29uos.sis.TMBEST_BM250.720.120.320.01−0.000.00
30waterloo.A-rank-normal0.930.280.710.00−0.00−0.01
31waterloo.B-rank-normal0.930.320.720.00−0.00−0.02
Table 3. CLEF TAR 2018 lab. The results shown are the ones recalculated in this experimental work, together with the differences between our results and the official results. We also sought to show the different values of the official R@k (recall at the number of documents reviewed) and our recall@k (recall at the number of relevant documents).
Table 3. CLEF TAR 2018 lab. The results shown are the ones recalculated in this experimental work, together with the differences between our results and the official results. We also sought to show the different values of the official R@k (recall at the number of documents reviewed) and our recall@k (recall at the number of relevant documents).
RunAUCMAPWSS@95MAP_diffWSS@95_diffR@kRecall@k
1auth_run10.950.400.770.00−0.021.000.40
2auth_run20.950.400.770.00−0.020.940.40
3auth_run30.950.390.76−0.00−0.030.940.39
4cnrs_RF_uni0.890.310.52−0.00−0.001.000.35
5cnrs_RF_bi0.920.310.63−0.00−0.011.000.33
6cnrs_comb0.930.340.680.00−0.021.000.34
7ECNU_RUN10.660.140.020.000.010.520.19
8ECNU_RUN20.590.08−0.030.000.050.370.12
9ECNU_RUN30.680.150.020.000.010.530.20
10unipd_t5000.910.320.620.00−0.010.860.36
11unipd_t10000.900.320.58−0.00−0.000.920.36
12unipd_t15000.890.320.55−0.00−0.000.940.36
13shef-feed0.920.610.63−0.000.011.000.56
14shef-general0.870.260.55−0.00−0.001.000.28
15shef-query0.850.220.51−0.00−0.011.000.24
16uic_model80.730.170.22−0.000.040.510.20
17uic_model70.740.180.230.000.030.580.21
18UWA0.940.360.76−0.00−0.010.990.37
19UWB0.950.380.770.00−0.010.930.39
Table 4. CLEF TAR 2019 lab. The results shown are the ones recalculated in this experimental work. We do not show any differences compared with the official results since, in the original evaluation campaign, there was a split among the different typologies of reviews.
Table 4. CLEF TAR 2019 lab. The results shown are the ones recalculated in this experimental work. We do not show any differences compared with the official results since, in the original evaluation campaign, there was a split among the different typologies of reviews.
RunAUCMAPWSS@95Recall@k
12018_stem_original_p10_t1000.820.250.510.26
22018_stem_original_p10_t10000.860.250.600.26
32018_stem_original_p10_t15000.860.250.600.26
42018_stem_original_p10_t2000.860.250.590.26
52018_stem_original_p10_t3000.860.250.590.26
62018_stem_original_p10_t4000.860.250.600.26
72018_stem_original_p10_t5000.860.250.600.26
82018_stem_original_p50_t1000.830.250.530.26
92018_stem_original_p50_t10000.880.260.610.26
102018_stem_original_p50_t15000.880.260.610.26
112018_stem_original_p50_t2000.870.260.580.26
122018_stem_original_p50_t3000.870.260.590.26
132018_stem_original_p50_t4000.870.260.610.26
142018_stem_original_p50_t5000.870.260.610.26
15abs-hh-ratio-ilps0.850.530.480.46
16abs-th-ratio-ilps0.820.490.460.43
17baseline_bm25_t1000.770.180.400.18
18baseline_bm25_t10000.760.180.380.18
19baseline_bm25_t15000.760.180.390.18
20baseline_bm25_t2000.770.180.400.18
21baseline_bm25_t3000.760.180.390.18
22baseline_bm25_t4000.760.180.390.18
23baseline_bm25_t5000.760.180.390.18
24distributed_effort_p10_t1000.790.230.460.24
25distributed_effort_p10_t10000.860.250.590.26
26distributed_effort_p10_t15000.860.250.600.26
27distributed_effort_p10_t2000.850.250.550.26
28distributed_effort_p10_t3000.860.250.570.26
29distributed_effort_p10_t4000.860.250.570.26
30distributed_effort_p10_t5000.860.250.580.26
31sheffield-baseline0.800.210.400.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Di Nunzio, G.M. Screening Smarter, Not Harder: Budget Allocation Strategies for Technology-Assisted Reviews (TARs) in Empirical Medicine. Mach. Learn. Knowl. Extr. 2025, 7, 104. https://doi.org/10.3390/make7030104

AMA Style

Di Nunzio GM. Screening Smarter, Not Harder: Budget Allocation Strategies for Technology-Assisted Reviews (TARs) in Empirical Medicine. Machine Learning and Knowledge Extraction. 2025; 7(3):104. https://doi.org/10.3390/make7030104

Chicago/Turabian Style

Di Nunzio, Giorgio Maria. 2025. "Screening Smarter, Not Harder: Budget Allocation Strategies for Technology-Assisted Reviews (TARs) in Empirical Medicine" Machine Learning and Knowledge Extraction 7, no. 3: 104. https://doi.org/10.3390/make7030104

APA Style

Di Nunzio, G. M. (2025). Screening Smarter, Not Harder: Budget Allocation Strategies for Technology-Assisted Reviews (TARs) in Empirical Medicine. Machine Learning and Knowledge Extraction, 7(3), 104. https://doi.org/10.3390/make7030104

Article Metrics

Back to TopTop