Next Article in Journal
AI-Driven Analysis for Real-Time Detection of Unstained Microscopic Cell Culture Images
Previous Article in Journal
Bio-Inspired Optimization of Transfer Learning Models for Diabetic Macular Edema Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reducing Annotation Effort in Semantic Segmentation Through Conformal Risk Controlled Active Learning

1
Department of Computer Engineering, Istanbul Technical University, Sariyer, Istanbul 34467, Türkiye
2
Department of Artificial Intelligence and Data Engineering, Istanbul Technical University, Sariyer, Istanbul 34467, Türkiye
*
Author to whom correspondence should be addressed.
AI 2025, 6(10), 270; https://doi.org/10.3390/ai6100270
Submission received: 4 September 2025 / Revised: 3 October 2025 / Accepted: 14 October 2025 / Published: 18 October 2025
(This article belongs to the Section AI Systems: Theory and Applications)

Abstract

Modern semantic segmentation models require extensive pixel-level annotations, creating a significant barrier to practical deployment as labeling a single image can take hours of human effort. Active learning offers a promising way to reduce annotation costs through intelligent sample selection. However, existing methods rely on poorly calibrated confidence estimates, making uncertainty quantification unreliable. We introduce Conformal Risk Controlled Active Learning (CRC-AL), a novel framework that provides statistical guarantees on uncertainty quantification for semantic segmentation, in contrast to heuristic approaches. CRC-AL calibrates class-specific thresholds via conformal risk control, transforming softmax outputs into multi-class prediction sets with formal guarantees. From these sets, our approach derives complementary uncertainty representations: risk maps highlighting uncertain regions and class co-occurrence embeddings capturing semantic confusions. A physics-inspired selection algorithm leverages these representations with a barycenter-based distance metric that balances uncertainty and diversity. Experiments on Cityscapes and PascalVOC2012 show CRC-AL consistently outperforms baseline methods, achieving 95% of fully supervised performance with only 30% of labeled data, making semantic segmentation more practical under limited annotation budgets.

Graphical Abstract

1. Introduction

Semantic segmentation, the task of assigning a semantic label to every pixel in an image, is fundamental to numerous computer vision applications, including autonomous driving [1], medical imaging [2], robotics [3], and aerial imaging [4]. Unlike image classification, which produces a single label per image, semantic segmentation requires dense, pixel-level predictions that capture fine-grained boundaries between objects and regions. This granularity is essential for applications requiring detailed scene understanding.
The remarkable success of deep learning in semantic segmentation comes at a significant cost: the need for vast amounts of meticulously labeled training data. Creating pixel-level annotations is extraordinarily labor-intensive compared to other vision tasks. For instance, annotating a single image from the Cityscapes dataset requires an average of 1.5 h of human effort [5], as each image contains urban street scenes with up to 30 object categories. This annotation bottleneck severely limits the scalability of supervised learning approaches and motivates methods that achieve high performance with minimal labeled data.
Active learning [6,7,8,9] offers a principled solution by strategically selecting which samples to annotate rather than relying on random sampling. The core insight is that not all training samples contribute equally to model performance. By iteratively selecting the most informative samples from a pool of unlabeled data, active learning can significantly reduce annotation requirements while maintaining model performance. The key challenge is designing an effective querying strategy that selects samples maximizing model improvement while minimizing annotation cost. The effectiveness of a querying strategy hinges on how it defines and measures sample informativeness.
Most strategies define informativeness through uncertainty estimates derived from model predictions, which makes the reliability of these estimates central to the effectiveness of active learning. A fundamental limitation of existing uncertainty estimation methods is their reliance on heuristic measures, such as those derived from softmax probabilities [10,11], or gradient magnitudes computed using pseudo-labels [12]. Despite their widespread use, these approaches fail to address the well-documented calibration problem in modern neural networks [13,14], where models tend to produce overconfident probability estimates even when their predictions are incorrect. This miscalibration is particularly severe in semantic segmentation for two reasons. First, the dense prediction setting requires thousands of per-pixel classifications for each image, with each classification being vulnerable to overconfidence. Second, severe class imbalance makes models systematically overconfident on frequent classes and underconfident on rare ones.
To address these fundamental calibration issues, conformal prediction [15,16,17] has been proposed as a theoretically grounded framework for uncertainty quantification with formal coverage guarantees. Unlike traditional approaches that output point estimates, conformal prediction constructs prediction sets that contain the true label with a specified confidence level. Conformal risk control [18] extends this framework beyond coverage by enabling guarantees on more general loss functions, thereby supporting complex outputs such as semantic segmentation. By calibrating the predictor on a held-out dataset, these approaches provide statistical guarantees independent of the underlying model architecture or data distribution. These properties make conformal prediction a strong foundation for developing reliable active learning strategies.
Despite these advances, only a handful of works [19,20,21] have applied conformal prediction to active learning, all focused exclusively on classification tasks. These techniques do not directly translate to the dense, pixel-level setting of semantic segmentation. A critical gap therefore remains: no existing method applies conformal prediction to active learning for semantic segmentation. The dense prediction nature of segmentation introduces unique challenges that classification-focused methods cannot address, including managing complex, interdependent pixel-level outputs and handling severe class imbalance. Our work fills this gap by introducing the first conformal prediction-based active learning framework specifically designed for semantic segmentation.
In this paper, we introduce Conformal Risk Controlled Active Learning (CRC-AL), the first active learning framework for semantic segmentation that provides statistical guarantees for uncertainty quantification through conformal prediction. Our key contribution is that conformal risk control, when applied independently to each semantic class, yields pixel-wise multi-class prediction sets that capture model uncertainty in a principled manner. Unlike heuristic uncertainty measures, these prediction sets provide statistical guarantees with calibrated, class-specific confidence. By calibrating classes independently at the same risk level, CRC-AL achieves more balanced uncertainty quantification across both frequent and rare classes, thus addressing intra-image class imbalance. Pixels with larger prediction sets correspond to regions of high uncertainty, where the model cannot confidently distinguish among classes.
CRC-AL transforms these prediction sets into two complementary representations for guiding sample selection. First, risk maps highlight uncertain image regions by identifying pixels where multiple classes remain plausible. Second, co-occurrence embeddings capture class confusion patterns, revealing which semantic categories the model struggles to distinguish. Together, these representations enable the identification of not only highly uncertain samples but also samples exhibiting diverse forms of uncertainty. To select informative yet diverse samples, we introduce the Top-Diverse-K algorithm, an extension of standard Top-K selection to high-dimensional space. Inspired by center-of-mass formulations in physics, it employs a barycenter-based distance metric that balances uncertainty weighting with spatial distribution in embedding space.
We validate our approach on Cityscapes [5] and PascalVOC2012 [22,23], where it consistently outperforms benchmark methods. CRC-AL achieves 95% of fully supervised performance using only 30% of the training data on both datasets, substantially reducing annotation requirements.
The main contributions of this work are as follows:
  • We introduce the first active learning framework for semantic segmentation based on conformal prediction, providing principled uncertainty quantification with statistical guarantees.
  • We propose a novel uncertainty representation that combines risk maps and class co-occurrence embeddings to capture both spatial and semantic uncertainty patterns.
  • We develop a physics-inspired selection algorithm that balances sample informativeness with diversity through a barycenter-based distance metric.
  • We provide comprehensive experimental validation, demonstrating significant improvements over benchmark methods across two fundamentally different datasets, and release our implementation to support future research.
The remainder of this paper is organized as follows. Section 2 reviews the related works. Section 3 introduces our CRC-AL framework. Section 4 describes experiments and benchmarking results, while Section 5 presents parameter sensitivity analysis. Section 6 provides an in-depth discussion of the findings, and Section 7 concludes the paper.

2. Related Works

2.1. Active Learning

Active learning methods are broadly categorized into uncertainty-based, diversity-based, and hybrid approaches. Uncertainty-based methods typically select samples using metrics such as entropy [24], margin [25], or least confidence [26]. While effective in classical settings with single-sample queries, deep neural networks require batch-mode selection, for which [10] proposed Top-K sampling. However, these strategies remain vulnerable to softmax overconfidence. Alternatives such as Monte Carlo dropout [27] and ensembles [28,29] attempt to mitigate this via predictive variance estimation, but these methods require dozens of forward passes per image, which is computationally prohibitive for high-resolution segmentation.
A second limitation of uncertainty-only methods is redundant sampling from similar high-uncertainty regions [30]. Diversity-based approaches attempt to mitigate this issue by ensuring that selected samples better represent the overall data distribution. CoreSet [31] formulates the objective as a K-Center clustering problem in the latent embedding space, while CDAL [32] introduces information-theoretic contextual diversity measures. However, these strategies often prioritize easy or uninformative samples that contribute little to model improvement.
Hybrid strategies attempt to balance both criteria. Ref. [33] proposed stochastic batch selection over groups of uncertain samples, with uncertainty computed using classical metrics. BADGE [12] employs gradient embeddings based on pseudo-labels, encoding uncertainty through gradient magnitudes while enforcing diversity via K-Means++ seeding. Nevertheless, these methods still rely on heuristic uncertainty measures and fail to address the underlying overconfidence problem.
In contrast, CRC-AL leverages conformal risk control to provide calibrated, class-wise uncertainty quantification with formal guarantees. Once calibrated, selection requires only a single forward pass per image, avoiding the high computational cost of ensemble- or dropout-based methods. The proposed Top-Diverse-K strategy further balances calibrated uncertainty with diversity, enabling more efficient and informative batch selection.

2.2. Active Learning for Semantic Segmentation

Active learning for semantic segmentation poses unique challenges due to its dense prediction nature and high annotation cost. Image-level methods query entire images for annotation [12,31,32,33,34,35,36], while region-level methods aim to improve efficiency by querying only parts of an image. Region-level methods consist of two groups. Patch-based approaches [11,37] divide images into fixed-size regions, whereas superpixel-based approaches [38,39] query irregular regions that better align with object boundaries. Although region-level selection can, in principle, reduce annotation costs, it introduces practical challenges such as interface complexity, boundary artifacts from partial labeling, and the need for specialized training procedures to handle incomplete annotations.
Despite extensive study, no consensus exists on which granularity is more cost-effective [40], partly due to differing annotation protocols and the lack of standardized cost metrics across studies. In this work, we adopt the standard image-level strategy for several reasons: (1) compatibility with conformal risk control, which requires complete class channels for calibration, (2) alignment with standard annotation workflows where full-image labels are the norm, and (3) avoiding the practical challenges of partial labeling such as boundary artifacts and specialized training procedures.

2.3. Active Learning with Conformal Prediction

Early work in this area explored transductive conformal prediction for classification tasks before the advent of deep learning [41,42,43]. Although effective for small-scale problems with simple classifiers, these methods required retraining for each unlabeled sample, making them computationally infeasible for deep neural networks.
The introduction of inductive conformal prediction with calibration sets enabled more practical implementations [15]. In this framework, the model is trained once on a training set; then, non-conformity scores are computed using a separate calibration set to construct prediction sets for unlabeled samples. ICP [20] applies this framework to pool-based active learning by using two key conformal metrics: credibility and confidence. ICP-CNN [19] extends this approach to convolutional networks, combining three selection criteria: informativeness, diversity, and information density. This multi-criteria approach balances uncertainty with representativeness. CPAL-LR [21] further refines sample relevance assessment on conformal scores. However, these methods are largely confined to classification tasks, leaving their adaptation to dense prediction problems such as semantic segmentation underexplored.
The dense prediction nature of semantic segmentation introduces unique challenges that traditional conformal prediction methods cannot address, particularly due to the strong correlations among pixels within an image. In this setting, the relevant notion of error is not simply miscoverage [17]. Conformal risk control addresses this by providing guarantees on expected loss rather than coverage, bounding errors at the image level (e.g., false negative rates) and producing outer confidence sets that exploit spatial structure. Yet no prior work has applied this framework to active learning. This gap motivates CRC-AL, which integrates conformal risk control into active learning for semantic segmentation.

3. Methodology

3.1. Problem Statement

Let D = { ( x i , y i ) } i = 1 N denote the complete dataset of N images, where x i R 3 × H × W is an RGB image with height H and width W, and  y i { 0 , 1 } K × H × W is the corresponding one-hot-encoded ground-truth segmentation mask with K classes. At any point in the active learning process, we partition D into the labeled and unlabeled subsets D L and D U , where D L contains image–label pairs { ( x , y ) } with revealed labels, and  D U contains images { x } whose labels remain hidden. The semantic segmentation task aims to learn a function Ψ : R 3 × H × W [ 0 , 1 ] K × H × W that maps each input image to per-pixel softmax probabilities. The model Ψ is trained by minimizing the pixel-wise cross-entropy loss over the labeled data.
The goal of active learning is to iteratively select the most informative batches to maximize model performance while minimizing annotation cost. At each iteration, a batch B D U is selected according to a querying strategy. This process is summarized in Algorithm 1. Initially, a learner model Ψ 0 is trained on the initial labeled dataset D 0 L . At iteration t, the current model Ψ t is used to select a batch B t D t U for annotation. After the oracle provides labels for images in B t , transforming them into labeled pairs B t * = { ( x , y ) : x B t } , the datasets are updated: D t + 1 L = D t L B t * and D t + 1 U = D t U B t . The model Ψ t + 1 is then retrained on the expanded labeled dataset D t + 1 L . This process continues until the annotation budget is exhausted or performance converges. The central challenge lies in developing a querying strategy that enhances model performance while keeping annotation costs low.
Algorithm 1 Active learning algorithm for model improvement via strategic batch selection and retraining
Input: 
Initial labeled set D 0 L , initial unlabeled set D 0 U
Output: 
Trained segmentation model Ψ  
  1:
Train initial model Ψ 0 on D 0 L
  2:
for  t = 0 , 1 , 2 , do
  3:
       Select batch B t D t U according to a querying strategy▹ (e.g., CRC-AL: Figure 1)
  4:
       Obtain labels from oracle: B t * = { ( x , y ) : x B t }
  5:
        D t + 1 L D t L B t *
  6:
        D t + 1 U D t U B t
  7:
       Retrain model Ψ t + 1 on D t + 1 L
  8:
       if annotation budget is exhausted or performance converges then
  9:
             break
10:
       end if
11:
end for
12:
return final model Ψ t + 1
To address this, we introduce Conformal Risk Controlled Active Learning (CRC-AL), a novel active learning framework that enhances the querying strategy through principled uncertainty quantification, as illustrated in Figure 1. Our method consists of three integrated stages. First, we apply conformal risk control to calibrate class-specific thresholds that transform softmax outputs into binary prediction masks, allowing multiple class predictions per pixel. Second, these masks generate both image-level uncertainty scores through risk maps and embeddings from co-occurrence patterns, capturing class confusion structure. Finally, our Top-Diverse-K algorithm selects samples that balance high uncertainty with spatial diversity in the embedding space, incorporating previously labeled samples to ensure good coverage across iterations.
Figure 1. Overview of the proposed CRC-AL method. (a) Conformal risk control calibration with risk level α determines class-specific thresholds, which transform softmax probabilities into multi-class predictions (Section 3.2). (b) Uncertainty quantification through risk map generation and embedding construction from prediction co-occurrence patterns, capturing both image-level uncertainty and class confusion structure (Section 3.3). (c) Sample selection via the Top-Diverse-K algorithm, which jointly optimizes for high uncertainty and spatial diversity in the embedding space (Section 3.4).
Figure 1. Overview of the proposed CRC-AL method. (a) Conformal risk control calibration with risk level α determines class-specific thresholds, which transform softmax probabilities into multi-class predictions (Section 3.2). (b) Uncertainty quantification through risk map generation and embedding construction from prediction co-occurrence patterns, capturing both image-level uncertainty and class confusion structure (Section 3.3). (c) Sample selection via the Top-Diverse-K algorithm, which jointly optimizes for high uncertainty and spatial diversity in the embedding space (Section 3.4).
Ai 06 00270 g001

3.2. Class-Wise Calibration

We formulate the multi-class semantic segmentation problem as K independent binary segmentation tasks, each distinguishing class k from all others. For each class k { 1 , , K } , we define the binary segmentation output θ k : R 3 × H × W [ 0 , 1 ] H × W by extracting the k-th channel from the multi-class network, given by θ k ( x ) h , w = Ψ ( x ) k , h , w , where Ψ ( x ) k , h , w denotes the probability that pixel ( h , w ) belongs to class k. Thus, θ k ( x ) [ 0 , 1 ] H × W represents the complete pixel-wise probability map for class k. The predicted binary mask M λ ( k ) ( x ) { 0 , 1 } H × W is obtained by thresholding these probabilities with parameter λ [ 0 , 1 ] :
M λ ( k ) ( x ) h , w = 1 , if θ k ( x ) h , w 1 λ , 0 , otherwise .
Pixels with scores above 1 λ are classified as belonging to class k. As  λ increases, the threshold 1 λ decreases, resulting in more pixels being classified as positive and thus expanding the predicted mask M λ ( k ) ( x ) .
Let D cal = { ( x , y ) } denote the calibration dataset, where x R 3 × H × W is an RGB image and y { 0 , 1 } K × H × W is its one-hot-encoded ground-truth mask with K classes. For each class k, we define the binary ground-truth mask y ( k ) { 0 , 1 } H × W by y h , w ( k ) = y k , h , w . Each element of y ( k ) equals 1 if pixel ( h , w ) belongs to class k and 0 otherwise. It is important to note that D cal is kept separate from the complete dataset D , and the choice of calibration set is further examined in Section 5.3.
In calibration, we determine a threshold λ k for each class such that the expected loss over the calibration set does not exceed a predefined risk level α ( 0 , 1 ) :
E ( x , y ) D cal L y ( k ) , M λ k ( k ) ( x ) α ,
where L : { 0 , 1 } H × W × { 0 , 1 } H × W [ 0 , 1 ] is a loss function that is non-increasing in λ . We adopt the false negative rate (FNR) as the loss function:
L FNR y ( k ) , M λ ( k ) ( x ) = 1 h , w M λ ( k ) ( x ) h , w · y h , w ( k ) h , w y h , w ( k ) ,
where the numerator counts the number of true positive pixels, and the denominator counts the total number of positive pixels in the ground truth.
To satisfy the risk constraint with finite-sample correction, we compute the optimal threshold λ ^ k as
λ ^ k = inf λ [ 0 , 1 ] : R ^ k ( λ ) α 1 α | D cal ( k ) | ,
where D cal ( k ) = { ( x , y ) D cal : h , w y h , w ( k ) > 0 } is the set of calibration images containing at least one pixel of class k, and the empirical risk is
R ^ k ( λ ) = 1 | D cal ( k ) | ( x , y ) D cal ( k ) L FNR y ( k ) , M λ ( k ) ( x ) .
The correction term ( 1 α ) / | D cal ( k ) | ensures, with high probability, that the empirical risk on the calibration set does not exceed the target level α [17]. Figure 2 illustrates conformal risk control in practice.
Calibrating a single threshold across all classes yields suboptimal performance due to class imbalance—frequent classes (e.g., Road) dominate calibration, while rare classes (e.g., Bus, Train) may be poorly calibrated. By calibrating each class independently at the same risk level α , we ensure balanced performance across all classes regardless of frequency.
As illustrated in Figure 1a, we first determine the optimal thresholds λ ^ k for all classes k { 1 , , K } using the class-wise calibration procedure described above, ensuring that each class satisfies the target risk level α . Given these thresholds, we generate multi-class predictions P ( x ) { 0 , 1 } K × H × W for each image by stacking the binary prediction masks, defined as P ( x ) k , h , w = M λ ^ k ( k ) ( x ) h , w . At each pixel location ( h , w ) , the prediction vector may contain multiple positive entries, indicating that several classes may simultaneously exceed their respective thresholds. Unlike standard argmax segmentation, these calibrated masks may overlap, reflecting conformal prediction sets that allow multiple plausible classes per pixel.
It is important to note that while each class-specific prediction mask M λ ^ k ( k ) ( x ) satisfies the conformal risk control guarantee independently, the stacked prediction masks P ( x ) do not provide joint statistical guarantees across all classes simultaneously. For active learning, however, we use these prediction sets primarily as indicators of uncertainty rather than for their formal coverage properties. Calibrating each class independently with the same risk level α helps ensure more balanced uncertainty quantification across frequent and rare classes. Pixels with larger prediction set sizes (i.e., k = 1 K P ( x ) k , h , w > 1 ) indicate regions of high model uncertainty where the model cannot confidently distinguish among classes (Figure 2). This observation forms the foundation of our image-level uncertainty quantification strategy, as detailed in the following section.

3.3. Uncertainty Quantification

Not all classes contribute equally to image-level uncertainty within an image. For instance, pixels predicted as Bicycle often co-occur with Rider predictions, while Road predictions rarely co-occur with other classes. Here, co-occurrence refers to pixels ( h , w ) satisfying k = 1 K P ( x ) k , h , w > 1 , where the conformal predictor includes multiple classes in its prediction sets.
Figure 1b illustrates the process of computing uncertainty scores and embeddings. We first construct the co-occurrence matrix C ( x ) N 0 K × K by counting pixel-wise class co-occurrences:
C ( x ) k , k ^ = h , w P ( x ) k , h , w · P ( x ) k ^ , h , w ,
where k , k ^ { 1 , , K } . The matrix C ( x ) is symmetric, and each entry C ( x ) k , k ^ counts the number of pixels where classes k and k ^ are simultaneously predicted. In particular, the diagonal elements C ( x ) k , k count the number of pixels where class k is predicted, regardless of whether other classes are also predicted at those pixels.
To account for class imbalance within images, we normalize the co-occurrence matrix to obtain the co-occurrence density matrix D ( x ) [ 0 , 1 ] K × K :
D ( x ) k , k ^ = C ( x ) k , k ^ = 1 K C ( x ) k , , if = 1 K C ( x ) k , > 0 , 0 , otherwise .
The k-th row of D ( x ) forms a categorical probability distribution over the class set { 1 , , K } , representing the relative likelihood that class k co-occurs with each other class. Diagonal entries D ( x ) k , k represent the self-association strength of class k, equaling 1 when class k never co-occurs with other classes and decreasing as co-occurrences increase.
When the conformal predictor assigns multiple classes to a pixel location, indicating higher uncertainty, the resulting distributions become less peaked. We quantify this spread using Shannon entropy to compute class-wise confusion weights w c ( x ) k [ 0 , 1 ] :
w c ( x ) k = 1 log K k ^ = 1 K D ( x ) k , k ^ log D ( x ) k , k ^ , if = 1 K C ( x ) k , > 0 , 0 , otherwise .
with the convention 0 log 0 = 0 . Higher values of w c ( x ) k indicate greater uncertainty for class k. In particular, w c ( x ) k = 0 when class k either does not appear in the image or appears without any co-occurrences, and increases as co-occurrences become more diverse.
The risk map R ( x ) [ 0 , 1 ] H × W aggregates class-wise uncertainty at each pixel using these weights:
R ( x ) h , w = 1 K k = 1 K P ( x ) k , h , w · w c ( x ) k .
This produces a spatial uncertainty map highlighting regions where the model is most uncertain, as shown in Figure 3. Pixels with multiple predicted classes and those containing classes with high confusion weights contribute more to the risk map, effectively capturing both prediction ambiguity and class-specific uncertainty.
The image-level uncertainty score u ( x ) [ 0 , 1 ] is the spatial average of the risk map:
u ( x ) = 1 H W h , w R ( x ) h , w .
This aggregation provides a single interpretable score that summarizes the overall uncertainty across the entire image, enabling direct comparison and ranking of unlabeled samples.
When the co-occurrence density matrix D ( x ) is diagonal (no co-occurrences), the uncertainty score u ( x ) = 0 , indicating high model confidence. Such images require no further annotation and can be excluded from selection. However, two images may have similar uncertainty scores while exhibiting different class confusion patterns. To capture these differences, we construct image embeddings v ( x ) [ 0 , 1 ] K 2 by flattening D ( x ) :
v ( x ) = D ( x ) 1 , 1 , , D ( x ) 1 , K , D ( x ) 2 , 1 , , D ( x ) K , K .
These embeddings represent images in a K 2 -dimensional space based on their class confusion structure, enabling diversity-aware sample selection.

3.4. Sample Selection

A common strategy in uncertainty-based active learning is Top-K selection, which iteratively selects the most uncertain samples [10,36]:
x * = arg max x D U B u ( x ) ,
where u ( x ) is defined in Equation (10), D U denotes the set of unlabeled images and B is the batch of already selected images (initially B = ). After each selection, we update B B { x * } and repeat until | B | = q , where q { 1 , , | D U | } is the desired batch size. If multiple images attain the same maximum uncertainty score, the  arg max operator returns one of them arbitrarily.
While computationally efficient, this approach suffers from redundant sampling, particularly in datasets with limited diversity. Similar samples may all receive high uncertainty scores, leading to inefficient use of the annotation budget [30]. Conversely, pure diversity-based selection often queries easy, non-informative samples that contribute minimally to model improvement. To address these limitations, we extend Top-K selection to operate in the embedding space, balancing uncertainty and diversity through a principled distance metric.
To realize this idea, we adopt a physics-inspired formulation: each embedding v i = v ( x i ) represents a particle with mass u i = u ( x i ) , where v ( x ) and u ( x ) are given in Equations (11) and (10). Drawing from the two-body problem in classical mechanics [44], the distance from particle i to the barycenter (center of mass) of the two-particle system ( i , j ) is given by v i v j 2 · u j u i + u j . This distance is asymmetric except when u i = u j . Specifically, when particle i has higher uncertainty than particle j (i.e., u i > u j ), the distance from i to the barycenter is smaller than when u i < u j .
Inspired by this physical analogy, we define a distance metric that quantifies the informativeness of sample j relative to sample i:
d i j = v i v j 2 τ · u j u i + u j + ϵ 1 τ ,
where τ [ 0 , 1 ] is a trade-off parameter and ϵ > 0 is a small positive constant to prevent division by zero. This formulation yields a weighted geometric mean between spatial distance and relative uncertainty. When τ = 0 , the metric reduces to pure uncertainty weighting, disregarding spatial distance. Conversely, when τ = 1 , only spatial distance matters. The metric d i j increases with both the spatial separation between samples and the relative uncertainty of sample j. Figure 4 provides a practical example of this metric.
Building on this distance metric, Algorithm 2 introduces the Top-Diverse-K selection algorithm, which extends traditional Top-K selection by incorporating spatial diversity through the embedding space. The algorithm iteratively selects samples that maximize a combined distance criterion balancing average distance and minimum distance to previously selected samples. Starting with the images from the labeled set D L marked as traversed, the algorithm selects each subsequent sample according to
x * = arg max x j D U B 1 | T | x i T d i j + min x i T d i j ,
where d i j is defined in Equation (13) and T denotes the set of traversed samples (initially T = { x : ( x , y ) D L } and B = ). After each selection, we update B B { x * } and T T { x * } . The process continues until | B | = q . The first term promotes samples that are, on average, distant from the traversed set, encouraging diversity. The second term ensures a minimum separation from all traversed samples, preventing clustering around high-density regions. This dual criterion prevents scenarios where outliers dominate the average distance while ensuring selected samples span the embedding space effectively.
Algorithm 2 Top-Diverse-K sample selection algorithm
Input: 
Labeled set D L , unlabeled set D U , batch size q
Output: 
Selected batch B  
  1:
Initialize traversed set: T { x : ( x , y ) D L }
  2:
Initialize selected batch: B
  3:
while  | B | < q   do
  4:
       x * arg max x j D U B 1 | T | x i T d i j + min x i T d i j                     ▹ (Equation 14)
  5:
       T T { x * }
  6:
       B B { x * }
  7:
end while
  8:
return selected batch B
In the first iteration, when T initially contains the images from D L , the algorithm evaluates the combined distance metric for all samples in D U relative to the samples in D L . It selects the sample x * D U that maximizes this metric, where larger values of d i j indicate greater informativeness. Once selected, x * is added to both the traversed set T and the selected batch B . By updating T T { x * } , we treat x * as effectively labeled for subsequent selections. This ensures that future samples are chosen with respect to both the original labeled set and previously selected samples, maintaining a well-distributed selection across the embedding space. In the second iteration, the algorithm selects from the remaining unlabeled samples D U B (where now B = { x * } ), evaluating distances to all samples in the updated traversed set T . This iterative process continues, with each selected sample influencing subsequent selections to achieve both high uncertainty and spatial diversity.
The parameter τ [ 0 , 1 ] controls the trade-off between uncertainty and diversity in sample selection. When τ = 0 , the algorithm reduces to standard Top-K selection based solely on uncertainty, as the spatial term v i v j 2 τ = 1 becomes constant. Conversely, as τ approaches 1, spatial diversity increasingly dominates the selection criterion.
Figure 5 illustrates this behavior empirically. For small τ values, selected samples cluster in high-uncertainty regions of the embedding space. As τ increases, the selected batch becomes more spatially distributed while still favoring uncertain samples, achieving an effective balance between uncertainty scores and diversity. This adaptive behavior enables tuning the selection strategy based on dataset characteristics and annotation requirements.

4. Experiments

In this section, we present benchmarking experiments on two semantic segmentation datasets to demonstrate the effectiveness of our method compared to state-of-the-art baseline methods.

4.1. Datasets

Cityscapes [5] consists of real-world street scene images annotated with 19 semantic categories, including Road, Sidewalk, Building, Car, and Person. The training set contains 2975 images with fine-grained annotations, while the validation set includes 500 images. Cityscapes provides structured urban scenes with systematic class transition patterns (e.g., Road–Car, Building–Sky) and a moderate degree of visual similarity between images.
PascalVOC2012 [22] includes images with diverse object sizes, occlusions, and variations in lighting and backgrounds. It comprises 20 foreground object categories plus 1 background class, totaling 1464 training images and 1449 validation images. We use the extended version [23], which expands the training set to 10,582 images. PascalVOC2012 offers a wide range of object categories with varied contexts and exhibits severe class imbalance due to the dominant background class, making it particularly challenging for active learning methods.
Together, Cityscapes and PascalVOC2012 represent fundamentally different characteristics essential for validating active learning methods in terms of uncertainty- and diversity-aware sample selection. Used in combination, they provide a comprehensive evaluation across diverse data distributions and class structures that are representative of real-world semantic segmentation applications.

4.2. Implementation Details

We conduct experiments on a workstation equipped with an NVIDIA Tesla V100 GPU. Our codebase is built on top of the open-source MMSegmentation framework [46]. We evaluate two representative segmentation models: DeepLabV3 [47], a widely adopted CNN-based benchmark, and SegFormer [48], a transformer-based architecture chosen to ensure architectural diversity.
For Cityscapes, we employ DeepLabV3 [47] with a ResNet-18 [49] encoder. Training uses batch size 4 for 50 epochs with the Adam optimizer (learning rate: 5 × 10 4 , weight decay: 2 × 10 4 ). We apply polynomial learning rate decay (power = 0.9, minimum = 10 5 ) with ϵ = 10 7 and β = ( 0.9 , 0.999 ) . These hyperparameters are selected through a simple grid search, starting from MMSegmentation defaults and tuning on the fully labeled training set. Images are resized to 688 × 344 .
For PascalVOC2012, we use SegFormer-B1 [48] with batch size 8, trained for 50 epochs using AdamW (initial learning rate: 6 × 10 5 , final: 2 × 10 7 , polynomial decay with power = 1.0, weight decay: 0.01, β = ( 0.9 , 0.999 ) ). These hyperparameters are adopted from the original paper and its implementation in MMSegmentation. Images are resized to 480 × 480 .

4.3. Evaluation Metrics

Following standard practice [47,48], segmentation performance is evaluated using the intersection-over-union (IoU) metric, defined as I o U = T P / ( T P + F N + F P ) , where T P , F N , and F P denote true positives, false negatives, and false positives, respectively. In semantic segmentation, these quantities are defined at the pixel level and aggregated across the entire dataset for each class: T P is the number of pixels correctly predicted as belonging to the class, F N is the number of pixels belonging to the class in the ground truth but predicted as another class, and F P is the number of pixels incorrectly predicted as that class but belonging to another in the ground truth. The overall metric mean-IoU (mIoU) is obtained by averaging the IoU values across all classes, giving each class equal weight regardless of frequency.
We further assess performance using a pairwise penalty matrix (PPM) with two-sided t-tests, following [12]. For each pair of methods ( i , j ) at each active learning iteration, we obtain seven per-seed mIoU scores, denoted by { s i 1 , , s i 7 } and { s j 1 , , s j 7 } . The t-statistic is computed as t = 7 μ / σ , where μ = ( 1 / 7 ) r ( s i r s j r ) and σ = ( 1 / 6 ) r ( s i r s j r μ ) 2 . At the 95% confidence level (critical interval: [ 2.447 , 2.447 ] ), two methods are considered significantly different when the corresponding statistic falls outside this range. For each pair ( i , j ) , performance is compared after every iteration. If method i significantly outperforms method j ( t > 2.447 ), a penalty score of 1 / t max is added to the PPM cell ( i , j ) , where t max = 6 is the total number of iterations. Conversely, if method j significantly outperforms method i ( t < 2.447 ), the penalty score is added to cell ( j , i ) . Larger values in cell ( i , j ) indicate that method i dominates j more consistently. The column-wise average Φ summarizes overall performance, with lower values indicating stronger methods, as they are dominated less frequently across iterations.

4.4. Active Learning Setup

All methods begin with the same randomly selected initial labeled dataset D 0 L and query an equal number of samples at each iteration t. We perform t max = 6 iterations following Algorithm 1, retraining the model Ψ t from scratch using ImageNet [50] pre-trained weights and applying only horizontal flipping as data augmentation. Identical hyperparameters are used across all iterations and methods. Performance is evaluated in terms of mIoU on the validation set of each dataset, reported as mean ± standard deviation over seven runs with different random seeds. Table 1 summarizes the active learning setup for each dataset. These iteration counts and sampling percentages are sufficient for the models to approach 95% of the fully supervised mIoU on both datasets.
Unlike baseline methods, CRC-AL requires a calibration set D cal for conformal risk control. Due to limited labeled data, we use the current iteration’s training set D t L for calibration. Section 5.3 analyzes the impact of this choice.

4.5. Baseline Methods

We compare our method against a comprehensive set of baseline methods including uncertainty-based, diversity-based, and hybrid approaches. These methods were selected for their state-of-the-art performance and their compatibility with standard training procedures. All methods were implemented from scratch and adapted for semantic segmentation where necessary.
  • Random: Selects samples uniformly at random, serving as a passive baseline.
  • Entropy [10]: Employs Top-K selection based on average pixel-wise entropy across each image, quantifying prediction uncertainty through the Shannon entropy of the softmax distribution.
  • CoreSet [31]: Solves the K-Center problem using 512-dimensional embeddings extracted from the network’s bottleneck layer via average pooling. Implements greedy selection where each sample maximizes the minimum Euclidean distance to previously selected samples.
  • CDAL [32]: Constructs contextual diversity vectors capturing the spatial distribution of predicted classes within each image. Sample selection uses the CoreSet algorithm with symmetric KL divergence as the distance metric.
  • BADGE [12]: Applies K-Means++ seeding in the gradient embedding space. Gradients are computed from the cross-entropy loss using pseudo-labels (argmax predictions) and dimensionality is reduced via average pooling to handle the large gradient space in segmentation tasks.

4.6. Benchmarking Results

We evaluate CRC-AL with parameters α = 0.05 (calibration risk level) and τ = 0.5 (uncertainty–diversity trade-off) against five baseline methods across six active learning iterations. Figure 6 presents the results on both datasets.
Cityscapes results: As shown in Figure 6a and detailed in Table 2, CRC-AL demonstrates superior performance throughout all iterations, achieving the steepest learning curve among all methods. Starting from a shared baseline of 47.07 ± 0.57 mIoU with just 10% of the data (300 images), CRC-AL rapidly improves performance at each iteration. By iteration t = 3 , it surpasses 56% mIoU, maintaining a clear margin over all baselines. The dashed line represents 95% of the mIoU achievable with the fully labeled training set, which CRC-AL approaches more rapidly than any competing method. Remarkably, CRC-AL achieves this 95% threshold using only 30% of the training data (900 images), while other methods require more extensive annotation to approach similar performance. The hierarchy is clearly established: CRC-AL consistently leads, followed by a competitive cluster of CoreSet and BADGE, entropy slightly behind, CDAL showing moderate effectiveness, and random sampling performing worst. These results confirm that conformal prediction-based uncertainty quantification identifies more informative samples than traditional softmax-based (entropy) or gradient-based (BADGE) methods, while the diversity-aware selection avoids redundancy and sustains performance improvements.
PascalVOC2012 results: Figure 6b and Table 3 reveal more complex dynamics on this diverse object-centric dataset, where the background class exhibits highly varied textures. Starting from 64.98 ± 1.01 mIoU with only 500 labeled images, all methods show rapid initial improvement. CRC-AL consistently maintains an advantage, though with a smaller margin than on Cityscapes, reaching 74% mIoU at 28% data utilization (3000 images), where it achieves the 95% fully supervised threshold. Unlike Cityscapes, the performance hierarchy is less pronounced, with random sampling emerging as a surprisingly strong baseline. Except for CRC-AL, all methods fall below random at certain iterations based on average mIoU, although their results largely overlap once standard deviations are taken into account. Among the baselines, BADGE (gradient-embedding diversity), CDAL (contextual diversity), and entropy (softmax uncertainty) perform comparably to random, while CoreSet (pure representativeness) performs consistently worse, especially in later iterations. By the final iteration t = 6 , all methods except CoreSet reaches or surpass the 95% fully supervised threshold, while CRC-AL reaches the threshold at t = 5 . These results underscore CRC-AL’s robustness, demonstrating that it is the only method to sustain a consistent advantage across both iterations and dataset complexities.
Statistical significance: Figure 7 presents the aggregated pairwise penalty matrix (PPM) across two fundamentally different datasets, obtained by element-wise averaging. The diagonal entries are zero by definition, as methods cannot dominate themselves. The CRC-AL row shows that our approach significantly outperforms all competing methods, yielding near-maximum penalties across iterations. The bottom row further confirms CRC-AL’s dominance with Φ = 0 , indicating that no method significantly outperforms it at any iteration under the 95% confidence level. BADGE ( Φ = 0.07 ) and entropy ( Φ = 0.1 ), both uncertainty-driven methods, form a second tier with competitive performance. The matrix shows that BADGE and entropy do not significantly outperform one another, although BADGE is less frequently dominated by other methods. This suggests that gradient embeddings offer marginal yet consistent gains over softmax entropy. CoreSet ( Φ = 0.17 ) and CDAL ( Φ = 0.28 ), focused solely on diversity, form a weaker third tier, outperforming only random sampling, which performs worst overall. CoreSet occasionally outperforms CDAL, while CDAL rarely dominates CoreSet, implying that visual representativeness can be more effective than contextual diversity in certain settings. Taken together, these results establish a stratified performance hierarchy—uncertainty-based methods outperform diversity-only strategies, while CRC-AL uniquely integrates both to deliver consistent superiority across datasets with markedly different characteristics.

5. Sensitivity Analysis

5.1. Uncertainty–Diversity Balance

The trade-off parameter τ [ 0 , 1 ] in the Top-Diverse-K algorithm governs the balance between uncertainty-driven and diversity-aware selection. Figure 8a illustrates its effect on Cityscapes with α fixed at 0.05. Pure uncertainty sampling ( τ = 0 ) performs poorly, indicating that uncertainty alone is insufficient and that diversity, captured through class confusion patterns, is essential for effective active learning. This poor performance arises from the dataset’s structured urban scenes where semantically related classes systematically co-occur, leading to redundant sampling of visually similar high-uncertainty regions without diversity awareness. While all τ values behave similarly in early iterations when the labeled set is small, their performance diverges as training progresses. Settings with τ < 0.5 degrade more noticeably, with τ = 0.25 performing better than τ = 0 but still suboptimal. This suggests that limited diversity incorporation helps, but stronger integration is required to maintain robust performance.
Figure 8b reveals contrasting behavior on PascalVOC2012. Here, all τ values consistently outperform random sampling, with even pure uncertainty sampling ( τ = 0 ) showing reasonable performance. This contrasting behavior reflects PascalVOC2012’s inherent diversity, naturally distributing high-uncertainty samples across different regions of the embedding space. To understand this difference, we include entropy—which also uses pure uncertainty with Top-K selection—as a direct comparison. While entropy achieves results comparable to random sampling (consistent with our main results in Table 3), our method with τ = 0 outperforms both. This crucial observation demonstrates that conformal prediction provides superior uncertainty quantification compared to traditional entropy measures, offering better-calibrated confidence estimates even without diversity.

5.2. Risk Level Sensitivity

The risk level α ( 0 , 1 ) controls the statistical guarantee in conformal risk control, representing the acceptable upper bound on the false negative rate. The choice of α involves a critical trade-off. With high α values, the conformal predictor tolerates more risk, producing empty or smaller prediction sets. This leads to sparse risk maps where only the most ambiguous regions appear uncertain, and embeddings become nearly diagonal, indicating minimal class confusion. Such oversimplification treats most images as equally certain, reducing the discriminative power needed for effective active learning. Conversely, low α values produce conservative predictions with larger prediction sets. This generates dense risk maps where extensive regions appear uncertain, potentially masking truly informative regions. The resulting embeddings capture spurious co-occurrences that may reflect statistical noise rather than meaningful class confusion patterns. Figure 9 shows example risk maps and co-occurrence density matrices for different α values.
Figure 8c empirically validates these theoretical considerations on Cityscapes. With τ = 0.5 , we compare three α values: 0.05, 0.1, and 0.2. The results reveal that α = 0.05 and α = 0.1 achieve similar performance throughout the active learning process, both substantially outperforming α = 0.2 . The degraded performance at α = 0.2 confirms that overly permissive risk levels fail to capture sufficient uncertainty for sample discrimination. Importantly, α = 0.05 demonstrates superior performance in early iterations. This suggests that tighter statistical guarantees, while potentially introducing some noise, better identify the subtle uncertainties that distinguish truly informative samples when labeled data is scarce. The consistent advantage of lower α values indicates that conservative calibration improves sample informativeness, with α = 0.05 emerging as the most effective setting for practical active learning.

5.3. Impact of Calibration Set

Active learning assumes that labeled data is limited. Our method, based on conformal risk control, requires an additional labeled set for calibration. This introduces a challenge: deciding whether to allocate labeled data for calibration or use it to improve model performance. To address this, we conduct experiments to analyze the impact of the calibration set on overall active learning performance.
We use the Cityscapes dataset with a randomly selected initial labeled dataset D 0 L containing 600 images and perform a single active learning iteration. The initial model Ψ 0 is trained on D 0 L . For the calibration set D cal , we compare two configurations: (1) using the current labeled data D cal = D 0 L , or (2) using a randomly sampled subset from the unlabeled pool D cal D 0 U with | D cal | = 600 . Although labels in D 0 U are typically hidden from the learner, we reveal them solely for calibration purposes to evaluate the conformal predictor.
After calibration, we select a batch B D 0 U containing 150 images for annotation. The model Ψ 1 is then retrained on the combined dataset D 1 L = D 0 L B * , where B * denotes the annotated batch. We evaluate the expected risk on both the Cityscapes validation split (500 images) and the remaining unlabeled data D 0 U D cal . Since these sets are not used during calibration, they remain unseen by the conformal predictor. Each experiment is repeated using five different random seeds.
Table 4 presents the mIoU scores and calibration performance, measured as the expected false negative rate (FNR) averaged across all classes. The differences in mIoU scores across configurations are minimal, although calibrating on unlabeled data performs slightly better when α = 0.05 . This indicates that our method is not highly sensitive to the calibration set choice in terms of active learning performance.
For calibration quality, using D cal D U achieves risk levels close to the target risk level α with minimal variation across both unseen evaluation sets. In contrast, calibrating on D cal = D L yields higher expected FNRs than α ; however, these remain consistent across different evaluation sets.
This consistency is the critical observation: when using D cal = D L , although the absolute risk deviates from α (0.182 vs. 0.05, and 0.280 vs. 0.10), the calibration produces stable predictions across both unseen evaluation sets, with nearly identical FNR values. This stability indicates that the calibration mechanism successfully learns relative uncertainty patterns even when the absolute thresholds are biased. The conformal predictor effectively identifies which pixels and classes are more uncertain than others, maintaining consistent uncertainty rankings across different data distributions.
For active learning, this relative ordering matters more than absolute calibration accuracy. The method correctly identifies uncertain regions and ambiguous class boundaries, enabling effective sample selection despite the shifted thresholds. The mIoU results confirm this, as both calibration strategies achieve similar performance (55.030 vs. 55.242 for α = 0.05 ), showing that uncertainty quantification provides reliable guidance for active learning.

6. Discussion

Our experiments establish a clear performance hierarchy across benchmark datasets: CRC-AL consistently surpasses existing active learning baselines, maintaining the steepest learning curve throughout the annotation budget. The persistent gap between CRC-AL and competing methods underscores the effectiveness of conformal risk control in producing statistically grounded uncertainty quantification. Unlike conventional softmax-based uncertainty (entropy) or gradient-based heuristics (BADGE), CRC-AL identifies fundamentally more informative samples by capturing both relative uncertainty and semantic confusion at the class level. The resulting prediction sets, risk maps, and co-occurrence embeddings provide richer guidance for sample selection, while Top-Diverse-K enables a principled balance between uncertainty and diversity.
Dataset-specific analyses reveal that the advantages of CRC-AL hold under varied conditions but also highlight unique vulnerabilities of competing methods. On Cityscapes, where structured urban scenes generate strong co-occurrence among semantically related classes (e.g., Road–Car, Building–Sky), purely context- or representativeness-driven strategies (CDAL, CoreSet) fail to match the effectiveness of CRC-AL. CDAL’s reliance on contextual diversity is ill-suited to environments where many classes systematically co-occur, while CoreSet’s visual representativeness overlooks the semantic structure of uncertainty. In contrast, CRC-AL effectively exploits co-occurrence embeddings to prevent redundant queries and sustain annotation efficiency. On PascalVOC2012, with its object-centric and visually heterogeneous images, CoreSet even underperforms random sampling, as the background class dominates pixel statistics and misleads visual diversity metrics. CDAL partially mitigates this effect through its contextual measures. In contrast, entropy and BADGE remain vulnerable to severe class imbalance, causing their performance to converge with random sampling, which unexpectedly emerges as a strong baseline. Although CRC-AL is not immune to such imbalance—uncertainty estimates and embeddings degrade in the presence of overwhelming background pixels—it consistently maintains advantages over baselines, demonstrating robustness under adverse dataset conditions. Notably, no single baseline performs well across both datasets. In contrast, CRC-AL consistently achieves strong results on both Cityscapes and PascalVOC2012, underscoring its superior generalization across datasets with markedly different characteristics.
Overall, entropy and BADGE, both uncertainty-driven methods, show stronger performance than diversity-only strategies but remain limited by their reliance on heuristic uncertainty measures. Entropy suffers from softmax overconfidence, leading to unreliable estimates in class-imbalanced regions, while BADGE’s gradient-based embeddings improve diversity but can be unstable and sensitive to pseudo-label noise. Consequently, both methods achieve moderate gains but fail to consistently close the gap with CRC-AL. By calibrating each class independently at the same risk level α , CRC-AL addresses these fundamental limitations through balanced uncertainty quantification across both frequent and rare classes. Compared to BADGE’s gradient embeddings, our co-occurrence embeddings introduce a different perspective by capturing statistically grounded semantic confusion patterns. These patterns encode both prediction ambiguity and class-specific uncertainty, enabling the selection of more informative samples in regions where the model struggles to distinguish between classes. CoreSet and CDAL lack any notion of uncertainty and select samples solely based on diversity. This limitation explains their weaker performance compared to uncertainty-driven methods, particularly on datasets with sufficient inherent diversity. Our Top-Diverse-K algorithm integrates co-occurrence embeddings with uncertainty weighting through a principled barycenter-based distance metric. In this way, it selects highly informative samples while maintaining a well-distributed batch, achieving an effective balance via the trade-off parameter τ . The clear stratification of Φ values in the pairwise penalty matrix (Figure 7) underscores the importance of principled uncertainty quantification, with diversity serving as a secondary, complementary factor. By combining conformal prediction-based risk maps with co-occurrence embeddings, CRC-AL overcomes the limitations of existing methods, providing statistically guaranteed uncertainty quantification and sample selection that adapts effectively to dataset-specific class structures.
The parameter sensitivity analysis provides practical insights for deployment. The trade-off parameter τ effectively controls the uncertainty–diversity balance, with pure uncertainty sampling ( τ = 0 ) leading to redundant selections and pure diversity losing informativeness. The robust performance at τ = 0.5 across both datasets suggests this balanced setting as a reliable default, while practitioners can adjust the parameter based on dataset characteristics—higher values for visually similar datasets (e.g., video sequences) and lower values for inherently diverse collections. Similarly, the calibration risk level α = 0.05 consistently outperforms looser settings, particularly in early iterations where sample selection has the greatest impact on learning trajectories. These findings indicate that tighter statistical control enhances active learning by better identifying the subtle uncertainties that distinguish truly informative samples. Taken together, the analysis highlights α = 0.05 and τ = 0.5 as robust defaults that can be applied across diverse semantic segmentation scenarios, reducing the need for extensive hyperparameter tuning.
An important practical consideration is the allocation of labeled data for calibration. Since conformal risk control requires a calibration set, practitioners must decide whether reserving part of the annotation budget for calibration is preferable to using it directly for training. Our calibration analysis indicates that even under this trade-off, CRC-AL remains highly competitive. Although absolute thresholds deviate from nominal values when calibration is performed on the labeled training set, the predictor preserves relative uncertainty rankings across unseen evaluation datasets. This stability ensures that informative samples are consistently prioritized, even when absolute risks diverge from theoretical expectations. For active learning, relative calibration quality is therefore more critical than strict adherence to error rates, aligning with the practical goal of guiding sample selection rather than providing formal risk certificates. The minimal performance difference between ideal and practical calibration further validates the practicality of CRC-AL in scenarios where separate calibration data may not be available.
Several limitations warrant discussion. First, extreme class imbalance still degrades performance—though CRC-AL handles it better than baselines, the background-dominated PascalVOC2012 results suggest room for improvement through adaptive reweighting or hierarchical calibration. Second, co-occurrence embeddings scale quadratically with the number of classes ( K 2 ), becoming computationally prohibitive for large K and vulnerable to the curse of dimensionality. For datasets with hundreds of classes, dimensionality reduction techniques or hierarchical class groupings may be necessary to maintain computational feasibility. In practice, dimensionality reduction techniques such as PCA can compress the K 2 embedding space to a manageable size while preserving the essential uncertainty structure, leveraging the observation that most class pairs seldom co-occur.
The proposed framework addresses a fundamental weakness of existing active learning approaches by embedding uncertainty quantification within a statistically grounded foundation. The empirical gains observed on two challenging benchmarks demonstrate both the methodological value and practical potential of CRC-AL, laying the groundwork for more reliable and annotation-efficient semantic segmentation systems.

7. Conclusions

In this paper, we introduced Conformal Risk Controlled Active Learning (CRC-AL), the first active learning framework for semantic segmentation based on conformal prediction theory. Our approach addresses a fundamental limitation in existing active learning methods: the reliance on poorly calibrated softmax probabilities that fail to reliably identify informative samples. By applying conformal risk control independently to each semantic class, we transform softmax outputs into statistically guaranteed prediction sets that reliably quantify model uncertainty. Our risk maps and co-occurrence embeddings capture both spatial and semantic uncertainty patterns, while the Top-Diverse-K algorithm effectively balances informativeness with diversity in sample selection.
Extensive experiments validate the effectiveness of our approach. CRC-AL consistently outperforms five state-of-the-art baselines across two fundamentally different datasets, achieving 95% of fully-supervised performance using only 30% of training data on both datasets. The superior performance over the benchmarking methods confirms that conformal prediction-based uncertainty quantification provides reliable guidance for active learning. The method’s robustness across structured street scenes and diverse object images demonstrates that statistical guarantees generalize better than heuristic confidence measures.
Future work will pursue theoretical studies on tighter bounds between conformal risk parameters and active learning performance, adaptive selection strategies, and extensions to other dense prediction tasks. Understanding the theoretical properties of class co-occurrence patterns under different data distributions could lead to more sophisticated embedding strategies. In addition, extensions toward joint calibration strategies could be explored to provide stronger formal guarantees across classes. We also plan to investigate adaptive schemes for the trade-off parameter τ that automatically adjust based on dataset characteristics and learning progress. The success of CRC-AL demonstrates that principled statistical methods can yield significant practical benefits, opening new avenues for uncertainty quantification in computer vision where annotation costs remain a critical bottleneck.

Author Contributions

Conceptualization, C.E. and N.K.U.; methodology, C.E.; software, C.E.; validation, C.E. and N.K.U.; visualization, C.E.; writing—original draft preparation, C.E.; writing—review and editing, C.E. and N.K.U.; supervision, N.K.U.; project administration, N.K.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available in publicly accessible repositories. The source code is openly available at https://github.com/ccerhan/AL-SEG (last accessed on 29 August 2025). The datasets can be accessed at the following links: Cityscapes (https://www.cityscapes-dataset.com/downloads, last accessed on 29 August 2025), PascalVOC2012 (https://www.kaggle.com/datasets/gopalbhattrai/pascal-voc-2012-dataset, last accessed on 29 August 2025), and PascalVOC2012 Augmented Patch (https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/semantic_contours/benchmark.tgz, last accessed on 29 August 2025). No special access restrictions apply.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Tian, J.; Xin, P.; Bai, X.; Xiao, Z.; Li, N. An Efficient Semantic Segmentation Framework with Attention-Driven Context Enhancement and Dynamic Fusion for Autonomous Driving. Appl. Sci. 2025, 15, 8373. [Google Scholar] [CrossRef]
  2. Altini, N.; Lasaracina, E.; Galeone, F.; Prunella, M.; Suglia, V.; Carnimeo, L.; Triggiani, V.; Ranieri, D.; Brunetti, G.; Bevilacqua, V. A Comparison Between Unimodal and Multimodal Segmentation Models for Deep Brain Structures from T1- and T2-Weighted MRI. Mach. Learn. Knowl. Extr. 2025, 7, 84. [Google Scholar] [CrossRef]
  3. Czajka, M.; Krupka, M.; Kubacka, D.; Janiszewski, M.R.; Belter, D. A Comparison of Segmentation Methods for Semantic OctoMap Generation. Appl. Sci. 2025, 15, 7285. [Google Scholar] [CrossRef]
  4. Formichini, M.; Avizzano, C.A. A Comparative Analysis of Deep Learning-Based Segmentation Techniques for Terrain Classification in Aerial Imagery. AI 2025, 6, 145. [Google Scholar] [CrossRef]
  5. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
  6. Settles, B. Active Learning Literature Survey; Technical Report 1648; University of Wisconsin-Madison: Madison, WI, USA, 2009. [Google Scholar]
  7. Wu, M.; Li, C.; Yao, Z. Deep Active Learning for Computer Vision Tasks: Methodologies, Applications, and Challenges. Appl. Sci. 2022, 12, 8103. [Google Scholar] [CrossRef]
  8. Tharwat, A.; Schenck, W. A Survey on Active Learning: State-of-the-Art, Practical Challenges and Research Directions. Mathematics 2023, 11, 820. [Google Scholar] [CrossRef]
  9. Li, D.; Wang, Z.; Chen, Y.; Jiang, R.; Ding, W.; Okumura, M. A survey on deep active learning: Recent advances and new frontiers. IEEE Trans. Neural Networks Learn. Syst. 2025, 36, 5879–5899. [Google Scholar] [CrossRef] [PubMed]
  10. Wang, D.; Shang, Y. A new active labeling method for deep learning. In Proceedings of the International Joint Conference on Neural Networks, Beijing, China, 6–11 July 2014; pp. 112–119. [Google Scholar]
  11. Golestaneh, S.A.; Kitani, K.M. Importance of Self-Consistency in Active Learning for Semantic Segmentation. arXiv 2020, arXiv:2008.01860v1. [Google Scholar] [CrossRef]
  12. Ash, J.T.; Zhang, C.; Krishnamurthy, A.; Langford, J.; Agarwal, A. Deep Batch Active Learning by Diverse, Uncertain Gradient Lower Bounds. arXiv 2019, arXiv:1906.03671. [Google Scholar]
  13. Han, Y.; Liu, D.; Shang, J.; Zheng, L.; Zhong, J.; Cao, W.; Sun, H.; Xie, W. BALQUE: Batch active learning by querying unstable examples with calibrated confidence. Pattern Recognit. 2024, 151, 110385. [Google Scholar] [CrossRef]
  14. Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On calibration of modern neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1321–1330. [Google Scholar]
  15. Vovk, V.; Gammerman, A.; Shafer, G. Algorithmic Learning in a Random World; Springer: New York, NY, USA, 2005. [Google Scholar]
  16. Balasubramanian, V.; Ho, S.S.; Vovk, V. Conformal Prediction for Reliable Machine Learning: Theory, Adaptations and Applications; Newnes: Boston, MA, USA, 2014. [Google Scholar]
  17. Angelopoulos, A.N.; Bates, S. A Gentle Introduction to Conformal Prediction and Distribution-Free Uncertainty Quantification. arXiv 2022, arXiv:2107.07511. [Google Scholar] [CrossRef]
  18. Angelopoulos, A.N.; Bates, S.; Fisch, A.; Lei, L.; Schuster, T. Conformal Risk Control. In Proceedings of the Twelfth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  19. Matiz, S.; Barner, K.E. Inductive conformal predictor for convolutional neural networks: Applications to active learning for image classification. Pattern Recognit. 2019, 90, 172–182. [Google Scholar] [CrossRef]
  20. Corrigan, A.; Hopcroft, P.; Narvaez, A.; Bendtsen, C. Batch mode active learning for mitotic phenotypes using conformal prediction. In Proceedings of the Ninth Symposium on Conformal and Probabilistic Prediction and Applications, PMLR, Online, 9–11 September 2020; Volume 128, pp. 229–243. [Google Scholar]
  21. Matiz, S.; Barner, K.E. Conformal prediction based active learning by linear regression optimization. Neurocomputing 2020, 388, 157–169. [Google Scholar] [CrossRef]
  22. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes (VOC) Challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  23. Hariharan, B.; Arbeláez, P.; Bourdev, L.; Maji, S.; Malik, J. Semantic contours from inverse detectors. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 991–998. [Google Scholar] [CrossRef]
  24. Lewis, D.D.; Catlett, J. Heterogeneous Uncertainty Sampling for Supervised Learning. In Machine Learning Proceedings; Morgan Kaufmann: San Francisco, CA, USA, 1994; pp. 148–156. [Google Scholar]
  25. Roth, D.; Small, K. Margin-Based Active Learning for Structured Output Spaces. In Proceedings of the Machine Learning: ECML 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 413–424. [Google Scholar]
  26. Joshi, A.J.; Porikli, F.; Papanikolopoulos, N. Multi-class active learning for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 2372–2379. [Google Scholar]
  27. Gal, Y.; Islam, R.; Ghahramani, Z. Deep Bayesian Active Learning with Image Data. In Proceedings of the 34th International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 1183–1192. [Google Scholar]
  28. Yang, L.; Zhang, Y.; Chen, J.; Zhang, S.; Chen, D.Z. Suggestive Annotation: A Deep Active Learning Framework for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 399–407. [Google Scholar]
  29. Beluch, W.H.; Genewein, T.; Nurnberger, A.; Kohler, J.M. The Power of Ensembles for Active Learning in Image Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9368–9377. [Google Scholar]
  30. Mittal, S.; Niemeijer, J.; Cicek, O.; Tatarchenko, M.; Ehrhardt, J.; Schäfer, J.P.; Handels, H.; Brox, T. Realistic evaluation of deep active learning for image classification and semantic segmentation. Int. J. Comput. Vis. 2025, 133, 4294–4316. [Google Scholar] [CrossRef]
  31. Sener, O.; Savarese, S. Active Learning for Convolutional Neural Networks: A Core-Set Approach. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  32. Agarwal, S.; Arora, H.; Anand, S.; Arora, C. Contextual Diversity for Active Learning. In Proceedings of the Computer Vision—ECCV 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 137–153. [Google Scholar]
  33. Gaillochet, M.; Desrosiers, C.; Lombaert, H. Active learning for medical image segmentation with stochastic batches. Med Image Anal. 2023, 90, 102958. [Google Scholar] [CrossRef]
  34. Sinha, S.; Ebrahimi, S.; Darrell, T. Variational Adversarial Active Learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 5971–5980. [Google Scholar]
  35. Kim, K.; Park, D.; Kim, K.I.; Chun, S.Y. Task-aware variational adversarial active learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 8162–8171. [Google Scholar]
  36. Yoo, D.; Kweon, I.S. Learning Loss for Active Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 93–102. [Google Scholar]
  37. Casanova, A.; Pinheiro, P.O.; Rostamzadeh, N.; Pal, C.J. Reinforced active learning for image segmentation. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  38. Siddiqui, Y.; Valentin, J.; NieBner, M. ViewAL: Active learning with viewpoint entropy for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  39. Cai, L.; Xu, X.; Liew, J.H.; Sheng Foo, C. Revisiting superpixels for active learning in semantic segmentation with realistic annotation costs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  40. Mittal, S.; Niemeijer, J.; Schäfer, J.P.; Brox, T. Best Practices in Active Learning for Semantic Segmentation. In Pattern Recognition; Springer: Cham, Switzerland, 2024; pp. 427–442. [Google Scholar]
  41. Ho, S.S.; Wechsler, H. Query by transduction. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 30, 1557–1571. [Google Scholar] [CrossRef]
  42. Balasubramanian, V.; Chakraborty, S.; Panchanathan, S. Generalized Query by Transduction for online active learning. In Proceedings of the IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 1378–1385. [Google Scholar]
  43. Makili, L.E.; Sánchez, J.A.V.; Dormido-Canto, S. Active Learning Using Conformal Predictors: Application to Image Classification. Fusion Sci. Technol. 2012, 62, 347–355. [Google Scholar] [CrossRef]
  44. Goldstein, H.; Poole, C.P.; Safko, J.L. Classical Mechanics, 3rd ed.; Pearson: Upper Saddle River, NJ, USA, 2001. [Google Scholar]
  45. van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  46. MMSegmentation: OpenMMLab Semantic Segmentation Toolbox and Benchmark. 2020. Available online: https://github.com/open-mmlab/mmsegmentation (accessed on 13 October 2025).
  47. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking Atrous Convolution for Semantic Image Segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar] [CrossRef]
  48. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. arXiv 2021, arXiv:2105.15203. [Google Scholar] [CrossRef]
  49. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  50. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
Figure 2. Conformal risk control demonstration with three classes: Sidewalk, Rider, and Bicycle. The graph shows softmax scores along a horizontal cross-section of the image. Hatched regions (above thresholds) indicate where the conformal predictor includes each class with risk level at most α . Although pixels around index 10 are classified as Bicycle via argmax, their softmax scores also exceed the Rider threshold, correctly including the true class in the prediction sets.
Figure 2. Conformal risk control demonstration with three classes: Sidewalk, Rider, and Bicycle. The graph shows softmax scores along a horizontal cross-section of the image. Hatched regions (above thresholds) indicate where the conformal predictor includes each class with risk level at most α . Although pixels around index 10 are classified as Bicycle via argmax, their softmax scores also exceed the Rider threshold, correctly including the true class in the prediction sets.
Ai 06 00270 g002
Figure 3. Example risk maps and co-occurrence density matrices with α = 0.05 . (Top) Cityscapes; (Bottom) PascalVOC2012. Risk maps highlight uncertain regions (warmer colors indicate higher uncertainty). Each matrix row shows the likelihood of a class co-occurring with others. Uncertainty scores are computed as the spatial average of risk maps, and embeddings are obtained by flattening the matrices.
Figure 3. Example risk maps and co-occurrence density matrices with α = 0.05 . (Top) Cityscapes; (Bottom) PascalVOC2012. Risk maps highlight uncertain regions (warmer colors indicate higher uncertainty). Each matrix row shows the likelihood of a class co-occurring with others. Uncertainty scores are computed as the spatial average of risk maps, and embeddings are obtained by flattening the matrices.
Ai 06 00270 g003
Figure 4. Illustration of the barycenter-based distance metric for sample selection. Sample A (blue) is labeled, while B (red), C (green), and E (purple) are unlabeled candidates. Circle sizes indicate uncertainty scores. Distances d a b , d a c , and d a e represent the metrics from A to each candidate. The metric balances uncertainty (circle size) and diversity (spatial distance) to select C as the most informative sample when τ 0.5 , avoiding the redundancy of nearby high-uncertainty sample B and the distant low-uncertainty sample E.
Figure 4. Illustration of the barycenter-based distance metric for sample selection. Sample A (blue) is labeled, while B (red), C (green), and E (purple) are unlabeled candidates. Circle sizes indicate uncertainty scores. Distances d a b , d a c , and d a e represent the metrics from A to each candidate. The metric balances uncertainty (circle size) and diversity (spatial distance) to select C as the most informative sample when τ 0.5 , avoiding the redundancy of nearby high-uncertainty sample B and the distant low-uncertainty sample E.
Ai 06 00270 g004
Figure 5. Visualization of the parameter τ ’s impact on sample selection. (Top row) t-SNE [45] projections of the selected batch B in the embedding space for increasing values of τ (left to right). Larger τ values yield more spatially distributed selections. (Bottom row) Uncertainty distributions of selected samples. The x-axis represents uncertainty scores and the y-axis shows sample counts per bin. As τ increases, the selection shifts from high-uncertainty clustering ( τ = 0 ) to a broader uncertainty range with greater spatial diversity ( τ > 0 ).
Figure 5. Visualization of the parameter τ ’s impact on sample selection. (Top row) t-SNE [45] projections of the selected batch B in the embedding space for increasing values of τ (left to right). Larger τ values yield more spatially distributed selections. (Bottom row) Uncertainty distributions of selected samples. The x-axis represents uncertainty scores and the y-axis shows sample counts per bin. As τ increases, the selection shifts from high-uncertainty clustering ( τ = 0 ) to a broader uncertainty range with greater spatial diversity ( τ > 0 ).
Ai 06 00270 g005
Figure 6. * denotes the proposed method. Comparison of baseline methods with CRC-AL ( α = 0.05 , τ = 0.5 ). (a) Cityscapes and (b) PascalVOC2012 mIoU performance across active learning iterations. Initial scores ( t = 0 ) are shown above each plot. The dashed line indicates 95% of fully supervised performance; shaded regions show standard deviation over 7 runs.
Figure 6. * denotes the proposed method. Comparison of baseline methods with CRC-AL ( α = 0.05 , τ = 0.5 ). (a) Cityscapes and (b) PascalVOC2012 mIoU performance across active learning iterations. Initial scores ( t = 0 ) are shown above each plot. The dashed line indicates 95% of fully supervised performance; shaded regions show standard deviation over 7 runs.
Ai 06 00270 g006
Figure 7. * denotes the proposed method. Comparison of baseline methods with CRC-AL ( α = 0.05 , τ = 0.5 ), shown as the pairwise penalty matrix (PPM) aggregated across Cityscapes and PascalVOC2012. Lower Φ scores indicate better performance. CRC-AL achieves Φ = 0 , demonstrating consistent superiority at the 95% confidence level.
Figure 7. * denotes the proposed method. Comparison of baseline methods with CRC-AL ( α = 0.05 , τ = 0.5 ), shown as the pairwise penalty matrix (PPM) aggregated across Cityscapes and PascalVOC2012. Lower Φ scores indicate better performance. CRC-AL achieves Φ = 0 , demonstrating consistent superiority at the 95% confidence level.
Ai 06 00270 g007
Figure 8. Sensitivity analysis of CRC-AL parameters. (a) Effect of uncertainty–diversity trade-off τ on Cityscapes with α = 0.05 . (b) Effect of τ on PascalVOC2012 with α = 0.05 . (c) Effect of calibration risk level α on Cityscapes with τ = 0.5 . The dashed line indicates 95% of fully supervised mIoU. Shaded regions show standard deviation over 7 runs. Random sampling and entropy are shown as baselines. The analysis identifies α = 0.05 and τ = 0.5 as the most effective settings for practical active learning.
Figure 8. Sensitivity analysis of CRC-AL parameters. (a) Effect of uncertainty–diversity trade-off τ on Cityscapes with α = 0.05 . (b) Effect of τ on PascalVOC2012 with α = 0.05 . (c) Effect of calibration risk level α on Cityscapes with τ = 0.5 . The dashed line indicates 95% of fully supervised mIoU. Shaded regions show standard deviation over 7 runs. Random sampling and entropy are shown as baselines. The analysis identifies α = 0.05 and τ = 0.5 as the most effective settings for practical active learning.
Ai 06 00270 g008
Figure 9. Risk maps and co-occurrence matrices for α { 0.05 , 0.1 , 0.2 } . Low α yields dense patterns, while high α produces sparse, near-diagonal matrices, reducing discriminative power for active learning.
Figure 9. Risk maps and co-occurrence matrices for α { 0.05 , 0.1 , 0.2 } . Low α yields dense patterns, while high α produces sparse, near-diagonal matrices, reducing discriminative power for active learning.
Ai 06 00270 g009
Table 1. Active learning setup for Cityscapes and PascalVOC2012 datasets.
Table 1. Active learning setup for Cityscapes and PascalVOC2012 datasets.
DatasetConfigurationNotes
CityscapesInitial labeled set: 300 images Batch size: 150 per iteration Total: 1200 images (40% of dataset)Ignored regions (e.g., ego-vehicle hood) are treated as known only in labeled data to ensure fair comparison.
PascalVOC2012Initial labeled set: 500 images Batch size: 500 per iteration Total: 3500 images (33% of dataset)Background class is included in training, yielding 21 classes overall for a realistic use-case scenario.
Table 2. Cityscapes mIoU scores across iterations (initial mIoU: 47.07 ± 0.57 ). Bold values indicate the highest score at iteration t. Results show mean ± standard deviation over 7 runs.
Table 2. Cityscapes mIoU scores across iterations (initial mIoU: 47.07 ± 0.57 ). Bold values indicate the highest score at iteration t. Results show mean ± standard deviation over 7 runs.
Method t = 1 t = 2 t = 3 t = 4 t = 5 t = 6
Random49.35 ± 0.8751.44 ± 0.7152.85 ± 0.5753.69 ± 0.5054.57 ± 0.6955.59 ± 0.38
Entropy52.58 ± 0.5454.17 ± 0.4754.87 ± 0.4156.07 ± 0.5656.39 ± 0.6256.93 ± 0.64
CoreSet52.04 ± 0.6354.50 ± 0.5955.69 ± 0.5556.65 ± 0.5557.00 ± 0.4157.70 ± 0.16
CDAL49.79 ± 0.7151.99 ± 0.3853.67 ± 0.4954.56 ± 0.3055.60 ± 0.3756.41 ± 0.62
BADGE51.66 ± 0.8653.55 ± 0.8854.85 ± 0.3755.86 ± 0.4656.64 ± 0.3257.61 ± 0.32
CRC-AL *52.76 ± 0.4654.97 ± 0.4156.18 ± 0.3857.02 ± 0.3257.48 ± 0.5457.98 ± 0.49
* The proposed method.
Table 3. PascalVOC2012 mIoU scores across iterations (initial mIoU: 64.98 ± 1.01 ). Bold values indicate the highest score at iteration t, while underlined values denote results below random sampling. Results show mean ± standard deviation over 7 runs.
Table 3. PascalVOC2012 mIoU scores across iterations (initial mIoU: 64.98 ± 1.01 ). Bold values indicate the highest score at iteration t, while underlined values denote results below random sampling. Results show mean ± standard deviation over 7 runs.
Method t = 1 t = 2 t = 3 t = 4 t = 5 t = 6
Random68.70 ± 0.7870.41 ± 0.3872.00 ± 0.3172.63 ± 0.4473.59 ± 0.3674.09 ± 0.44
Entropy69.27 ± 0.6170.72 ± 0.5771.59 ± 0.8172.51 ± 0.3673.16 ± 0.3673.99 ± 0.42
CoreSet68.52 ± 0.6370.45 ± 0.5671.80 ± 0.5372.35 ± 0.3072.78 ± 0.3573.71 ± 0.35
CDAL68.85 ± 0.5070.75 ± 0.6571.92 ± 0.6273.10 ± 0.2073.66 ± 0.3274.09 ± 0.37
BADGE69.09 ± 0.4770.68 ± 0.3571.69 ± 0.2872.80 ± 0.6573.41 ± 0.6574.17 ± 0.42
CRC-AL *69.34 ± 0.5471.16 ± 0.4072.13 ± 0.6473.17 ± 0.4174.00 ± 0.4474.30 ± 0.43
* The proposed method.
Table 4. Impact of calibration set choice on active learning performance. Calibration is performed on unseen unlabeled set ( D cal D U ) or labeled set ( D cal = D L ) with α { 0.05 , 0.1 } . Reported values are mIoU after one iteration and empirical risk (FNR) on both validation and unlabeled test sets.
Table 4. Impact of calibration set choice on active learning performance. Calibration is performed on unseen unlabeled set ( D cal D U ) or labeled set ( D cal = D L ) with α { 0.05 , 0.1 } . Reported values are mIoU after one iteration and empirical risk (FNR) on both validation and unlabeled test sets.
Risk Level ( α )Calibration SetmIoU * ( Ψ 1 )Risk on Val. SetRisk on D U D cal
0.05 D cal D U 55.242 ± 0.4290.056 ± 0.0140.055 ± 0.011
0.05 D cal = D L 55.030 ± 0.3420.182 ± 0.0910.182 ± 0.092
0.1 D cal D U 55.424 ± 0.4090.105 ± 0.0210.104 ± 0.011
0.1 D cal = D L 54.682 ± 0.3960.280 ± 0.1210.279 ± 0.126
* Random sampling baseline for comparison: 52.838 ± 0.579.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Erhan, C.; Ure, N.K. Reducing Annotation Effort in Semantic Segmentation Through Conformal Risk Controlled Active Learning. AI 2025, 6, 270. https://doi.org/10.3390/ai6100270

AMA Style

Erhan C, Ure NK. Reducing Annotation Effort in Semantic Segmentation Through Conformal Risk Controlled Active Learning. AI. 2025; 6(10):270. https://doi.org/10.3390/ai6100270

Chicago/Turabian Style

Erhan, Can, and Nazim Kemal Ure. 2025. "Reducing Annotation Effort in Semantic Segmentation Through Conformal Risk Controlled Active Learning" AI 6, no. 10: 270. https://doi.org/10.3390/ai6100270

APA Style

Erhan, C., & Ure, N. K. (2025). Reducing Annotation Effort in Semantic Segmentation Through Conformal Risk Controlled Active Learning. AI, 6(10), 270. https://doi.org/10.3390/ai6100270

Article Metrics

Back to TopTop