Next Article in Journal
Nonlinear Simulation of Terminal Maneuvers Including Landing Gear Dynamics, Crosswind and Ground Effect
Next Article in Special Issue
Machine Learning and Soft Computing: Current Trends and Applications
Previous Article in Journal
Estimations of Production Capacity Based on Simulation Models: A Case Study of Furniture Manufacturing Systems
Previous Article in Special Issue
Calibration of a Multiphase Poroelasticity Model Using Genetic Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fair Classification Without Sensitive Attribute Labels via Dynamic Reweighting

1
Department of Artificial Intelligence, Inha University, Incheon 22212, Republic of Korea
2
Department of Computer Science and Engineering, Incheon National University, Incheon 22012, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2026, 16(4), 1684; https://doi.org/10.3390/app16041684
Submission received: 6 January 2026 / Revised: 1 February 2026 / Accepted: 5 February 2026 / Published: 7 February 2026
(This article belongs to the Special Issue Machine Learning and Soft Computing: Current Trends and Applications)

Abstract

Fairness-aware classification with respect to sensitive attributes, such as gender and race, is one of the most important topics in machine learning. Although numerous studies have made outstanding progress through various approaches, one key limitation is that they necessarily require additional labels of sensitive attributes for training. This poses a significant challenge since sensitive attributes typically correspond to personal information. To this end, we propose a novel reweighting method that dynamically gives more weights to underrepresented groups across potential sensitive attributes. Without auxiliary networks or strong assumptions about sensitive attributes, the proposed method significantly improves fairness under various scenarios on benchmark datasets, outperforming the existing state-of-the-art methods.

1. Introduction

The abundance of training data and the development of sophisticated deep learning models have allowed us to solve various real-world problems with high accuracy and efficiency [1,2,3,4]. This has enabled the widespread use of AI models in real-world systems but has also brought forth a new issue, namely AI fairness, which has been overshadowed by its remarkable performance [5,6,7]. For instance, AI models deployed by national institutions and global corporations, such as COMPAS (Correctional Offender Management Profiling for Alternative Sanctions) [8], Google Photos [9], and Face App [10], sometimes exhibit unethical behaviors related to race, resulting in significant social repercussions. Specifically, the COMPAS algorithm, which predicts the risk of recidivism, was found to more frequently judge black defendants as being at a higher risk of recidivism than white defendants [8]. On the other hand, Google Photos and Face App provided discriminatory image detection and translation services with regard to certain races [9,10]. These issues have motivated researchers to investigate the topic of fairness in AI [11,12,13,14,15].
To this end, researchers established the definition of sensitive attributes (e.g., gender, race, age, and region) as the characteristics by which people should not be discriminated, and introduced the definition of fairness (e.g., demographic parity [11], equal opportunity, and equalized odds [16]). On top of that, researchers have proposed several methods to improve fairness by generating a fair dataset [17,18,19], excluding (or suppressing) sensitive attribute information in representation learning [20,21,22,23], and reweighting sample importance [24,25].
However, most of these methods require additional labels for sensitive attributes to ensure fairness, which leads to several limitations [26,27]. First, many popular benchmark datasets in the research community do not contain labels for sensitive attributes [1,28,29]. Moreover, since sensitive attributes are often related to personal information, labeling them on the existing datasets can be ethically risky and may require the data providers’ consent [26]. This is a significant obstacle to the deployment of existing fairness-aware methods in the real world.
To this end, recent studies have endeavored to alleviate model bias without demographics. Many of these works, particularly those based on debiasing approaches, tackle this challenge by capitalizing on specific assumptions about the characteristics of sensitive attributes or unwanted biases. For instance, they assume texture or color biases [30,31,32], a malignant attribute [33,34,35,36], and a high correlation with their proxies [27,37,38], allowing the effective mitigation of model bias without demographics. However, since sensitive attributes can be diversely defined based on social consensus and context, they may neither be malignant attributes nor have a substantial correlation with proxies. Consequently, these assumptions impose limitations on achieving universal fairness for potential sensitive attributes. On the other hand, some studies [39,40,41,42] do not rely on such assumptions; however, they suffer from significant performance degradation [39,40] and additionally require a small validation set with sensitive attribute labels [41] or a complex teacher network [42].
To address these limitations, we introduce a simple but effective reweighting method that enhances fairness without strong assumptions regarding sensitive attributes. To ameliorate fairness, conventional reweighting approaches [25,39,43,44] utilize predefined sensitive information to identify underrepresented groups and assign higher weights to them than to others. However, our method can assign higher weights to underrepresented groups without any knowledge of sensitive attributes.
In contrast to defining underrepresentation based solely on group size, we define it as performance-wise underrepresentation. For instance, a minority subgroup with clear and consistent features may achieve a low error rate and thus is not considered underrepresented under our definition. Conversely, a majority subgroup with substantial intra-group variability may exhibit poor performance and is treated as underrepresented. Motivated by this performance-based view, we propose to dynamically cluster samples into two groups, i.e., misclassified and correctly classified samples, based on the outputs of the current model, and subsequently assign a higher weight to the misclassified group. Since performance-wise underrepresented groups inherently contain a higher proportion of misclassified samples than their counterparts, a higher average weight is effectively given to such groups across any potential sensitive attributes.
On benchmark datasets for fair classification (i.e., CelebA [2], UTK Face [45], and COMPAS [8]), we extensively validate the effectiveness of the proposed method. Our method not only achieves state-of-the-art performances under diverse experimental settings but exhibits greater improvement in more severely unfair environments. The main contributions of this paper are summarized as follows.
  • We propose a novel method that dynamically groups data samples based on the current outputs and assigns more weights to the group with misclassified samples.
  • Regardless of the choice of sensitive attributes, the proposed method gives more weights to underrepresented groups, resulting in enhancing fairness across unknown sensitive attributes.
  • Without strong assumptions about sensitive attributes and auxiliary networks, our method significantly outperforms state-of-the-art methods on the benchmark datasets.

2. Related Work

2.1. Fairness-Aware Classification

In the literature [20,23,46,47], it has been indicated that learning sensitive attributes during the training phase is a major cause of unfairness. To solve the problem, prior works have attempted to suppress the learning of unnecessary biases in the training process or eliminate them in the test phase. For instance, some approaches employed the gradient reversal layer (GRL) to prevent the model from learning sensitive attributes [46,48,49]. Meanwhile, others proposed to directly disentangle the representation for target and sensitive attributes in the feature space [20,21,47]. On the other hand, generative approaches have been proposed [17,18,19] that synthesize a fair dataset based on a generative adversarial network (GAN) [50]. Lastly, some works enhanced fairness by upweighting or upsampling the data of minority groups [24,25,43,44].

2.2. Fairness/Debiasing Without Bias Supervision

To mitigate model bias without prior knowledge, many studies have relied on certain assumptions about the characteristics of sensitive attributes and biases. Some of them [30,31,32] assumed specific types of biases (e.g., texture and color) and designed an auxiliary network specialized to those biases. Others [33,34,35,36] grouped biases into two categories, i.e., malignant and benign biases, based on their relationship with the target attribute. They aimed at handling only the malignant biases, while the benign ones were left untouched. Meanwhile, others [27,37,38] assumed that proxies, such as clusters and feature representations, have a high correlation with sensitive attributes, and leveraged them as surrogates for the labels.
On the other hand, several works [39,40,51] adopted minimum–maximum fairness that minimizes the risk of the worst group. While they effectively enhance fairness by mitigating the performance gaps across all groups, they may experience a degradation in performance. Just Train Twice (JTT) [41] is similar to ours in that it upweights the misclassified samples. However, it constructs a fixed dataset by oversampling the misclassified samples with a pre-trained classifier. Subsequently, a new classifier is trained fairly using this dataset. Since the dataset fails to reflect the evolving classifier, it may not guarantee assigning more weight to minority groups for potential sensitive attributes. Furthermore, it needs a small validation set with sensitive attribute labels to tune the hyperparameters. Chai, Jang, and Wang [42] improved fairness without demographics by distilling knowledge from a complex teacher. Some works [52,53] proposed new approaches to enhance fairness in self-supervised settings.
Additionally, recent studies have proposed diverse perspectives to improve fairness without explicit bias supervision. Shared Latent Space-based Debiasing (SLSD) [54] leverages sensitive attribute information in an auxiliary source domain to learn a shared latent space between the source and target domains, and applies adversarial debiasing so that sensitive information is not encoded in the target model. However, since it relies on sensitive attribute labels in the source domain, it differs from methods that assume no access to sensitive labels. VFair [55] adopts a Rawlsian perspective that aims to improve the worst group performance without demographic labels, mitigating latent group disparities by minimizing the variance of losses across samples. Nonetheless, it reports that its effectiveness may be limited in classification tasks due to quantized utility measurements. The Graph of Gradients (GoG) [56] constructs a graph of gradients to learn sample weights by propagating weights to harder instances through graph neighborhoods. However, it incurs additional overhead due to gradient extraction and graph construction, and the reported improvements in the fairness–accuracy trade-off are sometimes modest.

3. Method

  • Problem Definition. Consider a data sample x comprising a target attribute y { c 1 , c 2 , , c n } and unknown sensitive attributes S { s 1 , s 2 , , s m } . During the training of a classifier parameterized by θ to predict y, the model may capture biased features associated with sensitive attributes as a shortcut to minimize the average training loss. Our goal is to train a classifier that is invariant to the sensitive attribute s S , without access to sensitive attribute labels, as follows.
    arg min θ E ( x , y ) D [ L ( θ ( x ) , y ) ] s . t . I ( θ ( x ) ; s ) = 0
    Here, D denotes the training data distribution and E ( x , y ) D [ · ] denotes the expectation over samples ( x , y ) from D . L ( · , · ) denotes the classification loss and I ( · ; · ) denotes mutual information.
In this work, we adopt equalized odds (EO) [16] for the definition of fairness, as follows:
P s ˙ ( y ˜ = c ˜ | y = c ) = P s ¨ ( y ˜ = c ˜ | y = c ) , s ˙ , s ¨ , c , c ˜ ,
where s ˙ , s ¨ S and c , c ˜ y . Here, y ˜ denotes the predicted label of the classifier. P s ˙ ( · ) and P s ¨ ( · ) denote the conditional probabilities for samples in the subgroups with s ˙ and s ¨ , respectively. This condition requires that, for any two sensitive groups s ˙ and s ¨ , the conditional distribution of predictions given y = c is identical across groups.
  • Overall Framework. As illustrated in Figure 1, we design the overall framework using Bias Pseudo-Attribute (BPA) [27] as a baseline. The major difference between the two frameworks is that BPA calculates fixed weights for samples using clusters from the pre-trained classifier, whereas ours dynamically updates the sample weights based on the outputs y ˜ of the current training model.

3.1. Strategy for Sample Grouping

In the previous work [27], BPA clusters training samples in the feature space of a biased model to estimate sensitive pseudo-attributes, assuming a high correlation between the target and sensitive attributes. Specifically, they pre-train a biased classification model θ ¯ with target labels y and perform k-means clustering for samples with the same target class c, which is formulated as k y = k-means ( θ ¯ , x , y ) . Subsequently, the samples are assigned into groups g i = { ( x , y ) | i = y × k y } based on the obtained clusters. Within their assumption, the groups effectively identify sensitive pseudo-attributes; however, they may encounter challenges in identifying them when the target and potential sensitive attributes are not sufficiently correlated.
To overcome this limitation, we first suppose the existence of underrepresented groups U ( y , S ) and overrepresented groups O ( y , S ) for the target attribute y and unknown sensitive attributes S. We define the underrepresented groups in terms of performance-wise underrepresentation, referring to subgroups that underperform under the current model, rather than being defined solely by their sample counts. Since U ( y , S ) exhibits lower classification performances than O ( y , S ) , U ( y , S ) has a higher proportion of misclassified samples. Therefore, assigning more loss weights to misclassified samples than to the other samples leads to attributing greater loss weights to U ( y , S ) than O ( y , S ) , regardless of the characteristics of the sensitive attributes S. From this motivation, we propose a new strategy for grouping samples. Specifically, our approach dynamically clusters samples into binary groups g i ( i { 0 , 1 } ) based on the prediction y ˜ of a training classification model θ as follows.
g i = { ( x , y ) | i = 1 ( y ˜ y ) } ,
where 1 ( · ) is the Boolean indicator function. Subsequently, we assign higher loss weights to the group with misclassified samples (i.e., g 1 ) than the other group with correctly classified samples (i.e., g 0 ). Consequently, we can upweight the underrepresented groups U ( y , S ) by utilizing the misclassified samples as a proxy. This leads to enhancing fairness by diminishing the accuracy gaps between U ( y , S ) and O ( y , S ) for any potential sensitive attributes.

3.2. Calculation of Group-Wise Weight

With regard to reweighting the grouped samples, we employ a smoothed version of Group Distributionally Robust Optimization (Group DRO) [24] proposed in [27]. When a target attribute y and sensitive attributes S are given, Group DRO [24] defines groups g G = { 1 , 2 , , m } , where m = | y | × | S | . Here, | · | denotes the cardinality of a set. It assumes that the overall data distribution P is a mixture of the group distributions { P g } g G , and minimizes the empirical worst group risk R ( θ ) as follows.
θ ^ D R O = arg min θ R ( θ ) : = max g G E ( x , y ) P ^ g [ l ( x , y ; θ ) ] ,
where P ^ g is the empirically approximated distribution over training data for the group distribution P g , and l ( · ) is a loss function.
On the other hand, BPA [27] cannot access S during training, and it substitutes the bias with the clusters. In addition, instead of minimizing the worst group risk, it assigns different weights w g i to the groups as follows:
θ ^ B P A = arg min θ R ( θ ) : = E ( x , y ) P [ w g i l ( x , y ; θ ) ] .
The importance weight w g i is proportional to the average loss of the group and inversely proportional to the size of that as follows.
w g i = E ( x , y ) g i [ l ( x , y ; θ ) ] | g i | ,
where | g i | denotes the cardinality of group g i .
In our method, the group with misclassified samples (i.e., g 1 ) has a higher average loss than the other group (i.e., g 0 ). It is intuitive and we provide empirical support in Figure 2. In addition, as the classification model becomes more proficient during the training process, the size of g 1 decreases relatively in comparison to that of g 0 . Therefore, both of these factors are required to ensure the effective reweighting of the groups and assign a greater weight to g 1 than g 0 . This adaptive mechanism eliminates the need for the tuning of the weighting ratio, which is typically infeasible when sensitive attribute labels are unavailable. Consequently, our method automatically balances the relative emphasis between groups without requiring additional hyperparameter tuning.

3.3. Algorithm for Overall Flow

To clarify the overall procedure of the proposed method, we present Algorithm 1. Here, D denotes the training dataset, E is the total number of epochs, B is the batch size, i t e r is the mini-batch iteration index within each epoch, η is the learning rate, and l ( · ) denotes the loss function. y ˜ k denotes the predicted label for sample x k . g i ( i { 0 , 1 } ) denotes the binary groups defined by misclassification, and w g i denotes the corresponding group weight.
Algorithm 1 Overall Flow
Require: 
Training data D, classifier θ , loss function l, total epochs E, batch size | B | , learning rate η
1:
Initialize θ
2:
for   e = 1 , 2 , , E   do
3:
   while  i t e r | D | | B |  do
4:
     Randomly sample a batch { ( x j , y j ) } j = 1 B
5:
     if  e = 1  then
6:
         θ θ η 1 B j = 1 B l ( x j , y j ; θ )
7:
     else
8:
         θ θ η 1 B j = 1 B w g i l ( x j , y j ; θ )
9:
     end if
10:
      i t e r = i t e r + 1
11:
   end while
12:
    g i = { ( x k , y k ) | i = 1 ( y k ˜ y k ) } for k = 1 , , | D |
13:
    w g i = E ( x , y ) g i [ l ( x , y ; θ ) ] | g i |
14:
end for
During the first epoch, we train the classification model with the standard cross-entropy loss. With the updated model θ , we group the training samples into g i and calculate the group-wise loss. Based on the loss and the size of groups, we calculate the weight w g i . In the next epoch, we fairly train the classification model with the weighted loss for the groups. This process is repeated until the end of training.

4. Experiments

4.1. Datasets

  • CelebA contains about 200k face images with 40 facial attributes. Following the previous works [23,27,33], we set Male as the sensitive attribute for evaluation. Following the convention [27], among the remaining attributes, we exclude 5 o’clock Shadow, Bald, Rosy Cheeks, Sideburns, Goatee, Mustache, and Wearing Necktie, for which minority groups contain few or no samples.
  • UTK Face includes about 20k face images where Gender, Race, and Age attributes are annotated. We set Gender as the target attribute and the others as the sensitive attributes. Following the previous work [23], we convert Race and Age into binary attributes and construct an imbalanced training set. Specifically, we construct three subsets according to the severity of data imbalance γ . Here, γ represents the ratio between the majority and minority groups, ranging from two to four with a step size of one. We construct the validation and test sets to be fully balanced.
  • COMPAS includes about 7k samples with 11 attributes. Following the convention [42], we only utilize Caucasian and African-American samples and set Race and Sex as sensitive attributes.
  • Cat and Dog has 40k images of dogs or cats. Following the previous method [22], we set species as the target attribute and color as the bias attribute. We construct a training set in which dogs and cats are correlated with white and black color, respectively, following the setup of the previous work [23]. The validation and test sets are fully balanced.

4.2. Evaluation Metrics

We evaluate the performance of the models based on three metrics. The first is balanced accuracy (BAcc.) to measure classification performance, which is defined as the average accuracy across all groups defined by both target and sensitive attributes. This metric evaluates the generalized performance of the models [22,27,33]. Regarding fairness, there are a range of conventional metrics for fairness, such as demographic parity [11], equal opportunity, and equalized odds [16]. Among them, we adopt equalized odds (EO) to consider the overall distribution of errors across diverse demographic groups. It is noted that demographic parity and equal opportunity are not chosen since they concentrate only on the true positive rates and predicted positive rates, respectively. Lastly, we also report the standard deviation of group-wise accuracy as a supplementary metric of EO. Specifically, we can detect undesirable situations, such as the disproportionate favoritism of the model toward a single class, via an extremely high standard deviation.

4.3. Comparison on COMPAS

We report the comparison results on COMPAS for sensitive attributes Race and Sex in Table 1 and Table 2, respectively. Focal Loss [57] effectively improves accuracy by upweighting hard examples. However, since it does not explicitly address performance gaps across sensitive groups, it yields only limited improvements in equalized odds (EO). Distributionally Robust Optimization (DRO) [39] and Adversarially Reweighted Learning (ARL) [40] significantly improve fairness over the baseline without strong assumptions about sensitive attributes. However, they show a lower trade-off performance between accuracy and EO than the other methods for fairness. FairRF [38], which is a proxy-based approach, and Chai, Jang, and Wang [42] exhibit superior trade-off performances. Without strong assumptions and auxiliary networks, ours achieves the best trade-off performance.
  • Comparison with Boosting. Boosting methods, such as Adaboost [58] and Gradient Boosting [59], have a similarity with ours in that they upweight misclassified samples during the training phase. However, there are significant differences. First, our method trains a single model by dynamically reweighting samples based on the outputs of the previous training stage while boosting methods iteratively train new weak learners with sample weights based on the outputs of previous learners. Second, the sample weights for each weak learner are not dynamically updated in boosting methods. Lastly, the methods for calculating weights are different from each other. We compare ours with the boosting approaches in Table 1. While boosting methods effectively improve classification accuracy over the baseline, they fail to enhance fairness.

4.4. Comparison on CelebA

To validate the proposed method under various scenarios, we conduct experiments with 32 target attributes on CelebA, excluding five attributes for reliable evaluation. In Table 3, we report the summarized results, along with the cases where EO of ResNet [3] exceeds 10 and 20 to analyze the effectiveness in an environment with severe unfairness. Group DRO [24] utilizes sensitive attribute labels during the training phase, and thus it serves as an upper bound for the other comparative methods. Learning from Failure (LfF) [33], Disentangled Feature Augmentation (DFA) [34], Bias Pseudo-Attribute (BPA) [27], and ours significantly improve both balanced accuracy (i.e., BAcc.) and equalized odds (i.e., EO) without sensitive attribute labels. Ours not only achieves the fairest results among them but also shows larger gains as unfairness becomes more severe (i.e., EO ≥ 10 and EO ≥ 20). In contrast to those, Just Train Twice (JTT) utilizes additional sensitive attribute labels during the validation process; nevertheless, ours outperforms it in terms of BAcc. and EO. In addition, we report the standard deviation (Std Dev.) to detect a shortcut, such as a model consistently predicting only one class, since it leads to extremely high standard deviation. For example, it is observed that the baseline shows extremely high standard deviation for several attributes, including Big Lips, Chubby, and Wearing Necklace. We include the full results tables in Table 4 and Table 5.

4.5. Comparison on UTKFace

In Table 6 and Table 7, we compare the effectiveness of models under different levels of data imbalance. As γ increases, ResNet shows a significant decrease in fairness. BPA and DFA significantly enhance fairness for the target attribute Age, while having a limited impact on the other target attribute Race. In contrast, LfF notably mitigates unfairness for Age; however, it suffers from a substantial performance drop. Ours largely improves fairness over the baseline and achieves the best trade-off performances in all the settings. As γ increases, ours exhibits a more significant improvement in fairness compared to the baseline. Similar to the results on CelebA, this indicates that our method has greater efficacy in addressing severe unfair scenarios.

4.6. Exploring Assumptions for Sensitive Attributes

As mentioned above, many previous works rely on assumptions about sensitive attributes to enhance fairness without demographics. In particular, it is popularly assumed that the sensitive attribute is a malignant attribute or is highly correlated with proxies. In this section, we analyze some scenarios where such assumptions do not hold. Specifically, we compare our method with the representative methods corresponding to each assumption, namely LfF [33] and BPA [27], for four target attributes: Attractive, Wearing Earrings, Chubby, and Brown Hair. Regarding Attractive, the sensitive attribute (i.e., Male) is a malignant attribute yet is less correlated with the proxy (i.e., target attribute). In contrast, for Wearing Earrings, it is not a malignant attribute and has a high correlation with the target attribute. For Chubby and Brown Hair, it is neither a malignant attribute nor highly correlated with the target attributes.
For all the scenarios, we measure the average weights of majority and minority groups for the comparative methods. It is noted that all the weights are normalized for better visualization. Figure 3 shows that LfF effectively gives more weight to minority groups for Attractive, where the sensitive attribute is a malignant attribute. However, it instead upweights majority groups for Wearing Earrings where it is not a malignant attribute. In contrast, BPA gives very similar weight to both majority and minority groups for Attractive, which has a low correlation with the sensitive attribute, while it preferentially upweights the minority group for Wearing Earrings. For Chubby and Brown Hair, both LfF and BPA fail to upweight the minority groups. As we claimed that the proposed method can attribute a higher average weight to the underrepresented groups regardless of the choice of sensitive attributes, it effectively assigns greater weight to minority groups in all the scenarios.

4.7. Effectiveness of Components

In this section, we analyze the effectiveness of our design through a comparison between different clustering strategies and an ablation study. For this purpose, we prepare several competitors using different clustering strategies.
  • BPA* is a modified version of BPA [27] that dynamically updates the clusters in the latent space of the debiased model with k-means clustering. In the first epoch, it generates clusters using k-means clustering in the latent space of the pre-trained model. In subsequent epochs, the clusters are updated with the same clustering method in the latent space of the current training model at each epoch. Based on the updated clusters, it utilizes the reweighting method such as BPA.
  • Random randomly clusters samples for each epoch. For each batch, it clusters the input samples into two groups randomly. For each cluster, samples are differently weighted based on the average loss and size of the respective groups, as in our method.
  • Loss sorts samples by loss and subsequently separates them into groups based on their ranking. At each epoch, it computes the training loss of all samples and separates them into two groups based on the loss, i.e., one group consists of samples with high loss, whereas the other group comprises samples with low loss. Specifically, we set the threshold to the median value of the loss and then assign weights to the samples in proportion to the average loss of the respective groups. If the groups are more finely separated, it converges toward Instance, as shown in Table 8.
  • EO first composes clusters by k-means clustering similar to BPA. In subsequent epochs, multiple candidate groups are generated by permuting samples between the groups with a probability of p = 0.3 . These candidate groups are evaluated in terms of equalized odds, and the groups with the largest equalized odds between them are selected. We then assign weights to the groups with our reweighting strategy.
  • Instance computes the training loss for all samples using the current training model and assigns a weight to each sample in proportion to the loss. These weights are then normalized by the batch so that their sum equals one.
  • Ours is an ablative version of our method. In the first epoch, it clusters samples based on whether they are correctly classified or not (i.e., the proposed method). However, it does not update the clusters.
For the comparison, we choose two target attributes that cause severe unfairness on CelebA, i.e., Attractive and Arched Eyebrows. As reported in Table 9, BPA improves balanced accuracy over the baseline for Arched Eyebrows, which is highly correlated with the sensitive attribute, but it is not very effective in terms of EO. Dynamic updating (i.e., BPA*) slightly improves both balanced accuracy and EO over BPA for both target attributes. The results of Random indicate that randomly reweighting samples can be somewhat effective in improving the generalized performance. On the other hand, other strategies (i.e., Loss, EO, and Instance) exhibit limited improvements in terms of fairness. In particular, Instance is vulnerable to outliers, such as mislabeled data, as it assigns the largest weights to extreme examples [40], as shown in Table 10. We note that this experiment is intended to compare sensitivity to label noise between the proposed method and instance-level weighting, rather than to claim absolute robustness. In addition, the results of ours suggest that dynamically updating the clusters is highly important in our method. Lastly, ours, with dynamic clustering, achieves the fairest performance among the competitors, demonstrating its effectiveness.
Table 8. Relationship between Loss and Instance. We respectively set Arched Eyebrows and Male as the target and sensitive attributes.
Table 8. Relationship between Loss and Instance. We respectively set Arched Eyebrows and Male as the target and sensitive attributes.
2 Groups4 Groups8 GroupsInstance
BAcc.EOBAcc.EOBAcc.EOBAcc.EO
76.732.975.929.374.128.471.423.8
Table 9. Analysis on clustering strategy and ablation study. We exploit two target attributes, Attractive and Arched Eyebrows, which respectively have low and high correlation with the sensitive attribute, Male. Updating indicates whether the initial clusters are fixed or dynamically updated.
Table 9. Analysis on clustering strategy and ablation study. We exploit two target attributes, Attractive and Arched Eyebrows, which respectively have low and high correlation with the sensitive attribute, Male. Updating indicates whether the initial clusters are fixed or dynamically updated.
MethodUpdatingAttractiveArched Eyebrows
Balanced AccuracyEqualized OddsStd Dev.Balanced AccuracyEqualized OddsStd Dev.
ResNet76.725.615.570.833.827.0
BPA76.424.116.673.634.331.9
BPA*77.721.213.577.128.521.6
Random77.721.513.774.938.923.6
Loss78.018.210.576.732.923.2
EO77.022.013.275.734.627.1
Instance76.123.219.971.423.824.2
Ours 77.423.614.274.438.222.2
Ours75.56.49.375.915.320.5
Table 10. Robustness to mislabeled samples. We set Race and Gender as the target and sensitive attributes on UTKFace. We deliberately assign incorrect target labels to 10% or 20% of the training data.
Table 10. Robustness to mislabeled samples. We set Race and Gender as the target and sensitive attributes on UTKFace. We deliberately assign incorrect target labels to 10% or 20% of the training data.
MethodClean10% Noised20% Noised
BAcc.EOBAcc.EOBAcc.EO
Instance79.416.475.425.068.233.3
Ours80.99.575.214.168.617.0

4.8. Fairness Improvement in Semi-Supervised Setting

When access to sensitive attributes is not completely prohibited, we demonstrate that a significant improvement in terms of fairness can be achieved by annotating sensitive attribute labels for a limited number of training data. In this semi-supervised setting, we train a classification model by utilizing Group DRO [24] for data with sensitive attribute labels and our method for unlabeled data. As shown in Table 11, our method shows a substantial improvement in fairness with just 10% of labeled data, achieving comparable EO to the fully-supervised Group DRO, as the amount of the labeled data increases.

4.9. Analysis on Cat and Dog

Differing from CelebA and UTK Face with sensitive facial attributes, Cat and Dog includes a more general bias, i.e., color. In Table 12, we demonstrate that the proposed method can effectively reduce different types of biases. For all the comparable methods, both balanced accuracy and equalized odds are significantly improved over the baseline, and ours achieves the best trade-off performance. In particular, the equalized odds of the proposed method is approximately half that of the second-best method.

4.10. Implementation Details

For a fair comparison, all comparative models are implemented using ResNet-18 [3] as the encoder network and a single fully connected layer as the classifier. We utilize the cross-entropy loss as the classification loss l ( · ) . Following the previous work [27], we adopt an improved version of LfF [33] for comparison. For DFA [34], we search the hyperparameters in the range of {0.1, 1, 5, 10}. For BPA [27], we set the number of clusters k to four since the performance saturates afterward. We follow the original implementation for the other hyperparameters. For JTT [41] and GDRO [24], we strictly follow the original settings from the papers. Our method is built based on BPA [27] and optimized by Adam [60] with a learning rate of 10 4 . For all models, we choose the best model regarding the balanced accuracy on the validation set for 30 epochs.

5. Limitations and Future Work

The proposed method significantly improves fairness across diverse datasets without demographic information. However, a performance gap still remains compared to fully supervised approaches that leverage sensitive attribute labels (e.g., Group DRO). In addition, since our debiasing strategy relies on whether a sample is misclassified, it can be sensitive to label noise. This limitation could be mitigated by incorporating robust learning methods for label noise. Lastly, since we dynamically regroup training samples and compute the corresponding group-wise weights at every epoch, our method incurs additional training time and computational overhead. For example, on UTKFace, the proposed method requires 1.08× the training time of the baseline (ResNet), on average. Nevertheless, our method provides a favorable trade-off by substantially improving fairness with a relatively small additional cost. Future work can explore more efficient group-wise weight update procedures to reduce this overhead.

6. Conclusions

In this paper, we aimed to enhance fairness in classification tasks without demographic information. First, we pointed out the limitations of prior approaches, including the strong assumption about the characteristics of sensitive attributes and reliance on the auxiliary information or networks. To overcome these limitations, we proposed a dynamic group-wise reweighting method that enhances fairness without these requirements. Our approach assigns greater weights to underrepresented groups for potential sensitive attributes by utilizing misclassified samples as a proxy. Through extensive experiments, we empirically demonstrated the effectiveness of the proposed method under various scenarios. Moreover, our method achieved state-of-the-art performance in terms of fairness on benchmark datasets.

Author Contributions

Conceptualization, S.P.; methodology, S.P. and P.L.; investigation, S.P. and P.L.; writing—original draft preparation, S.P.; writing—review and editing, S.P. and P.L.; and supervision, S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Incheon National University Research Grant in 2024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

This work was supported by Incheon National University Research Grant in 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. (IJCV) 2015, 115, 211–252. [Google Scholar] [CrossRef]
  2. Liu, Z.; Luo, P.; Wang, X.; Tang, X. Deep Learning Face Attributes in the Wild. In Proceedings of International Conference on Computer Vision (ICCV); IEEE: New York, NY, USA, 2015. [Google Scholar]
  3. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: New York, NY, USA, 2016; pp. 770–778. [Google Scholar]
  4. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive Growing of GANs for Improved Quality, Stability, and Variation. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  5. Brandao, M. Age and gender bias in pedestrian detection algorithms. arXiv 2019, arXiv:1906.10490. [Google Scholar] [CrossRef]
  6. Kärkkäinen, K.; Joo, J. FairFace: Face Attribute Dataset for Balanced Race, Gender, and Age. arXiv 2019, arXiv:1908.04913. [Google Scholar] [CrossRef]
  7. Burns, K.; Hendricks, L.A.; Darrell, T.; Rohrbach, A. Women also Snowboard: Overcoming Bias in Captioning Models. In Proceedings of the European Conference on Computer Vision (ECCV); IEEE: New York, NY, USA, 2018. [Google Scholar]
  8. Angwin, J.; Larson, J.; Mattu, S.; Kirchner, L. Machine bias: There’s software used across the country to predict future criminals. And it’s biased against blacks. ProPublica, 23 May 2016. [Google Scholar]
  9. Dougherty, C. Google photos mistakenly labels black people gorillas. Twitter, 1 July 2015. [Google Scholar]
  10. Lomas, N. FaceApp apologizes for building a racist AI. TechCrunch, 25 April 2017. [Google Scholar]
  11. Dwork, C.; Hardt, M.; Pitassi, T.; Reingold, O.; Zemel, R. Fairness through Awareness. In ITCS ’12: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference; ACM: New York, NY, USA, 2012; pp. 214–226. [Google Scholar] [CrossRef]
  12. Kusner, M.J.; Loftus, J.; Russell, C.; Silva, R. Counterfactual Fairness. In Proceedings of the Advances in Neural Information Processing Systems 30; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 4066–4076. [Google Scholar]
  13. Wang, T.; Zhao, J.; Yatskar, M.; Chang, K.W.; Ordonez, V. Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias in Deep Image Representations. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV); IEEE: New York, NY, USA, 2019; pp. 5309–5318. [Google Scholar]
  14. Wang, Z.; Qinami, K.; Karakozis, I.C.; Genova, K.; Nair, P.; Hata, K.; Russakovsky, O. Towards Fairness in Visual Recognition: Effective Strategies for Bias Mitigation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2020. [Google Scholar]
  15. Gong, S.; Liu, X.; Jain, A. Jointly De-biasing Face Recognition and Demographic Attribute Estimation. In Computer Vision—ECCV 2020; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  16. Hardt, M.; Price, E.; Srebro, N. Equality of Opportunity in Supervised Learning. In NIPS’16: Proceedings of the 30th International Conference on Neural Information Processing Systems; ACM: New York, NY, USA, 2016; pp. 3323–3331. [Google Scholar]
  17. Ramaswamy, V.V.; Kim, S.S.Y.; Russakovsky, O. Fair Attribute Classification Through Latent Space De-Biasing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2021; pp. 9301–9310. [Google Scholar]
  18. Sattigeri, P.; Hoffman, S.C.; Chenthamarakshan, V.; Varshney, K.R. Fairness GAN: Generating datasets with fairness properties using a generative adversarial network. IBM J. Res. Dev. 2019, 63, 3:1–3:9. [Google Scholar] [CrossRef]
  19. Xu, D.; Yuan, S.; Zhang, L.; Wu, X. FairGAN: Fairness-aware Generative Adversarial Networks. In 2018 IEEE International Conference on Big Data (Big Data); IEEE: New York, NY, USA, 2018; pp. 570–575. [Google Scholar] [CrossRef]
  20. Creager, E.; Madras, D.; Jacobsen, J.H.; Weis, M.; Swersky, K.; Pitassi, T.; Zemel, R. Flexibly Fair Representation Learning by Disentanglement. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 1436–1445. [Google Scholar]
  21. Park, S.; Hwang, S.; Kim, D.; Byun, H. Learning Disentangled Representation for Fair Facial Attribute Classification via Fairness-awareunbiased Information Alignment. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 2403–2411. [Google Scholar]
  22. Kim, B.; Kim, H.; Kim, K.; Kim, S.; Kim, J. Learning Not to Learn: Training Deep Neural Networks With Biased Data. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2019. [Google Scholar]
  23. Park, S.; Lee, J.; Lee, P.; Hwang, S.; Kim, D.; Byun, H. Fair Contrastive Learning for Facial Attribute Classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2022; pp. 10389–10398. [Google Scholar]
  24. Sagawa, S.; Koh, P.W.; Hashimoto, T.B.; Liang, P. Distributionally Robust Neural Networks. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
  25. Kim, D.; Park, S.; Hwang, S.; Byun, H. Fair classification by loss balancing via fairness-aware batch sampling. Rev. Econ. Dyn. 2023, 518, 231–241. [Google Scholar] [CrossRef]
  26. Jung, S.W.; Chun, S.; Moon, T. Learning Fair Classifiers with Partially Annotated Group Labels. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2022; pp. 10338–10347. [Google Scholar]
  27. Seo, S.; Lee, J.Y.; Han, B. Unsupervised Learning of Debiased Representations With Pseudo-Attributes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2022; pp. 16742–16751. [Google Scholar]
  28. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  29. Heilbron, F.C.; Escorcia, V.; Ghanem, B.; Niebles, J.C. ActivityNet: A large-scale video benchmark for human activity understanding. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2015; pp. 961–970. [Google Scholar] [CrossRef]
  30. Wang, H.; He, Z.; Lipton, Z.L.; Xing, E.P. Learning Robust Representations by Projecting Superficial Statistics Out. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  31. Bahng, H.; Chun, S.; Yun, S.; Choo, J.; Oh, S.J. Learning De-biased Representations with Biased Representations. In Proceedings of the 37th International Conference on Machine Learning; PMLR: Cambridge, MA, USA, 2020. [Google Scholar]
  32. Hong, Y.; Yang, E. Unbiased Classification through Bias-Contrastive and Bias-Balanced Learning. In Proceedings of the Advances in Neural Information Processing Systems; Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2021. [Google Scholar]
  33. Nam, J.; Cha, H.; Ahn, S.; Lee, J.; Shin, J. Learning from Failure: De-biasing Classifier from Biased Classifier. In Proceedings of the Advances in Neural Information Processing Systems; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2020; Volume 33, pp. 20673–20684. [Google Scholar]
  34. Lee, J.; Kim, E.; Lee, J.; Lee, J.; Choo, J. Learning Debiased Representation via Disentangled Feature Augmentation. In Proceedings of the Advances in Neural Information Processing Systems; Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2021. [Google Scholar]
  35. Lim, J.; Kim, Y.; Kim, B.; Ahn, C.; Shin, J.; Yang, E.; Han, S. BiasAdv: Bias-Adversarial Augmentation for Model Debiasing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2023; pp. 3832–3841. [Google Scholar]
  36. Zhang, Y.K.; Wang, Q.W.; Zhan, D.C.; Ye, H.J. Learning Debiased Representations via Conditional Attribute Interpolation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2023; pp. 7599–7608. [Google Scholar]
  37. Grari, V.; Lamprier, S.; Detyniecki, M. Fairness without the Sensitive Attribute via Causal Variational Autoencoder. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, Vienna, Austria, 23–29 July 2022; Raedt, L.D., Ed.; IJCAI: Vienna, Austria, 2023; pp. 696–702. [Google Scholar] [CrossRef]
  38. Zhao, T.; Dai, E.; Shu, K.; Wang, S. Towards Fair Classifiers Without Sensitive Attributes: Exploring Biases in Related Features. In WSDM ’22: Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining; ACM: New York, NY, USA, 2022; pp. 1433–1442. [Google Scholar] [CrossRef]
  39. Hashimoto, T.; Srivastava, M.; Namkoong, H.; Liang, P. Fairness Without Demographics in Repeated Loss Minimization. In International Conference on Machine Learning; PMLR: Cambridge, MA, USA, 2018; pp. 1929–1938. [Google Scholar]
  40. Lahoti, P.; Beutel, A.; Chen, J.; Lee, K.; Prost, F.; Thain, N.; Wang, X.; Chi, E.H. Fairness without Demographics through Adversarially Reweighted Learning. In Advances in Neural Information Processing Systems; NIPS’20; Curran Associates, Inc.: Red Hook, NY, USA, 2020. [Google Scholar]
  41. Liu, E.Z.; Haghgoo, B.; Chen, A.S.; Raghunathan, A.; Koh, P.W.; Sagawa, S.; Liang, P.; Finn, C. Just Train Twice: Improving Group Robustness without Training Group Information. arXiv 2021, arXiv:2107.09044. [Google Scholar] [CrossRef]
  42. Chai, J.; Jang, T.; Wang, X. Fairness without Demographics through Knowledge Distillation. In Proceedings of the Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2022. [Google Scholar]
  43. Chai, J.; Wang, X. Fairness with Adaptive Weights. In Proceedings of the 39th International Conference on Machine Learning; Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., Sabato, S., Eds.; PMLR: Cambridge, MA, USA, 2022; Volume 162, pp. 2853–2866. [Google Scholar]
  44. Kamiran, F.; Calders, T. Data Pre-Processing Techniques for Classification without Discrimination. Knowl. Inf. Syst. 2011, 33, 1–33. [Google Scholar] [CrossRef]
  45. Zhang, Z.; Song, Y.; Qi, H. Age Progression/Regression by Conditional Adversarial Autoencoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2017. [Google Scholar]
  46. Raff, E.; Sylvester, J. Gradient Reversal against Discrimination: A Fair Neural Network Learning Approach. In Proceedings of the 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA); IEEE: New York, NY, USA, 2018; pp. 189–198. [Google Scholar] [CrossRef]
  47. Sarhan, M.H.; Navab, N.; Eslami, A.; Albarqouni, S. Fairness by Learning Orthogonal Disentangled Representations. In Proceedings of the Computer Vision—ECCV 2020—16th European Conference, Glasgow, UK, 23–28 August 2020; Volume 12374, pp. 746–761. [Google Scholar] [CrossRef]
  48. Zhang, B.H.; Lemoine, B.; Mitchell, M. Mitigating Unwanted Biases with Adversarial Learning. In AIES ’18: Proceedings of the Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society; ACM: New York, NY, USA, 2018; pp. 335–340. [Google Scholar] [CrossRef]
  49. Wadsworth, C.; Vera, F.; Piech, C. Achieving fairness through adversarial learning: An application to recidivism prediction. arXiv 2018, arXiv:1807.00199. [Google Scholar] [CrossRef]
  50. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 2672–2680. [Google Scholar]
  51. Lokhande, V.S.; Sohn, K.; Yoon, J.; Udell, M.; Lee, C.Y.; Pfister, T. Towards Group Robustness in the Presence of Partial Group Labels. In Proceedings of the ICML 2022: Workshop on Spurious Correlations, Invariance and Stability, Baltimore, MD, USA, 22 July 2022. [Google Scholar]
  52. Chai, J.; Wang, X. Self-Supervised Fair Representation Learning without Demographics. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2022. [Google Scholar]
  53. Jang, T.; Wang, X. Difficulty-Based Sampling for Debiased Contrastive Representation Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: New York, NY, USA, 2023; pp. 24039–24048. [Google Scholar]
  54. Islam, R.; Chen, H.; Cai, Y. Fairness without Demographics through Shared Latent Space-Based Debiasing. Proc. Aaai Conf. Artif. Intell. 2024, 38, 12717–12725. [Google Scholar] [CrossRef]
  55. Wang, X.; Li, J.; Tsang, I.W.; Ong, Y.S. Towards Harmless Rawlsian Fairness Regardless of Demographic Prior. In Advances in Neural Information Processing Systems; Globerson, A., Mackey, L., Belgrave, D., Fan, A., Paquet, U., Tomczak, J., Zhang, C., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2024; Volume 37, pp. 80908–80935. [Google Scholar] [CrossRef]
  56. Luo, Y.; Li, Z.; Liu, Q.; Zhu, J. Fairness without Demographics through Learning Graph of Gradients. In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining, V.1, KDD 2025, Toronto, ON, Canada, 3–7 August 2025; pp. 918–926. [Google Scholar] [CrossRef]
  57. Lin, T.; Goyal, P.; Girshick, R.B.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. arXiv 2017, arXiv:1708.02002. [Google Scholar] [CrossRef]
  58. Schwenk, H.; Bengio, Y. Training Methods for Adaptive Boosting of Neural Networks. In Advances in Neural Information Processing Systems; Jordan, M., Kearns, M., Solla, S., Eds.; MIT Press: Cambridge, MA, USA, 1997; Volume 10. [Google Scholar]
  59. Badirli, S.; Liu, X.; Xing, Z.; Bhowmik, A.; Doan, K.D.; Keerthi, S. Gradient Boosting Neural Networks: GrowNet. arXiv 2020, arXiv:2002.07971. [Google Scholar] [CrossRef]
  60. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
Figure 1. Overall flow of the proposed method. The proposed method dynamically updates the groups based on the output of the classification model. Subsequently, the updated groups are utilized to fairly train the model. These procedures are iterated cyclically. The solid and dotted lines, respectively, depict the flow of the current and next training steps. All notations in the figure follow the same definitions as in the text, including y ˜ , g i , and w g i .
Figure 1. Overall flow of the proposed method. The proposed method dynamically updates the groups based on the output of the classification model. Subsequently, the updated groups are utilized to fairly train the model. These procedures are iterated cyclically. The solid and dotted lines, respectively, depict the flow of the current and next training steps. All notations in the figure follow the same definitions as in the text, including y ˜ , g i , and w g i .
Applsci 16 01684 g001
Figure 2. Empirical validation on the assumption. We set Age and Gender as the target and sensitive attributes, respectively, on UTK Face.
Figure 2. Empirical validation on the assumption. We set Age and Gender as the target and sensitive attributes, respectively, on UTK Face.
Applsci 16 01684 g002
Figure 3. Group-wise average weight on CelebA. We set Attractive (a), Wearing Earrings (b), Chubby (c), and Brown Hair (d) as the target attributes. The sensitive attribute, i.e., Male, is a malignant attribute with low correlation with Attractive, whereas it is not a malignant attribute but highly correlated with Wearing Earrings. For Chubby and Brown Hair, the sensitive attribute is not a malignant attribute and has a low correlation with them.
Figure 3. Group-wise average weight on CelebA. We set Attractive (a), Wearing Earrings (b), Chubby (c), and Brown Hair (d) as the target attributes. The sensitive attribute, i.e., Male, is a malignant attribute with low correlation with Attractive, whereas it is not a malignant attribute but highly correlated with Wearing Earrings. For Chubby and Brown Hair, the sensitive attribute is not a malignant attribute and has a low correlation with them.
Applsci 16 01684 g003
Table 1. Results on COMPAS for Race. We report a comparison with the previous methods for boosting and fairness. * indicates that the marked results are taken from the previous work [42]. For a fair comparison, we also measure conventional classification accuracy for the other methods.
Table 1. Results on COMPAS for Race. We report a comparison with the previous methods for boosting and fairness. * indicates that the marked results are taken from the previous work [42]. For a fair comparison, we also measure conventional classification accuracy for the other methods.
MethodAccuracy (↑)Equalized Odds (↓)
ResNet [3] *64.138.3
Focal Loss [57]66.433.7
Adaboost [58]65.631.3
GrowNet [59]66.138.9
DRO [39] *62.630.4
ARL [40] *63.229.5
FairRF [38] *63.325.7
Chai, Jang, and Wang [42] *63.320.3
Ours63.416.0
Table 2. Results on COMPAS for Gender. * indicates that the marked results are taken from the previous work [42].
Table 2. Results on COMPAS for Gender. * indicates that the marked results are taken from the previous work [42].
MethodAccuracy (↑)Equalized Odds (↓)
ResNet [3] *64.120.2
Focal Loss [57]66.422.2
Adaboost [58]65.126.3
GrowNet [59]66.822.2
DRO [39] *62.718.8
ARL [40] *63.219.1
FairRF [38] *63.318.7
Chai, Jang, and Wang [42] *63.414.3
Ours63.39.4
Table 3. Results on CelebA. We set the sensitive attribute as Male and report the summarized classification results for 32 target attributes. SA indicates whether sensitive attribute labels were used during training. The best and second-best scores are emphasized in bold and underlined, respectively.
Table 3. Results on CelebA. We set the sensitive attribute as Male and report the summarized classification results for 32 target attributes. SA indicates whether sensitive attribute labels were used during training. The best and second-best scores are emphasized in bold and underlined, respectively.
MethodSAAll ResultsEO ≥ 10EO ≥ 20
BAcc. (↑)EO (↓)Std Dev. (↓)BAcc. (↑)EO (↓)Std Dev. (↓)BAcc. (↑)EO (↓)Std Dev. (↓)
ResNet75.020.926.172.726.629.272.933.930.5
Group DRO81.54.54.679.64.95.278.06.16.8
JTT77.215.415.974.918.917.273.822.419.1
LfF78.117.815.476.120.117.174.722.818.9
DFA76.215.521.274.218.623.274.022.923.2
BPA79.916.713.778.220.315.376.723.617.3
Ours79.311.210.777.413.211.276.816.112.8
Table 4. BAcc. on CelebA. We set Male as the sensitive attribute. The reported results for 32 target attributes are separated based on whether algorithmic bias is present with the sensitive attribute (above) or not (below). We clarify that JTT [41] utilizes additional sensitive attribute labels during the validation procedure, and Group DRO [24] uses sensitive attribute labels during training.
Table 4. BAcc. on CelebA. We set Male as the sensitive attribute. The reported results for 32 target attributes are separated based on whether algorithmic bias is present with the sensitive attribute (above) or not (below). We clarify that JTT [41] utilizes additional sensitive attribute labels during the validation procedure, and Group DRO [24] uses sensitive attribute labels during training.
Target AttributeNo Sensitive LabelsSensitive Labels Required
ResNetLfFDFABPAOursJTTGroup DRO
Arched Eyebrows70.864.075.973.675.972.774.9
Bags Under Eyes73.172.369.475.672.561.175.5
Bangs89.293.590.991.293.094.396.8
Big Lips58.461.962.370.666.451.763.5
Big Nose67.368.470.370.471.068.672.5
Blond Hair79.287.684.183.586.885.691.1
Blurry76.483.976.988.587.688.689.2
Brown Hair75.486.382.182.584.083.984.0
Bushy Eyebrows77.183.478.783.880.983.883.6
Chubby64.175.466.076.970.381.481.6
Double Chin64.275.765.783.083.784.184.7
Gray Hair74.983.977.083.780.089.993.9
Heavy Makeup71.871.474.877.975.471.375.6
Narrow Eyes76.276.274.576.476.675.677.1
No Beard72.177.073.471.577.078.081.8
Oval Face62.460.162.964.162.162.664.6
Pale Skin71.686.479.089.586.387.791.1
Pointy Nose62.264.863.465.268.060.468.7
Receding Hairline74.082.678.082.079.284.466.0
Straight Hair68.665.964.373.772.073.176.0
Wavy Hair75.773.176.578.875.769.581.0
Wearing Hat87.993.192.289.889.998.097.5
Wearing Earrings74.270.473.281.181.261.383.0
Wearing Lipstick73.174.671.179.180.071.080.7
Wearing Necklace53.562.456.265.756.652.066.3
Young79.777.778.178.378.166.878.7
Attractive76.876.475.376.575.572.978.0
High Cheekbone83.684.582.884.383.175.583.6
Black Hair87.286.885.983.586.884.087.6
Mouth Slightly Open93.292.994.192.293.587.392.2
Eyeglasses96.897.197.294.498.098.498.8
Smiling90.690.891.190.492.389.490.4
Average75.078.176.279.979.377.281.5
Table 5. EO and STD Dev. on CelebA. We set Male as the sensitive attribute. We clarify that JTT [41] utilizes additional sensitive attribute labels during the validation procedure, and Group DRO [24] uses sensitive attribute labels during training.
Table 5. EO and STD Dev. on CelebA. We set Male as the sensitive attribute. We clarify that JTT [41] utilizes additional sensitive attribute labels during the validation procedure, and Group DRO [24] uses sensitive attribute labels during training.
Target AttributeEqualized Odds (↓)Std Dev. (↓)
No Sensitive LabelsSensitive Labels RequiredNo Sensitive LabelsSensitive Labels Required
ResNetLfFDFABPAOursJTTGroup DROResNetLfFDFABPAOursJTTGroup DRO
Arched Eyebrows33.932.730.334.415.328.16.627.031.923.231.920.517.44.2
Bags Under Eyes44.021.116.220.115.937.96.627.017.926.015.417.232.14.5
Bangs2.43.77.64.42.32.76.310.72.47.08.81.63.23.7
Big Lips11.738.913.216.013.820.99.846.227.441.012.010.745.014.6
Big Nose23.818.623.233.817.920.52.330.813.515.820.710.927.32.5
Blond Hair30.614.918.511.66.04.82.329.812.918.716.04.23.21.6
Blurry6.76.74.95.65.44.73.225.06.123.28.37.34.02.6
Brown Hair18.47.14.79.44.34.31.714.93.13.48.05.98.12.2
Bushy Eyebrows23.56.914.89.75.77.93.125.95.820.45.73.612.72.3
Chubby15.128.013.929.125.026.62.841.221.837.517.314.615.42.5
Double Chin15.225.011.422.420.018.25.141.223.138.214.111.710.93.6
Gray Hair13.116.916.111.26.513.73.829.917.927.520.38.38.02.9
Heavy Makeup44.347.245.039.338.827.825.636.632.126.122.722.621.823.9
Narrow Eyes27.13.64.83.43.97.51.932.02.226.29.56.619.92.0
No Beard44.932.739.850.332.437.014.341.625.733.129.218.721.415.1
Oval Face22.016.517.318.014.821.03.832.121.321.314.412.714.15.0
Pale Skin15.08.92.93.83.89.40.833.512.321.18.212.65.60.6
Pointy Nose22.524.721.229.010.533.42.736.332.131.823.39.733.35.4
Receding Hairline45.016.512.213.913.314.83.627.911.222.113.420.48.74.3
Straight Hair9.89.64.16.66.37.11.726.231.133.09.612.47.02.0
Wavy Hair18.44.217.814.93.314.31.622.013.024.310.54.710.16.8
Wearing Hat23.43.65.28.87.90.60.515.04.87.65.35.10.51.0
Wearing Earrings43.239.934.422.618.138.64.631.326.820.713.411.931.52.7
Wearing Lipstick47.441.338.236.625.733.78.033.727.232.922.415.523.810.7
Wearing Necklace0.857.721.114.717.422.010.852.734.145.618.942.137.09.1
Young17.517.417.018.13.723.41.316.110.217.712.26.919.11.9
Attractive25.711.521.624.16.45.00.915.510.116.516.69.36.11.6
High Cheekbone12.95.07.36.74.65.14.815.06.25.84.85.59.44.9
Black Hair5.05.05.611.64.29.51.13.04.36.716.05.28.42.1
Mouth Slightly Open2.01.90.90.62.30.50.22.91.61.22.91.58.00.1
Eyeglasses1.33.02.03.41.31.30.83.62.02.62.60.90.80.6
Smiling3.92.74.02.02.53.12.19.02.32.84.21.72.51.3
Average20.917.815.516.711.215.44.526.115.421.213.710.715.94.6
Table 6. Results on UTK Face for target attribute Age. We report the experimental results on three subsets of UTK Face. Severity ( γ ) represents the degree of data imbalance. We set Age and Gender as the target and sensitive attributes respectively. As γ increases, the fairness of ours is more improved over the baseline.
Table 6. Results on UTK Face for target attribute Age. We report the experimental results on three subsets of UTK Face. Severity ( γ ) represents the degree of data imbalance. We set Age and Gender as the target and sensitive attributes respectively. As γ increases, the fairness of ours is more improved over the baseline.
MethodSeverity ( γ ) = 2 Severity ( γ ) = 3 Severity ( γ ) = 4
BAcc. (↑)EO (↓)Std Dev. (↓)BAcc. (↑)EO (↓)Std Dev. (↓)BAcc. (↑)EO (↓)Std Dev. (↓)
ResNet82.117.710.381.921.012.380.326.016.1
LfF77.74.13.477.27.34.476.814.410.4
DFA82.115.89.381.818.110.479.620.213.8
BPA83.116.69.682.021.812.680.322.613.1
Ours80.33.12.280.54.05.379.28.27.1
Table 7. Results on UTK Face for target attribute Race. We set Race and Gender as the target and sensitive attributes respectively.
Table 7. Results on UTK Face for target attribute Race. We set Race and Gender as the target and sensitive attributes respectively.
MethodSeverity ( γ ) = 2 Severity ( γ ) = 3 Severity ( γ ) = 4
BAcc. (↑)EO (↓)Std Dev. (↓)BAcc. (↑)EO (↓)Std Dev. (↓)BAcc. (↑)EO (↓)Std Dev. (↓)
ResNet81.810.16.181.615.29.279.321.212.4
LfF81.84.82.981.69.26.680.512.99.8
DFA81.76.610.081.812.810.381.013.310.1
BPA81.96.53.981.810.210.780.611.58.7
Ours82.44.53.382.65.73.980.99.55.5
Table 11. Fairness improvement through semi-supervised learning. Labeled data represents the proportion of data with sensitive attribute labels in the entire training set.
Table 11. Fairness improvement through semi-supervised learning. Labeled data represents the proportion of data with sensitive attribute labels in the entire training set.
MethodLabeled DataBalanced AccuracyEO
Group DRO174.23.4
Ours1/272.34.2
1/471.64.7
1/1071.45.0
071.017.9
Table 12. Results on Cat and Dog. We set species and color to the target and sensitive attributes, respectively.
Table 12. Results on Cat and Dog. We set species and color to the target and sensitive attributes, respectively.
MethodBalanced AccuracyEqualized OddsStd Dev.
ResNet79.920.717.7
LfF81.614.08.4
DFA86.813.47.8
BPA87.710.77.9
Ours85.95.77.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, P.; Park, S. Fair Classification Without Sensitive Attribute Labels via Dynamic Reweighting. Appl. Sci. 2026, 16, 1684. https://doi.org/10.3390/app16041684

AMA Style

Lee P, Park S. Fair Classification Without Sensitive Attribute Labels via Dynamic Reweighting. Applied Sciences. 2026; 16(4):1684. https://doi.org/10.3390/app16041684

Chicago/Turabian Style

Lee, Pilhyeon, and Sungho Park. 2026. "Fair Classification Without Sensitive Attribute Labels via Dynamic Reweighting" Applied Sciences 16, no. 4: 1684. https://doi.org/10.3390/app16041684

APA Style

Lee, P., & Park, S. (2026). Fair Classification Without Sensitive Attribute Labels via Dynamic Reweighting. Applied Sciences, 16(4), 1684. https://doi.org/10.3390/app16041684

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop