Next Article in Journal
Exergy Analysis of a Convective Heat Pump Dryer Integrated with a Membrane Energy Recovery Ventilator
Previous Article in Journal
Revisiting Endoreversible Carnot Engine: Extending the Yvon Engine
Previous Article in Special Issue
SPIRIT: Structural Entropy Guided Prefix Tuning for Hierarchical Text Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gradient-Based Multiple Robust Learning Calibration on Data Missing-Not-at-Random via Bi-Level Optimization

1
Mogo Co., Ltd., Beijing 100000, China
2
Gaoling School of Artificial Intelligence, Renmin University of China, Beijing 100872, China
*
Authors to whom correspondence should be addressed.
Entropy 2025, 27(2), 196; https://doi.org/10.3390/e27020196
Submission received: 13 December 2024 / Revised: 24 January 2025 / Accepted: 10 February 2025 / Published: 13 February 2025
(This article belongs to the Special Issue Causal Inference in Recommender Systems)

Abstract

:
Recommendation systems (RS) have become integral to numerous digital platforms and applications, ranging from e-commerce to content streaming field. A critical problem in RS is that the ratings are missing not at random (MNAR), which is due to the users always giving feedback on items with self-selection. The biased selection of rating data results in inaccurate rating prediction for all user-item pairs. Doubly robust (DR) learning has been studied in many tasks in RS, which is unbiased when either a single imputation or a single propensity model is accurate. In addition, multiple robust (MR) has been proposed with multiple imputation models and propensity models, and is unbiased when there exists a linear combination of these imputation models and propensity models is correct. However, we claim that the imputed errors and propensity scores are miscalibrated in the MR method. In this paper, we propose a gradient-based calibrated multiple robust learning method to enhance the debiasing performance and reliability of the rating prediction model. Specifically, we propose to use bi-level optimization to solve the weights and model coefficients of each propensity and imputation model in MR framework. Moreover, we adopt the differentiable expected calibration error as part of the objective to optimize the model calibration quality directly. Experiments on three real-world datasets show that our method outperforms the state-of-the-art baselines.

1. Introduction

Recommendation systems (RS) is an effective tool to address the problem of information overload and has been widely used in e-commerce, social media, and entertainment [1]. RS aims to predict user preferences for items based on collected historical interaction data [2,3]. However, the collected data cannot include all ratings from users to items, and the ratings are missing not at random (MNAR) due to users’ self-selection behavior, i.e., users can choose the item to rate freely, which is also known as selection bias problem [4,5]. The MNAR problem indicates that the collected dataset is not representative of the target population of interest (all user-item pairs), and the training distribution differs from the target test distribution. Ignoring such distributional shift will inevitably lead to sub-optimal recommendation performance [2,3,6]. To address the MNAR problem, one line of previous research proposed to use error imputation-based (EIB) methods, which first impute the missing ratings and then train the prediction model based on both observed and imputed ratings [6,7]. Additionally, another category of methods leverages propensity scores, which computes the probability of an event being observed, to reweight the observed ratings and align the distribution of observed data with the target population [3,8]. Furthermore, Doubly Robust (DR) method combines the error imputation and the inverse propensity re-weighting to achieve double robustness, which means the DR estimator achieves unbiasedness if either the imputed errors or the learned propensities are correct [2,9,10]. Furthermore, the Multiple Robust (MR) method is proposed to mitigate inaccuracies in single-model propensity scores or error imputations found in DR method [11]. By considering multiple candidate propensity and imputation models, MR estimator achieves unbiasedness if any of the propensity models, imputation models, or a linear combination of these models accurately estimate the true propensities or prediction errors.
However, we argue that the imputed errors and estimated propensity scores are miscalibrated in the existing MR method, which cannot reflect the ground-truth likelihood of the correctness of the true error or true propensity. For instance, if we have 100 user-item pairs with estimated propensity scores equal to 0.2, there should be exactly 20 ratings being observed and 80 ratings being unobserved. Although previous study has proposed to adopt calibration experts to calibrate the single propensity model and imputation model in DR estimator [12], this approach cannot be directly extended to the MR estimator, as calibrating each model individually is expensive and unreasonable due to the unbiasedness condition of MR in terms of linear combinations is not considered. Furthermore, the calibration metric previously used in [12] is non-differentiable and cannot be directly optimized.
To fill this gap, we propose the calibrated multiple robust learning (Cali-MR) method to calibrate the linear combinations of multiple imputation models and propensity models using bi-level optimization, which aims to learn an ensemble model that simultaneously possesses strong prediction performance and calibration ability. In this bi-level optimization, we adopt differentiable expected calibration errors to quantify the calibration ability that allows it to be directly optimized. The calibrated linear combination of propensity and imputation models is then used to train the prediction model based on a joint learning algorithm. The contributions of this paper are summarized as follows.
  • We propose a novel MR calibration method using bi-level optimization via calibrating the ensemble imputation and propensity models and address the non-differentiable issue by adopting differentiable expected calibration errors.
  • We further propose a bi-level calibrated multiple robust learning algorithm to update the calibrated imputation models and the prediction model. To the best of our knowledge, this is the first work to perform calibration for the MR estimator.
  • We conduct extensive experiments on three real-world datasets, showing the effectiveness of our method compared to the state-of-the-art debiasing methods.

2. Related Works

2.1. Debiased Recommendation

The rating missing not at random (MNAR) problem results in the distribution of the observed population being different from the target population, hindering the prediction model from learning users’ true preferences [4,5,13,14,15]. There are many methods proposed to address this issue [16,17,18,19,20,21].
Specifically, methods including the Inverse Propensity Scoring (IPS), EIB, and DR were proposed to mitigate the MNAR problem in RS [6,22]. EIB methods might produce out-of-bound predictions; while the IPS method may suffer from large variance with small propensities [23]. DR methods combine the advantages of both the EIB and IPS methods, guaranteeing unbiasedness if either the error imputation model or propensity model is correctly specified.
There have been quantities of variants of DR methods to improve the debiasing performance, such as Multi-DR [24], BRD-DR [25], SDR [26], TDR [27], CDR [28], N-DR [29], DT-DR [30], UIDR [31], and OME-DR [32]. Besides, Multiple robust (MR) [11] combines multiple imputation models and propensity models, and is unbiased when there exists a linear combination of them is correct. In addition, Liu et al. [33] use an information bottleneck-based method and Yang et al. [34] and Wang et al. [35] use adversarial learning for debiasing. However, these methods fail to consider model calibration properties.
To mitigate this issue, DCE-DR [12] is proposed to calibrate the propensity and imputation model in the DR method. However, calibrating each imputation model and propensity model in MR is expensive and unreasonable, due to the unbiasedness condition of MR based on linear combinations is not taken into account. In this paper, we propose the Cali-MR method to calibrate multiple propensity and imputation models to further enhance the debiasing performance and reliability of MR method.

2.2. Model Calibration

Calibration means that the probability associated with the predicted class label should reflect its ground truth correctness likelihood [36,37], which plays an important role in building reliable, robust AI systems, especially in safety-critical fields such as medical diagnosis [38,39], self-driving [40,41], and financial decision making [42,43]. Early research demonstrated good calibration performance of simple neural networks on binary classification tasks [44]. However, with the rapid development of deep learning techniques, recent deep and complex neural networks are no longer well-calibrated [36].
Calibration methods can be divided into the following four categories [45]: post-hoc calibration, regularization methods, uncertainty estimation, and hybrid calibration methods. Post-hoc calibration methods aim to calibrate a model after training, including non-parametric calibration histogram binning [46], isotonic regression [47], and parametric methods such as Platt scaling [48]. Regularization methods adopt penalty terms such as the L 2 regularization [36], entropy regularization [49], difference between confidence and accuracy [50], and calibration errors [51,52] to ensure the calibration property. Uncertainty Estimation aims to alleviate model miscalibration by injecting randomness using Bayesian neural networks [53], model ensemble [54], Monte Carlo dropout [55], and Gumbel-softmax [56] based approaches. Hybrid calibration methods combine two or more methods to achieve calibration. For example, Zhang et al. [57] combines ensemble and temperature scaling and Laves et al. [58] adopts monte-carlo dropout with temperature scaling. We conclude the categorization of model calibration methods in Figure 1.
In this paper, we adopt a differentiable expected calibration error as part of the objective to ensure the model calibration. Compared to other techniques for calibration, our approach has the following advantages. First, unlike post-hoc methods that adjust calibration after training [47,48], our method based on calibration error regularization explicitly optimizes the calibration metric during the training process. Secondly, uncertainty estimation methods such as Monte Carlo Dropout typically rely on complex sampling processes [55], resulting in high computational costs, while hybrid calibration methods which combine multiple techniques, are similarly complex and challenging to implement [57,58]. By comparison, our proposed approach is both simple and computationally efficient. Furthermore, unlike other indirect regularization methods, such as L 2 or entropy regularization [36,49], our method directly incorporates differentiable expected calibration error as a loss function, enabling precise optimization of the calibration objective.

3. Preliminary

3.1. Debiased Recommendation

Let U = { u 1 , , u m } be the users set, I = { i 1 , , i n } be the item set, and  D = U × I be the set of all user-item pairs. The rating matrix is denoted as R R m × n with r u , i as element. Let o u , i { 0 , 1 } be the observation indicator, where o u , i = 1 indicates the rating r u , i is observed, otherwise is not. Define x u , i be the observed features. We denote the prediction model as f θ ( · ) parameterized by θ and the predicted ratings as r ^ u , i = f θ ( x u , i ) . The goal is to accurately predict r u , i for all user-item pairs, which can be achieved by minimizing the ideal loss
L ideal ( θ ) = 1 | D | ( u , i ) D L ( f θ ( x u , i ) , r u , i ) : = 1 | D | ( u , i ) D e u , i ,
where L ( · , · ) is the training loss function such as cross-entropy loss. However, in practice, we cannot obtain the complete rating matrix. We denote the set of user-item pairs with observed ratings as O = { ( u , i ) o u , i = 1 } . Thus, the naive method optimizes the average loss over the observed user-item pairs
L naive ( θ ) = 1 | O | ( u , i ) O e u , i .
Due to the MNAR problem, E [ L naive ( θ ) ] L ideal ( θ ) . Several methods were proposed to unbiasedly estimate the ideal loss, including the EIB, IPS, DR, and their variants. Because EIB and IPS can be regarded as special cases of DR, we only introduce the DR methods here. The loss function of the vanilla DR method is formulated as
L DR ( θ ) = 1 | D | ( u , i ) D e ^ u , i + o u , i ( e u , i e ^ u , i ) p ^ u , i ,
where p ^ u , i π ( x ; α ^ ) is the estimation of propensity score p u , i = Pr ( o u , i = 1 x u , i ) , and  e ^ u , i = L ( m ( x u , i ; β ) , r ^ u , i ) is the imputed error, while the imputation model is denoted as m ( x u , i ; β ^ ) . In addition, the multiple robust (MR) considers J propensity models π 1 ( x ; α ^ 1 ) , , π J ( x ; α ^ J ) and K imputation models m 1 ( x ; β ^ 1 ) , , m K ( x ; β ^ K ) . Let p ^ u , i j π j ( x u , i ; α ^ j ) , m ^ u , i k m k ( x ; β ^ k ) , the loss function of MR is shown below:
L MR ( θ ) = 1 | D | ( u , i ) D u T x u , i · η ^ ( θ ) ,
where u x u , i = 1 / p ^ u , i 1 , , 1 / p ^ u , i J , m ^ u , i 1 , , m ^ u , i K T and η ^ ( θ ) is the solution by minimizing
1 | D | ( u , i ) D o u , i e u , i u T x u , i · η 2 .
The MR estimator is unbiased when there exists a weight W = ( w 1 , w 2 , , w J , 0 , 0 , , 0 ) satisfying W u x u , i = 1 / p u , i or V = ( 0 , , 0 , v 1 , v 2 , , v K ) satisfying V u x u , i = e u , i for all user-item pairs.

3.2. Calibration

A model is calibrated if its output reflects the ground-truth likelihood of correctness [37]. For the propensity model π ( x ; α ^ ) and the observation indicator o, a formal definition is shown below:
E [ o π ( x ; α ^ ) = p ^ ] = p ^ p ^ [ 0 , 1 ] .
For instance, if we have 100 samples with estimated propensity scores equal to 0.2, there should be exactly 20 samples being observed. Similarly, the formal definition for the calibrated imputation model m ( x ; β ^ ) is formulated below:
E [ e m ( x ; β ^ ) = e ^ ] = e ^ e ^ R .
To measure the miscalibration of the model, the Expected Calibration Error (ECE) metric is proposed [59]. For a propensity model π ( x ; α ^ ) and imputation model m ( x ; β ^ ) , the ECE is defined as follows:
ECE ( α ^ ) = E p ^ [ E [ o π ( x ; α ^ ) = p ^ ] p ^ ] , ECE ( β ^ ) = E e ^ [ | E [ e m ( x ; β ^ ) = e ^ ] e ^ | ] .

4. Methodology

4.1. Distinctions from Previous Work

Previous studies have proposed to calibrate the single propensity model and imputation model in DR estimator [12], and they propose to use a binning strategy to estimate the ECE metric empirically, for example for the propensity model:
ECE ^ p ^ = m = 1 M B m N ( u , i ) B m o u , i B m ( u , i ) B m p ^ u , i B m ,
where B m is the predefined m-th bin and N is the corresponding number of samples in the bin.
However, how to properly calibrate multiple propensity and imputation models for MR estimators remains unexplored. A naive approach is individually calibrating each propensity and imputation model in the MR estimator. Note that this method is computationally expensive and overlooks the robust property of the MR estimator, that is, the MR estimator achieves unbiased if a linear combination of multiple candidate models is accurate.
Inspired by this, we propose to calibrate the linear combination of multiple models instead of calibrating each model individually. In addition, note that the previously used ECE ^ involves assigning each sample to a specific hard bin, making it non-differentiable and thus unsuitable for direct incorporation into the training objective. To address this issue, we employ a soft binning strategy to develop the differentiable expected calibration error metric and leverage it to construct a regularization term that constrains the model’s calibration error, which can be used for model training. Next, we will introduce the proposed Cali-MR in detail.

4.2. Differentiable Expected Calibration Error

To address the non-differentiable problem of ECE ^ , we leverage the soft binning strategy [52], using the following differentiable expected calibration error (DECE) that allows directly optimize calibration quality to mitigate the model miscalibration. For example, the DECE for a propensity model π ( x ; α ^ ) is defined as:
DECE α ^ = 1 | D | m = 1 M ( u , i ) D o m ( x u , i ; ϕ ) ( o u , i p ^ u , i ) ,
where o m ( x u , i ; ϕ ) = P ( x u , i B m | p ^ u , i ) denotes the probability that how likely it is that p ^ u , i belongs to the m-th bin. In practice, the  o m ( x u , i ; ϕ ) can be logistic regression or any other model [52].
In our Cali-MR, we adopt DECE as a regularization term and develop a gradient-based learning algorithm model training. Specifically, to calibrate the linear combination of multiple models, we formalize the propensity DECE loss L DECE p as follows:
L DECE p ( w ; ϕ p ; α 1 , , α J ) = 1 | D | m = 1 M ( u , i ) D o m ( x u , i ; ϕ p ) ( o u , i j = 1 J w j p ^ u , i j ) ,
where w = ( w 1 , , w J ) is a given set of weight coefficients. This loss measures the calibration error of the combination model, and minimizing this loss improves the calibration ability of the current propensity model combination under the current combination coefficients. We use a one-layer neural network with softmax activation function to model the propensity soft binning model o m ( x u , i ; ϕ p ) with parameters ϕ p , where m is a pre-defined hyper-parameter.
Similarly, the imputation DECE loss L DECE e under weight coefficients v = ( v 1 , , v K ) is formalized below:
L DECE e ( v ; ϕ m ; β 1 , , β K ) = 1 | D | m = 1 M ( u , i ) D o m ( x u , i ; ϕ m ) ( o u , i e u , i j = 1 J w j p ^ u , i j k = 1 K v k e ^ u , i k ) ,
where we similarly model the imputation soft binning model o m ( x u , i ; ϕ m ) with parameters ϕ m with a one-layer neural network with softmax activation function. Note that the e u , i is missing for user-item pairs with o u , i = 0 , we reweight the observed e u , i using the inverse of the linear combination of the multiple propensity models.
Based on the DECE loss L DECE p or L DECE e , we can measure the calibration quality of given multiple models and weight coefficients, and further improve the calibration ability of the current combined model using a gradient-based algorithm by minimizing such differentiable loss.

4.3. Calibrated Multiple Robust Learning

Note that in calculating the DECE loss L DECE p or L DECE e , the coefficients w or v need to be explicitly specified. However, these coefficients are unknown during the training process of the existing MR method. Therefore, we propose using a gradient-based bi-level optimization method to solve for the optimal coefficients, parameters of the soft binning model, and multiple imputation and propensity models. In addition, we alternatively update the prediction model and the calibrated imputation model based on a joint learning algorithm.
For propensity models, the optimization objective can be formalized as follows:
( α 1 * , , α J * ) = arg min α 1 , , α J 1 J j = 1 J L p j ( α j ) + λ L DECE p ( w * ) s . t . w * ( α 1 , , α J ) = arg min w * L p ( w * ( α 1 , , α J ) ) ,
where L DECE p ( w ) loss is the calibration constraints defined in Section 4.2, and  L p j ( α j ) is the training loss for a single propensity model π J ( x ; α ^ J ) ensuring the accuracy of each independent propensity model, which is shown below:
L p j ( α j ) = 1 | D | ( u , i ) D [ o u , i · log p u , i j ( 1 o u , i ) · log ( 1 p u , i j ) ] .
L p ( w ( α 1 , , α J ) ) is the loss for the combination coefficients w , aiming to learn a set of coefficients such that the linear combination j = 1 J w j p ^ u , i j can accurately predict observation indicator o u , i , which is shown below:
L p ( w ( α 1 , , α J ) ) = 1 | D | ( u , i ) D [ o u , i · log ( j = 1 J w j p ^ u , i j ) ( 1 o u , i ) · log ( 1 j = 1 J w j p ^ u , i j ) ] .
In this gradient-based bi-level optimization, we aim to train the propensity models such that each model performs well and their linear combination is well calibrated, where the coefficients also ensure the strong prediction performance of the combined model. For practical implementation, we first assumed update coefficients w through optimizing L p ( w ) . Using these coefficients, we calculate the DECE loss L DECE p ( w ) and combine it with the base prediction loss of each propensity model denoted as L p j ( α j ) to form the final loss. This final loss is then used to update the propensity models α 1 , , α J . After that, the loss L p ( w ) and L DECE p ( w ) are used to update the combination coefficients w and the soft binning model ϕ p sequentially.

Multiple Imputation Calibration

With the calibrated propensity model j = 1 J w j p ^ u , i j obtained from Equation (1) by the bi-level optimization, we can further calibrate the linear combination of multiple imputation models:
( β 1 * , , β K * ) = arg min β 1 , , β K 1 K k = 1 K L e k ( β k ) + λ L DECE e ( v * ) s . t . v * ( β 1 , , β K ) = arg min v * L e ( v * ( β 1 , , β K ) ) ,
where L DECE e ( v ) is the calibration constraints shown in Section 4.2. The naive training loss for each imputation model m K ( x ; β ^ K ) is expressed as
L e k ( β k ; θ ) = 1 | D | ( u , i ) D o u , i ( e u , i e ^ u , i k ) 2 j = 1 J w j p ^ u , i j ,
and the loss for the combination coefficients v is
L e ( v ( β 1 , , β K ) ) = ( u , i ) D o u , i ( e u , i k = 1 K v k e ^ u , i k ) 2 | D | j = 1 J w j p ^ u , i j ,
which aims to learn a set of weight coefficients such that the linear combination of the imputation models k = 1 K v k e ^ u , i k can unbiasedly estimate the prediction error e u , i .
Similar to updating the propensity models, we first use L e ( v ) to assume update coefficients v . Based on that, we compute the DECE loss L DECE e ( v ) , combine it with the base training loss of each imputation model L e k ( β k ; θ ) , and use the combined loss to update the imputation models β 1 , , β K . Then we adopt loss L e ( v ) and L DECE e ( v ) to update the combination coefficients v and the soft binning model ϕ e sequentially. After obtaining the updated imputation models and coefficients, we jointly train the prediction model using the standard multiple robust learning algorithm. Specifically, we use ridge regression to calculate η ^ in the MR estimator and update the prediction model based on the MR loss L MR using different samples. We summarize the above gradient-based bi-level learning algorithm in Algorithm 1.
Algorithm 1: Gradient-Based Bi-level Calibrated Multiple Robust Learning
Entropy 27 00196 i001

5. Experiments

5.1. Datasets

To evaluate the debiasing performance, we conduct experiments on three benchmark datasets Coat (https://www.cs.cornell.edu/~schnabts/mnar/, accessed on 15 January 2025) and Yahoo! R3 (https://webscope.sandbox.yahoo.com), and KuaiRec (https://github.com/chongminggao/KuaiRec, accessed on 15 January 2025) [60], which are widely used in debiased RS with both missing not at random (MNAR) and missing at random (MAR) data. Coat dataset consists of 6960 MNAR training samples and 4640 MAR test samples derived from 290 users rating on 300 items. The Yahoo! R3 dataset includes 311,704 MNAR training samples and 54,000 MAR test samples derived from 15,400 users rating on 1000 items. Both datasets are five-scale, and following previous works [61,62,63], we binarize the ratings greater than three to 1, and others to 0. The KuaiRec dataset is collected from a video-sharing platform and contains 4,676,570 video watching ratios derived from 1411 users evaluating 3327 videos. We binarize the continuous ratios greater than two to 1, otherwise to 0.

5.2. Baselines

We compare our method with the following baselines for comprehensive evaluations:
  • Naive method [64] naively optimizes the average loss over the observed user-item pairs.
  • IPS method [3] reweights the observed ratings with the inverse propensity scores.
  • SNIPS method [65] reweights the observed ratings with self-normalized propensity scores to further reduce the variance.
  • ASIPS method [22] generates reliable pseudo-ratings to mitigate propensity estimation bias and high variance problem.
  • DR method [10] combines error imputation and inverse propensity reweighting to construct a doubly robust estimator, where imputed errors are typically set based on label prior knowledge, such as the mean value of the labels.
  • DR-JL method [2] further proposes modeling error imputation with neural networks and jointly learns the prediction model and imputation model.
  • MRDR method [66] enhances the DR-JL method by explicitly controlling the variance of the DR estimator through imputation model learning.
  • DR-BIAS method [67] enhances the DR-JL method by further reducing the bias of the DR estimator through imputation model learning.
  • DR-MSE method [67] further combines MRDR method and DR-BIAS method to achieve bias-variance trade-off and control the generalization error.
  • MR method [11] adopts multiple candidate propensity and imputation models to mitigate inaccuracies in single-model propensity scores or error imputation in DR methods.
  • TDR and TDR-JL methods [27] correct the imputed errors with targeted learning to reduce the bias and variance simultaneously for existing DR approaches
  • StableDR method [26] constructs a stabilized DR estimator that has a weaker dependence on extrapolation and is robust to small propensities by learning constrained propensity scores.
  • IPS-V2 and DR-V2 methods [68] learn the propensity model which can balance some manually selected functions such as the first and second moments of the features.
  • KBIPS and KBDR methods [32] further propose to conduct causal balancing in the reproducing kernel Hilbert space (RKHS) and randomly select some kernel functions to balance for propensity model learning.
  • AKBIPS and AKBDR methods [32] adaptively select the kernel functions which contribute the most to reducing the estimation bias to balance for propensity model learning.
  • DCE-DR and DCE-TDR method [12] propose to calibrate the single propensity model and single imputation model in DR and TDR estimators through Mixture-of-Experts technique.

5.3. Experiment Protocols and Details

We evaluate the prediction performance using three widely adopted evaluation metrics: AUC (Area Under the ROC Curve), NDCG@T (Normalized Discounted Cumulative Gain), and F1@T.
  • AUC [69] is a performance metric for classifiers that measures the probability of a randomly chosen positive example being ranked higher than a randomly chosen negative one. A higher AUC score reflects better ranking performance in differentiating positive instances from negative ones.
  • NDCG@T [70] evaluates ranking performance by comparing the Discounted Cumulative Gain (DCG) of the top-T results to the Ideal DCG (IDCG), producing a normalized score between 0 and 1. A higher NDCG@T implies that more relevant items are ranked towards the top.
Let r i be the relevance of the item at rank i. We first compute the DCG at rank T as well as the IDCG@T by placing the most relevant items in the optimal (ideal) order:
DCG @ T = i = 1 T 2 r i 1 log 2 ( i + 1 ) , IDCG @ T = i = 1 T 2 r i * 1 log 2 ( i + 1 ) .
where r i is the relevance of the i-th item at rank i, and r i * denotes the relevance of the i-th item in the ideal ranking. NDCG@T is then defined as:
NDCG @ T = DCG @ T IDCG @ T .
  • F1@T [71] is the harmonic mean of precision and recall computed over the top-T predictions returned by a model. A higher F1@T indicates a better trade-off between precision and recall in the top-T results.
We set T = 5 on Coat and Yahoo! R3, and T = 20 on KuaiRec. In addition, we tune learning rate in { 0.01, 0.05} and weight decay in { 1 × 10 6 , 5 × 10 6 , 1 × 10 5 , , 1 × 10 3 , 5 × 10 3 } .We use the same hyperparameter search space and follow the results in Li et al. [32].

5.4. Performance Analysis

The experimental results are shown in Table 1, and we find that all the debiasing methods including both IPS-based and DR-based baselines outperform the Naive method, which demonstrates the importance of debiasing. Besides, among all baselines, AKBDR method introduces balancing kernel functions for propensity model training. This method adaptively identifies the most critical kernel functions to balance by fitting prediction errors, and well-trained propensity model effectively eliminates selection bias, thereby demonstrating strong performance on the Coat dataset. On the other hand, TDR introduces a targeted learning technique, which leverages propensity to enhance the error imputation, achieving reductions in both bias and variance. DCE-TDR further improves upon TDR by calibrating propensity model, utilizing the calibrated propensity for targeted learning, and achieving further performance gains. This calibration strategy enables DCE-TDR to deliver competitive results on the Yahoo! R3 and KuaiRec datasets. Furthermore, the proposed Cali-MR method exhibits superior overall performance and significantly outperforms existing methods on Yahoo! R3 and KuaiRec datasets. This shows that calibrating multiple propensity and imputation models in a Multiple Robust estimator with weight coefficient learned in bi-level optimization further enhances the debiasing performance.

5.5. In-Depth Analysis

We explore the difference between calibrating the ensemble model and calibrating individual models separately, with the number of propensity models J and imputation models K taking values from { 1 , 3 , 5 } . The results on the KuaiRec and Yahoo! R3 datasets are shown in Figure 2 and Figure 3, where darker colors in the heatmap represent higher values of the corresponding evaluation metrics, indicating stronger model performance. The proposed method for calibrating the linear combination of multiple candidate models is referred to as ‛joint calibration’, while the approach of calibrating each individual candidate propensity model and then randomly selecting one calibrated propensity model to calibrate the candidate imputation models is referred to as ‛individual calibration’. We find that directly calibrating the ensemble model outperforms calibrating a single model individually, especially on the Yahoo! R3 dataset. This demonstrates that the proposed Cali-MR method which takes into account the conditions for achieving unbiasedness in the MR estimator achieves better calibration and debiasing performance. On the other hand, examining the impact of the number of candidate models J and K on prediction performance, we find that when J = 1 or K = 1 , i.e., when using only one propensity model or one imputation model, the prediction performance is poor, validating the effectiveness of incorporating more candidate models in the MR estimator. Additionally, the optimal ( J , K ) combinations for Cali-MR on Yahoo! R3 and KuaiRec datasets are ( 5 , 5 ) and ( 3 , 3 ) , respectively.
We conduct an ablation study of the proposed Cali-MR on three benchmark datasets, with the experimental results shown in Table 2. It can be observed that the prediction performance declines if either the propensity model or the imputation model is not calibrated, demonstrating the necessity of model calibration. Furthermore, the performance drop is more significant when removing the calibration of the propensity model compared to the imputation model, indicating that propensity models are more miscalibrated in the existing MR estimator. This highlights that the current propensity model learning approach, based on simple fitting to the observed variable indicator o with naive binary cross-entropy loss, fails to produce well-calibrated models, emphasizing the need to incorporate calibration loss on top of it.
Figure 4 and Figure 5 investigate the impact of varying model calibration hyper-parameter λ in multiple propensity and imputation calibration on prediction performance on Coat and KuaiRec datasets. We record the results for different values of the hyper-parameter λ and fit a curve to show its variation trend, with a horizontal line representing the baseline method without calibration. We observe that when the hyper-parameter λ falls within an appropriate range, such as [ 0.1, 100 ] , the proposed method stably outperforms the baseline method, with the best performance achieved around moderate values, such as λ = 10 . Notably, on the Coat dataset, when the value of hyperparameter λ is set too large (e.g., 1000), the model fails to achieve good prediction performance, which indicates the propensity and imputation models overly emphasize calibration at the expense of prediction accuracy, leading to poor performance. This highlights the need to balance the prediction and calibration capabilities of propensity and imputation models during training.
Figure 6 and Figure 7 explore the effect of varying number of bins M in soft binning strategy on prediction performance on Coat and KuaiRec datasets. The scatter points represent the results for different values of M, and a fitted curve illustrates the trend of predictive performance as M changes. The results in the figure indicate that the optimal performance is achieved at moderate values of M, such as around 30 for Coat dataset and 15 for KuaiRec dataset. However, performance declines when M is either too large or too small. A possible reason is that when M is too small, the probability of samples with different prediction values ( p ^ or e ^ ) being assigned to the same bin increases. Conversely, when M is too large, the probability of samples with similar prediction values ( p ^ or e ^ ) being assigned to the same bin decreases. Both situations can lead to inaccurate DECE estimation, thereby reducing the calibration quality of the model.
The results in Table 3 demonstrate that proposed Cali-MR achieves comparable ECE performance to the current SOTA model, DCE-TDR. However, when evaluating the relative ECE reduction compared to the base models, we observe that DCE-TDR achieves a smaller relative improvement over TDR because TDR itself, leveraging targeted learning, already exhibits a relatively calibrated imputation model. Conversely, the original MR model shows a high ECE, indicating poor calibration, while the proposed Cali-MR significantly reduces the ECE, thereby enhancing calibration quality, which validates the effectiveness of the proposed approach.
We compare the training time and parameter size of different methods on Table 4. The experimental results show that the Naive method, which adopts a single model to fit observed ratings, has a parameter size of 1 × . The IPS method introduces an additional propensity model, resulting in a parameter size of 2 × , while the DR method further introduces an imputation model, with a parameter size of 3 × . In this experiment, the MR method adopts two candidate propensity and imputation models, leading to a parameter size of 5 × . As an improvement, our proposed Cali-MR (J = 2, K = 2) adds only a soft binning module, which increases the parameter count marginally compared to the original MR method, thus still maintaining a parameter size of 5 × . In other words, the proposed method achieves significant performance improvements over the existing MR method, while maintaining the same parameter size. Additionally, by comparing the training times for MR and Cali-MR (J = 2, K = 2), we observe that model calibration does not add substantial training time, for example, with times of 127.8 and 144.56 on Yahoo! R3 dataset, respectively. Furthermore, increasing the number of candidate models (J and K) for Cali-MR does not lead to a significant increase in training time. This is because as the number of candidate models increases, the ensemble model tends to overfit, and due to early stopping strategy, the number of epochs required for convergence varies with the number of candidate models. In summary, compared to the MR method, the proposed Cali-MR method improves model calibration and prediction accuracy with almost the same parameter size and acceptable training time.

6. Conclusions

In this paper, we explore how to properly calibrate multiple propensity and imputation models in a Multiple Robust (MR) estimator. First, we argue that calibrating each candidate model individually is too costly and unreasonable due to the unbiasedness condition of MR in terms of linear combinations is not considered. Based on this, we propose using a gradient-based bi-level optimization method to calibrate the linear combination of multiple candidate models. Specifically, in the bi-level optimization, we first assumed to update the combination coefficients to obtain the best-performing combination coefficients under the current candidate model parameters. Then, based on these coefficients, we update the candidate model parameters to ensure that each model maintains good prediction performance, while also ensuring that the combination model achieves strong calibration ability, using the differentiable expected calibration error metric that allows it to be directly optimized. Experimental results on three real-world datasets demonstrate that the proposed multiple propensity and imputation calibration method further enhances the prediction performance.
Regarding the broader impact, this is the first work to address the calibration of MR models, by introducing the differentiable expected calibration error (DECE) to directly optimize the objective, rather than relying on post-hoc adjustments in previous work. This provides a novel perspective on model calibration within the debiasing recommendation field. The main limitation of this work lies in the use of bi-level optimization to update parameters of multiple candidate models and the optimal combination coefficients within the MR method. This results in longer training times for the proposed Cali-MR method compared to the original MR method. A potential future work is to develop an alternative optimization algorithm to reduce the computational cost. Furthermore, in future works, we plan to investigate whether other calibration metrics, aside from ECE, might be more suitable for debiasing recommendation scenarios.

Author Contributions

Conceptualization, S.G.; Methodology, S.G.; Validation, C.M.; Investigation, S.G. and C.M.; Writing – original draft, S.G.; Writing – review & editing, C.M.; Visualization, C.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Our code is available at https://github.com/Yilu114/Cali-MR (accessed on 15 January 2025). Data were obtained from GitHub and are available https://github.com/RyanWangZf/CVIB-Rec with the permission of GitHub (accessed on 15 January 2025).

Conflicts of Interest

Author Shuxia Gong is employed by Mogo Inc. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Correction Statement

This article has been republished with a minor correction to Reference 14. This change does not affect the scientific content of the article.

References

  1. Ricci, F.; Rokach, L.; Shapira, B. Introduction to recommender systems handbook. In Recommender Systems Handbook; Springer: Boston, MA, USA, 2010; pp. 1–35. [Google Scholar]
  2. Wang, X.; Zhang, R.; Sun, Y.; Qi, J. Doubly Robust Joint Learning for Recommendation on Data Missing Not at Random. In Proceedings of the ICML, Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
  3. Schnabel, T.; Swaminathan, A.; Singh, A.; Chandak, N.; Joachims, T. Recommendations as Treatments: Debiasing Learning and Evaluation. In Proceedings of the ICML, New York, NY, USA, 19–24 June 2016. [Google Scholar]
  4. Chen, J.; Dong, H.; Wang, X.; Feng, F.; Wang, M.; He, X. Bias and Debias in Recommender System: A Survey and Future Directions. Acm Trans. Inf. Syst. 2022, 41, 1–39. [Google Scholar] [CrossRef]
  5. Wu, P.; Li, H.; Deng, Y.; Hu, W.; Dai, Q.; Dong, Z.; Sun, J.; Zhang, R.; Zhou, X.H. On the Opportunity of Causal Learning in Recommendation Systems: Foundation, Estimation, Prediction and Challenges. In Proceedings of the IJCAI, Vienna, Austria, 23–29 July 2022. [Google Scholar]
  6. Steck, H. Training and testing of recommender systems on data missing not at random. In Proceedings of the KDD, Washington, DC, USA, 25–28 July 2010. [Google Scholar]
  7. Chang, Y.W.; Hsieh, C.J.; Chang, K.W.; Ringgaard, M.; Lin, C.J. Training and testing low-degree polynomial data mappings via linear SVM. J. Mach. Learn. Res. 2010, 11, 1471–1490. [Google Scholar]
  8. Saito, Y.; Yaginuma, S.; Nishino, Y.; Sakata, H.; Nakata, K. Unbiased recommender learning from missing-not-at-random implicit feedback. In Proceedings of the WSDM, Houston, TX, USA, 3–7 February 2020. [Google Scholar]
  9. Morgan, S.L.; Winship, C. Counterfactuals and Causal Inference: Methods and Principles for Social Research, 2nd ed.; Cambridge University Press: Cambridge, UK, 2015. [Google Scholar]
  10. Saito, Y. Doubly robust estimator for ranking metrics with post-click conversions. In Proceedings of the RecSys, Virtual Event, 22–26 September 2020. [Google Scholar]
  11. Li, H.; Dai, Q.; Li, Y.; Lyu, Y.; Dong, Z.; Zhou, X.H.; Wu, P. Multiple Robust Learning for Recommendation. In Proceedings of the AAAI, Washington, DC, USA, 7–14 February 2023. [Google Scholar]
  12. Kweon, W.; Yu, H. Doubly Calibrated Estimator for Recommendation on Data Missing Not At Random. In Proceedings of the WWW, Singapore, 13–17 May 2024. [Google Scholar]
  13. Luo, H.; Zhuang, F.; Xie, R.; Zhu, H.; Wang, D.; An, Z.; Xu, Y. A survey on causal inference for recommendation. Innovation 2024, 5, 100590. [Google Scholar] [CrossRef]
  14. Li, M.; Sui, H. Causal Recommendation via Machine Unlearning with a Few Unbiased Data. In Proceedings of the AAAI Workshop on Artificial Intelligence with Causal Techniques, Philadelphia, PA, USA, 25 February–4 March 2025. [Google Scholar]
  15. Wang, W.; Zhang, Y.; Li, H.; Wu, P.; Feng, F.; He, X. Causal Recommendation: Progresses and Future Directions. In Proceedings of the SIGIR, Taipei, Taiwan, 23–27 July 2023. [Google Scholar]
  16. Saito, Y.; Nomura, M. Towards Resolving Propensity Contradiction in Offline Recommender Learning. In Proceedings of the IJCAI, Vienna, Austria, 23–29 July 2022. [Google Scholar]
  17. Wang, H.; Yang, W.; Yang, L.; Wu, A.; Xu, L.; Ren, J.; Wu, F.; Kuang, K. Estimating Individualized Causal Effect with Confounded Instruments. In Proceedings of the KDD, Washington, DC, USA, 14–18 August 2022. [Google Scholar]
  18. Zou, H.; Wang, H.; Xu, R.; Li, B.; Pei, J.; Jian, Y.J.; Cui, P. Factual Observation Based Heterogeneity Learning for Counterfactual Prediction. In Proceedings of the CCLR, Tübingen, Germany, 11–14 April 2023. [Google Scholar]
  19. Wang, H.; Kuang, K.; Lan, L.; Wang, Z.; Huang, W.; Wu, F.; Yang, W. Out-of-distribution generalization with causal feature separation. IEEE Trans. Knowl. Data Eng. 2024, 36, 1758–1772. [Google Scholar] [CrossRef]
  20. Wang, H.; Kuang, K.; Chi, H.; Yang, L.; Geng, M.; Huang, W.; Yang, W. Treatment effect estimation with adjustment feature selection. In Proceedings of the KDD, Long Beach, CA, USA, 6–10 August 2023. [Google Scholar]
  21. Wu, A.; Kuang, K.; Xiong, R.; Li, B.; Wu, F. Stable estimation of heterogeneous treatment effects. In Proceedings of the ICML, Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  22. Saito, Y. Asymmetric Tri-training for Debiasing Missing-Not-At-Random Explicit Feedback. In Proceedings of the SIGIR, Xi’an, China, 25–30 July 2020. [Google Scholar]
  23. Wang, H.; Chang, T.W.; Liu, T.; Huang, J.; Chen, Z.; Yu, C.; Li, R.; Chu, W. Escm2: Entire space counterfactual multi-task model for post-click conversion rate estimation. In Proceedings of the SIGIR, Madrid, Spain, 11–15 July 2022. [Google Scholar]
  24. Zhang, W.; Bao, W.; Liu, X.Y.; Yang, K.; Lin, Q.; Wen, H.; Ramezani, R. Large-scale Causal Approaches to Debiasing Post-click Conversion Rate Estimation with Multi-task Learning. In Proceedings of the WWW, Taipei, Taiwan, 20–24 April 2020. [Google Scholar]
  25. Ding, S.; Wu, P.; Feng, F.; He, X.; Wang, Y.; Liao, Y.; Zhang, Y. Addressing Unmeasured Confounder for Recommendation with Sensitivity Analysis. In Proceedings of the KDD, Washington, DC, USA, 14–18 August 2022. [Google Scholar]
  26. Li, H.; Zheng, C.; Wu, P. StableDR: Stabilized Doubly Robust Learning for Recommendation on Data Missing Not at Random. In Proceedings of the ICLR, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  27. Li, H.; Lyu, Y.; Zheng, C.; Wu, P. TDR-CL: Targeted Doubly Robust Collaborative Learning for Debiased Recommendations. In Proceedings of the ICLR, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  28. Song, Z.; Chen, J.; Zhou, S.; Shi, Q.; Feng, Y.; Chen, C.; Wang, C. CDR: Conservative Doubly Robust Learning for Debiased Recommendation. In Proceedings of the CIKM, Birmingham, UK, 21–25 October 2023. [Google Scholar]
  29. Li, H.; Zheng, C.; Ding, S.; Feng, F.; He, X.; Geng, Z.; Wu, P. Be Aware of the Neighborhood Effect: Modeling Selection Bias under Interference for Recommendation. In Proceedings of the ICLR, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  30. Zhang, H.; Wang, S.; Li, H.; Zheng, C.; Chen, X.; Liu, L.; Luo, S.; Wu, P. Uncovering the Propensity Identification Problem in Debiased Recommendations. In Proceedings of the ICDE, Utrecht, The Netherlands, 13–17 May 2024. [Google Scholar]
  31. Li, H.; Zheng, C.; Wang, S.; Wu, K.; Wang, E.; Wu, P.; Geng, Z.; Chen, X.; Zhou, X.H. Relaxing the Accurate Imputation Assumption in Doubly Robust Learning for Debiased Collaborative Filtering. In Proceedings of the ICML, Vienna, Austria, 21–27 July 2024. [Google Scholar]
  32. Li, H.; Zheng, C.; Xiao, Y.; Wu, P.; Geng, Z.; Chen, X.; Cui, P. Debiased collaborative filtering with kernel-based causal balancing. In Proceedings of the ICLR, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  33. Liu, D.; Cheng, P.; Zhu, H.; Dong, Z.; He, X.; Pan, W.; Ming, Z. Debiased representation learning in recommendation via information bottleneck. ACM Trans. Recomm. Syst. 2023, 1, 1–27. [Google Scholar] [CrossRef]
  34. Yang, M.; Dai, Q.; Dong, Z.; Chen, X.; He, X.; Wang, J. Top-n recommendation with counterfactual user preference simulation. In Proceedings of the CIKM, Virtual Event, 1–5 November 2021. [Google Scholar]
  35. Wang, J.; Li, H.; Zhang, C.; Liang, D.; Yu, E.; Ou, W.; Wang, W. Counterclr: Counterfactual contrastive learning with non-random missing data in recommendation. In Proceedings of the ICDM, Shanghai, China, 1–4 December 2023. [Google Scholar]
  36. Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On calibration of modern neural networks. In Proceedings of the ICML, Sydney, NSW, Australia, 6–11 August 2017. [Google Scholar]
  37. Kull, M.; Silva Filho, T.; Flach, P. Beta calibration: A well-founded and easily implemented improvement on logistic calibration for binary classifiers. In Proceedings of the AISTATS, Fort Lauderdale, FL, USA, 20–22 April 2017. [Google Scholar]
  38. Caruana, R.; Lou, Y.; Gehrke, J.; Koch, P.; Sturm, M.; Elhadad, N. Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. In Proceedings of the KDD, Sydney, NSW, Australia, 10–13 August 2015. [Google Scholar]
  39. Huang, Y.; Li, W.; Macheret, F.; Gabriel, R.A.; Ohno-Machado, L. A tutorial on calibration measurements and calibration models for clinical prediction models. J. Am. Med. Inform. Assoc. 2020, 27, 621–633. [Google Scholar] [CrossRef]
  40. Bojarski, M. End to end learning for self-driving cars. arXiv 2016, arXiv:1604.07316. [Google Scholar]
  41. Chen, Z.; Huang, X. End-to-end learning for lane keeping of self-driving cars. In Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), Los Angeles, CA, USA, 11–14 June 2017; pp. 1856–1860. [Google Scholar]
  42. Büchel, P.; Kratochwil, M.; Nagl, M.; Rösch, D. Deep calibration of financial models: Turning theory into practice. Rev. Deriv. Res. 2022, 25, 109–136. [Google Scholar] [CrossRef]
  43. Biagini, F.; Gonon, L.; Walter, N. Approximation rates for deep calibration of (rough) stochastic volatility models. SIAM J. Financ. Math. 2024, 15, 734–784. [Google Scholar] [CrossRef]
  44. Niculescu-Mizil, A.; Caruana, R. Predicting good probabilities with supervised learning. In Proceedings of the ICML, Bonn, Germany, 7–11 August 2005. [Google Scholar]
  45. Wang, C. Calibration in deep learning: A survey of the state-of-the-art. arXiv 2023, arXiv:2308.01222. [Google Scholar]
  46. Zadrozny, B.; Elkan, C. Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In Proceedings of the ICML, Williamstown, MA, USA, 28 June–1 July 2001. [Google Scholar]
  47. Zadrozny, B.; Elkan, C. Transforming classifier scores into accurate multiclass probability estimates. In Proceedings of the KDD, Edmonton, AB, Canada, 23–26 July 2002. [Google Scholar]
  48. Platt, J. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Adv. Large Margin Classif. 1999, 10, 61–74. [Google Scholar]
  49. Pereyra, G.; Tucker, G.; Chorowski, J.; Kaiser, Ł.; Hinton, G. Regularizing neural networks by penalizing confident output distributions. arXiv 2017, arXiv:1701.06548. [Google Scholar]
  50. Liang, G.; Zhang, Y.; Wang, X.; Jacobs, N. Improved trainable calibration method for neural networks on medical imaging classification. arXiv 2020, arXiv:2009.04057. [Google Scholar]
  51. Kumar, A.; Sarawagi, S.; Jain, U. Trainable calibration measures for neural networks from kernel mean embeddings. In Proceedings of the ICML, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  52. Bohdal, O.; Yang, Y.; Hospedales, T. Meta-Calibration: Learning of Model Calibration Using Differentiable Expected Calibration Error. Trans. Mach. Learn. Res. 2023. Available online: https://openreview.net/forum?id=R2hUure38l (accessed on 15 January 2025).
  53. Blundell, C.; Cornebise, J.; Kavukcuoglu, K.; Wierstra, D. Weight uncertainty in neural network. In Proceedings of the ICML, Lille, France, 6–11 July 2015. [Google Scholar]
  54. Lakshminarayanan, B.; Pritzel, A.; Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proceedings of the NeurIPS, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  55. Gal, Y.; Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the ICML, New York City, NY, USA, 19–24 June 2016. [Google Scholar]
  56. Jang, E.; Gu, S.; Poole, B. Categorical Reparametrization with Gumble-Softmax. In Proceedings of the ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  57. Zhang, J.; Kailkhura, B.; Han, T.Y.J. Mix-n-match: Ensemble and compositional methods for uncertainty calibration in deep learning. In Proceedings of the ICML, Virtual Event, 13–18 July 2020. [Google Scholar]
  58. Laves, M.H.; Ihler, S.; Kortmann, K.P.; Ortmaier, T. Well-calibrated model uncertainty with temperature scaling for dropout variational inference. arXiv 2019, arXiv:1909.13550. [Google Scholar]
  59. Naeini, M.P.; Cooper, G.; Hauskrecht, M. Obtaining well calibrated probabilities using bayesian binning. In Proceedings of the AAAI, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  60. Gao, C.; Li, S.; Lei, W.; Chen, J.; Li, B.; Jiang, P.; He, X.; Mao, J.; Chua, T.S. KuaiRec: A Fully-observed Dataset and Insights for Evaluating Recommender Systems. In Proceedings of the CIKM, Atlanta, GA, USA, 17–21 October 2022. [Google Scholar]
  61. Chen, J.; Dong, H.; Qiu, Y.; He, X.; Xin, X.; Chen, L.; Lin, G.; Yang, K. AutoDebias: Learning to Debias for Recommendation. In Proceedings of the SIGIR, Online, 11–15 July 2021. [Google Scholar]
  62. Li, H.; Wu, K.; Zheng, C.; Xiao, Y.; Wang, H.; Geng, Z.; Feng, F.; He, X.; Wu, P. Removing Hidden Confounding in Recommendation: A Unified Multi-Task Learning Approach. In Proceedings of the NeurIPS, Vancouver, BC, Canada, 10–15 December 2024. [Google Scholar]
  63. Li, H.; Xiao, Y.; Zheng, C.; Wu, P. Balancing Unobserved Confounding with a Few Unbiased Ratings in Debiased Recommendations. In Proceedings of the WWW, Austin, TX, USA, 30 April–4 May 2023. [Google Scholar]
  64. Koren, Y.; Bell, R.; Volinsky, C. Matrix factorization techniques for recommender systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  65. Swaminathan, A.; Joachims, T. The Self-Normalized Estimator for Counterfactual Learning. In Proceedings of the NeurIPS, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  66. Guo, S.; Zou, L.; Liu, Y.; Ye, W.; Cheng, S.; Wang, S.; Chen, H.; Yin, D.; Chang, Y. Enhanced Doubly Robust Learning for Debiasing Post-Click Conversion Rate Estimation. In Proceedings of the SIGIR, Online, 11–15 July 2021. [Google Scholar]
  67. Dai, Q.; Li, H.; Wu, P.; Dong, Z.; Zhou, X.H.; Zhang, R.; Zhang, R.; Sun, J. A generalized doubly robust learning framework for debiasing post-click conversion rate prediction. In Proceedings of the KDD, Washington, DC, USA, 14–18 August 2022. [Google Scholar]
  68. Li, H.; Xiao, Y.; Zheng, C.; Wu, P.; Cui, P. Propensity Matters: Measuring and Enhancing Balancing for Recommendation. In Proceedings of the ICML, Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  69. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef]
  70. Järvelin, K.; Kekäläinen, J. Cumulated gain-based evaluation of IR techniques. ACM Trans. Inf. Syst. (TOIS) 2002, 20, 422–446. [Google Scholar] [CrossRef]
  71. Lipton, Z.C.; Elkan, C.; Naryanaswamy, B. Optimal thresholding of classifiers to maximize F1 measure. In Proceedings of the Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2014, Nancy, France, 15–19 September 2014; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2014; pp. 225–239. [Google Scholar]
Figure 1. Categorization of model calibration methods: Post-hoc Calibration [46,47,48], Regularization [36,49,50,51,52], Uncertainty Estimation [53,54,55,56], and Hybrid Calibration [57,58].
Figure 1. Categorization of model calibration methods: Post-hoc Calibration [46,47,48], Regularization [36,49,50,51,52], Uncertainty Estimation [53,54,55,56], and Hybrid Calibration [57,58].
Entropy 27 00196 g001
Figure 2. Comparison of joint calibration and individual calibration on the Yahoo! R3 dataset, with different numbers of candidate propensity and imputation models.
Figure 2. Comparison of joint calibration and individual calibration on the Yahoo! R3 dataset, with different numbers of candidate propensity and imputation models.
Entropy 27 00196 g002
Figure 3. Comparison of joint calibration and individual calibration on the KuaiRec dataset, with different numbers of candidate propensity and imputation models.
Figure 3. Comparison of joint calibration and individual calibration on the KuaiRec dataset, with different numbers of candidate propensity and imputation models.
Entropy 27 00196 g003
Figure 4. Impact of model calibration hyper-parameter λ p r o p in multiple propensity calibration and λ i m p in multiple imputation calibration on Coat dataset.
Figure 4. Impact of model calibration hyper-parameter λ p r o p in multiple propensity calibration and λ i m p in multiple imputation calibration on Coat dataset.
Entropy 27 00196 g004
Figure 5. Impact of model calibration hyper-parameter λ p r o p in multiple propensity calibration and λ i m p in multiple imputation calibration on KuaiRec dataset.
Figure 5. Impact of model calibration hyper-parameter λ p r o p in multiple propensity calibration and λ i m p in multiple imputation calibration on KuaiRec dataset.
Entropy 27 00196 g005
Figure 6. Effect of varying M in soft binning strategy on prediction performance on Coat dataset.
Figure 6. Effect of varying M in soft binning strategy on prediction performance on Coat dataset.
Entropy 27 00196 g006
Figure 7. Effect of varying M in soft binning strategy on prediction performance on KuaiRec dataset.
Figure 7. Effect of varying M in soft binning strategy on prediction performance on KuaiRec dataset.
Entropy 27 00196 g007
Table 1. Performance on AUC, NDCG@T, and F1@T on Coat, Yahoo! R3 and KuaiRec. The best and the second best results are bolded and underlined, where * means statistically significant results ( p - value 0.05 ) using the paired-t-test.
Table 1. Performance on AUC, NDCG@T, and F1@T on Coat, Yahoo! R3 and KuaiRec. The best and the second best results are bolded and underlined, where * means statistically significant results ( p - value 0.05 ) using the paired-t-test.
CoatYahoo! R3KuaiRec
MethodAUCNDCG@5F1@5AUCNDCG@5F1@5AUCNDCG@20F1@20
Naive0. 703 ± 0.006 0. 605 ± 0.012 0. 467 ± 0.007 0. 673 ± 0.001 0. 635 ± 0.002 0. 306 ± 0.002 0. 753 ± 0.001 0. 449 ± 0.002 0. 124 ± 0.002
IPS0. 717 ± 0.007 0. 617 ± 0.009 0. 473 ± 0.008 0. 678 ± 0.001 0. 638 ± 0.002 0. 318 ± 0.002 0. 755 ± 0.004 0. 452 ± 0.010 0. 131 ± 0.004
SNIPS0. 714 ± 0.012 0. 614 ± 0.012 0. 474 ± 0.009 0. 683 ± 0.002 0. 639 ± 0.002 0. 316 ± 0.002 0. 754 ± 0.003 0. 453 ± 0.004 0. 126 ± 0.003
ASIPS0. 719 ± 0.009 0. 618 ± 0.012 0. 476 ± 0.009 0. 679 ± 0.003 0. 640 ± 0.003 0. 319 ± 0.003 0. 757 ± 0.005 0. 474 ± 0.007 0. 130 ± 0.005
IPS-V20. 726 ± 0.005 0. 627 ± 0.009 0. 479 ± 0.008 0. 685 ± 0.002 0. 646 ± 0.003 0. 320 ± 0.002 0. 764 ± 0.001 0. 476 ± 0.003 0. 135 ± 0.003
KBIPS0. 714 ± 0.003 0. 618 ± 0.010 0. 474 ± 0.007 0. 676 ± 0.002 0. 642 ± 0.003 0. 318 ± 0.002 0. 763 ± 0.001 0. 463 ± 0.007 0. 134 ± 0.002
AKBIPS0. 732 ± 0.004 0. 636 ± 0.006 0. 483 ± 0.006 0. 689 ± 0.001 0. 658 ± 0.002 0. 324 ± 0.002 0. 766 ± 0.003 0. 478 ± 0.009 0. 138 ± 0.003
DR0. 718 ± 0.008 0. 623 ± 0.009 0. 474 ± 0.007 0. 684 ± 0.002 0. 658 ± 0.003 0. 326 ± 0.002 0. 755 ± 0.008 0. 462 ± 0.010 0. 135 ± 0.005
DR-JL0. 723 ± 0.005 0. 629 ± 0.007 0. 479 ± 0.005 0. 685 ± 0.002 0. 653 ± 0.002 0. 324 ± 0.002 0. 766 ± 0.002 0. 467 ± 0.005 0. 136 ± 0.003
MRDR-JL0. 727 ± 0.005 0. 627 ± 0.008 0. 480 ± 0.008 0. 684 ± 0.002 0. 652 ± 0.003 0. 325 ± 0.002 0. 768 ± 0.005 0. 473 ± 0.007 0. 139 ± 0.004
DR-BIAS0. 726 ± 0.004 0. 629 ± 0.009 0. 482 ± 0.007 0. 685 ± 0.002 0. 653 ± 0.002 0. 325 ± 0.003 0. 768 ± 0.003 0. 477 ± 0.006 0. 137 ± 0.004
DR-MSE0. 727 ± 0.007 0. 631 ± 0.008 0. 484 ± 0.007 0. 687 ± 0.002 0. 657 ± 0.003 0. 327 ± 0.003 0. 770 ± 0.003 0. 480 ± 0.006 0. 140 ± 0.003
MR0. 724 ± 0.004 0. 636 ± 0.006 0. 481 ± 0.006 0. 691 ± 0.002 0. 647 ± 0.002 0. 316 ± 0.003 0. 776 ± 0.005 0. 483 ± 0.006 0. 142 ± 0.003
TDR0. 714 ± 0.006 0. 634 ± 0.011 0. 483 ± 0.008 0. 688 ± 0.003 0. 662 ± 0.002 0. 329 ± 0.002 0. 772 ± 0.003 0. 486 ± 0.005 0. 140 ± 0.003
TDR-JL0. 731 ± 0.005 0. 639 ± 0.007 0. 484 ± 0.007 0. 689 ± 0.002 0. 656 ± 0.004 0. 327 ± 0.003 0. 772 ± 0.003 0. 489 ± 0.005 0. 142 ± 0.003
StableDR0. 735 ± 0.005 0. 640 ± 0.007 0. 484 ± 0.006 0. 688 ± 0.002 0. 661 ± 0.003 0. 329 ± 0.002 0. 773 ± 0.001 0. 491 ± 0.003 0. 143 ± 0.003
DR-V20. 734 ± 0.007 0. 639 ± 0.009 0. 487 ± 0.006 0. 690 ± 0.002 0. 660 ± 0.005 0. 328 ± 0.002 0. 773 ± 0.003 0. 488 ± 0.006 0. 142 ± 0.004
KBDR0. 730 ± 0.003 0. 631 ± 0.005 0. 482 ± 0.006 0. 682 ± 0.002 0. 648 ± 0.003 0. 323 ± 0.002 0. 765 ± 0.004 0. 460 ± 0.006 0. 138 ± 0.003
AKBDR 0.745 ± 0.004 0. 645 ± 0.008 0.493±0.0070. 692 ± 0.002 0. 661 ± 0.002 0. 328 ± 0.002 0. 782 ± 0.003 0. 498 ± 0.008 0. 147 ± 0.003
DCE-DR0. 736 ± 0.006 0. 648 ± 0.007 0. 489 ± 0.005 0. 698 ± 0.002 0. 670 ± 0.002 0.333±0.0030. 795 ± 0.004 0. 512 ± 0.005 0. 153 ± 0.002
DCE-TDR0. 740 ± 0.004 0.651±0.0060. 489 ± 0.007 0.701±0.0020.672±0.0020. 331 ± 0.002 0.798 ± 0.005 0.514±0.0060.155±0.002
Cali-MR 0.741 ± 0.002 0.658 ± 0.004 * 0.495 ± 0.004 0.703 ± 0.002 * 0.678 ± 0.002 * 0.338 ± 0.004 * 0.798 ± 0.003 0.521 ± 0.005 * 0.158 ± 0.002 *
Table 2. Ablation study of Cali-MR on Coat, Yahoo! R3 and KuaiRec datasets.
Table 2. Ablation study of Cali-MR on Coat, Yahoo! R3 and KuaiRec datasets.
MethodCoatYahoo! R3KuaiRec
AUCNDCG@5F1@5AUCNDCG@5F1@5AUCNDCG@20F1@20
Cali-MR0.7410.6580.4950.7030.6780.3380.7980.5210.158
Cali-MR w/o prop0.7300.6400.4860.7000.6740.3360.7870.4830.151
Cali-MR w/o imp0.7360.6390.4830.7030.6740.3360.7930.5090.155
Cali-MR w/o imp & prop0.7270.6350.4770.6980.6670.3310.7830.4820.148
Table 3. Comparison of Expected Calibration Error (ECE) on Coat, Yahoo! R3 and KuaiRec datasets, where RD refers to the relative decrease value over the corresponding base model.
Table 3. Comparison of Expected Calibration Error (ECE) on Coat, Yahoo! R3 and KuaiRec datasets, where RD refers to the relative decrease value over the corresponding base model.
CoatYahoo! R3KuaiRec
MethodECERDECERDECERD
DR-JL0.1626-0.0589-0.0999-
DCE-DR0.14280.01980.05540.00350.04510.0548
TDR-JL0.1468-0.0481-0.0499-
DCE-TDR0.12700.01980.04760.00050.04880.0011
MR0.1512-0.1594-0.2367-
Cali-MR0.12390.02730.03490.12450.05190.1848
Table 4. Comparison of training time (seconds) and parameter size (Naive method denoted as 1×) on Coat, Yahoo! R3 and KuaiRec datasets.
Table 4. Comparison of training time (seconds) and parameter size (Naive method denoted as 1×) on Coat, Yahoo! R3 and KuaiRec datasets.
MethodCoatYahoo! R3KuaiRec
TimeParamsTimeParamsTimesParams
Naive4.0426.0711.67
IPS6.5432.8415.14
DR17.4143.1831.20
DR-JL18.81166.21110.01
TDR-JL21.13128.88101.09
MR (J = 2, K = 2)13.55127.8114.78
Cali-DR (J = 1, K = 1)21.03132.98128.31
Cali-MR (J = 2, K = 2)23.28144.56124.58
Cali-MR (J = 1, K = 5)21.57139.80130.34
Cali-MR (J = 1, K = 5)21.43136.77141.46
Cali-MR (J = 3, K = 3)19.55127.80106.66
Cali-MR (J = 3, K = 3)25.43151.97127.47
Cali-MR (J = 3, K = 5)23.59166.42163.02
Cali-MR (J = 5, K = 3)23.71143.76103.60
Cali-MR (J = 5, K = 3)29.35132.74128.03
Cali-MR (J = 5, K = 5)25.6511×153.7911×166.7311×
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gong, S.; Ma, C. Gradient-Based Multiple Robust Learning Calibration on Data Missing-Not-at-Random via Bi-Level Optimization. Entropy 2025, 27, 196. https://doi.org/10.3390/e27020196

AMA Style

Gong S, Ma C. Gradient-Based Multiple Robust Learning Calibration on Data Missing-Not-at-Random via Bi-Level Optimization. Entropy. 2025; 27(2):196. https://doi.org/10.3390/e27020196

Chicago/Turabian Style

Gong, Shuxia, and Chen Ma. 2025. "Gradient-Based Multiple Robust Learning Calibration on Data Missing-Not-at-Random via Bi-Level Optimization" Entropy 27, no. 2: 196. https://doi.org/10.3390/e27020196

APA Style

Gong, S., & Ma, C. (2025). Gradient-Based Multiple Robust Learning Calibration on Data Missing-Not-at-Random via Bi-Level Optimization. Entropy, 27(2), 196. https://doi.org/10.3390/e27020196

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop