PaperHub
6.8
/10
Poster4 位审稿人
最低3最高5标准差0.8
4
3
5
5
2.8
置信度
创新性3.0
质量3.0
清晰度2.5
重要性3.0
NeurIPS 2025

Unveiling Extraneous Sampling Bias with Data Missing-Not-At-Random

OpenReviewPDF
提交: 2025-05-11更新: 2025-10-29

摘要

关键词
Missing-Not-At-RandomSampling Bias

评审与讨论

审稿意见
4

The paper identifies a critical limitation in existing debiasing methods, which assume "exact matching" between training and test sets. It introduces a more realistic "super-population" setting, where samples are drawn independently from the same underlying distribution. Under this setting, even perfectly specified DR estimators are biased due to a new covariance term. To address this, the authors propose SuperDR, which corrects the imputation model by introducing a learnable parameter ε to minimize the empirical covariance, inspired by TMLE.

优缺点分析

Strengths:

  • The clear distinction between exact matching, cold-start, and the proposed super-population setting fills an important gap and better reflects real-world applications.

  • The paper formally derives the bias and shows how the correction term effectively targets the covariance source. Theoretical results (e.g., Theorem 4.3, 4.7) are elegant and insightful.

  • Experiments on multiple datasets, including the large-scale KuaiRec, demonstrate significant and consistent improvements. Analyses directly validate theoretical claims (e.g., performance vs. empirical covariance).

Weaknesses:

  • Lack of detail on whether stabilization techniques (e.g., propensity clipping) were used, which affects reproducibility and variance control explanation.
  • The alternating optimization scheme increases overhead but is not analyzed or discussed in terms of cost-performance trade-offs.
  • Lack of some important related work: [1] Uncertainty Calibration for Counterfactual Propensity Estimation in Recommendation [2] Cdr:Conservative doubly robust learning for debiased recommendation

问题

  • Was any propensity score stabilization used?
  • How were the inner loop steps and learning rates chosen in Algorithm 1?
  • Why apply correction to the imputation model rather than the prediction output?
  • Are MF-based features sufficient for generalization across disjoint user-item samples?

局限性

Yes

最终评判理由

My concerns have been adequately addressed, and I have decided to maintain my positive rating.

格式问题

No

作者回复

We thank the reviewer for efforts in evaluating our work and helpful suggestions. Below, we will try our best to address your concerns and questions.

W1 & Q1: Lack of detail on whether stabilization techniques (e.g., propensity clipping) were used, which affects reproducibility and variance control explanation.

Response to W1 & Q1: We thank the reviewer for pointing out these issues. Yes, we use a propensity clipping technique, and the clipping threshold is a hyperparameter tuned in [0.005, 0.05] on all three datasets. The best threshold is 0.0186 for Coat, 0.0127 for Yahoo! R3, and 0.0112 for the KuaiRec dataset. We will add these details in our revised version.

W2: The alternating optimization scheme increases overhead but is not analyzed or discussed in terms of cost-performance trade-offs.

Response to W2: We thank the reviewer for raising this meaningful concern.

  • Note that the proposed method subtly modifies the imputation loss e^\hat e by adding ϵ(op)\epsilon (o-p), which theoretically does not increase complexity compared to the original joint learning method, due to we only require one more gradient calculation and backward propagation for a single learnable parameter ϵ\epsilon.
  • Empirically, the model structurally is also not more complex. Besides a single learnable parameter ϵ\epsilon, no new learnable parameters are introduced.
  • We also add training time experiments on all three datasets under varying dataset scales (controlled by the ratio parameter bb, varying in {0.1, 0.3, 0.5, 0.7, 0.9}, i.e., we randomly sample 100b% of users and items). We take DR-JL as the baseline. Thus, all results are 1 for DR-JL. The results show that the time complexity of our method is close to the doubly robust joint learning method, and much less than the previous SOTA DR-based methods, demonstrating the efficiency of our method.
Coatb=0.1b=0.3b=0.5b=0.7b=0.9
DR-JL1.0001.0001.0001.0001.000
DR-MSE1.4351.9562.0081.6191.183
MR1.5882.2062.3651.5241.296
DR-V21.2982.0931.8501.4051.028
DCE-TDR2.2352.9063.2521.9121.596
Ours1.0781.1221.2731.1850.956
Yahoo! R3
DR-JL1.0001.0001.0001.0001.000
DR-MSE1.8312.5832.8663.4012.808
MR1.2532.4582.9072.6892.862
DR-V21.4291.9581.3201.9211.891
DCE-TDR2.0143.2713.2472.8643.075
Ours1.2551.4011.5361.4241.356
KuaiRec
DR-JL1.0001.0001.0001.0001.000
DR-MSE1.3711.7062.8711.8672.714
MR1.6042.2943.3552.0363.036
DR-V21.2661.6471.6771.0122.631
DCE-TDR2.7344.0005.0652.8313.607
Ours1.3541.4121.1290.8271.726

W3: Lack of some important related work: [1] Uncertainty Calibration for Counterfactual Propensity Estimation in Recommendation [2] Cdr:Conservative doubly robust learning for debiased recommendation.

Response to W3: We thank the reviewer for the useful suggestions. We will modify the related work section to include these two important works. In addition, we add experiments to compare our method to these two methods. The results are shown below:

MethodsCoatYahoo! R3KuaiRec
AUCNDCG@5Recall@5AUCNDCG@5Recall@5AUCNDCG@50Recall@50
DR-JL0.733±0.0030.650±0.0050.625±0.0070.667±0.0050.655±0.0040.449±0.0080.818±0.0030.620±0.0040.655±0.007
CDR [2]0.743±0.0040.652±0.0050.627±0.0050.672±0.0040.658±0.0030.454±0.0050.823±0.0050.623±0.0040.660±0.005
MRDR-CAL [1]0.744±0.0030.651±0.0040.630±0.0040.675±0.0030.660±0.0030.454±0.0050.822±0.0060.625±0.0050.665±0.006
DCE-TDR0.746±0.0050.654±0.0050.629±0.0060.679±0.0040.662±0.0050.459±0.0040.824±0.0030.632±0.0040.659±0.004
D-DR0.750±0.0040.654±0.0040.630±0.0050.678±0.0040.659±0.0040.456±0.0030.822±0.0040.630±0.0050.672±0.005
Ours0.757±0.0040.667±0.0050.637±0.0070.686±0.0030.667±0.0040.463±0.0030.828±0.0040.640±0.0050.680±0.005

We find that our method still outperforms the baselines.

Q2: How were the inner loop steps and learning rates chosen in Algorithm 1?

Response to Q2: We thank the reviewer for the question. Following the previous implementation in [1-3], the step is fixed to 1. In addition, as shown in lines 250-251 in the manuscript, the learning rate is tuned in {0.001, 0.005, 0.01, 0.02, 0.05} for parameters in prediction, imputation, and propensity model, and in {0.01, 0.05, 0.1, 0.15, 0.2} for ϵ\epsilon. The best choice is {0.01, 0.15} for Coat, {0.005, 0.15} for Yahoo! R3, and {0.005, 0.1} for KuaiRec.

Q3: Why apply correction to the imputation model rather than the prediction output?

Response to Q3: We thank the reviewer for the interesting question. The conclusion is, applying the correction to either the imputation model or the prediction model will lead to exactly the same results! Here are the reasons.

  • For updating the prediction model, we first recap the vanilla DR loss below:

    LDR(θ)=1D(u,i)D[e^u,i+ou,i(eu,ie^u,i)p^u,i].\mathcal L_{\text{DR}}(\theta)=\frac{1}{|\mathcal{D}|} \sum_{(u, i) \in \mathcal{D}}\left[\hat e_{u, i}+\frac{o_{u, i}\left(e_{u, i}-\hat e_{u, i}\right)}{\hat p_{u, i}}\right]. For updating the prediction model, note that whether adding the correction to the imputation model or the prediction model based on this DR loss will not affect the updating of the prediction model, because there is no gradient to the prediction model.

  • For updating the imputation model, we recap the vanilla imputation loss function below Le=1D(u,i)Dou,i(eu,ie^u,i)2p^u,i.\mathcal L_e = \frac{1}{|\mathcal{D}|} \sum_{(u, i) \in \mathcal{D}} \frac{o_{u, i}\left(e_{u, i}-\hat e_{u, i}\right)^2}{\hat p_{u, i}}.

Note that adding the correction to the imputation model or the prediction model based on this imputation loss also makes no difference when updating the imputation model and the parameter ϵ\epsilon.

Q4: Are MF-based features sufficient for generalization across disjoint user-item samples?

Response to Q4:

We thank the reviewer for raising this concern.

  • We add experiments on all three datasets, demonstrating that MF-based features are sufficient for generalization. Specifically, for all three datasets, we randomly split the training data (i.e., MNAR data) into training / validation / test data, with a ratio 0.8 / 0.1 / 0.1. Then we use an MLP as the backbone to train a prediction model using the naive loss, i.e., loss on the observed user-item pairs. Lnaive(θ)=1O(u,i)Oeu,i.\mathcal L_{\text{naive}}(\theta)=\frac{1}{|\mathcal{O}|} \sum_{(u, i) \in \mathcal{O}}e_{u,i}.

The test performance on all three datasets corresponding to MSE and AUC is shown below:

DatasetsAUCMSE
Coat0.9520.092
Yahoo! R30.9460.096
KuaiRec0.8890.117
  • The results show that simple naive loss-based neural networks demonstrate excellent performance without data MNAR on all three datasets, proving MF-based features are sufficient for generalization.
  • Furthermore, since KuaiRec is a more complicated industrial dataset, its lower results compared to Coat and Yahoo! R3 are in line with expectations.
  • Finally, naive methods perform poorly with data MNAR (Coat and Yahoo! R3 have around 20% lower AUC, and KuaiRec around 10% lower AUC), highlighting the need for debiasing in real-world datasets.

We believe the above discussion has thoroughly addressed your concerns regarding our work. We would be grateful if you could kindly consider raising your rating. Thank you!

评论

Thank you for your detailed response. My concerns have been adequately addressed, and I have decided to maintain my positive rating.

评论

We are happy to hear that your concerns have been adequately addressed. We are truly grateful for your dedicated time and constructive comments!

审稿意见
3

This paper investigates a practical and underexplored challenge in unbiased learning: selection bias in scenarios where training and test sets do not share identical units (e.g., different users/items in recommender systems), though the joint distribution of features and outcomes remains the same. The authors argue that existing doubly robust (DR) estimators, which assume exact matching, are insufficient in this "super-population" setting. They derive the theoretical bias of DR under this new scenario and identify a non-trivial covariance term that induces additional bias. To address this, the paper proposes SuperDR, a novel estimator with a corrected imputation model designed to control this covariance term. They provide theoretical guarantees on bias, variance, and generalization error, and validate their method on three real-world datasets, demonstrating superior performance over existing approaches.

优缺点分析

Strengths

S1. Principled Solution: The proposed SuperDR method is well-motivated and theoretically grounded. The use of imputation correction inspired by targeted learning is elegant and practical.

S2. Strong Empirical Results: Experiments across three datasets (Coat, Yahoo! R3, KuaiRec) demonstrate that SuperDR consistently outperforms strong DR baselines, particularly in low-overlap scenarios between train and test data.

Weaknesses

W1. Assumption of Equal Joint Distributions: The super-population scenario assumes Ptrain(x,r)=Ptest(x,r)P_{\rm train}(x,r)=P_{\rm test}(x,r); in practice, covariate drift or nonstationarity could violate this and undermine unbiasedness.

W2. Dependence on Propensity Pre-training: Requires a pre-trained propensity model (e.g., logistic regression) whose miscalibration could propagate through the correction step, yet sensitivity to propensity errors is not deeply explored.

W3. Computational Overhead: Jointly learning three components (prediction, imputation, and ε) may increase training time and tuning complexity compared to standard DR methods; scalability to very large item × user spaces warrants further study.

W4. Limited Domain Scope: Experiments focus on recommender-system datasets; applicability to other MNAR settings (e.g., causal inference, medical missingness) remains to be demonstrated.

问题

Please refer to the Weaknesses.

局限性

Yes.

格式问题

N/A

作者回复

We thank the reviewer for efforts in evaluating our work and helpful suggestions. Below, we will try our best to address your concerns and questions.

W1. Assumption of Equal Joint Distributions: The super-population scenario assumes Ptrain(x,r)=Ptest(x,r)P_{\rm train}(x,r)=P_{\rm test}(x,r); in practice, covariate drift or nonstationarity could violate this and undermine unbiasedness.

Response to W1: We thank the reviewer for the question. However, we think there is a huge misunderstanding.

  • Note that we assume Ptrain(x,r)=Ptest(x,r)P_{\rm train}(x,r)=P_{\rm test}(x,r) means that the target data distribution (not the observed data distribution) in training is the same as the distribution in test data. Then, due to the MNAR issue, the observed data distribution Pobs(x,r)P_{\rm obs}(x,r) is different from the test data.
  • For example, the Coat dataset includes 290 users and 300 items with 6,960 MNAR observations, then we only assume target data distribution in training Ptrain(x,r)P_{\rm train}(x,r) (the distribution of 290 * 300 = 87,000 user-item pairs) the same as the test distribution Ptest(x,r)P_{\rm test}(x,r), and we do not assume distribution on 6,960 observations Pobs(x,r)P_{\rm obs}(x,r) the same as the test distribution. We will make it clearer in our revised manuscript.

W2. Dependence on Propensity Pre-training: Requires a pre-trained propensity model (e.g., logistic regression) whose miscalibration could propagate through the correction step, yet sensitivity to propensity errors is not deeply explored.

Response to W2: We thank the reviewer for the question. The existing DR-based training algorithms are mainly divided into two types:

  • The first is based on the joint training algorithm, which pretrains the propensity model, and then trains the imputation model and the prediction model, such as DR-JL [1], MRDR-JL [2], DCE-DR [3], etc.
  • The second is based on the multi-task learning algorithm, which learns the propensity model, the prediction model, and the imputation model together, such as ESCM2-DR [4], SDR [5], DR-V2 [6], etc.

As suggested by the reviewer, we added experiments on all three datasets, Coat, Yahoo, and Kuairec, where the propensity model is 1 layer (i.e., logistic regression), 2 layers, and 3 layers neural networks, respectively, to explore the sensitivity to propensity errors of our method. In addition, we also implemented both joint learning and multi-task learning as the learning algorithm to investigate whether our method needs a pre-trained propensity model. The results are as follows:

Num of layers in propensity modelCoatYahoo! R3KuaiRec
AUCNDCG@5Recall@5AUCNDCG@5Recall@5AUCNDCG@50Recall@50
Layer1_MTL0.7540.6630.6380.6880.6690.4670.8290.6330.670
Layer1_JL0.7530.6610.6490.6740.6540.4500.8260.6310.667
Layer2_MTL0.7520.6630.6360.6880.6680.4640.8220.6370.674
Layer2_JL0.7570.6670.6370.6860.6670.4630.8280.6400.680
Layer3_MTL0.7540.6690.6380.6780.6590.4550.8300.6410.681
Layer3_JL0.7400.6550.6240.6840.6650.4640.8290.6360.675
Best baseline0.7500.6540.6300.6790.6620.4590.8240.6320.671

The results show that our method is stable to the choice of propensity model and learning algorithm.

W3. Computational Overhead: Jointly learning three components (prediction, imputation, and ϵ\epsilon) may increase training time and tuning complexity compared to standard DR methods; scalability to very large item × user spaces warrants further study.

Response to W3: We thank the reviewer for raising this meaningful concern.

  • Note that the proposed method subtly modifies the imputation loss e^\hat e by adding ϵ(op)\epsilon (o-p), which theoretically does not increase complexity compared to the original joint learning method, due to we only require one more gradient calculation and backward propagation for a single learnable parameter ϵ\epsilon.
  • Empirically, the model structurally is also not more complex. Besides a single learnable parameter ϵ\epsilon, no new learnable parameters are introduced.
  • We also add training time experiments on all three datasets under varying dataset scales (controlled by the ratio parameter bb, varying in {0.1, 0.3, 0.5, 0.7, 0.9}, i.e., we randomly sample 100b% of users and items). We take DR-JL as the baseline. Thus, all results are 1 for DR-JL. The results show that the time complexity of our method is close to the doubly robust joint learning method, and much less than the previous SOTA DR-based methods, demonstrating the efficiency of our method.
Coatb=0.1b=0.3b=0.5b=0.7b=0.9
DR-JL1.0001.0001.0001.0001.000
DR-MSE1.4351.9562.0081.6191.183
MR1.5882.2062.3651.5241.296
DR-V21.2982.0931.8501.4051.028
DCE-TDR2.2352.9063.2521.9121.596
Ours1.0781.1221.2731.1850.956
Yahoo! R3
DR-JL1.0001.0001.0001.0001.000
DR-MSE1.8312.5832.8663.4012.808
MR1.2532.4582.9072.6892.862
DR-V21.4291.9581.3201.9211.891
DCE-TDR2.0143.2713.2472.8643.075
Ours1.2551.4011.5361.4241.356
KuaiRec
DR-JL1.0001.0001.0001.0001.000
DR-MSE1.3711.7062.8711.8672.714
MR1.6042.2943.3552.0363.036
DR-V21.2661.6471.6771.0122.631
DCE-TDR2.7344.0005.0652.8313.607
Ours1.3541.4121.1290.8271.726

W4. Limited Domain Scope: Experiments focus on recommender-system datasets; applicability to other MNAR settings (e.g., causal inference, medical missingness) remains to be demonstrated.

Response to W4: We thank the reviewer for the question. As suggested by the reviewer, we further explore the effectiveness of our method on two scenarios, the medical field and business field, with two datasets, IHDP and Jobs.

  • The semi-synthetic IHDP dataset [7] is constructed from the Infant Health and Development Program (IHDP) with 747 individuals and 25 covariates.
  • The real-world Jobs dataset [8] is based on the National Supported Work program with 3,212 individuals and 17 covariates.

In this scenario, the DR estimator has the following form:

τ(x)=E[T(Ym1(X))π(X)(1T)(Ym0(X))1π(X)+m1(X)m0(X)X=x],\tau(x)=E[\frac{T(Y-m_1(X))}{\pi(X)}-\frac{(1-T)(Y-m_0(X))}{1-\pi(X)}+m_1(X)-m_0(X) | X=x],

where TT \in {0,1} is the binary treatment indicator, YY is the outcome, π(X)=P(T=1X)\pi(X) = P(T=1|X) is the propensity model, and m0(X),m1(X)m_0(X), m_1(X) are the imputation models for T=0T = 0 and T=1T = 1, respectively.

Following the previous study [9], we split the data into training/validation/testing sets with ratios 63/27/10 and 56/24/20, with 10 repeated times on the IHDP and the Jobs datasets, respectively. Note that each sample is already disjoint, and the training distribution is the same as the test distribution due to we randomly split the data. For evaluation, we use ϵPEHE\sqrt{\epsilon_{\text{PEHE}}} and ϵATE\epsilon_{\text{ATE}} for IHDP, and RpolR_{\text{pol}} and ϵATT\epsilon_{\text{ATT}} for Jobs. The definitions are shown below:

  • ϵPEHE=1Ni=1N[(Y^i(1)Y^i(0))(Yi(1)Yi(0))]2\epsilon_{\text{PEHE}} = {\frac{1}{N} \sum_{i=1}^N [(\hat {Y}_i(1) - \hat {Y}_i(0)) - (Y_i(1) - Y_i(0))]^2}, where Y^i(1)\hat {Y}_i(1) and Y^i(0)\hat {Y}_i(0) are the predicted values for the corresponding true potential outcomes of unit ii.
  • ϵATE=1Ni=1N((Y^i(1)Y^i(0))(Yi(1)Yi(0))).\epsilon_{\text{ATE}}=\frac{1}{N}| \sum_{i=1}^N ((\hat {Y}_i(1) - \hat {Y}_i(0)) - (Y_i(1) - Y_i(0)))|.
  • RPol=1(E[Y(1)Y^(1)Y^(0)>0,X=1]P(Y^(1)Y^(0)>0)+E[Y(0)Y^(1)Y^(0)0,X=0]P(Y^(1)Y^(0)0)){R}_{\text{Pol}}= 1-(\mathbb{E}[Y(1) \mid \hat {Y}(1) - \hat {Y}(0) >0, X=1] \cdot \mathbb{P}(\hat {Y}(1) - \hat {Y}(0)>0)+\mathbb{E}[Y(0) \mid \hat {Y}(1) - \hat {Y}(0) \leq 0, X=0] \cdot \mathbb{P}(\hat {Y}(1) - \hat {Y}(0) \leq 0)), where T,C,ET, C, E are the indexes of treatment, control, and randomized sample set, respectively.
  • ϵATT=ATT1TiT(Y^i(1)Y^i(0)\epsilon_{\text{ATT}} = |\text{ATT} - \frac{1}{|T|} \sum_{i \in T} (\hat {Y}_i(1) - \hat {Y}_i(0)|

where ATT=1TiTYi1CEi(CE)Yi\text{ATT} = |\frac{1}{|T|} \sum_{i \in T}{Y_i} - \frac{1}{|C \cap E|} \sum_{i \in (C \cap E)}{Y_i}|.

The lower the metrics, the better the performance. See [9] for a more detailed definition of evaluation metrics. The performance on the test set is shown below:

Results on IHDP:

MethodsϵPEHE\sqrt{\epsilon_{\text{PEHE}}}ϵATE\epsilon_{\text{ATE}}
DR-JL1.37±0.300.22±0.15
Ours0.95±0.110.14±0.09

Results on Jobs:

MethodsRPol{R}_{\text{Pol}}ϵATT\epsilon_{\text{ATT}}
DR-JL0.27±0.060.05±0.03
Ours0.23±0.070.05±0.02

The results show that our method can be applied to those datasets, and the performance is better than the DR-JL method.

We believe the above discussion has thoroughly addressed your concerns regarding our work. We would be grateful if you could kindly consider raising your rating. Thank you!


References

[1] Doubly Robust Joint Learning for Recommendation on Data Missing Not at Random. ICML 2019.

[2] Enhanced Doubly Robust Learning for Debiasing Post-Click Conversion Rate Estimation. SIGIR 2021.

[3] Doubly Calibrated Estimator for Recommendation on Data Missing Not At Random. WWW 2024.

[4] ESCM2^2: Entire Space Counterfactual Multi-Task Model for Post-Click Conversion Rate Estimation. SIGIR 2022.

[5] StableDR: Stabilized Doubly Robust Learning for Recommendation on Data Missing Not at Random. ICLR 2023.

[6] Propensity Matters: Measuring and Enhancing Balancing for Recommendation. ICML 2023.

[7] Bayesian Nonparametric Modeling for Causal Inference. Journal of Computational and Graphical Statistics 2011.

[8] Evaluating the Econometric Evaluations of Training Programs with Experimental Data. The American Economic Review 1986.

[9] Estimating Individual Treatment Effect: Generalization Bounds and Algorithms. ICML 2017.

评论

Dear Reviewer gNuC,

Thanks again for your insightful comments and interesting questions. During the rebuttal period, we have validated time efficiency, insensitivity to the propensity model, and applicability to other domains (such as medical and business) of our method. In addition, we highlight that the previous method requires “exact match user and item” between the training set and test set, while our method only assumes Ptrain(x,r)=Ptest(x,r)P_{\rm train}(x,r)=P_{\rm test}(x,r). To further validate our method's applicability in real-world scenarios (may have covariate shift or nonstationarity), we show our method is still valid and outperforms baselines if the assumption Ptrain(x,r)=Ptest(x,r)P_{\rm train}(x,r)=P_{\rm test}(x,r) is violated.

We add experiments on the industrial dataset KuaiRec. To process the data, we first randomly split 80% users and items to construct the training set to make sure we do not see all users and items during training. Then we consider the scenario with Ptrain(x,r)Ptest(x,r)P_{\rm train}(x,r) \neq P_{\rm test}(x,r):

  • We assign a propensity according to the item frequency and the user's active level (measured by the number of rated items) to split the dataset. This will lead to the first training set containing popular users and items, and the second containing non-popular users and items. We set the threshold to make the number of user-item pairs the same in the two datasets.

For all methods, we test on the MAR dataset. The results are shown below:

AUCRecall@50NDCG@50
First set (popular users & items)DR-JL0.8190.6700.633
Ours0.8230.6660.629
Second set (non-popular users & items)DR-JL0.7050.4800.478
Ours0.7170.5020.507

The results show that our method still outperforms the baseline DR-JL, especially under the second set (when the covariate shift is large). One possible reason is that our method deals with a super-population scenario, which is more robust toward a huge distribution shift.

A more practical scenario is that data may arrive sequentially with Ptrain(x,r)Ptest(x,r)P_{\rm train}(x,r) \neq P_{\rm test}(x,r), and we need to retrain our model efficiently on new data. To further show the applicability of our method, we evaluate whether our method can be efficiently applied in this scenario. We propose two ways to update our predictions efficiently:

  • We store the parameters in models and optimizers at the end of training, and fine-tune them when new data comes (named "continue training").
  • We freeze the parameters in prediction, imputation, and propensity model, and add a learnable shallow residual structure, which is zero-initialized when new data comes (named "adding residual").
  • Also, we consider the full retraining of all parameters on both old and new data as the oracle baselines.

We also add experiments on the industrial dataset KuaiRec and first randomly split 80% users and items. Then we consider the following scenario:

  • We randomly split the user set and item set into two parts with a ratio of 50/50, and train our method on the first dataset. Then we assign a propensity according to the item frequency and the user's active level, sampling popular users and items into the first time retraining, and the remaining into the second time retraining. Also, we adjust the threshold to make the number of user-item pairs the same in each retraining.

The results on the MAR test data are shown below:

DataInconsistent in distribution
AUCNDCG@50Recall@50
Train Set1 (50%)DR-JL0.8170.6230.663
Ours0.8250.6360.678
Train Set2 (25%)Continue training on DR-JL0.8110.6310.666
Continue training on ours0.8210.6330.669
Adding residual on DR-JL0.8160.6200.653
Adding residual on ours0.8210.6400.680
Full training on DR-JL0.8210.6260.665
Full training on ours0.8270.6400.679
Train Set3 (25%)Continue training on DR-JL0.7120.4570.449
Continue training on ours0.7290.4920.504
Adding residual on DR-JL0.7150.4500.449
Adding residual on ours0.7250.4790.506
Full training on DR-JL0.8220.6270.668
Full training on ours0.8280.6400.679
  • We find that our methods can efficiently adapt our predictions as new data becomes available.
  • In addition, when there is a large gap between the new data and the original training data (such as non-popular users and items, as shown in the "Train Set3”), though the performance of our method will drop, our method significantly outperforms DR-JL.

By now, we hope we have fully addressed your concerns! We sincerely expect that the added extensive experiments and discussions on time efficiency, insensitivity to the propensity model, and applicability to other domains (such as medical and business) as well as the assumption Ptrain(x,r)=Ptest(x,r)P_{\rm train}(x,r)=P_{\rm test}(x,r) is violated, will enable you to raise your rating!

评论

Thanks for the rebuttal. However, the explanations regarding W1 and W2 are still unclear.

For W1, first and foremost, the covariate drift or nonstationarity of data distribution means the preference shift of users or items becoming outdated, rather than the observed data. Second, the rebuttal contains many elusive concepts. For example, what is 'target data distribution in training distribution' ? Why is there a distribution in a distribution? what is 'popular user'? I assume the authors use this term to mean 'active user'. Besides, the results on the first set (popular users & items) of KuaiRec show DR-JL outperforms the proposed method w.r.t. Recall and NDCG. In addition, it is elusive why testing on popular and unpopular items separately can demonstrate that the proposed method can handle nonstationarity between training and test sets. Last but not least, what is the necessity of introducing a new baseline, DR-JL?

For W2, my concern is that the errors made by the propensity model may degrade the final accuracy. However, the authors compare two types of methods, both of which involve a propensity model, and the results show that these two types of methods achieve comparable accuracy. I don't understand how such results can address my concern.

Besides, the above rebuttals are merely built upon experimental results, which are not sufficiently convincing.

评论

Thank you for your thoughtful review and valuable questions! We address your questions point-to-point in the following.

For W1, first and foremost, the covariate drift or nonstationarity of data distribution means the preference shift of users or items becoming outdated, rather than the observed data.

Thank you for the great question! We would like to clarify that covariate drift or nonstationarity of data distribution is orthogonal to our studied problem.

  • We fully agree that the covariate drift or nonstationarity of data distribution means the preference shift of users or items becoming outdated, rather than the observed data, which is widely exist and has been well studied in the out-of-distribution (OOD) community, e.g., OOD in recommender system [1].

  • However, our paper focus on the sample selection bias, which is formally equivalent to the semi-supervised learning (SSL) with missing-not-at-random (MNAR) labels. In SSL articles, people are more concerned with covariate drift Pobs(X,Y)=P(X,YO=1)P(X,Y)P_{obs}(X, Y)=P(X, Y\mid O=1)\neq P(X, Y) due to MNAR labeled data, in which OO is the labeled data indicator. They usually assume that the OOD doesn't exist, because the missing-label problem more emphasize on the (non-random) sampling or labeling process, e.g., user will not randomly rate his/her interacted items, but selectively rate some).

  • In summary, the reviewer mentioned OOD problem is definitely practical and important, but is orthogonal to our studied SSL with MNAR labels problem. To avoid misunderstanding that we use a strong assumption, we feel the necessity to change PtrainP_{train} to PidealP_{ideal} (please kindly find more details below). We consider the combination of OOD and selection bias to be an interesting research direction for the future work.

Next, we would like to briefly discuss, What makes our unique contribution?

  • To our best knowledge, all previous work studied selection bias in recommender system are limited to the fixed user-item pairs.
    • That is, given a finite user-item dataset Dideal=u1,u2,,um×i1,i2,,inD_{ideal}=\\{u_1, u_2, \ldots, u_m\\} \times \\{i_1, i_2, \ldots, i_n\\} as the ideal training set, due to the MNAR labels, we can only observe the ratings from a subset of user-item pairs, i.e., DobsDidealD_{obs}\subset D_{ideal}.
    • In this paper, we find that previous widely-used inverse propensity scoring (IPS) and doubly robust (DR) methods can only achieve unbiasedness when Dtest=DidealD_{test}=D_{ideal}, not even under Ptest=PidealP_{test}=P_{ideal}, that is, cannot generalize well even for a new user um+1u_{m+1} or a item in+1i_{n+1} that sampling from the same distribution of PidealP_{ideal}.
    • Technically, this is because previous work only consider the randomness of Ou,iO_{u, i}, not for the rest variables Xu,i,Yu,iX_{u, i}, Y_{u, i}. By treating all variables are random, we find an interesting theoretical gap between previous finite-population DR and our super-population DR (see Lemma 4.1 v.s. Theorem 4.3, and Corollary 4.2 v.s. Corollary 4.4), as the core contribution of our paper.

Second, the rebuttal contains many elusive concepts. For example, what is 'target data distribution in training distribution'?

Thank you for raising this concern. We acknowledge that this is vague, but what we intend to say is that we only need to assume that the ideal distribution (note that the training distribution with MNAR labels is sampled from this ideal distribution) and the test distribution are the same, i.e., Ptest=PidealP_{test}=P_{ideal}, without requiring the datasets to be exact the same under finite population, i.e, Dtest=DidealD_{test}=D_{ideal}. In short, we extend the same dataset requirement to same distribution requirement, and we consider further incorporating OOD to allow PtestPidealP_{test}\neq P_{ideal} is natural, by combining the existing methods for overcoming distribution shifts.

What is 'popular user'? I assume the authors use this term to mean 'active user'.

Thank you for pointing out this issue! You are absolutely right that it should be 'active user' rather than 'popular user'.

评论

Besides, the results on the first set (popular users & items) of KuaiRec show DR-JL outperforms the proposed method w.r.t. Recall and NDCG.

Thank you so much for your careful reading. We checked the experimental log and found that this is actually true result. We are honest to say testing on the first set (popular users & items), by tuning both DR-JL and our method entirely, their performance are comparable. Please also kindly note that our method outperform the DR-JL in almost the remaining test scenarios.

In addition, it is elusive why testing on popular and unpopular items separately can demonstrate that the proposed method can handle nonstationarity between training and test sets.

  • Thank you for the question. This is a misunderstanding - we never testing on popular and unpopular items separately, but instead training on popular and unpopular items separately, which aims to demonstrate our method can perform well even in the violation of Ptrain=PtestP_{train}=P_{test}.

Last but not least, what is the necessity of introducing a new baseline, DR-JL?

  • Thanks for the comment. DR-JL is exactly the 'DR' in our original manuscript - we are sorry for the ambiguity, so we not introduce new baseline. More specifically, DR-JL is the most popular baseline developed upon the doubly robust (DR) estimator, but as discussed before, DR-JL limits to the finite population. Our paper extend the widely used DR-JL to the super-population, named Super-DR. That's why we compared DR-JL here to make a fair comparison.

For W2, my concern is that the errors made by the propensity model may degrade the final accuracy. However, the authors compare two types of methods, both of which involve a propensity model, and the results show that these two types of methods achieve comparable accuracy. I don't understand how such results can address my concern.

  • Thank you for the great question! During our rebuttal, we intend to say - under varying propensity training paradigms, our method stably outperform DR-JL, showing that our performance gain is insensitive to the specific propensity training method. To fully address your concern, we add new experiments with clearly printed propensity learning accuracy (marked as AUCp_p, Precisionp_p, Recallp_p, F1p_p) and final accuracy of the prediction model (marked as AUC, NDCG@5, Recall@5), in which Ours (1-layer), Ours (2-layer), Ours (3-layer) indicates the propensity model in our method is trained via 1-layer, 2-layer, and 3-layer NN, respectively.

Coat

Num of layersAUCp_pPrecisionp_pRecallp_pF1p_pAUCNDCG@5Recall@5
Ours (1-layer)0.7040.1290.6670.2160.7530.6610.649
Ours (2-layer)0.7110.1720.5940.2670.7570.6670.637
Ours (3-layer)0.6900.1140.7280.1960.7400.6550.624
Best baseline0.7090.1530.5820.2420.7500.6540.630

Yahoo

Num of layersAUCp_pPrecisionp_pRecallp_pF1p_pAUCNDCG@5Recall@5
Layer1_JL0.7580.1330.4020.2000.6740.6540.450
Layer2_JL0.7520.1810.3160.2300.6860.6670.463
Layer3_JL0.7570.2220.2630.2410.6840.6650.464
Best baseline0.7640.1870.3180.2360.6790.6620.459

KuaiRec

Num of layersAUCp_pPrecisionp_pRecallp_pF1p_pAUCNDCG@50Recall@50
Layer1_JL0.6320.0490.8790.0930.8260.6310.667
Layer2_JL0.6470.0610.6710.1120.8280.6400.680
Layer3_JL0.6530.0680.5430.1210.8290.6360.675
Best baseline0.6480.0800.3880.1330.8240.6320.671

Note: the subscript pp means that this metric evaluates the propensity model.

From above, we find our methods stably outperform the best baseline, even with a larger error on the trained propensity model.


We are eager to hear your feedback. We’d deeply appreciate it if you could let us know whether your concerns have been addressed.

评论

Dear Reviewer gNuC,

As the discussion deadline approaches, we are wondering whether our responses have properly addressed your concerns? Your feedback would be extremely helpful to us. If you have further comments or questions, we hope for the opportunity to respond to them.

Many thanks,

19687 Authors

评论

Dear reviewer gNuC,

Since the discussion period will end in a few hours, we will be online waiting for your feedback on our rebuttal, which we believe has fully addressed your concerns.

We would highly appreciate it if you could take into account our response when updating the rating and having discussions with AC and other reviewers.

Thank you so much for your time and efforts. Sorry for our repetitive messages, but we're eager to ensure everything is addressed.

Authors of # 19687

审稿意见
5

This paper introduces SuperDR, a novel doubly robust estimator for debiasing recommendation models under a more realistic super-population scenario, where the joint distribution of features and ratings remains the same across training and test sets, but the specific user-item pairs may not overlap. This setting contrasts with the conventional exact matching assumption commonly made in prior work. The key insight is that, under the super-population setting, existing DR estimators remain biased due to an uncontrolled covariance term, even when the imputation and propensity models are correctly specified. To address this, the authors propose a correction mechanism through a learnable parameter ϵ\epsilon, which is added to the imputation model. This adjustment aims to reduce the empirical covariance Cov^(p^u,iou,ip^u,i,eu,ie^u,i)\hat{\text{Cov}}\Big({\hat{p}_{u, i} - o_{u, i} \over \hat{p}_{u, i}}, e_{u, i} - \hat{e}_{u, i}\Big). The paper presents supporting theoretical results—including a generalization error bound—and demonstrates, through experiments on three datasets (Coat, Yahoo! R3, and KuaiRec), that SuperDR outperforms existing debiasing methods.

优缺点分析

  • Strengths

    • The shift from exact matching to super-population is realistic and meaningful, especially for large-scale industrial recommendation systems, where user-item overlap between training and test sets is uncommon.
    • The paper establishes a strong theoretical foundation, with clearly articulated lemmas, corollaries, and proofs. Notably, the identification and treatment of the covariance term addresses an important and underexplored aspect of doubly robust (DR) estimation.
    • The correction to the imputation model using ϵ(ou,ip^u,i)\epsilon (o_{u, i} - \hat{p}_{u, i}) is elegant and well-motivated by targeted maximum likelihood estimation (TMLE).
    • The derivation of the generalization error bound incorporating the new covariance term and empirical Rademacher complexity is both novel and informative.
    • Experiments on multiple datasets consistently demonstrate performance improvements over a broad set of DR- and IPS-based baselines.
    • The sensitivity study on the learning rate for ϵ\epsilon and sample ratio shows the method’s robustness under different bias severities and hyperparameter configurations.
  • Weaknesses

    Nothing

问题

Nothing

局限性

The authors properly address the limitations in Section 6.

最终评判理由

The paper is well-organized, and the proposed method shows strong performance compared to the baselines. The authors also responded sincerely and appropriately during the discussion period to the weaknesses and related questions raised by other reviewers. In particular, I believe the authors provided a sufficiently convincing response to the concern raised by another reviewer regarding the Assumption of Equal Joint Distributions. However, the concern about Dependence on Propensity Pre-training still appears to remain partially unaddressed. That said, the motivation behind the problem, the proposed solution, and the logical progression presented by the authors offer a meaningful contribution to the field.

格式问题

There is no major formatting issue.

作者回复

Response: Thank you for your time and dedication in reviewing our manuscript! We are really glad that you are enjoying our paper and give our paper very positive comments! We believe that our study makes an important contribution to the field of debiasing in RecSys and causal inference, and we hope that you will support our research in subsequent discussions.

审稿意见
5

The paper proposes a novel doubly robust estimator with a joint learning algorithm to address the limitation of selection bias in designing the recommender systems.

优缺点分析

The paper is well-organized and easy to follow. It presents strong results on large-scale datasets, which highlight the effectiveness and practical relevance of the proposed method. Additionally, the inclusion of a generalization bound provides valuable theoretical support, offering insights into the method's expected performance beyond the training data. However, further clarification or empirical validation of the bound would strengthen the contribution even more.

问题

Can this method be applied in a real-time setting where data arrives sequentially? In other words, is the proposed approach capable of efficiently adapting or updating its predictions as new data becomes available, without requiring full retraining?

局限性

See above.

格式问题

None

作者回复

We thank the reviewer for efforts in evaluating our work and helpful suggestions. Below, we will try our best to address your concerns and question.

W1: Further clarification or empirical validation of the bound would strengthen the contribution even more.

Response to W1: We thank the reviewer for the useful suggestions.

  • Further clarification of error bounds: Given the propensity model p^u,i\hat p_{u, i} and imputation model e^u,i\hat e_{u, i}, satisfying p^u,iKψ\hat p_{u, i}\geq K_{\psi} and eu,ie^u,iKϕ|e_{u, i}-\hat e_{u, i}|\leq K_{\phi}, then with probability at least 1η1-\eta, we have

Lideal(θ)1D(u,i)D[e^u,i+ou,i(eu,ie^u,i)p^u,i]empirical DR lossL_{\text{ideal}}(\theta) \leq \underbrace{{\frac{1}{|\mathcal{D}|} \sum_{(u, i) \in \mathcal{D}}\big[\hat e_{u, i}+\frac{o_{u, i}(e_{u, i}-\hat e_{u, i})}{\hat{p}_{u, i}}\big]}} _{\text{empirical DR loss}}

+ 1D(u,i)D1E[ou,ip^u,ixu,i]E[eu,ixu,i]E[e^u,ixu,i]empirical risks of imputation and propensity models + ~\underbrace{\frac{1}{|\mathcal{D}|} \sum_{(u, i) \in \mathcal{D}}\left|1-\mathbb{E}\left[\frac{o_{u, i}}{\hat p_{u, i}} | x_{u, i}\right]\right| \cdot\left|\mathbb{E}\left[e_{u, i} \mid x_{u, i}\right]-\mathbb{E}\left[\hat e_{u, i} \mid x_{u, i}\right]\right|} _{\text{empirical risks of imputation and propensity models}}

+ 1D(u,i)DCov(ou,ip^u,ip^u,i,eu,ie^u,i)the empirical covariance + ~\underbrace{\left|\frac{1}{|\mathcal{D}|} \sum_{(u, i) \in \mathcal{D}} \operatorname{Cov}\left(\frac{o_{u, i}-\hat p_{u, i}}{\hat p_{u, i}}, e_{u, i}-\hat e_{u, i}\right)\right|} _{\text{the empirical covariance}}

+(1+1Kψ)(2R(F)+Kϕ18log(4/η)D)tail bound controlled by empirical Rademacher complexity and sample size, + \underbrace{\left(1+\frac{1}{K_\psi}\right)\left(2 \mathcal{R}(\mathcal{F})+K_\phi \sqrt{\frac{18 \log (4 / \eta)}{|\mathcal{D}|}}\right)} _{\text{tail bound controlled by empirical Rademacher complexity and sample size}},

which includes four terms: the empirical DR loss, the empirical risks of imputation and propensity models, the empirical covariance, and the tail bound. Even if the p^u,i=pu,i\hat p_{u,i} = p_{u,i} or e^u,i=eu,i\hat e_{u,i} = e_{u,i}, the empirical covariance does not vanish in the super-population scenario. The previous DR-based method can only control the first two terms, while our method can further control the empirical covariance by subtly modifying the imputation model (introducing one learnable parameter), without adding new modules and increasing time complexity.

  • Further empirical validation of error bounds:
    • In our manuscript, we have demonstrated that our method can control empirical covariance on the observed samples on three widely used real-world datasets.
    • Furthermore, we validate empirical covariance control on the whole dataset by conducting a semi-synthetic experiment on the Movielens 100k dataset. Following previous studies [1-4], we use matrix factorization to generate the full binary rating matrix as the ground truth, the propensities for generating MNAR observed data, as well as the embedding for each user and item. Then we randomly select 50% / 70% users and items into our training data and train DR-JL, DCE-DR, and our method on the training data, finally calculating the empirical covariance (EC) on the whole dataset.

The results are shown below:

Ratio = 50%Ratio = 70%
MethodAUCEC on Whole DataAUCEC on Whole Data
DR-JL0.8855-3.34290.9104-3.0932
DCE-DR0.8941-1.95070.9169-2.0315
Ours0.9017-0.94280.9214-1.5318

The results show that our method can control EC the most, which verifies the validity of the Controllability of Empirical Covariance (Thm 4.7), as well as we may a tighter error bounds.

Q1: Can this method be applied in a real-time setting where data arrives sequentially? In other words, is the proposed approach capable of efficiently adapting or updating its predictions as new data becomes available, without requiring full retraining?

Response to Q1: We thank the reviewer for the insightful question. This is a very common scenario in practice. For example, in the CTCVR scenario, the new MNAR data comes daily. To evaluate whether our method can be efficiently applied in this scenario, we propose two ways to efficiently update our predictions:

  • We store the parameters in models and optimizers at the end of training, and fine-tune them when new data comes (named "continue training").
  • We freeze the parameters in prediction, imputation, and propensity model, and add a learnable shallow residual structure, which is zero-initialized when new data comes (named "adding residual").
  • Also, we consider the full retraining of all parameters on both old and new data as the oracle baselines.

We add experiments on the industrial dataset KuaiRec. To process the data, we first randomly split 80% users and items to construct the training set to make sure we do not see all users and items during training. Then we consider the following two scenarios:

  • The new data has the same distribution compared to the training data (Consistent in distribution). We randomly split the user set and item set into three parts with a ratio of 50/25/25. First, we train our method on the data with 50% users and 50% items, and then retrain the model on the remaining data sequentially, i.e., twice in total.

  • The new data has a different distribution compared to the training data (Inconsistent in distribution). We randomly split the user set and item set into two parts with a ratio of 50/50, and train our method on the first dataset. Then we assign a propensity according to the item frequency and the user's active level to sample the first half of users and items into the first time retraining, and the remaining into the second time retraining. We control the sampling process to ensure the sample number is the same in the two retrainings.

For all methods, we test on the MAR dataset. The results on the KuaiRec dataset are shown below. We bold the better results between ours and DR-JL.

DataConsistent in distributionInconsistent in distribution
AUCNDCG@50Recall@50AUCNDCG@50Recall@50
Train Set1 (50%)Ours0.8250.6360.6780.8250.6360.678
DR-JL0.8170.6230.6630.8170.6230.663
Train Set2 (25%)Continue training on ours0.8190.6280.6640.8210.6330.669
Continue training on DR-JL0.8080.6170.6470.8110.6310.666
Adding residual on ours0.8110.6140.6480.8210.6400.680
Adding residual on DR-JL0.8050.5990.6390.8160.6200.653
Full training on ours0.8260.6380.6790.8270.6400.679
Full training on DR-JL0.8190.6250.6650.8210.6260.665
Train Set3 (25%)Continue training on ours0.8190.6290.6600.7290.4920.504
Continue training on DR-JL0.8170.6170.6540.7120.4570.449
Adding residual on ours0.8240.6270.6580.7250.4790.506
Adding residual on DR-JL0.8230.6140.6530.7150.4500.449
Full training on ours0.8280.6400.6790.8280.6400.679
Full training on DR-JL0.8220.6270.6680.8220.6270.668

We find that our methods can efficiently adapt our predictions as new data becomes available. An interesting finding is that when there is a large gap between the new data and the original training data (such as cold users and items, as shown in the "Train Set3" in the "Inconsistent in distribution" scenario), though the performance of our method will drop, our method significantly outperforms DR-JL. One possible reason is that our method deals with a super-population scenario, which is more robust toward a huge distribution shift.

We sincerely thank you for your feedback, and welcome any further technical advice or questions on this work. We will do our best to address your concerns.


References

[1] Recommendations as Treatments: Debiasing Learning and Evaluation. ICML 2016.

[2] Doubly Robust Joint Learning for Recommendation on Data Missing Not at Random. ICML 2019.

[3] Enhanced Doubly Robust Learning for Debiasing Post-Click Conversion Rate Estimation. SIGIR 2021.

[4] Be Aware of the Neighborhood Effect: Modeling Selection Bias under Interference for Recommendation. ICLR 2024.

最终决定

This paper considers the problem of selection bias. They go beyond the setting of exact matching and instead assume only that the joint distribution of the feature and the rating is the same in the training and test set. They show that in this broader setting, the previous DR estimator is biased due to an additional covariance term. They propose SuperDR which adds a learnable “correction” parameter. They provide sound and nontrivial theoretical guarantees and demonstrate improved performance on a variety of datasets. This paper is novel, well-written, and a nice contribution.