Conformal Prediction with Cellwise Outliers: A Detect-then-Impute Approach
This paper proposes a general detect-then-impute conformal prediction framework to address the cellwise outliers in test data.
摘要
评审与讨论
Conformal Prediction (CP) provides prediction intervals with guaranteed coverage for black-box models under exchangeability assumptions. However, cellwise outliers isolated contaminated entries in test features break this exchangeability, leading to unreliable PI. This paper addresses this challenge by introducing a detect-then-Impute conformal prediction framework to robustly handle cellwise outliers.
This paper proposed two novel algorithms of conformal prediction, PDI-CP and JDI-CP, and provided a distribution-free coverage analysis under detection and imputation procedures. This paper establishes coverage error bounds for PDI-CP and JDI-CP achieves a finite sample 1 − 2α coverage guarantee. In section 6 and section 7, author experiments on synthetic and real datasets to demonstrate that proposed algorithms are robust and efficient.
=============
I thank the reviewer for the responses which clarified my concerns. I have read them and keep my score unchanged.
给作者的问题
NA
论据与证据
yes
方法与评估标准
yes
理论论述
Theorems are correct under assumptions, but the assumptions are restrictive. Assumption 3.1 is critical for ODI-CP but unrealistic in practice. Assumption 3.2 may not hold for correlated features.
实验设计与分析
The experiment is reasonable overall, but a new experiment with setting linear, homoscedastic, Light-tailed could be added in simulation.
补充材料
There is no supplementary material so far.
与现有文献的关系
The key contributions of the paper are positioned at the intersection of cellwise outlier detection, conformal prediction, and missing data imputation, building on and extending prior work in these areas.
遗漏的重要参考文献
NA
其他优缺点
Weakness
- Computational Cost: JDI-CP requires O(n) pairwise operations, which could be inefficient for large datasets.
- Logic: The logic of the article introduction is confusing and should highlight the research issues of the article. Author should highlight the purpose of this paper is using DI method to solve cellwise outlier issues instead of giving the notation in introduction.
- Innovation: This paper seems simply combining DI and CP, I don’t see any theoretical innovation. Author cites many theories which are proved in prior work like theorem 3.5 and focus too much on detection outlier instead of CP.
- Application: Assumption 3.1 assume all the cellwise outliers are detected, but it sounds impossible in real world.
其他意见或建议
NA
Q1: Author should highlight the purpose of this paper instead of giving the notation in introduction.
We sincerely appreciate this constructive critique of our manuscript's organization. In response to your suggestion, we will comprehensively restructure the introduction in the future revision to:
-
Problem Focus
- Lead with the critical challenge of cellwise contamination in conformal prediction
- Highlight gaps in existing methods' ability to handle this data corruption
-
Solution Framework
- Clearly articulate our methodological innovations
- Emphasize how we address previously unsolved problems
-
Presentation Strategy
- Defer technical notation to methodology sections
- Maintain narrative flow while preserving rigor
These adjustments will prioritize readability and better highlight the scientific significance of our approach.
Q2: Assumption 3.1 is unrealistic in practice.
We appreciate your inquiry regarding Assumption 3.1 and we acknowledge its imperfection.
- We'd like to stress that Assumption 3.1 is theoretically necessary for model-free and distribution-free coverage guarantees and can be satisfied by choosing a small detection threshold in practice.
- Additionally, we also present a total variation coverage gap of our method when Assumption 3.1 is violated. If the total variation between processed calibration data and test data is small, the coverage can still be approximately controlled. Please refer to Reviewer vdr2's Q1 for details and an additional experiment about Assumption 3.1 (due to the space limit).
We would be happy to discuss this further if additional clarification would be helpful.
Q3: Assumption 3.2 may not hold for correlated features.
Thanks for your question and we’d like to make the following explanations. For correlated features:
- Assumption 3.2 can be satisfied if we use one-class SVM method to learn the score for each coordinate.
- In simulations, we also used detection methods do not satisfy Assumption 3.2, such as DDC in Figure 3 and cellMCD in Figure 4(b).
- In Section 5, we provided theoretical results for our method when Assumption 3.2 does not hold.
Q4: A new experiment with setting linear, homoscedastic, light-tailed could be added.
Thank you for your advice. In fact, we have included the experimental result of the setting you suggested in Appendix D.2 but didn't mention it in the main text due to the space limitation. We apologize for any possible reading incompleteness and will add this setting to the main text when we have an additional page in the revision.
Q5: JDI-CP requires O(n) pairwise operations, which could be inefficient for large datasets.
We appreciate this insightful observation about our method's computational aspects.
- JDI-CP Design Considerations:
- Primary objective: Finite-sample coverage guarantee
- Trade-off: Achieves robustness (demonstrated in Figure 7) at some computational cost
- Reference: Similar trade-offs exist in Barber et al. (2021)
- Future Directions:
- Actively researching more efficient implementations
- Will optimize computational performance while maintaining theoretical guarantees
[1]Barber R F, Candes E J, Ramdas A, et al. Predictive inference with the jackknife+. The Annals of Statistics, 2021.
Q6: This paper seems simply combining DI and CP. Author cites many theories which are proved in prior work like theorem 3.5. Focus too much on detection instead of CP.
Thanks for your insightful comments on the contribution of our approach. We’d like to provide some clarifications to highlight the contribution of our work.
- Firstly, our approach fundamentally differs from naive DI+CP combinations. As discussed and shown in Appendix A.1, such a direct combination cannot achieve coverage control. Therefore, we proposed new techniques to adaptively deploy DI on calibration data and proposed PDI-CP and JDI-CP.
- Regarding Theorem 3.5, this is a negative result we obtained to illustrate the necessity of Assumption 3.1, and we have not found similar results in other papers. Additionally, we have provided a new counterexample under the CQR score, please refer to Reviewer vdr2's Q2.
- Finally, we believe that detection is essential to cope with cellwise outliers in predictive inference tasks. The classic CP methods under exchangeability assumption have been extensively studied. Here, we are concerned with the nonexchangeable problem caused by cellwise outliers, which is challenging and has not been studied before. The detection and imputation steps are used to identify and remove those outliers, so we can construct informative conformal prediction intervals. The detection is key to constructing exchangeable processed features in ODI-CP and JDI-CP, which rebuilds exchangeability between calibration data and test data and enables coverage control.
Hope these explanations are acceptable for you!
This paper addresses conformal prediction with feature-wise outliers in the test sample. It assumes access to a detection oracle satisfying the sure detection and isolated detection assumption, and impute the values of the outlier features. After the detection and imputation procedure, split conformal prediction and Jackknife+ are applied, respectively, resulting in two kinds of prediction sets. A distribution-free finite sample coverage guarantee is proved for the latter. Beyond isolated detection, the coverage guarantee worsens with the difference of two detection set, which is empirically shown small. Experiments on synthetic and real datasets with synthetic perturbations show valid coverage and controllable sizes of the prediction set.
给作者的问题
Are there contamination models other than Equation 1 that are worth consideration?
论据与证据
The claim lacks support in L48 which states weighted conformal prediction is unsuitable for CP with outliers because distribution shift cannot be estimated. For example, localized conformal prediction reweigh the calibration samples according to their distance to the test sample, without estimating the distribution shift. The WCP is clearly applicable, since the experiments also report Tibshirani et al.'s WCP method. There could be more discussion why weighted conformal prediction is less competitive in this setting.
方法与评估标准
-
The major concern is over the strength of assumption 3.1 and 3.2. They assume oracle access to an outlier detector that has zero false negative rate. And assumption 3.2, which is necessary for finite length prediction sets, additionally excludes outliers in the joint space of features. Figure 2 has shown that the false discovery rate is around 0.4, which indicated that the assumptions are not approximately satisfied.
-
The form of non-conformity score is restricted to absolute deviation.
理论论述
For theorem 4.3, the coverage lower bound can be vacuous for large calibration size since max E_i is non decreasing.
I have not checked the proofs.
实验设计与分析
-
The experiments show that the proposed method is insensitive to different detection and imputation methods, which indicates the flexibility of the design.
-
Results with more recent weighted conformal prediction methods will consolidate the claim, such as localized conformal prediction.
-
Datasets with real outliers will be more convincing. The current results on real datasets introduce artificial perturbations.
补充材料
I have reviewed section D.1.
与现有文献的关系
The most significant contribution of this paper is the first to consider conformal prediction with cellwise outliers, and proposes an algorithm with valid distribution-free coverage guarantee for finite samples (Theorem 4.4).
遗漏的重要参考文献
Related works on weighted conformal prediction are listed but not fully addressed on why they are not applicable with outliers. Additionally, localized conformal prediction as a special form of WCP is not discussed. They are free of estimation of distribution shift which is claimed in the paper as a reason for incompetence.
[1] Leying Guan. Localized conformal prediction: A generalized inference framework for conformal prediction. Biometrika, 110(1):33–50, 2023.
[2] Rohan Hore and Rina Foygel Barber. Conformal prediction with local weights: randomization enables local guarantees. arXiv preprint arXiv:2310.07850, 2023.
其他优缺点
NA
其他意见或建议
NA
Q1:Concerns focus on Assumptions 3.1 and 3.2, assuming detection with FNR=0. Assumption 3.2, crucial for finite length prediction interval, also excludes joint feature space outliers. Figure 2 shows FDR is around 0.4 , indicating unmet assumptions.
Thanks for your valuable questions! We acknowledge the imperfection of Assumption 3.1 and explain as follows.
-
Assumption 3.1(FNR=0), also used in Liu et al.(2022) and Wasserman&Roeder(2009), can be met by choosing a small detection threshold in practice and is theoretically essential for model-free coverage guarantee. When Assumption 3.1 doesn't hold, we derived a new bound showing the coverage gap depends on the total variation between processed calibration and test data (see Reviewer vdr2’s Q1 due to space limit). We also added experiments by varying DDC detection thresholds (adjusting ) based on Setting C in Appendix D.2. When FNR0, our method can still approximately control the coverage. ||0.5|0.7|0.9|0.99| |-|-|-|-|-| |FNR |0|0.005|0.008|0.013| |FDR|0.793|0.669|0.340|0.035| |PDI coverage|0.909|0.907|0.901|0.902| |JDI coverage|0.904|0.904|0.899|0.895|
-
We'd like to clarify the primary purpose of Assumption 3.2 is to ensure exchangeability of (see Lemma 3.3), not for finite length prediction intervals (PIs). In Section 5, we also provide theoretical coverage results when Assumption 3.2 doesn’t hold. In simulations, we also used detection methods that unsatisfy Assumption 3.2, such as DDC in Figure 3 and cellMCD in Figure 4(b), which don't lead to infinitely wide PIs.
-
Actually, Assumption 3.1 requires FNR equal to zero, while the frequency of in Figure 2 means the false discovery (positive) rate (FDR) is zero.
Q2:Why weighted conformal prediction is less competitive?
WCP requires precise likelihood ratio estimation, but arbitrary cellwise outliers in our setting render this impossible. Figures 3 and 11 highlight this limitation, showing WCP fails to meet target coverage.
Q3:Localized conformal prediction as a special form of WCP is not discussed.
Thanks for your advice and we apologize for overlooking these references.
- LCP controls conditional coverage under i.i.d. test and calibration data, differing from our focus on distribution shift caused by cellwise outliers in test data, which LCP cannot cope with. We’ll add a detailed discussion of LCP and these references in the revision.
- We also compare our method with baseLCP, calLCP (Guan,2023), and RLCP (Hore&Barber, 2023) through experiments. Clean data follows Section 5.1.1 of Hore&Barber(2023), and cellwise outliers are generated via Equation (10) in our paper with . Results show LCP methods fail to provide informative PIs. ||Bandwidth|0.1|0.2|0.4|0.8|1.6| |-|-|-|-|-|-|-| |baseLCP|Coverage|0.90|0.90|0.90|0.90|0.95| ||Infinite PI%|0.90|0.90|0.90|0.90|0.81| |calLCP|Coverage|0.94|0.94|0.94|0.94|0.93| ||Infinite PI%|0.88|0.88|0.88|0.87|0.66| |RLCP|Coverage|0.90|0.90|0.90|0.90|0.90| ||Infinite PI%|0.90|0.90|0.90|0.86|0.48| |PDI|Coverage|0.90||||| || Infinite PI%|0||||| |JDI|Coverage|0.89||||| ||Infinite PI%|0|||||
Q4:The form of non-conformity score is restricted to absolute deviation.
For simplicity, we use the absolute residual score, but our method supports various non-conformity scores like CQR score. Please refer to Reviewer vdr2's Q2 for a discussion about CQR.
Q5:Theorem 4.3's coverage lower bound may be vacuous with large calibration size.
After checking our proof, we found the expansion to is unnecessary and the last inequality in (B.10) (Appendix B.3) should be removed. The modified coverage gap is given by:\mathbb{P}\left\\{\hat{q}\_{\alpha}^+(\\{R_i^*-S_{\hat{\mu}}\cdot E_i\cdot|\tilde{\cal{O}}\_{n+1}\setminus\mathcal{O}^*|\\}\_{i=n_0}^n)<R_{n+1}^*\leq\hat{q}\_{\alpha}^+(\\{R_i^*+S_{\hat{\mu}}\cdot E_i\cdot|\tilde{\cal{O}}\_{n+1}\setminus\cal{O}^*|\\}\_{i=n_0}^n)\right\\},which will not be vacuous for large calibration size.
Thank you for catching this key technical nuance!
Q6:Dataset with real outlier will be more convincing.
To further demonstrate robustness, we test our method on a riboflavin gene expression dataset(Liu et al.,2022) with confirmed cellwise outliers. Our method maintain coverage above the target while LCP methods () fail to provide meaningful PIs.
| SCP | WCP | baseLCP | calLCP | RLCP | PDI | JDI | |
|---|---|---|---|---|---|---|---|
| Coverage | 0.83 | 0.85 | 0.90 | 0.96 | 0.90 | 0.93 | 0.95 |
| Length | 1.82 | Inf | Inf | Inf | Inf | 3.08 | 3.29 |
Q7:Are other contamination models worth considering?
The Tukey-Huber Contamination Model generates casewise outliers, assuming most samples follow a target distribution , while the others follow an arbitrary distribution :where is the contamination ratio.
Thank the author for their response. Most of my concerns and questions are addressed. I have raised the rating from 3 to 4.
Many thanks for the review and raising your rating! If you have any other questions, concerns, and comments, please let us know. We would like to provide our responses and address them in the future revision. Thank You!
When some entries of the test features are contaminated, the paper introduces a detect-then-impute conformal prediction framework. This framework first applies an outlier detection procedure to identify contaminated entries in the test features and then uses an imputation method to fill in the identified outliers. Moreover, the authors apply the detection and imputation procedures to the calibration set, ensuring the construction of exchangeable features for the conformal prediction interval of the test label. Two practical algorithms including PDI-CP and JDI-CP are proposed, with lower bounds on marginal coverage probability established under certain conditions. Numerical experiments on both synthetic and real datasets are provided to demonstrate the performance of the proposed algorithms.
给作者的问题
-
The validity of Assumption 3.1 appears to depend heavily on the quality of the detection rule, which may be restrictive in practice. In Theorem 3.5, the negative result is derived using a specific form of the prediction set. Would the conclusion change if other adaptive prediction sets, such as those based on conformalized quantile regression (CQR), were used instead?
-
How does the contamination rate in the test features affect the length of the prediction sets?
论据与证据
Yes.
方法与评估标准
The paper evaluates the proposed method on synthetic data and two real datasets and compares its performance against several relevant baseline methods.
理论论述
I checked the theoretical proofs of Lemma 3.3 and Proposition 3.4 and believe they are correct.
实验设计与分析
I have reviewed all experimental parts in the paper.
补充材料
I reviewed the parts related to numerical studies in the supplementary material.
与现有文献的关系
The paper proposes a framework to construct prediction sets with marginal coverage guarantees when the test input feature is contaminated.
遗漏的重要参考文献
No critical references appear missing.
其他优缺点
No.
其他意见或建议
No.
Q1: The validity of Assumption 3.1 (sure detection) depends on quality of detection, which may be restrictive in practice.
Thanks for your valuable question! We acknowledge the imperfection of Assumption 3.1 and make the following explanations.
- Sure detection/screening conditions are commonly used in existing works (Wasserman&Roeder,2009; Liu et al.,2022) to ensure all relevant variables are retained. We adopt a similar condition in Assumption 3.1 for model-free coverage guarantee.
- Assumption 3.1 is essential for meaningful prediction intervals (PIs): if outliers persist in , Theorem 3.5 shows PIs can become infinitely wide in expectation.
- Violating Assumption 3.1 impacts our method by introducing a total variation coverage gap: if outliers persist in the test point after detection, Lemma 3.3 fails due to the exchangeability between and is broken, where are the ODI features. At this point, there is a total variation coverage gap of the PI:\mathbb{P}\\{Y_{n+1}\in\hat{C}^{\rm{ODI}}(\tilde{X}\_{n+1})\\}=\mathbb{P}\left\\{|\hat{\mu}(\check{X}\_{n+1}^{\rm{DI}})-Y_{n+1}|\le\hat{q}\_{\alpha}^+(\\{R_i^*\\}\_{i=n_0}^n)\right\\}\ge 1-\alpha-\frac{1}{n-n_0+1}\sum_{i=n_0}^n d_{\rm{TV}}(\check{X}\_{n+1}^{\rm{DI}},\check{X}\_i^*).The impact is mild if is small; similar bounds apply to PDI-CP and JDI-CP.
- In practice, Assumption 3.1 (TPR=1) can be satisfied when the detection threshold is small; our method still maintains target coverage in simulations and real examples even if some outliers were not completely detected. For your convenience, we also added a new experiment under different DDC detection thresholds (varying ) based on Setting C in Appendix D.2, which shows Assumption 3.1 is exactly satisfied (TPR=1) at . When TPR<1, our method can still achieve approximate control of coverage. ||0.5|0.7|0.9|0.99| |-|-|-|-|-| |TPR|1|0.995|0.992|0.987| |FDR|0.793|0.669|0.340|0.035| |PDI coverage|0.909|0.907|0.901|0.902| |JDI coverage|0.904|0.904|0.899|0.895|
[1] Wasserman, L. and Roeder, K. High dimensional variable selection. The Annals of Statistics, 2009.
[2] Liu, Y., Ren, H., Guo, X., Zhou, Q., and Zou, C. Cellwise outlier detection with false discovery rate control. Canadian Journal of Statistics, 2022.
Q2: Would the conclusion in Theorem 3.5 hold with other adaptive prediction sets like conformalized quantile regression?
Thank you for your question! We prove a similar result for the CQR score, and add new simulation results based on CQR.
- The PI constructed from CQR is , where and are the lower and upper quantile regression models, and is the quantile of empirical distribution of CQR computed on the calibration set.
- Following the proof of Theorem 3.5 in Appendix B.2, where for . Suppose where , and the test point wherefor some large positive value . If still contains and covers the true label, we havewhich means .
- We also conduct an experiment using CQR to construct PIs, where the Baseline used in our simulation can be considered as the optimal method for constructing split conformal PI for cellwise outlier, which masks calibration and test features by . This experiment will be added in future revisions. ||Baseline|ODI|PDI|JDI| |-|-|-|-|-| |Coverage| 0.905|0.902|0.900|0.885| |Length|4.366|4.291|4.282|5.544|
Q3: How does the contamination rate affect the length of the prediction sets?
- Figure 6 in Section 6.3 shows the coverage and length across cell contamination probabilities with DDC and Mean Imputation, while Figures 12-13 in Appendix D.4 display results for kNN and MICE imputation.
- Results indicate the length of our method remains stable across varying contamination rates, whereas that of Baseline increases with higher contamination. Figure 3 demonstrates our method's competitive length and target coverage, surpassing the classical WCP method.
Hope these explanations can ease your doubts!
This paper proposes a DI-CP framework to handle cellwise outliers in conformal prediction. The key idea is first to detect outliers in the test feature vector and then impute them before applying conformal prediction. To maintain exchangeability, a similar detection-imputation process is used to calibration samples. The authors propose two methods: 1, PDI-CP (Proxy Detection-Imputation CP), which applies detection and imputation separately. 2, JDI-CP (Joint Detection-Imputation CP), which modifies detection rules to ensure theoretical coverage guarantees. The paper provides theoretical guarantees, including a finite-sample coverage bound for JDI-CP and empirical results showing that DI-CP performs robustly under contamination.
给作者的问题
-
The method assumes perfect outlier detection. How does coverage degrade when detection is imperfect?
-
If choosing a larger detection threshold ensures Assumption 3.1, doesn’t this break exchangeability by misclassifying inliers as outliers? How does this affect empirical coverage?
-
Theorem 4.3 and Theorem 4.4 assume that imputation does not change the residual distribution. How do results hold when imputations introduce bias?
-
How does the method perform when detection is imperfect?
-
Have you tested other imputation strategies?
-
Can you provide a sensitivity study showing how conformal coverage varies under different imputation and detection methods?
论据与证据
This paper presents a novel and relevant problem in conformal prediction, but the theoretical guarantees depend on overly strong assumptions about detection accuracy. The method may fail if detection is imperfect or if too many inliers are misclassified as outliers.
方法与评估标准
There exist two problems:
-
Strong assumptions on detection accuracy (Assumption 3.1 is unrealistic). The method assumes perfect outlier detection. This is impractical because most real-world detection methods have false negatives. If detection is imperfect, theoretical guarantees do not hold, making the results less applicable to real-world data.
-
Excessive false positives break the method. The authors claim that choosing a larger detection threshold guarantees Assumption 3.1. However, this results in too many false positives. Excessive imputation may change the calibration distribution, violating exchangeability assumptions (Lemma 3.3 fails). There is no analysis of how misclassified inliers affect conformal validity.
-
Appendix A.1 considers Direct-ODI and Direct-PDI, which do not modify calibration samples. This setting contradicts the main method’s justification that calibration samples must be processed to maintain exchangeability.
理论论述
- Theoretical results assume that imputation does not introduce significant bias. However, imputation methods systematically shift feature distributions, leading to biased conformity scores.
实验设计与分析
Empirical evaluation lacks robustness analysis. The paper does not study sensitivity to different detection thresholds and imputation choices.
补充材料
I did not review the supplementary code as part of my evaluation. My review is based on the theoretical justifications, experimental results, and clarity of the main paper.
与现有文献的关系
This paper reviews the literature of cellwise outliers, conformal prediction without exchangeability, predictive inference with missing data, and conformal inference for outlier detection. It also lists the key difference of settings or tasks for works in the related area.
遗漏的重要参考文献
I think authors have achieved a fairly complete literature review.
其他优缺点
Strength:
Novel problem setting: The problem of cellwise outliers in conformal prediction is important and underexplored. The paper provides a structured approach to address it.
Exchangeability considerations: The idea of applying detection and imputation jointly to calibration and test samples is a novel extension that ensures that conformal prediction remains valid despite outliers.
其他意见或建议
There are a lot of math notations proposed in this paper. Could the author provide the math notation table in the appendix?
Q1: Assumption 3.1 is impractical. How does coverage degrade when detection is imperfect?
Thank you for your insightful question!
- We acknowledge the imperfection of Assumption 3.1 and have obtained a new coverage gap bound in total variation: if there are still outliers in test point after the detection procedure, Lemma 3.3 will not hold because the exchangeability between and is broken, where are the ODI features. At this point, there is a total variation coverage gap of the prediction interval (PI):\mathbb{P}\\{Y_{n+1}\in\hat{C}^{\rm{ODI}}(\tilde{X}\_{n+1})\\}=\mathbb{P}\left\\{|\hat{\mu}(\check{X}\_{n+1}^{\rm{DI}})-Y_{n+1}|\le\hat{q}\_{\alpha}^+(\\{R_i^*\\}\_{i=n_0}^n)\right\\}\ge 1-\alpha-\frac{1}{n-n_0+1}\sum_{i=n_0}^n d_{\rm{TV}}(\check{X}\_{n+1}^{\rm{DI}},\check{X}\_i^*).Notice that, we can still have a approximate coverage control if is small. A similar bound can be obtained for PDI-CP and JDI-CP and we will add these results in the revision.
- In addition, we add a new experiment to show the influence of imperfect detection. Please see Q6 for details.
Q2: If choosing a larger detection threshold ensures Assumption 3.1, doesn’t this break exchangeability? How does this affect empirical coverage?
- According to the construction, we know that false positives do not break the data exchangeability in ODI-CP and JDI-CP. Since we use to approximate in PDI-CP, it will break the exchangeability between processed test data and calibration data because depends only on test data. As we stated in Theorem 4.3, the coverage gap is affected by the number of false discoveries .
- In Appendix D.3, we summarized the empirical false discovery rate (FDR) and true positive rate (TPR) of the detection methods in our simulation. We will include this discussion in future revision, using the table from Q6 as an illustrative example. As the threshold decreases, the number of discoveries increases, leading to a higher FDR. Notably, the table demonstrates the empirical coverage of our method remains robust to variations in FDR.
Q3: Theorem 4.3 and 4.4 assume imputation doesn't change the residual distribution.
We apologize for any confusion. Here, denotes the distribution function of the ODI-CP residuals , not the residuals from the raw calibration data before the DI procedure. Thus, our method doesn't assume "imputation preserves the residual distribution." We appreciate your attention and will clarify this notation in the revision.
Q4: Have you tested other imputation strategies? How coverage varies under different imputation and detection methods?
- In addition to Mean Imputation, we evaluated two other imputation methods in Section 6.2: kNN and MICE. Figure 5 shows our method maintains robust empirical coverage across all imputation strategies.
- Our analysis in Sections 6.1-6.2 compares coverage and length of PI across various detection and imputation methods. Figures 4-5 show our method consistently maintains robust coverage control across all configurations with stable interval lengths.
Q5: Direct-ODI and Direct-PDI contradict the main method’s justification.
We apologize for any ambiguity and would like to clarify:
- Direct-ODI and Direct-PDI in Appendix A.1 are naive combinations of DI with CP, which are not our proposed methods.
- Figure 8 shows Direct-PDI fails to maintain proper coverage while our PDI-CP successfully achieves target coverage. This empirical evidence confirms simply combining DI with CP (without our proposed modifications) cannot guarantee valid coverage.
We appreciate your careful review and will revise Appendix A.1 for clarity.
Q6: The paper does not study sensitivity to different detection thresholds and imputation choices.
Regarding the sensitivity of our method:
- Imputation methods: Figure 5 in Section 6.2 demonstrates our method maintains robust performance across different imputation choices.
- Detection thresholds: We conduct additional experiments by varying the DDC detection threshold (adjusting ) based on Setting C in Appendix D.2. As shown in the table, our method demonstrates robust coverage performance when threshold changes. ||0.5|0.7|0.9|0.99| |-|-|-|-|-| |FNR |0|0.005|0.008|0.013| |FDR|0.793|0.669|0.340|0.035| |PDI coverage|0.909|0.907|0.901|0.902| |JDI coverage|0.904|0.904|0.899|0.895|
Please let us know if you would like any additional details.
Q7: Could you provide a math notation table?
Thanks for your helpful advice. We will add a math notation table to the appendix of the revision.
Hope these explanations are acceptable for you!
Thank the author for their response. Most of my concerns are addressed. I have raised my score from 2 to 3.
We greatly appreciate your efforts in reviewing our work and raising your score! If there are any additional insights or suggestions you would like to share, we are eager to hear them. Thank you once again for your support!
This paper studies a scenario in conformal prediction when some entries of the test features are contaminated, which leads to a violation of the exchangeability assumption that is critical for the correct coverage probability. The proposed approach is first to apply an outlier detection method to identify contaminated entries in the test features and then impute these outliers before applying conformal prediction.
Two practical algorithms are proposed, with lower bounds on marginal coverage probability established under certain conditions. Numerical experiments on both synthetic and real datasets are provided to demonstrate the performance of the proposed algorithms. A bonus is that the same approach can also be applied to the calibration set. Since the calibration data presumably came from the same source as the training data, it would seem that this approach needs to be extendable to the setting when the training set contains outliers, or realistically, a worse scenario when both the training and calibration data are contaminated. It is important that the authors address this question in the final version.
The problem is highly relevant but one reviewer thinks the theoretical guarantees depend on overly strong assumptions about detection accuracy. The method may fail if detection is imperfect or if too many inliers are misclassified as outliers.