Enhancing Treatment Effect Estimation via Active Learning: A Counterfactual Covering Perspective
Data valuation for selection under limited budget for treatment effect estimation
摘要
评审与讨论
This paper proposes a data-efficient method to construct a regression model for estimating the mean difference between the treatment and control responses. Specifically, the author upper-bounds the evaluation error and proposes two radius-based approaches to minimize the upper bound with a limited label budget.
给作者的问题
Please see the Theoretical Claims.
论据与证据
My primary concern regarding the paper is that the proposed approaches bear a stronger resemblance to coreset learning rather than active learning. In active learning, users interact with the real world, acquire responses (i.e., labels), and utilize these responses to refine the learning process. However, as seen in Algorithm 1 and Eq. (8), the core of the proposed approach focuses on finding a coreset that covers the samples in the treatment and control groups. In other words, there is no interactive learning in which the treatment/control responses are contributory.
方法与评估标准
The authors evaluate the performance of the proposed algorithm across various label budgets and compare them with multiple baseline. This makes sense.
理论论述
I did not check the correctness of proofs. However, I have a couple of concerns regarding the theory.
(1) \mathbf{Theorem 3.3}: What is ? What is the hypothesis class comprised of? Does contain regression functions which are used to model among and ?
(2) Corollary 3.4: Shouldn’t this be a probability-type result? I am thinking the response is noisy and there is chance even is small, could also be big.
实验设计与分析
I have a bare-minimum check on the experimental results, as the methodology part raises significant concerns.
补充材料
I reviewed Algorithm 1 in Appendix to try to figure out the data query scheme.
与现有文献的关系
The paper states that the proposed method can be applied to evaluate the difference between treatment and control responses in a label-efficient manner.
遗漏的重要参考文献
I am not aware of.
其他优缺点
I think that lines 6 and 7 in Algorithm 1, which involve searching for points, are critical. However, the authors have included these parts in the appendix, and it seems they did not provide the intuition behind these queries in the main paper. It would be helpful to include an explanation of what data should be selected to help reduce the maximum radius.
其他意见或建议
In assumption 2.3, should use to indicate a random variable and and use to indicate the realization., e.g., .
Active Learning vs. Coreset Learning: The problem setting we described in lines 114-127 (left column) is an active learning loop that consists of 5 steps, where in Step 4 we identify a subset from the completely unlabeled pool set for the oracle to obtain the labels (i.e., treatment outcome ). The newly labeled samples are then used to refine the estimation model in Step 5. As such, our setting differs from coreset learning that operates with all samples' labels . As our labels are actively acquired, Algorithm 1 corresponds to Step 4 (unlabeled subset identification from ) of our problem setting. In Eq.(8), though it is formulated using the labeled training sets and , the calculation of Eq.(8) only requires covariate and its corresponding treatment assignment (lines 202-203 right column), hence are applicable to all samples in the unlabeled pool set . In both Eq.(8) and Algorithm 1, we slightly abused the notation to make the notations less cluttered, as their calculations do not actually rely on the access to treatment outcome labels . To avoid potential confusion, we will update the description and notation in corresponding parts to highlight the active learning setting.
, and Class : Thanks for pointing out the typo "'', we will correct it to the feature space "''. The family of functions ={} does contain the regression model as part of the function . As per Theorem 3.3, we have ( is a constant for any given ), which contains the loss function to evaluate regression model . The rationale for defining is that the constant in our derived bound is dependent on . As per line 966, the term in Eq.(32c) is further upper-bounded by Integral Probability Metric (IPM) under the premise that is in , which is a useful bounding technique also seen in (Shalit et al., 2017).
Corollary 3.4 and Noisy Response : Thanks for improving the completeness of the Corollary, which should be claimed with the probability of least since it is a further simplification of Theorem 3.3. Its main idea is to reveal what optimizable terms are closely related to risk convergence. For the gap , we follow the common practice in (Sener & Savarese, 2018) by assuming zero training loss and noiseless observation for the theoretical analysis purpose. Under the active learning setting, it is reasonable that the labels obtained from the oracle are sufficiently reliable, i.e., has minimal noise.
How We Query in Algorithm 1: As stated in lines 215-219 (right column), the intuition behind the query process is to greedily prioritize the reduction of the largest radius, which is a commonly used intuition also seen in (Tsang et al., 2005; Sener & Savarese, 2018) for the -Center problem. As per Eq.(8), each of the four radii is defined in lines 202-204 (right column), thus, we just need to query the point (no label) that constitutes the radius , and the query is explicitly defined in line 6 of the full Algorithm 1 (Appendix C) if we need to reduce . The visualizations in Figure 1 are a graphical expression of Algorithm 1's intuition, i.e., the point at the other end of the radius is the next acquisition goal. Notably, Algorithm 1 is a naive instantiation of the covering radius-based active learning and comes with a strong (yet impractical) assumption on fully overlapping data distributions (lines 232-234 right column). Thus, we made its description relatively brief to pave the way for our proposed main algorithm FCCM (Algorithm 2), which relaxes the full coverage constraint to achieve stronger radius reduction under the same labeling budget. We will add this explanation when introducing Algorithm 1 to provide more intuitions.
Random Variable : Thanks for pointing out the typo, we will correct to where needed.
Thank you for your response. I will withhold my judgment on the score for now, primarily because the labels have not been incorporated into the acquisition function in Eq. (8). While selecting informative points based solely on covariates and treatment and response assignment may provide improvement over the random querying, it is far less effective than actively involving the labels in the decision process. In fact, a key advantage of active learning lies in its ability to avoid querying points that are unlikely to provide additional value based on the observed labels. I think important consideration is missing from the proposed approach
We greatly appreciate reviewer CkNS's responsive reply and valuable time during the discussion phase. We address the raised concerns as follows:
Unlabeled data selection with existing labels in the training set: We would firstly address this concern by distinguishing the difference between the Uncertainty-based Querying and the Diversity-based Querying. As in the uncertainty-based querying, e.g., BALD (Gal et al., 2017), BAIT (Ash et al., 2021), Causal-BALD (Jesson et al., 2021b), the label is used for querying further unlabeled samples because the proposed methods are model-dependent, which requires training an uncertainty-aware model based on the existing labeled training set, thus further querying on the unlabeled sample involves the uncertainty output from the trained model. Also, the model-dependent method requires accurate uncertainty estimation, please refer to Q1 for reviewer hAfw for elaborating on the drawback in model-dependent methods. However, the diversity-based querying is model-independent, e.g., Core-Set (Sener & Savarese, 2018), ProbCover (Yehuda et al., 2022), MACAL (Wen et al., 2024), which queries the unlabeled samples from the diverse regions of the data distribution purely based on the covariates (and treatment if for causal inference) and Eq. (8) is a classical form of the core-set querying without the labels. Note, that our task focuses on the regression as stated in line 025 (right column), where labels are distributed in the continuous space, such that each unique sample is assigned with a unique label , so the diversity-based querying on the feature space natural guarantees the diversity in the label space. Why Diversity-based querying in our paper? We would like further to address the method of choice in our paper. It is noted that our paper is theory-based. Our main theorem -- Theorem 3.3 suggests that the incalculable estimation risk PEHE is upper-bounded by four computable radii, which unveils the independence from the label information under our theoretical framework and the proposed theorem deeply motivates our algorithm designed to reduce the covering radii. Thus, to reduce the key terms in the risk upper bound, we first propose a straightforward radius reduction method in Algorithm 1. By considering the limitation of Algorithm 1, we then further propose Algorithm 2 to accelerate the radius reduction specifically tailored for the treatment effect estimation setting. In summary, incorporating the label information for querying depends on the choice of the method, and most importantly, our proposed interesting and intuitive theorem for the treatment effect estimation problem with active learning unveils the independence of the label information in the risk upper bound.
We will incorporate the discussion to update our paper regarding the use of the label information and we kindly welcome any further concerns from reviewer CkNS.
This paper tackles causal effect estimation in active learning, where only features and treatments are known, but outcome labeling is costly and incomplete. To ensure balanced label expansion, it bounds generalization error using factual and counterfactual covering radius. Starting with a greedy approach, it develops the FCCM algorithm, optimizing coverage in real-world data. Experiments on benchmarks show state-of-the-art performance in causal effect estimation.
给作者的问题
1 The Introduction mentions a real-world application for “collecting customer preferences from those who have already received different services”. The applicability of the studied problem to this example is not very clear to me; can you explain with more details? 2 How does FCCM’s performance look like when using a finer-grained increment of the labeling budget? 3 Any insights into sharing the same hyperparameter value among all four radii? In other words, in what scenarios will different radii values help? 4 Is there a reason that results for IBM with 0% budget are omitted? 5 Can the authors provide the PEHE results of FCCM and FCCM- from the ablation study?
论据与证据
The claims are supported by the evidence.
方法与评估标准
The method is developed by extending the covering radius concept to the counterfactual sample setting, which in my view is a neat and intuitive solution to the problem. The evaluation adapts major benchmark datasets in causal effect estimation to the active learning setting, and covers a variety of strong baselines. Thus, the methodology and experiment sections are rigorous.
理论论述
(1) In the proof of Theorem 3.3, it would be useful to explicitly describe the magnitude of , such that the probability of is sufficiently large to be practical.
(2) In the proof of Lemma A.3, the acronym SUTVA should be expanded as it is not introduced elsewhere.
实验设计与分析
The experimental results show a promising level of performance gain from the proposed FCCM algorithm. The settings are rational and the radius hyperparameters are provided, where the algorithm is made available for reproducibility. On top of the quantitative results, I find the visualizations in Figure 6 also helpful for showing FCCM’s advantages.
补充材料
All materials are checked and correct.
与现有文献的关系
This paper resolves the issue of limited labeled data in the causal inference space, which aligns with real-world challenges when building data-driven models under data scarcity due to cost and/or privacy concerns. Hence, the relevance to the broader of causal inference is high.
遗漏的重要参考文献
N/A
其他优缺点
Strengths
1 This paper clearly differentiates the problem it studies with another line of work in data-efficient causal inference, which is active experiment design. The practical scenarios where only the treatment effect (but not treatment) is unknown is well motivated and presents some important challenges unique to the problem.
2 The solution approaches treatment-aware data selection/labeling from an inter- and intra-group covering radius perspective, which is a technically sound and innovative solution.
Weakness
1 In terms of the experimental setting, as each dataset has 50 acquisition steps, it might help demonstrate the performance trend of FCCM if the results at more budget levels can be shown. For example, a figure similar to Figure 5 can be added to Appendix to showcase the prediction performance with a finer-grained increment (e.g., 10% or even 2% percent).
2 The presentation of the ablation study results in Table 2 is somewhat untraditional. It would be more intuitive to directly report the PEHE values of both variants, with the calculated percentage difference as an additional reference.
3 The active learning strategy needs to maximize a synergic objective with four covering radii (two inter-group and two intra-group) combined, where the authors state that all four radii share the same setting. It is worth discussing whether or not using a uniform hyperparameter setting for all four radii is the best possible practice.
4 In Figure 5, the reported results for both TOY and CMNIST datasets start from 0% budget, while IBM starts from 20%. Is it because the error at zero budget is too large for IBM? Since the causal estimator is the same across all active learning methods, the results at 0% are in fact less informative and can be removed for TOY and CMNIST.
其他意见或建议
Typos/presentation issues: 1 Line 380, “a fine granularity results” should be “fine-granularity results” 2 Line 845 in Appendix, “distributions” is mis-spelled 3 Line 957 in Appendix, the meaning of “by change of variable reverting y to x” is unclear
Q1: For instance, when performing A/B tests to compare two software versions, the service provider needs to assign two distinct user groups to different versions (i.e., treatments). To understand which software version offers a better experience for a specific user profile (i.e., features), it is necessary to collect users' feedback via customer surveys. Though the features and treatments associated with each user are known to the service provider, collecting all treatment outcomes -- users' satisfaction levels is time-consuming and economically unrealistic (e.g., paid surveys). In this case, active learning can be used to select a small subset of the most informative users, from whom the treatment outcomes can be obtained at a much lower cost.
Q2+W1: We provide the figures under the 2% increment evaluation scheme, which is the most fine-grained setting based on the 50 query steps in total in the anonymous link https://anonymous.4open.science/r/To-Reviewer-FYYH (click into folder ``To-Reviewer-FYYH''). Note that on the IBM dataset, all methods yield an impractically large prediction error at the initial acquisition steps (<20% budget), thus we show the results from 20% for better clarity.
Q3+W3: Our main rationale is to avoid making the search space prohibitively large -- up to if each radius has candidate settings. The key insight to reduce the search complexity is that, because , each radius is independent and each of the sub-terms increases monotonically with the radius, then so does the mean coverage . Thus, by the independence and monotonicity, the search space greatly reduces to while the search of four radii stays in the same direction, thus we let the four radii share the same setting. When will different radii help? Though the shared value among four radii allows for effective and efficient hyperparameter tuning, it could potentially fail if the distribution discrepancy between the two treatment groups is very large. This is because the counterfactual covering radius can be for the counterfactual coverage to get close to full, lowering the utility of the uniform radii value. As such, if the distribution discrepancy between treatment groups is large, we can then identify a different value for each covering radius to maintain the high coverage under the linear complexity due to the independence and monotonicity. We will add this discussion to an updated version of our paper.
Q4+W4: The reason for omitting 0% in IBM is that the estimation risk at the first few steps is too large to be practical for all methods. Thus, the reported results start from 20% for clarity. We will append a more explicit explanation on the figure caption to eliminate potential confusion.
Q5+W2: Thanks for the suggestion, we will update the table to show the PEHE results for both FCCM and FCCM in our paper. Please check the anonymous link https://anonymous.4open.science/r/To-Reviewer-FYYH (click into folder ``To-Reviewer-FYYH'') for the detailed results.
Probability Threshold ( in Our Paper): We will add a discussion of the magnitude of the probability threshold . It is noted that higher confidence gives a larger upper bound due to the increase in the third term of the bound, but it remains a constant during our optimization.
SUTVA: We will update it with the full spell, namely Stable Unit Treatment Value Assumption (SUTVA).
Typos: Thanks for pointing them out, and we will correct them correspondingly. For ``by change of variable reverting y to x'', we intended to express that the integral over the domain is changed to the domain by the change of variable. We will modify the description to clear the confusion.
This paper proposes an active learning (AL) approach tailored to enhance the treatment effect estimation from observational data, when labeling outcomes is costly. The authors introduce a theoretical formulation using the concepts of factual/counterfactual covering radii, to upper bound the (fundamentally incalculable) population-level risk --- expected precision in estimation of heterogeneous effect (PEHE), defined in Eq. (3) --- effectively by the sum of four factual/counterfactual covering radii, up to a multiplicative constant and lower-order terms, cf. Eq. (4) and Corollary 3.4.
Two distinct greedy algorithms to minimize the covering radii are introduced: (1) A direct radius-reduction algorithm (Algorithm 1 in Section 4.1) motivated by traditional core-set techniques, and (2) the factual and counterfactual coverage maximization (FCCM) method (Algorithm 2 in Section 4.2), which seeks a more balanced and effective radius reduction through maximizing coverage under given radii constraints. Empirical experiments conducted on synthetic, semi-synthetic, and image-based datasets demonstrate that the FCCM approach notably outperforms established baselines by systematically prioritizing samples from regions of high density and significant overlap between treatment groups.
给作者的问题
-
Could you elaborate on the following sentence you wrote in Section 5.1: "For example, although BALD bias the data acquisition toward the overlapping region, it does not further embed the property to query from the high-density region, and thus accounts for less risk" and explain the mechanism how FCCM specifically addresses distributional discrepancies compared to existing methods like BALD, particularly in scenarios with significant non-overlap? I am asking this question because (1) Figure 6-(c) seems to suggest that BALD indeed places some points outside the overlapping region, instead of "biasing the data acquisition toward the overlapping region," and (2) I want to gain intuitive understanding on how FCCM (Algorithm 2) addresses the challenge.
-
How sensitive is FCCM to the choice of hyperparameters, particularly the radii parameter 's and the weight parameter ? More details or experiments here would significantly enhance confidence in the robustness of your method.
-
Can you discuss the implications on your theoretical guarantees and empirical results if key assumptions, such as strong ignorability and Lipschitz continuity, do not hold in practice?
论据与证据
The main claims---(1) a counterfactual covering perspective provides a tractable risk upper bound, and (2) the proposed FCCM enhances estimation performance and improves theoretical risk upper bounds---are well-supported by empirical evidence provided through experiments. The choice of datasets (TOY, IBM, and CMNIST) effectively demonstrates the general applicability and performance advantages of the FCCM approach. Nonetheless, a deeper discussion linking theoretical guarantees to empirical outcomes, especially considering variations in dataset properties such as the degree of overlap between treatment groups, would enhance the validity of these claims. While the empirical evidence is strong, a more explicit acknowledgment of potential failure modes or limitations of the proposed methods could strengthen the authors’ analysis and presentation.
方法与评估标准
The chosen methods and evaluation criteria---e.g., the PEHE metric---are appropriate and commonly accepted within the causal inference community. The diversity in datasets (synthetic, semi-synthetic, and image-based) effectively demonstrates methodological versatility. However, more explicit explanations and motivations for the chosen datasets in the main text would improve clarity. While Appendix C.2 provides some context, essential details such as dataset characteristics (e.g., types and ranges of covariantes and response variables) and preprocessing steps remain insufficiently clear.
理论论述
I reviewed the main theoretical claim (Theorem 3.3) and associated lemmas (A.2, A.3, and A.5). While I did not verify every line in the proofs, the theoretical derivations appear generally sound, utilizing standard techniques. However, assumptions such as strong ignorability and Lipschitz continuity, though commonly used in the literature, deserve more explicit justification regarding their practicality. Moreover, clarifying and simplifying the presentation of assumptions and key theorem statements would significantly improve readability and comprehension.
实验设计与分析
The experimental designs are sound, leveraging suitable baselines including QHTE, Causal-BALD, MACAL, LCMD, BADGE, and BAIT, effectively situating the proposed method within existing literature. Nevertheless, the robustness of the FCCM algorithm could be more comprehensively assessed through sensitivity analyses, particularly concerning hyperparameters such as the radii parameter (for ) and the weight parameter .
补充材料
I reviewed the supplementary materials briefly, particularly the theoretical proofs in Section A. The provided proofs appear thorough and rigorous, but clearer linkage to the main text, as well as additional intuitive explanations, would enhance accessibility. The literature review in Section B looks fine.
与现有文献的关系
The paper situates its contributions within the broader literature, linking clearly to existing AL methods such as BALD and core-set approaches, as well as contemporary causal inference methods like QHTE and MACAL.
遗漏的重要参考文献
The discussion of related literature provided in Appendix B seems mostly sufficient.
其他优缺点
-
Strength: The theoretical framework based on factual and counterfactual covering radii is innovative and potentially valuable for future methodological development.
-
Weakness: The paper's presentation could be significantly improved to enhance clarity and readability, particularly in defining key concepts and motivating algorithmic choices.
其他意见或建议
The paper presents an interesting idea and a decent contribution; however, I believe its readability can be significantly enhanced. Below I list some suggestions.
-
Definitions could be streamlined to improve clarity and reduce redundancy. For example, Definitions 3.1 and 3.2 could be merged to define covering radius generically. which can then be specialized to define factual/counterfactual radii. The same comment also applies to Definitions 4.2 and 4.3
-
The statement of Theorem 3.3 is currently dense and challenging to follow. Explicitly separating and clearly stating assumptions outside the theorem (e.g., by defining "Assumption" environment) would significantly aid comprehension. I would also suggest the authors further discuss the practical implications and limitations of the theoretical assumptions explicitly.
-
The description of Section 4.2 is unclear and confusing. I would suggest the authors revise description more precisely and compactly, e.g., by defining quantities and sets, e.g., , , explicitly using mathematical expressions instead of vaguely defining them in English.
-
Sentences throughout the manuscript often tend to be overly long and verbose. Breaking these into shorter, clearer sentences would enhance readability and improve interpretability.
-
Additional Minor comments
- Please consider revising the captions more informatively. For example: (1) please add "by Algorithm 1" in Figure 2; (2) I think the caption of Figure 3 should contain "Algorithm 2" instead of "Algorithm 1"; and (3) please add "by Algorithm 2" in Figure 4.
- I think Algorithm 2 is missing an input parameter .
Potential Failure Modes: FCCM is designed to better handle the partially overlapped data for a quicker bound reduction while maintaining high coverage (Figure 3). As such, in scenarios where the two treatment groups have no overlapping regions (e.g., biased treatment assignment), the data acquisition of FCCM will be challenged as there are no overlapping, counterfactual pairs to be identified. One possible remedy is to operate FCCM in the latent space where the inter-group distributions are aligned by methods like (Zhang et al., 2020; Wang et al., 2024), but it also offsets the model-independent advantage of FCCM. We will add this discussion in an updated version.
Practicality of Strong Ignorablity (SI): It is noted that the focus of this paper is active label acquisition, it can then integrate with the selection bias-aware method to make predictions. The validity of the SI can be approximated by carefully selecting sufficient relevant covariates and constructing a more balanced dataset.
Practicality of Lipschitz Continuity: Theorem 3.3 assumes the Lipschitzness of the , this can be a practical assumption if the regression model , e.g., a well-regularized neural network (NN), learns a smooth mapping from to , which implies the Lipschitzness for . Also, the NN is differentiable w.r.t. , thus the squared loss is also bounded and sufficiently differentiable w.r.t. , further implying the Lipschitzness of .
Q1: Elaboration on The Sentence: Despite that -BALD () highlights searching for samples within the non-overlapping region, its acquisition criterion naturally prefers samples that lead to higher estimated uncertainty by its criterion. However, the criterion design of is not density-aware. For example, in early acquisition steps, two unseen data points and can be of the same uncertainty, while is an outlier yet data belongs to a dense cluster (not seen). Then, prioritizing the acquisition of over can help generalize the estimator to more samples, which means training on can eliminate more estimation risk when data is limited. However, does not differentiate the importance of and , and thus "account for less risk'' in such cases. How FCCM addresses this challenge: Pertaining to the previous example, the factual covering from FCCM can prioritize the acquisition of over , as has more neighbors (i.e., higher out-degree) than in the graph constructed in Algorithm 2. To enhance the distribution alignment, FCCM makes use of the weight imposed on the counterfactual edges (line 308 left column). For example, assume is in the same treatment group as , while and have similar numbers of neighbors in the graph. Then, if has more neighbors from the counterfactual group than does, then the priority of is higher than owing to the additional bonus (with ) on counterfactual edges when calculating the out-degree. For cases with significant non-overlap (almost no overlap), please refer to our response to potential failure modes. Why queries non-overlapping samples in Figure 6(c): As stated in line 85, relies heavily on the accuracy of estimated uncertainty (model-dependent), which can be erroneously high for certain samples outside the overlapping region, boosting the acquisition metric score. Thus, a few such samples are selected as per Figure 6(c). In contrast, FCCM is model-independent, and the acquisition score is robustly computed via the out-degree of nodes (Algorithm 2).
Q2: Please see the anonymous link: https://anonymous.4open.science/r/To-Reviewer-hAfw for sensitivity analysis results on and (click into folder "To-Reviewer-hAfw''). Note that the acquisition on treatment sample is insensitive on and in our setting, as all control samples () are seen. For , our setting of has an overall lower error across different acquisition budgets. We will include all results in an updated version.
Q3: Theorem 3.3 (Lipschitz continuity): If the Lipschitzness does not hold, the multiplicative constant will be unbounded, and thus the reduction of the radii may not help control the risk upper bound. Theorem 4.1 (SI): SI provides an ideal scenario for acquiring the counterfactual samples, where a quick bound reduction by Algorithm 1 is seen in Figure 2(a). If it cannot be guaranteed in real-world data, e.g., CMNIST, the reduction of the bound will be significantly slower and less effective for Algorithm 1, as shown in Figure 2(b). Thus, it motivates us to propose Algorithm 2 which can handle compromised data distributions more effectively via a slight trade-off in coverage.
We will adopt all the constructive suggestions to update the paper.
Thank you for the response. With the proviso that the authors will incorporate appropriate revisions to clarify the discussed aspects, I have increased my evaluation rating from 2 to 3.
We again deeply appreciate Reviewer hAfw’s valuable time and constructive suggestions, which have been instrumental in helping us further improve the paper. We will update the paper accordingly.
This paper tackles the problem of active learning for treatment effect estimation, where the task is to label a treated or untreated unit given a fixed budget. The authors take the core set approach and extend it by introducing a counterfactual covering radius along with the factual covering radius. An algorithm FCCM is then introduced that chooses to label data points to minimises the expected factual and counterfactual coverings. Through quantitive experiments, the authors show that their method reduces the PEHE metric faster that other methods. Qualitatively, the authors also show that their method is better at choosing data points from the overlap of treatment groups, which is more informative of the treatment effect.
update after rebuttal
I thank the authors for their response, I will keep my already positive score.
给作者的问题
- What is Gamma in L228? I don't believe this is defined anywhere.
- I don't really understand what section 4.2 is trying to solve? It is shown that the covering depends on the data distribution in section 4.4 but statements such as "approximating the full coverage given the relatively smaller fixed covering radius" (L238 RHS) does not make sense to me.
- This problem is compounded as I dont believe P(A^t_F) (and other variants) are properly defined, are these the normalised means over the corresponding covering radiuses?
论据与证据
The claims seem to be supported as their method does outperform previous active learning schemes. It is also clear why it is performing better (better acquisition in the overlap region).
方法与评估标准
The benchmarks are standard for these tasks and PEHE is a standard metric for ITE.
理论论述
The proofs seem correct.
实验设计与分析
The experiments seem sound.
补充材料
The proofs and the experimental details.
与现有文献的关系
Although the idea of using covering sets for active learning is not new, using it for the counterfactual covering seems to be novel. It is also demonstrated that this leads to lear performance improvements.
遗漏的重要参考文献
N/A
其他优缺点
N/A
其他意见或建议
N/A
Q1: is a pseudo-operator defined in the "Input'' section of the Algorithm 1, i.e., . We use as a shorthand for the distance-based radius defined underneath Eq. (8) (line 203 right column) to keep Algorithm 1 concise. Take line 6 of Algorithm 1 as an example, to reduce radius , by applying as the selection criteria, the selected sample from is , where and are respectively the unlabeled and labeled sets for .
Q2: As per line 273 (left column), the informative bound in Theorem 3.3 consisting of four covering radii is derived under the full coverage, i.e., all actively acquired samples should together cover 100% of the full data space (we will mathematically explain the full coverage in Q3). Thus, the straightforward solution in Algorithm 1 strictly maintains 100% coverage requirement using currently the smallest radii (in Definition 3.1 and 3.2) at every acquisition step, while progressively minimizing the four radii via newly acquired samples. However, due to this full coverage requirement, Algorithm 1 cannot work effectively on real-world datasets where inter-group distributions do not align well (Figure 2 and lines 255-269 left column). Thus, in Section 4.2, we further build Algorithm 2 (FCCM) upon Algorithm 1 as a solution. Clarifying the statement: Instead of computing the minimal radii in Algorithm 1, Algorithm 2 fixes the radii by a small constant when performing data acquisition. Though the 100% coverage requirement cannot be fully satisfied in early acquisition steps, as the acquisition expands, we can effectively increase the coverage of the data space. So as shown in Figure 3(b), FCCM can achieve up to 25% bound reduction compared to Algorithm 1 with only a negative coverage gap (99% coverage). Thus, in line 238 (right column) we state ``given the relatively smaller fixed covering radius'', we can '' approximate (approaching) the full coverage''. We will add more explanations to those parts to better motivate Section 4.2 in an updated version.
Q3: For each treatment group , is the proportion of the data points from the unlabeled pool set that has been covered by the training set in the covering ball , i.e., . Analogously, , then the mean coverage rate . We will add this formal definition to complement our textual descriptions in lines 266-271 (right column) to make it easier to follow. A further note is that, for Algorithm 1, we have the full coverage, i.e., , strictly satisfied at every acquisition step, while Algorithm 2 -- FCCM is to maximize the mean coverage rate to approximate the full coverage () given the four fixed small covering radii to constitute a lower bound value.
This paper proposes an active learning approach tailored to enhance the treatment effect estimation from observational data, when labeling outcomes is costly. The reviewers found the paper to be novel, with both the theoretical analysis and experimental results being sound. Most of the reviewers' concerns were satisfactorily addressed in the rebuttal. However, the authors are encouraged to incorporate the reviewers' suggestions in the final version to enhance the clarity of this paper.