Prediction-Powered Adaptive Shrinkage Estimation
We introduce Prediction-Powered Adaptive Shrinkage (PAS), a novel method to estimate multiple means using few labeled data points and black-box model predictions.
摘要
评审与讨论
This paper proposes a "shrinkage estimator" for estimating multiple problems at once using PPI (prediction powered inference). They discuss the various methods by which one can reduce variance in such an estimator and how their method takes advantage of each, demonstrating theoretically that they can get better estimators by optimizing the CURE (correlation based risk estimate) which they introduce. Empirically they show on synthetic and real datasets that this improves on multi-estimation over both classical and PPI baselines.
######### UPDATE AFTER REBUTTAL: I remain positive about this paper - keeping my score at 4.
给作者的问题
n/a
论据与证据
I find the claims + evidence shown in this paper to be convincing. The authors provide both theoretical and empirical evidence for their claim, which is that PAS is an improved estimator over PPI in the multi-prediction setup. I find the background and intuition given in sections 2-3 to be quite helpful.
方法与评估标准
Yes, the evaluation setup is fairly standard for PPI and adapted to their specific setting (multi-prediction).
It would be nice to see experiments shown on different sizes of labelled data to demonstrate at what value of n we start seeing improvements from PAS.
理论论述
I did not check closely but I did look at the proofs in Appendix A.
实验设计与分析
fairly standard setup and showing improved risk on two metrics
补充材料
I looked at the proofs in Appendix A although I did not check them carefully
与现有文献的关系
connection to existing literature is well laid out in this paper - discusses very clearly in Section 3 exactly how various methods from the PPI and Bayesian space connect to this work
遗漏的重要参考文献
n/a
其他优缺点
Main critique: I would love a little more intuition about what “parallel problems” are supposed to be, in particular early on in exposition and then tied to the examples in 2.2 and methods in 4.2. I was a little unclear on how "borrowing information across problems" is done (as stated on line 42). I can see there is shrinkage going on but that is calculated on what seems to be a mostly problem-by-problem level as in (15). So a little more explanation here would help my understanding of what's going on - for instance I don't know see why we need the machinery from Eq 5 of defining the meta-distribution over problem parameters, it doesn't seem there's any assumption made on how problems are connected or distributed.
Quick note on the assumption of correlation/covariance statistics being known - Var(Y) and corr(Y, f) can be tough to estimate is labelled data is small. Would be good to discuss how this is implicitly an assumption that labelled data is big enough, and/or what regime of n you expect this estimator to work well in.
其他意见或建议
L204: “cl” in classical is underlined?
L170 right: typo on “unlabeled”
Relatedly to the "sharing information" point, I would be curious to know if there's any type of covariate shift type assumption that you think is helping here- for instance, it is roughly (not exactly) satisfied in the synthetic example Ex 2.2.
We thank the reviewer for considering our idea convincing and thoughtful comments.
-
Sharing information across problems: We agree with the reviewer that further motivation of why and how it is possible to share information across problems will improve our exposition. In our revision we will add further motivation. Below is an alternative attempt to further explain and elaborate on Prop 5.3. Suppose as in that proposition that is the same across all problems, , and that second moments are identical across all problems. We make these assumptions throughout the remainder of our response. Then we could ask: What is the best convex combination of and in the following sense:
By direct calculation (since the RHS is a convex quadratic in ) we find that: This implies the following intuitive result: the larger the MSE , the less weight we should assign to . Note that if we have a single problem , then if is sufficiently large, we could estimate accurately. However, we cannot estimate accurately since we only have a single (a single problem). At best, we can compute an unbiased estimate of this quantity since:
Now suppose we have multiple problems, then we can now learn how good the ML predictor is for estimating the by sharing information across problems. To wit, as
where we emphasize that the mean squared on the RHS also integrates with respect to the meta-distribution that models the distribution of the . Thus we can also estimate the optimal . In words: sharing information across problems allows us to learn how good the ML predictor is and to then decide how much to shrink toward it. Our implementation shares information in a similar way but is more involved due to the heteroscedasticity across problems (not all second moments are identical) and also for heterogeneity in , .
-
Role of meta-distribution: See above. We note that we could state our result in a frequentist setting wherein all parameters are deterministic, similar to the classical result of James-Stein. We preferred to state results in terms of a meta-distribution, since we thought this would be a more familiar setup for audience at ICML.
-
Assumption on known second moments: We agree that our assumption on second-moment parameters are not easily satisfied in practice. In Sec 2 of paper, such treatment is more of a theoretical convenience, and sample-based estimates are used for real-world datasets. Therefore, it is true that an implicit requirement for our theory of PAS to work well in many practical settings is that the sample-based estimates are good enough. While our regime (, finite) cannot yield asymptotic result on these sample-based estimates, we make the following remarks:
-
In practice, PAS works well even when is very small (and so the estimates are noisy). To highlight this point further and motivated by your suggestion, we reran our real data examples with different labeled/unlabeled splits, going from 1%-40%. PAS still dominates other baselines even when only 1% of the data is labeled (e.g. for Amazon Review). [Link to the plots]
-
In our response to Reviewer Tcxk, we propose a new variant of PAS called UniPAS, whose asymptotic guarantee does not require knowledge of second moments. UniPAS also has competitive empirical performance with PAS.
-
-
Other comments:
- L204: The underline is intentional to denote the abbreviation "cl" for classical estimator. We can clarify this.
- L170: Thanks for catching the typo; we will fix it.
- Covariate Shift: We interpret this as potential differences in across problems . In the synthetic model, the mean of varies with , but that is not an assumption that PAS makes. What makes "information sharing" more effective, as mentioned above, is when is small on average across problems. If the question refers to train/test covariate shift within problems, this violates the PPI assumption.
Thanks for the rebuttal. I find these clarifications helpful - the connection of E[theta - Z] to information shrinking as well as to the distributional assumptions piece are both useful for me. The experiments on small n and with UniPAS are also nice! I'm already positive about this paper and continue to be so.
This paper proposes a method for adaptively combining ML predictions with gold-standard labels, to estimate a multivariate parameter (e.g., the mean across several partitions of the data) with small mean-squared error. The paper builds upon the PPI++ estimator, while proposing to additionally perform global shrinkage using an adaptively chosen shrinkage parameter. Experimental results show that the approach approves over the classical estimator in MSE more often than a series of baseline methods.
Update after rebuttal: I have reviewed the response and will maintain my score.
给作者的问题
My questions are essentially derivative from my comments above, but to recap (in priority order):
- How is the PPI++ baseline implemented in the experiments, and in particular, how does the "multivariate" form of PPI++ compare to the PT baseline shown here? Is the multivariate form of PPI++ essentially the same, but with only a single fixed parameter lambda instead of a per-coordinate parameter?
- Where do existing shrinkage methods fit into the baselines shown in the experiments? For instance, does the "Shrinkage" baseline correspond to the method of Rosenman et al. 2023 or some variant of that method?
- As noted in Section 2.2., all of the theoretical work starts from the assumption that are known. For instance, do Proposition 5.1 and Theorem 5.2 still hold as both (number of tasks) and (number of samples) tend to infinity, i.e., is there at least an asymptotic argument that this does not matter?
- Could you explain the conceptual reason why the analog in Equation (16) of to makes sense? I think I get the intuition at a very rough level (it's additional "prior" information), but see my comments above in "other strengths and weaknesses"
论据与证据
The claimed contributions in the introduction are (in my opinion) well-scoped and well-supported, clearly distinguishing asymptotic claims from finite-sample claims, and justifying the benefits of the method with empirical evidence.
方法与评估标准
The method certainly makes sense, and has a clear intuitive basis. Regarding the evaluation criteria, the theoretical development is mainly (in my view) for intuition, given some limitations (e.g., knowledge of certain parameters, see "other strengths and weaknesses"). So I view the empirical evaluation as the main "evaluation" component, where the evaluation approach appears sound to me---I particularly appreciated the use of the "percentage of problems improved" metric as a thoughtful counter to typical concerns with shrinkage-type estimators.
理论论述
Unfortunately, I did not check the proofs in the supplement due to a lack of time.
实验设计与分析
As stated in "methods and evaluation criteria", the evaluation appears sound to me, mainly focusing on the real data analysis which I believe presents the most robust empirical evidence for the method. Given that the focus of this paper is on estimation (and not inference / uncertainty quantification), it is reasonable to measure performance by MSE, and I appreciated the extra inclusion of the "% improved" metric, since (as the authors note) MSE across an entire vector can be improved while sacrificing performance on some dimensions.
补充材料
While I briefly skimmed the supplement, I did not read any particular section in depth.
与现有文献的关系
As noted in part by the authors, there is a long line of work on combining imputed labels (which may be biased) with gold-standard labels. PPI++ is one such idea, where PPI constructs an unbiased estimator using imputed and gold-standard labels, while "power tuning" is added in PPI++ to improve efficiency by estimating the optimal degree of reliance on imputed labels. Shrinkage is another idea, which can be shown to provably improve estimation (in a particular sense, namely MSE across the entire vector of parameters) when there are 3 or more parameters to estimate, an idea going back to the James-Stein estimator. This shrinkage idea has also been applied in (other) areas of combining "potentially biased data" with "unbiased data", such as in causal inference, where the "biased" data is observational and the "unbiased" data experimental (see Rosenman et al. 2023, cited among other shrinkage-style estimators in this work).
In my understanding, this paper brings together these two lines of work with perhaps an additional twist (e.g., adopting a particular Empirical Bayes perspective on shrinkage, which I am less familiar with), and demonstrates that they work well together empirically, while giving some theoretical intuition.
遗漏的重要参考文献
I believe much of the relevant related work is cited, but I would appreciate more discussion of how the proposed approach relates to some of the (cited) alternatives, e.g., Fisch et al. 2024 (who similarly considered stratified PPI++ across different subpopulations) and Rosenman et al. 2023 (who similarly consider shrinkage towards a biased predictor in a related causal inference problem). For instance, does the "Shrinkage" baseline correspond to the method of Rosenman et al. 2023 or some variant of that method?
其他优缺点
The paper is very clearly written, the synthetic experiments are a nice intuition pump, and the real-data experiments were compelling in my view.
However, a few points of clarifications / "weaknesses":
- How is the PPI++ baseline implemented in the experiments? As currently presented, I imagine that PPI++ is implemented (separately?) for each dimension of the target parameter to estimate. However, PPI++ has more general formulations than mean estimation, e.g., one could view the present problem as a more general M-estimation problem, or even the task of learning a generalized linear model with a fixed coefficient for each stratum, both of which are covered in Sections 4 and 3 respectively of [1]. I'm not sure that actually results in a different estimator from the formulation presented here, but it would be helpful to clarify.
- As noted in Section 2.2., all of the theoretical work starts from the assumption that are known. However, a major challenge in this area of research is that these must be estimated from the (same) limited data that we have for estimating . Typically this deficit is handled somewhat crudely by appealing to asymptotics (i.e., convergence of similar terms in probability to their true values), but I don't see any of that here. Of course, the ultimate validation is empirical, but I'm curious how the authors would address this weakness in the theoretical development: For instance, do Proposition 5.1 and Theorem 5.2 still hold as both (number of tasks) and (number of samples) tend to infinity?
- I didn't follow the analog between the main method derived in Section 4 and the exposition in Section 3. For instance, it seems in Section 3 that and are known to be correlated with the same mean, and in any case we should expect that is a better predictor of than . Moving from Equation (12) to Equation (16), we use in place of by "analogy", but I don't understand the conceptual connection, given that e.g., could be arbitrarily biased, while is not.
[1] A. N. Angelopoulos, J. C. Duchi, and T. Zrnic. Ppi++: efficient prediction-powered inference. arXiv preprint (2311.01453v2), 11 2023. URL: http://arxiv.org/abs/2311.01453v2, arXiv:2311.01453v2.
其他意见或建议
As a minor note, I found the use of distinct notation in Section 3 and Assumption 2.1 to be confusing, and it was a little hard for me to see the connections between the two.
We thank the reviewer for a very helpful report, which has helped us improve on this work.
-
We agree that we should expand on connections to related work, which will be added as a new section in the appendix after our revision. Briefly:
- Fisch et al. 2024 (StratPPI): The starting point is similar to ours with multiple parameters . However, in StratPPI, these parameters are not of interest per-say. Instead there exists known weights such that the parameter of true interest is . The ultimate goal is to come up with an unbiased and low variance estimator of .
- Rosenman et al (2023): When their stratum weights are the same, their estimator corresponds to our shrinkage baseline up to differences in the one-dimensional family of shrinkage weights (our Eq. (15)).
- PPI++: Suppose for all problems, then indeed we can cast our setting into PPI++ with a parameter vector. Moreover, PT is then using the same for all problems (more of which below).
-
Previously, we were comparing to PPI++ applied separately to each problem (with its own ). PPI++ itself requires asymptotics with . It would be possible to extend to asymptotics with ; however, based on the reviewer comments we now have a more compelling methodological alternative: if we only seek to compare ourselves to PPI++ with a single , then we can learn the optimal in asymptotics with fixed, (a result complementary to the PPI++ paper which takes fixed and ). In closed form, we have which coincides with the optimal weight selected by PPI++. We further clip to be within .
Without knowing , we can still plug in their sample-based estimates (see Point 2 for Reviewer Tcxk) and obtain (after clipping it as well), which has the property that as (this is different from the per-problem case when is needed). We denote this method as UniPT.
-
Taking UniPT as a starting point, we develop UniPAS. This has an asymptotic guarantee ( fixed, ) analogous to PAS replacing PT with UniPT (that is, we try to be asymptotically at least as good as PPI++ with a single , while PAS was trying to at least as good as PPI++ with a separate-per problem ). The upshot is that UniPAS does not require knowledge of . In empirical results UniPAS is competitive with PAS (although PAS is slightly better).
-
In addition to including two new methods (UniPT, UniPAS), we have reran our real data examples with labeled/unlabeled split ratios ranging from 1% to 40%. [Link to plots]
-
On the analogy of and : We agree that the analogy is somewhat loose and we will provide more details in the revision. Our goal is to provide a heuristic motivation for the one-dimensional parameterized family of weights (whose ultimate success is judged by the empirical results). To elaborate on this: in Eq. (12) of Sec 3, we find that the best weights take the form: for . If we instead take the best convex combination of and (instead of ), the optimal weights again take the form above, now with . Now suppose that we ask for the best convex combination (not necessarily a Bayes predictor) between and where is some fixed function. Then we can show that the optimal convex combination is given by the form above with
(The above expression is interesting as it forces us to inflate "", i.e., to shrink less toward in a way that depends on how close is to .) The takeaway is that for many possible predictors, the optimal weights have the same parameterized form up to the single parameter that varies according to the quality of the predictor. This motivates our one-dimensional family of weights. Once this family has been motivated, we learn in (15) in a way that does not depend on the above analogy at all by minimizing CURE.
I appreciate the clarifications and new comparisons (which I believe should go into the main paper), and I'm glad my review was helpful for improving upon the work. I will retain my score, but I'm still positive on the paper (somewhere between a 3 and 4)
The paper proposes the Prediction-Powered Adaptive Shrinkage (PAS) method to enhance estimation accuracy for multiple means. PAS integrates Prediction-Powered Inference (PPI) with empirical Bayes shrinkage, first debiasing noisy machine learning (ML) predictions within each task and then leveraging information across tasks for improved estimation. Theoretically, the authors establish the asymptotic optimality of the tuning strategy and prove that PAS achieves a lower asymptotic mean squared error (MSE) than any baseline estimator. Experimental results on synthetic and real-world datasets demonstrate that PAS consistently outperforms existing methods.
给作者的问题
How does the violation of the assumption of boundness of finite moment affect the optimality?
How does PAS perform when the ML model is biased, thereby the estimation is incorrect? If there is a way to screen and exclude those poor estimations?
If one of the estimations is nearly optimal (close to the true value), is the shrinkage method over-shrink it?
论据与证据
Please see my comments for each category below.
方法与评估标准
The proposed approach leverages Prediction-Powered Inference (PPI) and empirical Bayes methods to enhance statistical estimation, which I find to be a novel contribution. The theoretical guarantee of asymptotic optimality and the strong performance in numerical experiments further support its effectiveness.
However, the proposed approach's performance is expected to depend heavily on the quality of the predictor used. If I understand the mechanism correctly, PAS does not have a built-in mechanism to correct or exclude misleading predictors, potentially impacting its effectiveness. Additionally, the computational complexity may be high due to the power tuning and adaptive shrinkage steps, which require optimization and could introduce additional overhead.
理论论述
Although the asymptotic optimality of PAS provides a strong theoretical guarantee, one of my main concerns is the assumption of finite moments. The authors state that “PAS inherits the flexibility of PPI in working with any black-box predictive model and makes minimal distributional assumptions about the data.” However, the boundedness of moments may not always hold.
实验设计与分析
The numerical experiments demonstrate the superior performance of the proposed method compared to the baselines. However, given that PAS's performance critically depends on the quality of the predictor, it would be interesting to analyze its sensitivity to heavily biased predictors—for example, by intentionally using a biased predictor to observe its impact. Additionally, since PAS involves power tuning and adaptive shrinkage, the computational cost is expected to be high. A detailed comparison of computational efficiency, such as running time benchmarks, would be helpful to assess the trade-off between accuracy improvement and computational expense.
补充材料
Yes, I briefly went through the proof sketch of the theorem, but there is a possibility that I may have missed something.
与现有文献的关系
This paper could contribute to estimation, particularly for parallel ML tasks, which look for accurate estimation.
遗漏的重要参考文献
NA
其他优缺点
NA
其他意见或建议
NA
We thank the reviewer for acknowledging the novelty and contribution in our paper, as well as the many constructive comments. We hope our response below addresses all the concerns directly.
Background on PPI
The core idea of Prediction-Powered Inference (PPI) is as follows. Given existing black-box ML predictors , how can we use them to enhance classical statistical procedures and improve their efficiency? The focus in this literature is on employing safeguards so that if the ML predictor is good, then statistical gains can be large, while if the ML predictor is bad, then one still retains some form of statistical guarantees (such as consistency). The focus on this literature is not on how to train the best possible ML models (most results are agnostic to that), but rather how to wrap around an existing arbitrary to improve statistical procedures. In this paper, we follow this established perspective in the PPI literature.
Here, arbitrary typically means that can be of any functional form or a total black-box (e.g., API calls to LLMs), and should thus be considered given and fixed. While adapting to the quality of is important, an “arbitrarily bad” is rarely considered in practice, since prediction-powered methods are generally deployed only when is expected to provide at least some signal about the response. Nevertheless, PAS does include mechanisms to adapt to both poor and high-quality predictors. To demonstrate this, we revisit the synthetic model and consider the following predictors:
- a (newly added) very poor predictor , which outputs a value for all inputs,
- the near-optimal predictor from the paper (in fact, closely approximates ).
Table 1: MSE (± se) in synthetic model ( replicates)
| Classical | Prediction Avg | PPI | PAS | |
|---|---|---|---|---|
| 3.14 ± 0.03 | 549.17 ± 2.57 | 24.05 ± 0.22 | 3.14 ± 0.03 | |
| 3.14 ± 0.03 | 0.27 ± 0.00 | 2.69 ± 0.03 | 0.27 ± 0.00 |
In the extreme case where is very biased and provides no useful information, PAS effectively defaults to the classical estimator, showing robustness against poor predictions. On the other hand, with a very good , PAS tracks the predictive mean closely and perform best. Importantly, no over-shrinkage is observed. In words, PAS learns how good the predictor is in a data-driven way.
Bounded moment assumptions
We appreciate the reviewer's question regarding the finite moment assumptions. We first remark that these assumptions are satisfied when the response and prediction are bounded (as in the Amazon and Galaxy datasets), or when their joint distribution is reasonably well-behaved (as in our synthetic setting).
We note that some moment assumptions are standard across the PPI literature. Specifically, finite second moments of the joint model are generally needed for basic procedures like power tuning. These are typically viewed as mild assumptions, although we admit that they preclude heavy-tailed distributions.
Our work extends PPI to compound mean estimation using shrinkage principles, particularly risk minimization via an unbiased estimate (CURE). To prove the asymptotic optimality of PAS, we require finite fourth moments. This is a technical condition used to control the variance of the risk estimate itself.
Finally, addressing the reviewer's question about violations: if the fundamental second moment assumption fails, the variance-reduction premise of PPI itself becomes ill-defined. If only the fourth moment assumption is violated, PAS may lose its formal asymptotic guarantee to outperform power tuning and the predictive mean. However, CURE remains an unbiased risk estimate provided that second moments exist.
A light-weight approach
A key strength of PAS is its low computational overhead compared to both classical and PPI estimators. Although PAS uses both power tuning and adaptive shrinkage, the first stage yields a closed-form expression for the optimal tuning parameter ; the second stage involves optimizing CURE over a one-dimensional parameter space, and each evaluation is inexpensive since CURE has an analytic form as well. This makes the optimization very fast in practice.
We first precompute all ML predictions. After this step, all estimators (including PAS) are very fast to compute. Below we benchmark the runtimes for three estimators on the Amazon Review dataset with the precomputed ML predictions. In each trial, the task is to estimate the mean product ratings for products. We then record the time to construct each estimator by taking the average of 100 repeated constructions. The table reports the mean and max time (in milliseconds) to construct each estimator across 10 such trials.
| Estimator | Mean Time (ms) | Max Time (ms) |
|---|---|---|
| Classical | 7.4 | 8.4 |
| PT | 21.2 | 22.0 |
| PAS | 34.5 | 35.4 |
Paper proposes prediction-powered adaptive shrinkage (PAS), an extension of prediction-powered inference (PPI) that uses empirical Bayes ideas to further reduce estimates when multiple related estimation problems are solved together. The paper is well written and well thought out, with good theoretical and empirical results.
给作者的问题
- What assumptions or methods might be required to strengthen the asymptotic results to a finite m one?
- It would be good to more explicitly work out what happens if the variance parameters tau_j, sigma_j, rho_j are unknown and need to be estimated, and the impact of this additional estimation step on the resulting method and theoretical results.
- Please elaborate on how these variance parameters are estimated from data in the empirical results (particularly the Amazon and Galaxy Zoo datasets).
- It would be nice to look at performance of the method and baselines for smaller n_j/N_j, and also for a wider range of the label/unlabeled split beyond 20-80. The method should be able to work with a wider range.
论据与证据
The claims are well supported by both theory and empirical results. The paper is well written and explains the ideas clearly with good intuition building examples.
方法与评估标准
Yes. Evaluation on both synthetic and real world examples demonstrate the efficacy of the method.
理论论述
Theoretical claims are asymptotic in nature, holding as the number of problems m->infty.
They also assume the variance parameters tau_j, sigma_j, rho_j are known. While paper refers to this as of secondary concern, and referred to existing papers that took the same approach, this would seem to me to be quite a significant assumption and would be good to work out implications of having to estimate these from data on the theoretical developments.
实验设计与分析
The experimental design looks sensible, with clear separation of datasets used to estimate the different quantities needed (e.g. the Bert fine-tuning). It is nice to see the method working across both textual and image domains with different predictors etc.
补充材料
Skimmed through proofs.
与现有文献的关系
The relation to broader literature is clearly set out through the paper.
遗漏的重要参考文献
None.
其他优缺点
None.
其他意见或建议
None.
We thank the reviewer for the careful assessment. The thoughtful questions have helped us improve our work. For better exposition, we slightly reordered our responses to the questions.
-
Finite results. Our current theoretical analysis permits for finite sample bounds. We will track these results more explicitly in the revision. For instance, the term in Prop 5.1 is actually .
-
How are variance parameters estimated in practice? For real-world datasets, we use sample-based estimates for the variance parameters. We estimate and with the standard unbiased estimators
using labeled data. For the prediction variance , we use predictions in both labeled/unlabeled data:
where .
-
Impact of estimating variance parameters (with Point 2). We agree with the reviewer that our current theory does not account for estimation errors due to plugging in sample-based estimates of the variance parameters. One possible remedy would be to consider asymptotics in which all of . However, we prefer asymptotics with fixed and to represent the regime in which we have a lot of very noisy/difficult individual problems and also to distinguish results from standard setup in the PPI literature that keeps fixed and takes . We note PAS with plug-in variance estimates empirically works well even for very small , see point 4 below.
Motivated by the reviewer's comment, we now consider two further methods: UniPT (that is, power-turning with the same for all problems, see response to reviewer qLFS) and UniPAS, which is similar to PAS but uses the same power-tuning parameter across problems. We can prove that UniPAS has asymptotic (, fixed) risk less or equal to that of UniPT, PPI, the classical estimator, and the prediction mean. The result for UniPAS accounts for the data-based estimation of the variance parameters. Here is a sketch of UniPAS:
-
We start with UniPT by estimating a single power-tuning parameter (as in our response to reviewer qLFS) that has the property that in as (with fixed) where is the optimal single tuning parameter. Call the resulting estimator .
-
We come up with stable working estimates of by pretending are the same across all (but sample sizes can vary). The common values can be estimated accurately by averaging the estimates in Point 2 over all . Plugging these in, we derive working estimates of . Then we consider the one-dimensional parameterized family of weights This family retains the property that for it recovers prediction mean and for it recovers UniPT.
-
Pretend momentarily that and are deterministic. (This is not actually needed; by steps 1 and 2 above asymptotically these converge to a deterministic limit.) Suppose we consider the family of estimators
An unbiased estimate of risk is given by CURE, which is analogous to the expression of CURE for PAS. Now here comes the punchline. If above we replace and by unbiased estimates (analogous to Point 2), then we still retain an unbiased estimate of risk. Moreover, since we are averaging over with , we can establish asymptotic uniform consistency of our objective to the true risk, and thus we can establish asymptotic optimality for UniPAS.
-
-
Different and split ratios. Following the reviewer's suggestion we reran our real data analyses with the new methods and different labeled/unlabeled splits, going from 1%-40%. [Link to the plots]
The paper presents an extension of a recent statistical approach called Prediction-Powered Inference. The extension enables to non-trivially exploit this approach in the context of several parallel and possibly related tasks.
All reviewers were very positive regarding the contribution, novelty and importance of the problem. So I am happy to accept this paper.
- The reviewers have made several useful suggestions for improving the presentation and providing intuition, please take these into consideration