Sample-Adaptivity Tradeoff in On-Demand Sampling
We study the tradeoff between sample complexity and round complexity in on-demand sampling.
摘要
评审与讨论
The paper studies the problem of adaptive data collection for multi-distribution binary classification. Multi-distribution refers to the learner being evaluated on the error for data coming from across distributions. The adaptive data collection is formalised by a procedure over rounds where in each round the learner can choose the number of samples to collect from each of the distributions (and this can be based on samples collected in previous rounds). The aim of the paper is to understand the trade-off between and the total number of samples needed for accuracy (on the multi-distribution error in a PAC sense). In the realizable case (where a perfect predictor exists), they provide upper and lower bounds that characterise the (near) optimal trade-off. In particular, O(log k) adaptive rounds is optimal to achieve the optimal sample complexity. In the more challenging agnostic case (where a perfect predictor does not exist), an algorithm is provided that achieves the near-optimal sample-complexity in a number of rounds that is independent of . The optimization via on-demand sampling framework is introduced whose connection to the multi-distribution problem allows to motivate the difficulties in improving the round-complexity of the agnostic case, as well as establish general results on the trade-off between sample complexity and adaptivity.
优缺点分析
Strengths:
- The paper is generally well-written and easy to follow (except a few parts – see weaknesses). The algorithms are well explained and the proof sketches provided in the main text are nice and provide some useful intuitions.
- Realizable case: An upper-bound is established that captures a trade-off between adaptivity and sample-complexity. In particular, I believe the main point that can be concluded from the upper and lower bound is about the optimality (sufficiency + necessity) of rounds to achieve the optimal sample-complexity, completing the characterization of optimality for adaptive sample-complexity.
- Agnostic case: An upper-bound is established that achieves the near-optimal sample-complexity with a round-complexity that is independent of the accuracy , improving on prior work whose algorithms had round-complexity that depended polynomially on .
- The algorithms are non-trivial extensions of algorithms from prior work.
Weaknesses:
- Some of the technical description is a bit unclear. There seems to be some missing notation in the description of Algorithm 1, e.g. is not defined (I guess this means m samples from ?), / is not defined (only is). Additionally, it is not clear where and are used in the algorithm, which then makes the following discussion of the technical overview a bit unclear. In particular, my understanding is that the learner is called to to learn a predictor that minimizes the fraction of distributions (as weighted by ) where the weights have been adjusted in previous rounds according to the errors being more or less than but I don’t see how the equation below l.142 fits in. The proof sketch of Theorem 2 for provides some nice intuition but the explanation of the way as uniformly random permutations is not clear. It is not clear to me what is meant by a uniformly random permutation of a copy of . Is a sample from a permutation of ? Is itself a permutation, though my understanding is that is the dimension of subspaces that more or less partition the space, I.e.\ an integer. Also, if possible I think it would be clearer to avoid using notation here and explicitly put the values.
- The lower-bound in Theorem 2 has a dependence while the upper bound is . However, this is enough to establish the necessity of the rounds, which in my understanding is the main contribution this result allows to establish for the realizable case (and perhaps should be stressed more ?), especially since known algorithms with optimal sample-complexity already had a round complexity of . Indeed, for the case where is a small constant, the upper-bound from Theorem 1 is an interesting result, but does not match the lower-bound.
- The high-level connection between optimization via on-demand sampling and MDL is discussed in lines 260-265 (and again in Section 5.3), however it is unclear how the results from each section are connected. What is the connection between Theorem 4 and proposition 1 ? And with this absence of connection, it is a bit unclear how the results from Section 5 fit in with the rest of the paper. In particular, I understand how Theorem 9 provides some insight into how well lazy-hedge can perform for MDL but it is unclear how Theorem 4 does this from an upper-bound perspective. More generally, while Section 3 has a clear structure that is easy to follow, it becomes a bit harder to follow the narrative through Sections 4 and 5. Also, in Section 4 a bit more discussion could be provided as currently the results are simply stated without much discussion.
- Proposition 1 provides an algorithm with near-optimal sample-complexity with a poly(k) round-complexity but no lower-bound is provided.
问题
- In the discussion following Theorem 1, I am not sure about the claim at l.130 that is the “most sample-efficient”. Given the upper-bound in Theorem 1, the most sample efficient is when - I guess what was meant is that already allows to recover the order of the optimal sample complexity. But really if we ignore what we know to be the optimal sample-complexity, the bound in Theorem 1 suggests that we should set to match the second term of the bound. And at a higher level, what it says is that the adaptivity level (the value of ) should be set as a function of k and d and in particular will behave somewhat differently when k>d or d>k. Have you thought about this ? And also how this might be relevant to settings where may be unknown, how hard would it be to extend these results ?
- In l.132, when is chosen in Theorem 1, why is it in the sample complexity and not , is this a typo ? I think there is also a typo l.157: it should be not , right?
- In weakness 4., it was mentioned that no lower-bound is provided for the round-complexity in the agnostic case. In section 5.3, it was discussed the improving the round-complexity required a different approach. Could you comment on whether you believe the round-complexity could be improved and if so what the optimal round-complexity could be?
局限性
yes
最终评判理由
The paper is well-written and presents a substantial set of results, significantly improving upon prior work. There are a few weaknesses to the work but they are sufficiently minor to not be an obstacle to acceptance.
格式问题
NA
Thank you for your detailed review! We start by clarifying the notations that you mentioned in the "weaknesses" part:
-
We will define the notations and in Algorithm 1. Thank you for your comment!
-
Parameters and are used for setting the learning rate (Lines 1 and 8) and the sample size for each round (Lines 2 and 5).
-
The equation below Line 142 is a consequence of Markov's inequality when has an error on the mixture distribution .
-
The notation stands for a length- vector , which is in turn a uniformly random permutation of the following numbers: one copy of , copies of , and copies of . (The constant factors are chosen such that these numbers sum up to at most , and they are suppressed by the notation on Lines 161--162.)
We answer your questions in the following:
- Thanks for catching this! Yes, we meant that setting is sufficient for the factor to be , so larger values of will not help (modulo constant factors).
Regarding setting so that the two terms match, that is a very nice observation and thanks for suggesting it! For example, when , setting would be sufficient for the first term to be dominated by the second. What we had in mind was the regime (e.g., the setup of Theorem 16 (formal version of Theorem 2)), in which case the first term would dominate.
We were unsure about the setting with an unknown : how would the learner sample from the distributions if the number is unknown?
-
Yes, it is a typo (or at least, a loose bound) on Line 132: rounds should be sufficient. And yes, it should have been on Line 157---thank you catching that!
-
We believe that the right round complexity for the agnostic setting should be , mirroring the realizable setting. That is, our lower bounds for the OODS framework suggest a barrier in achieving this round complexity if the learning algorithm treats MDL as an optimization problem in a black-box way. A concrete starting point towards breaking this barrier would be the alternative MDL algorithms (e.g., the one of [Peng, COLT'24]) that use the samples in a more sophisticated way (beyond running ERM or evaluating empirical errors) and sidestep the OODS barrier.
Thank you for your detailed response. Most of my points have been cleared up. I still don't fully understand the eq. below line 142, is it saying the weights distribution errors should be less than p rather than just minimized ? I would encourage the authors to add some more detail explaining what it means / represents in the final version. I also would encourage the authors to think about clarifying the part mentioned in the third weakness. For the unknown , I guess you are right it does not make sense. Otherwise, I am happy to keep my score.
Thank you for your suggestions! For the equation below Line 142, the following is a more detailed explanation, which we will incorporate into the final version:
Algorithm 1 calls the learner to minimize the error of predictor on a mixture of distribution , which can be viewed as a weighted average of the errors of on the distributions. This would then imply that, on most of the distributions, the error of is low.
In more detail, the learner learns predictor for the mixture . By our choice of the sample size (on Line 2), has an error on (except with probability ). Equivalently, if we sample a random index according to , the expected error of on ------is at most . By Markov's inequality, we have , which is equivalent to the inequality below Line 142.
This paper works on reducing the round complexity for multi-distributional learning problem. Multi-distribution learning (MDL) focuses on learning multiple distributions with on-demand sampling. Recent work established near-optimal sample complexity for MDL. In this work, the authors try to reduce the round complexity with lazy updates. They provide a trade-off between the round complexity and the sample complexity for the realizable case, and prove a sub-linear round complexity for the agnostic case.
优缺点分析
Strengths:
- The major contribution of this work is to establish the trade-off between the adaptivity and the sample complexity. This represents a valuable contribution to related fields.
- In technique, the major contribution is a novel application of a variant of the AdaBoost algorithm with a particular notion of margin. The construct of the lower bound is also a significant technical contribution.
Weakness:
- The results in the agnostic setting are relatively weak due to the absence of statistical lower bounds.
问题
- Can you provide some intuition for why the lower bound is not polynomial in ? Is there any method to combine your arguments with the classical lower bound (without limitation on round complexity) to obtain a better rate?
- I wonder whether there is substantial difficulty in extending the upper bounds for realizable setting to the agnostic setting. In particular, fix , could you prove a sample complexity upper bound better than the trivial bound for the agnostic setting?
局限性
The theoretical results only apply to the specific parameter regimes.
最终评判理由
My major concerns are resolved, so I would like to keep my evaluation as 4.
格式问题
I did not find any significant formatting issues.
Thank you for your review! We answer your questions in the following:
- While the lower bound in Theorem 2 is stated for constant (and thus does not have an dependence), it is easy to derive an lower bound for general via the standard argument below. (The dependence would be instead of since the hard instance that we use is realizable.) We focused on the constant- regime for brevity. We will add the following discussion to our paper and point to it after the statement of Theorem 2.
We start with the construction for constant accuracy parameter . Then, we obtain a new MDL instance by "diluting" each data distribution by a factor of : we scale the probability mass on each example by a factor of , and put the remaining mass of on a "trivial" example , where is a dummy instance that is labeled with by every hypothesis. Learn this new instance up to error is equivalent to learning the original instance (before diluting) up to error . Intuitively, the example provides no information, and the learning algorithm has to draw samples in expectation to see an informative example. This leads to a lower bound with a factor.
- Yes, Theorem 20 and the discussion in Appendix C.4 imply an -round algorithm for agnostic MDL with sample complexity for any . For example, setting gives an -round algorithm with sample complexity , which is generally better than the trivial bound of .
Thank you for the detailed response. I would like to keep my score at the moment.
This paper studies an adaptive learning process in the Multi-Distribution Learning (MDL) setting, and characterizes a tradeoff between the sample complexity and round complexity. Specifically, it studies to what extent an adaptive sample collection can achieve a near-optimal sample complexity while maintaining a beneficial round complexity. Under the realizable setting, it characterizes the sample complexity as , where is the round complexity, using a variant of the classical AdaBoost algorithm. In the agnostic case, it shows that with rounds, a sample complexity of can be achieved through an revised algorithm from that of [ZZC+24] (by allowing laze sampling with a cap). From a theoretical perspective, the paper also introduces the framework of Optimization via On-Demand Sampling that naturally connects to the multi-distributional learning problems, which implies both the upper and lower bounds for the round complexity with specified sample overhead.
优缺点分析
The paper considers an important problem of multi-distributional learning under the adaptive setting, and provides a theoretical foundation for studying the tradeoff between the sample and round complexity. It studies the problem in both realizable and agnostic settings, providing both upper (through well-designed algorithms) and lower bounds (through rigorous analysis). The paper is very well presented with original and insightful discussions. The work is very influential in adaptive learning settings, and could benefit in this area of research, including on-demanding sampling, multi-arm bandit problems, continual learning, to name a few.
问题
Did authors consider other boosting algorithms that might further reduce the sample complexity?
局限性
Yes.
最终评判理由
The paper is of high quality. During the discussion period, the authors provided further clarification on its related work. I have checked other reviews, and all reviewers showed support for acceptance. Hence, I believe the paper is definitely above the acceptance bar.
格式问题
No concerns.
Thank you for your review! For the realizable setting, in Appendix A (Lines 584--607), we discuss using other boosting algorithms based on recursive boosting [Schapire, 1990] and boost-by-majority [Schapire and Freund, 2012]. They yield similar quantitative bounds in terms of tradeoff between sample complexity and round complexity, but can be favorable in the regime of constant rounds. For example, as mentioned in Lines 597--598, only 3 rounds of adaptive sampling suffice to achieve a sample complexity of .
This work studies on-demand sampling, where there are distributions and the goal of Learner is to learn a hypothesis that minimizes the worst-case error. They consider two classical settings in statistical learning: the realizable setting and the agnostic setting. They present both the upper and lower bounds results. For this work, they study the fundamental trade-off between sample complexity and round complexity, which is very important.
优缺点分析
Strength: the studied trade-off is very fundamental.
Weakness: for some part about the presented Algorithm 1, it is not clear. The details are in the box below.
问题
-
What is adaptive sampling, or how to decide in each round, whether it is adaptive sampling or non-adaptive sampling? And what is full adaptivity and full non-adaptivity? I hope to see more background knowledge of adaptive sampling.
-
Is my understanding right the provided results in the realizable setting can only improve the existing non-adaptive sampling results? From the content below Theorem 1, it seems it cannot improve the results shown in [BHPQ17, CZZ18, NZ18].
-
Can you give more explanations and intuitions about the agnostic results such as how to get rid of the in your new results? And why the round complexity does not need to be dependent on ? For me, the existing results make more sense.
-
For Algorithm 1, how to decide the number of samples that are distributed according to each of the K distribution in each round? Can it be varied among distributions?
-
The discussion below Theorem 1, it will make it look nicer if, the two special cases, the case where and , are presented in Table 1.
-
Also, for the discussion below Theorem 1, it seems that when , it cannot recover the existing results. Why claim it is almost recovered? I think it only improves the non-adaptive sampling results.
-
Algorithm 1 inputs round , can it be chosen adaptively, for example, based on the past collected data?
-
Maybe I missed something in Algorithm~1. what are and ?
-
In algorithm 1, in step 3, it seems we do not know right? So, how to compute ?
-
For Theorem 2, it is a minimax lower bound or a lower bound for this specific algorithm? If it is a minimax lower bound, it seems to be contradicted with the existing results in [BHPQ17, CZZ18, NZ18] in the realizable setting.
局限性
N/A
最终评判理由
There are some space for the improvement in terms of presentation. So, I keep my current score.
格式问题
No
Thank you for your review! We answer your questions in the following:
- The term adaptive sampling refers to that the sampling is done in multiple rounds (see Lines 104--107). The learner does not get to "decide" whether each round is adaptive or not---every round counts toward the round complexity.
Put in a different way: If the learner wants the second round of sampling to be non-adaptive (i.e., independent of the samples drawn in the first round), they could have merged the first two rounds into a single round.
"Full adaptivity" means that there is no limit on the number of rounds. "Full non-adaptivity" means that there is only one round.
-
The algorithms of [BHPQ17, CZZ18, NZ18] all require rounds. In comparison, our results for the realizable setting (Theorems 1 and 2) apply to a general value of (between and ) and not just the fully non-adaptive case. As mentioned on Lines 47--50, little was known in the regime that .
-
To see why the -dependence can be removed, it would be the most informative to check Proposition 2 and its proof sketch, which we outline in the following. The round complexity of the algorithm in [ZZC+24] has an -dependence because the Hedge algorithm needs iterations, and each iteration requires one round of sampling. In comparison, the box version of our algorithm only samples when the weight assigned to one of the distributions substantially increases (say, by a factor of ). Since the weight of each distribution starts at and is at most , each weight can double at most times. There are distributions in total, so the round complexity of the algorithm is at most , which is independent of .
-
In each round , the numbers of samples to be drawn from the distributions are computed from the mixing weights . On Line 5 of Algorithm 1, samples are drawn from the mixture distribution . Drawing a single sample from is equivalent to the following two steps: (1) Sample according to the distribution specified by , i.e., is set to each with probability ; (2) Draw a sample from . By repeating this times, Algorithm 1 obtains the number of samples that need to be drawn from each .
-
Thank you for this comment!
-
Setting in Theorem 1 gives a sample complexity of , recovering the existing results in the first row of Table 1. (Note that the factor reduces to when .)
-
As stated, is a parameter of the algorithm and is specified before drawing any samples. It is an interesting direction for future work to design learning algorithms that set adaptively, and thereby using as few rounds as possible on each MDL instance.
-
is the mixture distribution specified by the mixing weights , i.e., . For a distribution and positive integer , denotes the product distribution formed by copies of . That is, is a shorthand for drawing independent samples from . We will clarify these notations---thank you for catching these!
-
Note that is only used on Line 5, where Algorithm 1 draws independent samples from . As mentioned in the answer to Question 4, the algorithm does not need to know to sample from ---to sample from , it suffices to draw an index from uniformly at random and sample from .
-
Theorem 2 is a minimax lower bound that holds for all learning algorithms. It does not contradict the results of [BHPQ17, CZZ18, NZ18], since the algorithms in these work all require at least rounds. At , the lower bound in Theorem 2 reduces to , which is consistent with the prior results (namely, the sample complexity for constant ).
Thank you for the detailed response. I would like to keep my score.
The reviewers agree that the paper studies a fundamental tradeoff. Moreover, the paper is well-written and presents a substantial set of results. The Algorithms are well explained as well as the proof sketches. And the results altogether for the realizable and agnostic case make an interesting and deep contribution, which is very relevant to the field. I will therefore accept.