Adversarially Robust Dense-Sparse Tradeoffs via Heavy-Hitters
摘要
评审与讨论
The authors present an algorithm for adversarially robust -estimation estimation in the turnstile streaming model, which improves on the space-complexity of existing algorithms in the regime . They achieve this improvement via an alteration of the dense-sparse tradeoff technique of the existing state-of-the-art algorithm where they show it is sufficient to track the heavy-hitters while keeping track of an estimated residual. In addition to the improved Lp-estimation algorithm, the results include an adversarially-robust streaming algorithm for the heavy-hitters problem along with an adversarially robust algorithm for estimating the residual of the frequency vector. The authors also include experiments providing support for their theoretical improvement guarantees.
优点
- The authors do a great job of situating themselves in prior work, and explaining exactly how their results differ and improve.
- While I am not well-versed in this area, the results seem correct and were clearly explained and decomposed.
缺点
-
The actual improvement over the previous work seems very minimal, i.e. a tiny reduction in the space complexity on a very small portion of possible choices of p. However, the relation to heavy-hitters seems like an interesting and non-trivial insight used to achieve this improvement, and the algorithms for heavy hitters and residual estimation seem to me to be of independent interest.
-
The authors do a great job at explaining the intuition behind their techniques and how they fit with existing works. However, I found the actual preliminaries to be extremely confusing. The paper would greatly benefit from a clearer introduction on the exact setting and definitions of the various variables. I found the introduction of https://arxiv.org/pdf/2003.14265 to be very helpful in clarifying my confusion.
问题
- What is the variable b that is referred to in Algorithm 3 (in the parameters for LZeroEst and ResidualEst)? Perhaps I am missing something but I don't see where it is defined.
- I was a bit confused on the purpose of the empirical evaluations. It seems that you demonstrate that empirically in one real-world instance, the flip number of the residual vector can be significanly smaller than the flip number of the entire datastream. However, if I am understanding correctly, your algorithmic guarantees depend on the worst-case flip-number of the residual vector, and are not adaptive to the actual flip number. How should I interpret these empirical results with respect to the performance of your proposed algorithm?
局限性
I think the authors have adequately addressed limitations and the assumptions made in their results.
The actual improvement over the previous work seems very minimal, i.e. a tiny reduction in the space complexity on a very small portion of possible choices of p. However, the relation to heavy-hitters seems like an interesting and non-trivial insight used to achieve this improvement, and the algorithms for heavy hitters and residual estimation seem to me to be of independent interest.
Indeed, we do agree that in some settings, the quantitative improvement is not large. However, we emphasize that optimal bounds for adversarial insertion-deletion streams is a major open question and therefore, the main strength of our work is the conceptual message that the despite the lack of recent progress, previous results are not optimal. Thus we hope our work will inspire future research in this direction.
The authors do a great job at explaining the intuition behind their techniques and how they fit with existing works. However, I found the actual preliminaries to be extremely confusing. The paper would greatly benefit from a clearer introduction on the exact setting and definitions of the various variables. I found the introduction of https://arxiv.org/pdf/2003.14265 to be very helpful in clarifying my confusion.
Thanks for the suggestion. We will expand the preliminaries to reiterate the model described in Section 1 (lines 50-67) with the discussion specifically centered around the heavy-hitter problem and/or the norm estimation problem.
What is the variable b that is referred to in Algorithm 3 (in the parameters for LZeroEst and ResidualEst)? Perhaps I am missing something but I don't see where it is defined.
Similar to Algorithm 1, is the number of queries made to the oblivious algorithms. Since the stream has length and the stream is split into blocks of length , then there are at most queries made to each of the algorithms, as in Algorithm 1. We will explicitly clarify the setting of in Algorithm 3.
I was a bit confused on the purpose of the empirical evaluations. It seems that you demonstrate that empirically in one real-world instance, the flip number of the residual vector can be significanly smaller than the flip number of the entire datastream. However, if I am understanding correctly, your algorithmic guarantees depend on the worst-case flip-number of the residual vector, and are not adaptive to the actual flip number. How should I interpret these empirical results with respect to the performance of your proposed algorithm?
Yes, that's a fair point. However, we remark that the algorithmic guarantees of the previous results also depend on the worst-case flip-number of the entire vector, so in some sense we are still comparing apples to apples by comparing the empirical flip-number of the residual vector to the empirical flip-number of the entire vector. In particular, the algorithm must still commit to some budget for the flip-number before the algorithm is initialized and our empirical results show that the same budget can handle significantly more updates for our algorithm.
Thanks for your response! I will tentatively raise my score to a 7: Accept.
In adversarially robust streaming, one wants to design streaming algorithms that work well even in the interactive setting, in which the stream is not fully fixed in advance, but is constructed element by element by an adversary who can see the current solution provided by the streaming algorithm. If the algorithm is randomized, each intermediate solution may reveal some information about the internal randomness of the car, which could allow the adversary make updates that break the estimates of the algorithm.
The paper concerns two fundamental problems in this setting: heavy hitters and moment estimation. In particular, the paper improves on the best known algorithm for moment estimation that considers two regimes, sparse and dense vectors. The paper shows that estimating moments for can be done in smaller space than previously known.
The achievement follows by leveraging known deterministic algorithms for heavy hitters and using them for tracking significant changes to the frequency vector.
优点
-
Insightful contribution to a very active research area.
-
Non-trivial improvement for some of the most popular streaming problems: heavy hitters and moments. The combination of different ideas to make this work is not easy.
缺点
The moments/norms for which the paper improves the space requirements are not the most important ones, which are I think are and .
问题
"Unusual" values of moments can be used for computing entropy (see Harvey, Nelson, Onak 2008). Do you think this could be used here to provide further motivation for your results?
Your result heavily relies on Ganguly's heavy hitters result. I tried looking at this and related papers but I didn't have enough time to read them. Those are not the most frequently cited papers. Did you maybe verify their correctness?
Are your results in the general turnstile model (where frequencies can get negative) or the strict turnstile modeli (where deletions are allowed, but all frequencies are always non-negative)? In particular, this goes back to the results of Ganguly. Which of the models are they for? Because if they are only for strict turnstile, then the case of is trivial.
局限性
No direct negative social impact.
The moments/norms for which the paper improves the space requirements are not the most important ones, which are I think are and .
In general, a common goal is to find the heavy-hitters above a desired threshold that may be arbitrary. In this case, we want for some threshold . We can then choose the values of and accordingly, so that so that the -heavy hitters correspond to the items whose frequency are above the threshold.
"Unusual" values of moments can be used for computing entropy (see Harvey, Nelson, Onak 2008). Do you think this could be used here to provide further motivation for your results?
Yes, entropy estimation is certainly an additional application for our results, particularly since the previous streaming algorithm utilizes moment estimation for . Thanks for the pointer!
Your result heavily relies on Ganguly's heavy hitters result. I tried looking at this and related papers but I didn't have enough time to read them. Those are not the most frequently cited papers. Did you maybe verify their correctness?
Are your results in the general turnstile model (where frequencies can get negative) or the strict turnstile modeli (where deletions are allowed, but all frequencies are always non-negative)? In particular, this goes back to the results of Ganguly. Which of the models are they for? Because if they are only for strict turnstile, then the case of is trivial.
Ganguly and Majumder's heavy-hitters algorithm uses Chinese remainder codes in their Lemma 9 statement and can be stated for the general turnstile model. Importantly, their Lemma 11 does not give an lower bound for the general turnstile model for the heavy-hitter problem because it only applies for heavy-hitters problem "with parameter " with , where the additive error for each estimate is and the universe has size .
Furthermore, their work is subsequently extended to general error correcting codes in [NNW12], which actually achieves a slightly better space bound in logarithmic dependencies. Specifically, [NNW12] constructs a deterministic matrix with Johnson-Lindenstrauss properties, so that maintaining for the underlying frequency vector is amenable even in the general turnstile model.
[NNW12] Jelani Nelson, Huy L. Nguyên, David P. Woodruff: On Deterministic Sketching and Streaming for Sparse Recovery and Norm Estimation. APPROX-RANDOM 2012: 627-638
Thank your for the response and reassuring me about the correctness of papers by Ganguly et al. I'm happy to improve my score to 7 (Accept). I still stand behind my conviction that the cases of or would be more interesting.
As for the entropy application question, let me just warn you about a subtle point if you decide to mention it in your paper. For a fixed constant , you usually do not care if there is a multiplicative factor that depends only on in the complexity of estimating . In the entropy application, you are not considering, however, a fixed , but you select , where is some function of and perhaps parameters of the stream. Then the final complexity will have an additional multiplicative factor that depends on and you have to make sure it's not prohibitively large. This is usually not what people focus on for moment estimation, but for standard moment estimation algorithms, you get at a factor of at least I think. You would have to check the literature devoted to fractional moment estimation and see what impact (if any) this has on the complexity of deterministic heavy hitters you use.
This paper focuses on the L_p estimation (of the frequencies of items in a stream) in the adversarially robust streaming setting. The previous work by Ben-Eliezer, Eden and Onak achieved an space, which is slightly better than the space bound due to the flip number technique by considering a sparse-dense tradeoff. The authors of this work question whether the above bound is tight or if it can be improved.
The authors present an algorithm that beats the bound of . This is obtained by first building an adversarially robust streaming algorithm for L_p heavy hitters, utilizing deterministic turnstile heavy-hitter algorithms with better tradeoffs. This is then combined with another algorithm that estimates the frequency of the tail-vector, and has additive error and space independent of the size of the tail.
优点
As the paper mentions, this is a conceptual breakthrough, showing that the previous bound of is not tight. The paper explains all the ideas pretty well, and it was easy to read.
The novel insight here is that in the dense-sparse tradeoff of [BEO22], in order to change the p-th moment by at least a factor, most of the updates are to a small number of coordinates, which then naturally must have either already been heavy-hitters, or became so after the update. The authors are able to effectively handle the had case of [BEO22], resulting in a better improvement.
缺点
The only obvious weakness is that the result is a very slight improvement, sometimes in the third decimal place. While the insight is nice, the techniques are a linear combination of different algorithms for different cases. While I did not find the techniques very exciting, I think this passes the bar for NeurIPS.
问题
Do you see any potential barriers to this approach? A discussion on what the limits are (is the bottleneck the heavy hitters algorithm or the tail-estimation algorithm?) would be greatly appreciated. This will also give the reader the assurance that this minor improvement is still significant as it pushes a new idea to a reasonable extent.
局限性
Apart from the limitation described in the theorem statement, the authors could discuss limitations of this approach, and what the current barrier is.
The only obvious weakness is that the result is a very slight improvement, sometimes in the third decimal place. While the insight is nice, the techniques are a linear combination of different algorithms for different cases. While I did not find the techniques very exciting, I think this passes the bar for NeurIPS.
We do agree that in some settings, the quantitative improvement is not large. However, we emphasize that optimal bounds for adversarial insertion-deletion streams is a major open question and therefore, the main strength of our work is the conceptual message that the despite the lack of recent progress, previous results are not optimal. Thus we hope our work will inspire future research in this direction.
Do you see any potential barriers to this approach? A discussion on what the limits are (is the bottleneck the heavy hitters algorithm or the tail-estimation algorithm?) would be greatly appreciated. This will also give the reader the assurance that this minor improvement is still significant as it pushes a new idea to a reasonable extent.
The relatively larger space used by the deterministic heavy-hitter algorithm compared to the space-optimal randomized heavy-hitter algorithms is a major bottleneck on further improving the bounds for this approach. One hope would be to improve the space usage of the deterministic heavy-hitter algorithm. However, it is known that the existing bounds are optimal, i.e., there are no deterministic heavy-hitter algorithms that use less space. Thus, the current analysis we use is tight for this approach and any further improvements likely require significantly new or different algorithmic designs.
Thank you for the response.
The work develops new algorithms in the adversarially robust streaming model. In this model, the algorithm observes updates to some vector arrive in the form of a data stream and maintain estimates to some property of the vector as it changes. At each time-step i, the algorithm receives the update of the vector at this time-step. Specifically, it receives an update of the form that means that the -th coordinate of the vector changes by . Also, at every time-step i the algorithm outputs an estimate of some property of this vector.
The work focuses on the adversarially robust setting, where the updates in subsequent time-steps can potentially depend on the estimates that the algorithm gave in previous time-steps. This is an important property of the streaming algorithm, because when decisions are made based on the estimates given by a streaming algorithm this will likely influence the data received by this algorithm in the future. At the same time, notably, many previously studied streaming algorithms are not adversarially robust. Additionally, the paper focuses on the turnstile setting, i.e. the updates to the vector can be both positive and negative.
The length of the data stream is denoted by , and in the adversarially robust streaming setting the paper mainly studies the problems of -heavy hitter estimation and the problem of estimating the norm of the vector.
The problem of adversarially robust -heavy-hitter estimation requires the streaming algorithm maintain a list of elements that contribute a lot to the norm of the vector. Formally, it requires the list to contain all elements for which and all elements in the list should satisfy .
Up to constantants and logarithmic factors in m and , the -heavy-hitter estimation algorithm requires , where . This improves on the previous algorithm from the literature that achieves . This constitutes an improvement for all in , with the most pronounced improvement at where the space usage is improved from polynomial in to polylogarithmic.
The work also presents new algorithms for estimating the -norm of the underlying vector (up to a multiplicative factor of ). Up to polylogarithmic factors and constants, the amount of space used by the algorithm is of the form . The exact dependence of on the value of is . As the paper notes, for we have whereas the best previous algorithm has .
The work improves on the dense-sparse decomposition method of the previous work.
优点
-
The adversarially robust streaming framework is natural and well-motivated.
-
For the L_1-heavy-hitter problem the work gives qualitative improvement on the previous algorithms by improving the space use from polynomial in the stream length to polylogarithmic.
缺点
-
Other than the -heavy hitter estimation, many of the improvements are extremely incremental, for instance the aforementioned improvement of the dependence on the length of the stream for -norm estimation to from .
-
Other than the results for L_1-heavy-hitter estimation, my understanding is that there is no evidence of optimality for the results given in this work. For example, to the best of my understanding, it can conceivably be the case that the above-mentioned dependence of could in the future be improved to polylog. If this happens, the impact of this work could potentially be small.
问题
In subsection 1.1 what is the relationship between vectors and ? In the rest of the paper is used to denote the frequency vector of the stream, whereas the vector does not appear to be invoked anywhere again. In any case, it would be helpful if the vectors and are clearly defined in the beginning of subsection 1.1.
Subsection 1.1. would be easier to read if the paper used \paragraph to separate the parts of the subsection that address different problems.
局限性
It would be helpful if the authors added a dedicated paragraph in the paper that indicated what are the main limitations of the work according to the authors.
Other than the -heavy hitter estimation, many of the improvements are extremely incremental, for instance the aforementioned improvement of the dependence on the length of the stream for -norm estimation to from .
Other than the results for L_1-heavy-hitter estimation, my understanding is that there is no evidence of optimality for the results given in this work. For example, to the best of my understanding, it can conceivably be the case that the above-mentioned dependence of could in the future be improved to polylog. If this happens, the impact of this work could potentially be small.
We do agree that in some settings, the quantitative improvement is not large. However, we emphasize that optimal bounds for adversarial insertion-deletion streams is a major open question and therefore, the main strength of our work is the conceptual message that the despite the lack of recent progress, previous results are not optimal. Thus we hope our work will inspire future research in this direction.
In subsection 1.1 what is the relationship between vectors and ? In the rest of the paper is used to denote the frequency vector of the stream, whereas the vector does not appear to be invoked anywhere again. In any case, it would be helpful if the vectors and are clearly defined in the beginning of subsection 1.1.
Due to the problems of heavy-hitters and moment estimation, we use both and to denote the same underlying vector. We will unify this notation with the same symbol in the updated version.
Subsection 1.1. would be easier to read if the paper used \paragraph to separate the parts of the subsection that address different problems.
Thanks for the suggestion, we will add paragraph notation to demarcate the discussion on heavy-hitters and norm estimation in the updated version.
Thank you for your response. Tentatively, I keep my rating at 6: Weak Accept
We thank the reviewers for their thoughtful comments and valuable insight. We also appreciate the positive feedback, such as:
- The adversarially robust streaming framework is natural and well-motivated. (Reviewer KENe)
- For the L_1-heavy-hitter problem the work gives qualitative improvement on the previous algorithms by improving the space use from polynomial in the stream length to polylogarithmic. (Reviewer KENe)
- As the paper mentions, this is a conceptual breakthrough, showing that the previous bound of is not tight. (Reviewer pbML)
- The paper explains all the ideas pretty well, and it was easy to read. (Reviewer pbML)
- The authors are able to effectively handle the had case of [BEO22], resulting in a better improvement. (Reviewer pbML)
- Insightful contribution to a very active research area. (Reviewer VAPi)
- Non-trivial improvement for some of the most popular streaming problems: heavy hitters and moments. The combination of different ideas to make this work is not easy. (Reviewer VAPi)
- The authors do a great job of situating themselves in prior work, and explaining exactly how their results differ and improve. (Reviewer B1cC)
- The results seem correct and were clearly explained and decomposed. (Reviewer B1cC)
We provide specific responses to the initial comments of each reviewer below. We hope our answers addresses all points raised by the reviewers and we will be most happy to answer any remaining or additional questions during the discussion phase!
This paper studies the problem of heavy-hitters estimation in the robust streaming model. The results of the paper improve polynomially the memory requirements for this task, and the gains are more pronounced for , where the memory bound becomes poly-logarithmic, as opposed to previous works where the corresponding bound is . These results are based on novel innovations on previous works which introduced a dense-sparse tradeoff technique and techniques from differential-privacy. Reviewers were consistently positive about the merits of the work, while there were some concerns about the broader implications (or lack thereof) of these results, as well as the lack of matching lower bounds.