PaperHub
6.8
/10
Poster4 位审稿人
最低6最高8标准差0.8
6
7
6
8
2.5
置信度
正确性3.3
贡献度3.0
表达2.8
NeurIPS 2024

The Limits of Differential Privacy in Online Learning

OpenReviewPDF
提交: 2024-05-13更新: 2024-11-06

摘要

关键词
online learningdifferential privacy

评审与讨论

审稿意见
6

The paper studies online learning of concept classes under DP (differential privacy) constraint. The paper makes progress towards understanding mistake bounds (mostly) in the realizable case in a few settings. Concretely, The paper shows that:

  1. If the adversary is oblivious, then PAC pure DP learnability implies online pure DP learnability.
  2. On the other hand, if the adversary is adaptive, then PAC pure DP learnability does not imply online pure DP learnability.
  3. In contrast with the standard online learning model, for every non-trivial hypothesis class, the mistake bound depends on TT.

优点

  1. Assuming that the results are correct, the paper contains a significant improvement towards understanding mistake bounds in online learning under DP constraint.
  2. The questions and results are interesting and in the scope of NeurIPS.
  3. The proof techniques are explained in detail.

缺点

1.It is a bit hard to digest the results and understand the remaining gaps, because there are many settings considered in the paper (DP/non-DP, oblivious/adaptive, approximate/pure...), and no figure/meta-theorem that neatly explains all the relationships. Such a figure/meta-theorem would significantly improve the presentation of the paper.

  1. As a consequence of the above, it is not exactly clear how tight the results are. If you prove a lower bound (as in Section 4), I think it is better to formally state, right after it, the best known lower bound (and the dual statement for proving a lower bound).

问题

Questions:

  1. I think that the type of classes you require in Section 4 is related (or equivalent?) to the notion of "non-trivial classes" which is used in lower bounds for learning with data poisoning. See, e.g., [1,2]. This setting is also related to some notions of "stability", like DP (see [2]).
  2. What is the best known upper bound for the result of Section 4?

Suggestions:

  1. Line 84: you wrote ϵ\epsilon before defining DP.
  2. If I understand correctly, the sentence in line 190 is supposed to say: "Our result reveals that, against oblivious adversaries, pure private online learnability is equivalent to pure private offline learnability in both realizable and agnostic settings." If so, I would recommend clarifying it.
  3. The whole paragraph that begins at line 192 is unclear.  It involves not yet stated results and definitions. Either remove it or move it to a suitable place.
  4. It could be helpful to add a figure of the landscape of separations between the many settings discussed in the paper (DP/non-DP, oblivious/adaptive, approximate/pure...).

Typos/english:

  1. line 146: "not known by" --> "unknown to"
  2. line 264: "an lower" --> "a lower"
  3. line 272: "inpupt"

[1] Nader H Bshouty, Nadav Eiron, and Eyal Kushilevitz. Pac learning with nasty noise. TheoreticalComputer Science, 288(2):255–275, 2002.

[2] S. Hanneke, A. Karbasi, M. Mahmoody, I. Mehalel, and S. Moran. On optimal learningunder targeted data poisoning. In Advances in Neural Information Processing Systems36, 2022.

局限性

Yes.

作者回复

We thank the reviewer for the constructive comments, and our detailed responses are listed below.

figure/meta-theorem that neatly explains all the relationships.

We will add a figure to explain the relationships in the revision.

Our main results imply that there exists a hypothesis class such that:

Oblivious adversary, finite mistakesOblivious adversary, learnableAdaptive adversary, learnable
no DPYesYesYes
approximate DPNoYesYes
pure DPNoYesNo

the best known upper bound

For pure DP, the best known upper bound is OH(log2T)O_{\mathcal{H}}(\log^2 T) as in our Theorem 3.3. This is larger than our lower bound by a factor of logT\log T. We show that OH(logT)O_{\mathcal{H}}(\log T) is achievable for some specific hypothesis classes (Appendix F, page 20-21). Whether the logT\log T dependence is attainable for generic hypothesis classes remains open (page 12, line 480-483).

For approximate DP, Golowich and Livni [30] proposed an algorithm with OH(logT)O_{\mathcal{H}}(\log T) mistakes against oblivious adversaries. Thus, our lower bound is tight assuming a constant Littlestone dimension. However, their algorithm exhibits an OH(T)O_{\mathcal{H}}(\sqrt{T}) upper bound against adaptive adversaries. It remains open whether this can be improved to OH(logT)O_{\mathcal{H}}(\log{T}).

We will include the above discussion in the revision.

[30] Noah Golowich and Roi Livni. Littlestone classes are privately online learnable. In Advances in Neural Information Processing Systems, volume 34, pages 11462–11473, 2021.

I think that the type of classes you require in Section 4 is related (or equivalent?) to the notion of "non-trivial classes" which is used in lower bounds for learning with data poisoning. See, e.g., [1,2]. This setting is also related to some notions of "stability", like DP (see [2]).

Yes, the type of classes we require in Section 4 is equivalent to the notion of non-trivial classes in [1, 2]. We will add some comments and related works on this in the revision. Thanks for pointing this out.

[1] Nader H Bshouty, Nadav Eiron, and Eyal Kushilevitz. Pac learning with nasty noise. Theoretical Computer Science, 288(2):255–275, 2002.

[2] S. Hanneke, A. Karbasi, M. Mahmoody, I. Mehalel, and S. Moran. On optimal learning under targeted data poisoning. In Advances in Neural Information Processing Systems36, 2022.

Suggestions: Line 84: you wrote ϵ\epsilon before defining DP.

We will add a forward reference in the revision.

Suggestions: If I understand correctly, the sentence in line 190 is supposed to say: "Our result reveals that, against oblivious adversaries, pure private online learnability is equivalent to pure private offline learnability in both realizable and agnostic settings." If so, I would recommend clarifying it.

Yes, your interpretation is precise. We will clarify this sentence in the revision.

Suggestions: The whole paragraph that begins at line 192 is unclear. It involves not yet stated results and definitions. Either remove it or move it to a suitable place.

We will rewrite the paragraph to make it more accessible and move it to an appropriate place in the revision. The paragraph is about the gap between our upper bound (OH(log2T)O_{\mathcal{H}}(\log^2 T)) and lower bound (OH(logT)O_{\mathcal{H}}(\log T)) on pure DP online learning.

Typos/English:

We will fix the typos in the revision. Thank you for pointing them out.

评论

Thank you very much for addressing my comments and questions.

审稿意见
7

The paper demonstrates that any function class that is offline PAC learnable with pure DP is also online learnable with pure DP against an oblivious adversary. In this context, a hypothesis class is considered online learnable in the realizable setting if there exists an algorithm with a sublinear mistake bound.

The paper also establishes a distinction between online learning with pure privacy for oblivious and adaptive adversaries. Specifically, it shows that the hypothesis class pointNpoint_N is privately online learnable against an oblivious adversary but not against adaptive adversaries. This finding also indicates a separation between pure and approximate private online learnability, as pointNpoint_N is online learnable with approximate DP against an adaptive adversary.

Additionally, the paper presents a general lower bound on DP online learning against an oblivious adversary for non-complementary function classes.

优点

  • The paper offers general results on private online learnability, identifying conditions under which a hypothesis class is online learnable with pure DP against an oblivious adversary. This connection to Representation Dimension links to existing results on DP PAC learnability.
  • It also explores different layers of separation using the function class pointNpoint_N, contributing to a deeper understanding of the cost of privacy in online learning.

缺点

The proof of Theorem 4.3 is not clear

问题

It looks like Lemma E.2 shows the concentration assumption in in Dmitriev et al. Can you give some intuition on how the proof of Theorem 4.3 get rid of this assumption? Also, are there examples of DP online learning algorithms that do not satisfy their concentration assumption?

局限性

No significant limitations.

作者回复

We thank the reviewer for the constructive comments, and our detailed responses are listed below.

The proof of Theorem 4.3 is not clear

We will polish the proof of Theorem 4.3 to make it more clear in the revision.

It looks like Lemma E.2 shows the concentration assumption in Dmitriev et al. Can you give some intuition on how the proof of Theorem 4.3 get rid of this assumption? Also, are there examples of DP online learning algorithms that do not satisfy their concentration assumption?

Our Lemma E.2 is essentially different from their assumption.

  • In their assumption, it is required that Pr[t[T],A(S0)t(u1)=f1(u1)]1β\Pr[\forall t\in[T], A(S_0)_t(u_1) = f_1(u_1)]\ge 1 - \beta. Roughly speaking, they require that with high probability, the predictions (on u1u_1) are the same for the entire time span when running on S0S_0.
  • In our Lemma E.2, we only require that Pr[A(S0)t(u1)=f1(u1)]1/2\Pr[A(S_0)_t(u_1) = f_1(u_1)]\ge 1/2 for each tt. We allow the predictions to be different at different rounds, which is more general.

A straightforward example is an algorithm that predicts uniformly randomly at each time-step. The algorithm is (0,0)(0, 0)-DP, and satisfies our Lemma E.2. But it does not fit the assumption made by Dmitriev et al. since the probability that all predictions are the same is only 1/2T11/2^{T-1}.

Our proof strategy is also completely different from theirs and does not rely on the assumption. In our proof, we first construct a series of sequences S1,,SmS_1, \dots, S_m, where each of them is obtained from S0S_0 by replacing logT\log T (Algorithm 2 on page 8) data points. By the property of DP, any DP algorithm should output similarly on every SiS_i. We then use a binary search approach (Algorithm 3 on page 9) to show that the outputs cannot be too similar if the algorithm makes less than logT\log T mistakes. Combining the two steps leads to the desired lower bound.

评论

Thanks for your answers. Though the algorithm that predicts randomly at each time step does not satisfy the β\beta-concentration assumption, it has poor utility.

However, I appreciate the explanation that clarifies the proof strategy, and I think the paper has great contribution in providing an algorithm-independent lower bound for DP online learning. I raised my score accordingly.

审稿意见
6

This paper studies limits of pure DP and approximate DP in the context of online learning (with oblivious and adaptive adversaries).

优点

The research questions are interesting and the results may have fundamental value. I'm not an expert in all topics covered by the paper, but the contributions seem novel to me. The text is in general well-written and understandable.

缺点

The paper has a lot of theorems and lemmas, which is interesting. Still, the paper has no conclusions, discussion, further work, description of limitations or illustrative experiments or extensive examples. This puts the task of understanding the value and applicability of the work to a large extent to the reader.

问题

.

局限性

The authors don't discuss limitations or societal impact. I believe there are no ethical concerns with this theoretical work.

作者回复

We thank the reviewer for the constructive comments, and our detailed responses are listed below.

no conclusions, discussion, further work, limitations

We state our main conclusions in Section 1.1, which are the separation results between no DP, pure DP, and approximate DP online learning. The discussion on the limitations and future work were put in Appendix A due to the space restrictions.

We will polish the structure of the paper to make it more accessible in the revision.

illustrative experiments or extensive examples.

We will provide some examples in the revision.

评论

Thanks for your answers.

审稿意见
8

Various results are proved about online learning of private learning algorithms, contrasting no DP, pure DP and approximate DP.

优点

The paper present some interesting new results on online learning with DP. I read up to section 3 and the writing is very clear and the results are important and purportedly novel. (I don't have background or recent experience in the area of online learning so I cannot independently confirm their novelty.)

缺点

I can't point to any weaknesses, but this paper is outside of my area, and I was only able to follow up to Section 3, so it is certainly possible there is something I missed. I am basically taking the paper at its word on the claims made in Sections 1 and 2.

问题

No questions.

局限性

No potential negative societal impact.

作者回复

Thanks for the review and comments, and we have further clarified a few places to make it more accessible.

最终决定

This paper derives basic results in online differentially private (DP) learning. I recommend acceptance.