PaperHub
5.0
/10
Rejected4 位审稿人
最低3最高8标准差2.1
8
6
3
3
3.3
置信度
ICLR 2024

Optimized Tradeoffs for Private Majority Ensembling

OpenReviewPDF
提交: 2023-09-22更新: 2024-02-11
TL;DR

Optimizing privacy-utility tradeoffs in private ensemble learning

摘要

We study the problem of computing an $(m\epsilon, \delta)$-differentially private majority of $K$ $(\epsilon, \Delta)$-differentially private algorithms for $m < K$ and $\delta \geq \Delta \geq 0$. Standard methods, such as subsampling or randomized response, are widely used but do they provide optimal privacy-utility tradeoffs? Surprisingly, we show that an $(m\epsilon, \delta)$-private majority algorithm with maximal utility can be computed tractably for any $m < K$. Specifically, we introduce Data-dependent Randomized Response Majority (DaRRM), a general privacy framework characterized by a data-dependent noise function $\gamma$ that allows for efficient utility optimization over the class of all private algorithms subject to privacy constraints. By deriving a structural understanding of DaRRM, our novel learning approach is made tractable by critically reducing infinitely many privacy constraints into a polynomial set. Theoretically, we show DaRRM enjoys a privacy gain of a factor of 2 over common baselines under i.i.d. teachers and $\delta = 0$. Lastly, we demonstrate the empirical effectiveness of our first-of-its-kind privacy-constrained utility optimization for ensembling labels and gradients from private teachers through applications of private semi-supervised knowledge transfer and private distributed Sign-SGD, highlighting the outstanding performance of our DaRRM framework with an optimized $\gamma$ against several baselines.
关键词
differential privacyensemble learning

评审与讨论

审稿意见
8

Authors show that a private majority algorithm with maximal utility can be computed tractably under certain assumptions. They introduce a privacy framework characterized by a data-dependent noise function called "Data-dependent Randomized Response Majority" (DaRRM) that allows for efficient utility optimization subject to privacy constraints. Considerable theoretical results and some empirical evidence is presented.

优点

The proposed framework called "Data-dependent Randomized Response Majority" (DaRRM) is interesting and innovative. There seems to be some significant breakthroughs arising from this framework. Designing the tuning parameter γ\gamma with provable privacy amplification, and optimization for γ\gamma are important developments as well.

Significant theoretical details have been established (although I dd not check the proof in details). The reported results from the experiments are plausible and seem reasonable.

缺点

The writing is dense in some parts. The technical assumptions and the mathematical details are not clearly stated in the main paper (although they can presumably be found in the supplementary materials), and hence the various theoretical results referring to Problem~1.1 lack adequate discussion and contextualization.

问题

While realizing that there is limited space, I would request authors to discuss the main technical assumptions they need for their theoretical results. This looks like a very solid work otherwise.

评论

We would like to thank the reviewer for providing valuable and detailed feedback. The setting and the problem we consider are formally stated in Problem 1.1. The assumptions for theoretical results presented in section 4 are clearly stated at the beginning of this section, that is, we consider the setting when all mechanisms are i.i.d. under pure differential privacy guarantee.

评论

Thank you authors, for your response.

审稿意见
6

This paper focuses on the problem of exploring the optimal utility of an (m\epsilon,\delta)-DP mechanism to compute the majority function under mild conditions. In this paper, the authors proposed the Data-dependent Randomized Response Majority (DaRRM) framework that approaches the problem of interest by improving the classical Randomized Response (RR) mechanism on the subsampling probability. The authors provides theoretical guarantees and compare the mechanism to the baselines empirically.

优点

  1. The paper is sound in theory and supported by empirical comparison with the state-of-art benchmarks.

  2. The paper is well organized and written, and lays out its contributions clearly.

  3. Empirical results in the paper showed that the proposed DaRRM framework outperforms the state-of-art benchmarks for different tasks.

缺点

Although the authors proposed the optimization procedure to tackle the problem of designing γ\gamma in general DP setting, it is in general computationally intractable to optimize a set of O(K7)O(K^7) constraints in the (ϵ,δ)(\epsilon,\delta)-DP setting.

问题

How will different priors of pip_i affect the result both theoretically and computationally? Have the authors try any experiments without assuming uniform distribution?

评论

We would like to thank the reviewer for providing valuable and detailed feedback.

Our results hold for any prior pip_i. The key motivation in developing the DaRRM\text{DaRRM} framework is to automatically calibrate the noise added, i.e., the design of our γ\gamma function, so that the utility can be maximized under the same privacy guarantee, without knowing the underlying distribution of pip_i's. We choose the uniform prior distribution of pip_i in our experiments to present results in the most general case when one does not have prior knowledge of pip_i's. If we have prior knowledge of the pip_i's and by incorporating the prior distribution of pip_i in the optimization framework, we can get even higher utility with the same privacy guarantee. Computationally, since we approximate the optimization objective (note this does not affect the privacy guarantee) by subsampling pip_i's from its distribution, using a different prior of pip_i does not slow down the run time.

We added more experiment results in section F.1.3 to compare the error of DaRRMγ\text{DaRRM}_\gamma when γ\gamma is optimized using the objective under

piUniform([0,1])p_i \sim \text{Uniform}([0, 1])

as we did before and under another distribution TUniform([0,0.3][0.7,1])\mathcal{T} \sim \text{Uniform}([0, 0.3]\cup [0.7, 1]), which represents one's prior belief that there is a clear majority among the mechanisms. We then compute DaRRMγ\text{DaRRM}_\gamma's error in three settings with three different actual pip_i distributions: 1. piUniform([0,1])p_i \sim \text{Uniform}([0, 1]),i[K]\forall i\in [K], 2. piUniform([0,0.1])p_i \sim \text{Uniform}([0, 0.1]), i[K]\forall i\in [K], implying a clear majority of 0 among the mechanisms, and 3. pi=0.5,i[K]p_i = 0.5, \forall i\in [K], implying there is no clear majority.

As one can see from the results, if the prior distribution of pip_i is close to the actual distribution, optimizing γ\gamma under this prior distribution leads to additional utility gain (i.e., less error); while if the prior distribution of pip_i is far from the actual distribution, optimizing γ\gamma under the prior distribution is not as optimal but still better than naive methods. We find that using a uniform prior generally leads to good performance.

Please see the updated draft for more results.

评论

Thanks for the response. After reading the reviews and responses, I decide to keep my current score.

审稿意见
3

The paper studies an intriguing question: If we have K (ϵ,δ)(\epsilon, \delta)-DP mechanisms, can we release their majority vote with a privacy guarantee that's better than (Kϵ,Kδ)(K\epsilon, K\delta)-DP? This paper focuses on the binary voting scenario, and presents a data-dependent randomized response mechanism where the probability of releasing the true majority vote is based on the count of the actually majority vote. To maximize the utility, the paper identifies the worst-case distribution pair and reduces the problem into a constraint optimization problem which can be solved in acceptable runtime.

优点

This work identifies a very interesting problem. Releasing the majority voting for an ensemble of DP mechanisms is a very good extension for the well-known private selection problem (release the index of the maximum among an ensemble of DP mechanisms.

The formulation of a semi-infinite program for maximizing the utility sounds a very interesting technique, and the author shows it's practical through Gurobi.

缺点

The writing of the paper can be improved. For example:

  • in the last line of Problem 1.1, SiS_i does not need (D)(D).
  • line 6 of Algorithm 1 should be SiS_i instead of MiM_i.
  • Lemma 3.2: have you defined ff?
  • Lemma 3.3: there is another ff.

I am quite worried about the experiment. For Figure 3, could you plot privacy-utility tradeoff instead? (i.e., change the x-axis from communication rounds to ϵ\epsilon). The same thing also applies to Table 1. For DP experiments, the comparison is only fair when the ϵ\epsilon of different techniques are aligned to be the same.

What is the dimension of the gradient? And how is the privacy parameter for releasing a single dimension being composed? For sign-sgd experiment, the total number of composition is gradient dimension x number of rounds, which seems to be very very large and I am not sure what's the final privacy parameter.

Also, it seems the author assumes each client trains DP models, but for PATE each teacher does not need to be trained differentially private. Therefore, I am not sure whether the comparison in the experiment is fair given that each teacher is already DP but still adds the same noise as what is stated in PATE. Can the author provide a comparison with the original version of PATE for both experiments in Section 6.2 and 6.3?

问题

Is it possible to extend the proposed framework to a multi-class setting? (this does not affect the score given the technical contribution of the paper is rich enough, but I am curious about authors' opinion)

Could you provide some intuition for Theorem 4.1? Especially why when m>K/2m > K/2 the privacy improvement seems to be automatically applied?

评论

We would like to thank the reviewer for providing valuable and detailed feedback. We have uploaded an updated version the draft to reflect changes to improve the presentation. All modified places in the draft are colored in blue.

ResponsetoWeakness.**Response to Weakness. **

Writing.**Writing.**

  • Point 1 & 2 & 3 See the updated draft.
  • Point 4: The ff is defined in Lemma 3.3. To make it clearer, we changed f(...)=f(...)= to be f(...):=f(...) :=.

Experiments.**Experiments. **

In the experiments on private Sign SGD, we fix the total privacy loss of each communication around privacy loss to be mϵ=0.3m\epsilon = 0.3. The total privacy loss for each coordinate in the gradient vector is composed across TT iterations, and the resulting total privacy loss across all communication rounds, i.e., the overall ϵ\epsilon parameter, is the same for each private majority ensembling technique we use in the experiments. We used two simple CNN models on the two datasets MNIST and CIFAR10 with 2893828938 and 90742189074218 parameters Hence, the dimension of the gradient vector here is d=28938d = 28938 on MNIST and d=9074218d = 9074218 on CIFAR10. Since the dimension dd for both datasets are different, we think making a plot with x-axis as the # communication rounds is clearer.

You are correct that the total number of composition is gradient dimension dd ×\times # rounds, which can be very large. This is indeed a drawback of private Sign SGD (which guarantees pure DP), compared to the more widely applied DP-SGD (which only comes with approximate DP guarantees), although people use private Sign SGD instead of DP-SGD to combat Byzantine adversaries. Designing Byzantine resilient and private optimization algorithm with good privacy-utility trade-offs is out of the scope of this work, but can be interesting to think about.

We emphasize that we only use the base version of private Sign SGD as an application to demonstrate the usage of our theory and the DaRRM\text{DaRRM} framework; more advanced versions of Sign SGD with higher utility that exploit sparsity and round skipping will also benefit from our framework. For example, as we have mentioned in the footnote on page 8, to limit the privacy loss, one can apply sparsification techniques to reduce the number of gradient coordinates sent by the clients.

As the reviewer mentioned, our setting and PATE's setting are different and are not directly comparable. PATE aggregates outputs from non-private teachers while we consider private teachers, while our DaRRM\text{DaRRM} framework does not apply when the teachers are non-private. In our setting, when the teachers are private, the ensembling guarantee of PATE still applies as it relies on the typical sensitivity analysis, yet gives much worse utility guarantee, since it cannot exploit the fact the teachers are private themselves. Therefore, we show that the naive PATE extension is suboptimal but there is no way to extend our algorithms into the original PATE setting, and comparing to ensembling non-private teachers in term of downstream accuracy is outside the scope of our setting. Note that the main goal is to show that the typical sensitivity-based noise injection, via "Laplacian" and "GNMax" which are two subroutines used in PATE, are suboptimal in terms of ensembling utility and can be used as naive baselines for majority ensembling in the experiments.

评论

ResponsetoQuestions.**Response to Questions. **

Extensiontomulticlassoutput.**Extension to multi-class output.**

Extension to multi-class output is feasible but there are a few challenges.

Suppose we have KK mechanisms, each outputting a label in TT classes, i.e., Mi:D{0,1,,T1}M_i: \mathcal{D} \rightarrow \{0,1,\dots,T - 1\}. In the binary output case, the support of γ\gamma is simply {0,1,,K}\{0,1,\dots, K\}, which indicates the number mechanisms voting for class 1. However, if there are TT classes, the support of γ\gamma needs to be (v1,v2,,vT1)NT1(v_1, v_2,\dots,v_{T-1}) \in \mathbb{N}^{T-1}, which specifies the number of outputs in T1T-1 classes. That is, γ:NT1[0,1]\gamma: \mathbb{N}^{T-1} \rightarrow [0,1].

The DaRRM\text{DaRRM} algorithm generally follows the one for binary output, except that after we obtain samples S={S1,,SK}\mathcal{S} = \{S_1, \dots, S_K\} from the mechanisms, we compute a vector v=(v1,,vT1)\mathbf{v} = (v_1,\dots, v_{T-1}), where vt=i=1KIv_t = \sum_{i=1}^{K} \mathbb{I} {Si=tS_i = t}, t[T1]\forall t\in [T-1]. The probability pγp_\gamma is set by γ(v)\gamma(\mathbf{v}), and with probability pγp_\gamma we output the true majority; otherwise, we output a random class.

The privacy objective ff, which is later used as the privacy constraint in optimizing γ\gamma, is harder to compute when the output is multi-class. Specifically, one needs to compute

Pr[DaRRMγ(D)=t]\Pr[\text{DaRRM}_\gamma(\mathcal{D}) = t] for some class tt on dataset D\mathcal{D}.

This is now

all possible (v1,v2,,vT1)Pr[DaRRMγ(D)=t\sum_{ \text{all possible } (v_1, v_2,\dots,v_{T-1})} \Pr[ \text{DaRRM}_\gamma(\mathcal{D}) = t \mid

v=(v1,v2,,vT1)]Pr[v=(v1,v2,,vT1)]\mathbf{v} = (v_1, v_2, \dots, v_{T-1}) ] \cdot \Pr[\mathbf{v} = (v_1,v_2,\dots,v_{T-1})].

Also, since the support of γ\gamma is now (T1)(T-1)-dim, we will now need to optimize (T1)(K+1)(T-1)\cdot (K+1) variables, instead of (K+1)(K+1) variables as in the binary output case.
For simplicity, here we do not take into account the symmetry property of γ\gamma.

Designing more efficient optimization algorithms to find the best γ\gamma in the multi-class output case is an interesting future direction.

IntuitionofTheorem4.1.**Intuition of Theorem 4.1. **

We have replaced the proof sketch of Theorem 4.1 with more intuition (in blue). Please see the updated draft.

评论

Thanks for the response!

In your response, you write "we fix the total privacy loss of each communication around privacy loss to be 0.3". However, in your Figure 3's caption, you write "ensuring that each coordinate of the aggregated sign gradient is 0.3-differentially private". So does it mean for each round, the privacy loss is 0.3*d-DP? Then even just 1 round will be more than 8000-DP? I would say such an epsilon is just meaningless. Or you mean each coordinate aggregation is 0.3/d-DP? Given the magnitude of d, it's almost 0-DP and it's hard to believe there will be any utility.

Furthermore, in Appendix, I saw "each client computes its gradient based on the entire local dataset at each communication around", so is the privacy calculation correct in the paper? What does it even mean by "each ensemble itself is ε\varepsilon-DP" in this case? (if the dataset for training different ensemble is different, the naive privacy loss is no longer mεm \varepsilon-DP)

Although I understand that the paper is considering the case that each teacher is already (ε,δ)(\varepsilon, \delta)-DP, I think it's beneficial to add a comparison with original PATE. If the original PATE significantly outperforms aggregating private teachers, then what is the main motivation for this work? E.g., are there scenarios where we must have differentially private teachers and cannot have non-private teachers?

审稿意见
3

This paper studies how differentially private is the majority vote of K (epsilon,delta)-DP classifiers.

优点

The paper studies the interesting question of reducing the cost of majority voting in terms of privacy budget, in particular, a very naive approach would say that KK classifiers are computed, and hence we can say (by the post-processing property) the majority vote is (Kϵ,δ)(K\epsilon,\delta)-DP. The paper studies the interesting question whether we can have a noisy majority vote which is mϵm\epsilon-DP with m<Km<K.

缺点

While the problem studied is interesting, the text is insufficiently rigorous.

The main idea of the text is mostly clear, but sometimes overloading of notations and other elements lead to confusion. For example:

  • In Algorithm 1 line 4, γ\gamma is a real number
  • In Eq (1), γ\gamma is a function γ:N{0,1}\gamma:\mathbb{N}\to \{0,1\}
  • In Lemma 3.2, γ\gamma is a function taking as input S{0,1}KS\in\{0,1\}^K
  • In Theorem 4.1, γ\gamma is a function γ:N{0,1}\gamma:\mathbb{N}\to\{0,1\}
  • In Eq (3), γ\gamma is a tuple of (K+1)/2(K+1)/2 (not KK) real numbers (not booleans), i.e., γ[0,1](K+1)/2\gamma\in[0,1]^{(K+1)/2}

While the γ\gamma confusion is not very harmful, some issues are more problematic, e.g., the formulation of Theorem 4.1:

  • Theorem 4.1 says: "Consider Problem 1.1 when pi=pp_i = p, ..." but Problem 1.1 does not feature a variable pip_i, pip_i^\prime,
  • Theorem 4.1 says: "Given a privacy budget m[K]m \in [K]", but usually ϵ\epsilon is called a privacy budget and ϵ\epsilon is rarely restricted to be an integer. Is mm really to be interpreted as a privacy budget? In fact, according to Algorithm 1, mϵm\epsilon (not just m) is the "target privacy cost".
  • Theorem 4.1 says "Given a privacy budget mm, if one sets γ(l)=...\gamma(l) = ... when m(K+1)/2m\ge (K+1)/2". The sentence seems ill-structured, I suppose you mean "if" rather than "when", and it becomes easier to break the sentence down into smaller parts, i.e., "Let m[K]m\in[K]. If m(K+1)/2"thensetm\ge (K+1)/2" then set \gamma(l) = ...$, else ....".
  • Theorem 4.1 mentions γ\gamma, which does not occur in Problem 1.1, while Problem 1.1 mentions the majority function g not used in Theorem 4.1.

Other notation and related issues:

  • Lemma 3.1 says γsubsampling\gamma_{subsampling} depends on L\mathcal{L} and next defines \gamma_{subsampling(l) rather than γsubsampling(L)\gamma_{subsampling}(\mathcal{L}).
  • Eq 1 defines both γ\gamma and γsubsampling\gamma_{subsampling}. What is the difference in meaning between the two notations?
  • In Lemma 3.3, what are D\mathcal{D} and D\mathcal{D}' ? Are these arbitrary datasets, or do you assume they are adjacent datasets?
  • In Lemma 3.3, γ\gamma is required to be a "symmetric function". What does this mean exactly? I guess that the meaning of "symmetric" depends on the signature of γ\gamma, for example, if γ\gamma is a function of a single variable, we could call a function for which γ(l)=γ(Kl)\gamma(l)=\gamma(K-l) symmetric. If γ\gamma is a function of tuples SS, then we could call γ\gamma symmetric if it is invariant under permuting SS.

There are a few minor language issues, e.g.,

  • Title of Section 5: "Optimizing γ\gamma-function" -> "Optimizing the function γ\gamma"
  • Section 5: "On the other hand, one can to optimize for such γ\gamma^* but this involves solving “Semi-infinite Programming”, due to the infinitely many privacy constraints.": Especially the first part of the sentence doesn't make sense grammatically. Also, you probably mean 'solving a semi-infinite programming problem' rather than 'solving "semi-infinite programming"'.

Some claims are too optimistic. For example, Lemma 3.3 says " ... is (mϵ,δ)(m\epsilon, \delta)-differentially private if and only if ...". There doesn't seem to be a proof (in particular, Appendix B is about preliminaries while appendix C is already about Section 4). I doubt a proof can exist for this statement, since under the provided conditions DaRRM_\gamma could be (mϵ,δ(m\epsilon,\delta-DP even if Eq (2) is not satisfied, for example if the individual mechanisms MiM_i are ϵ\epsilon'-DP for some ϵ<ϵ\epsilon' < \epsilon, or if the mechanisms MiM_i satisfy other favorable conditions. So while I believe one can prove "if", I don't believe one can also prove "and only if".

Similar issues arise in the appendices, which provide a number of proofs but with insufficient rigor to easily determine what is the exact claim nor to easily verify them.

问题

--

伦理问题详情

--

评论

We would like to thank the reviewer for providing valuable and detailed feedback. We have uploaded an updated version the draft to reflect changes to improve the presentation. All modified places in the draft are colored in blue.

Overloadingof**Overloading of \gammanotations. notations. **

  • We updated Algorithm 1 by introducing a different notation, pγp_\gamma, as the coin-flipping probability.
  • In most places in the draft, γ\gamma is the function γ:{0,1,,K}[0,1]\gamma: \{0,1,\dots,K\} \rightarrow [0, 1], except in Lemma 3.2, where we show the generality of DaRRM\text{DaRRM} holds with an even more general γ:{0,1}K+1[0,1]\gamma: \{0, 1 \}^{K+1} \rightarrow [0, 1].
  • To make the γ\gamma notation consistent, we rephrased the optimization problem in Eq.3 so that the variable is now γ[0,1]K+1\gamma \in [0, 1]^{K+1}, with an additional constraint saying γ\gamma is symmetric around K+12\frac{K+1}{2}.

TheformulationofTheorem4.1.**The formulation of Theorem 4.1.**

  • Variables pi,pip_i, p_i' are defined in "Notations", which presents all notations used throughout the paper as is stated in this section "throughout the paper, we use ...", in section 2.1 "preliminaries".
  • To avoid confusion, we now call mm the privacy allowance and have updated this in the draft, including the Appendix. mϵm\epsilon now refers to the usual notion of privacy budget. Just to be clear, ϵ\epsilon is never an integer in the context. mm does not have to be an integer in our context either. However, the total privacy loss of a subsampling algorithm is an integer multiplier of ϵ\epsilon by nature. For a privacy allowance m(i,i+1)m \in (i, i+1), iN+i\in \mathbb{N}^{+}, we can only subsample ii items to ensure the privacy loss is within the total budget. Hence, to simplify the presentation, we consider an integer mm in the context of the subsampling algorithm, as in Theorem 4.1.
  • We have slightly re-structured Theorem 4.1 to make it clearer. See the updated draft.
  • The goal is to design γ\gamma in our proposed DaRRM\text{DaRRM} framework to solve Problem 1.1. We have updated the theorem statement to make it clearer.

Othernotationandrelatedissues.**Other notation and related issues. **

  • Point 1 & 2, we updated the statement of Lemma 3.1, changing Ll\mathcal{L} \Rightarrow l and γγSubsampling\gamma \Rightarrow \gamma_{Subsampling} as suggested.
  • D\mathcal{D} and D\mathcal{D'} are defined as the adjacent datasets in the "notation" section in section 2.1 "preliminaries". We also re-stated that D\mathcal{D} and D\mathcal{D'} are adjacent datasets in the paragraph right before introducing Lemma 3.3, in the original draft.
  • In Lemma 3.3, it is stated "For γ\gamma that is a symmetric function of the sample sum", which implies "symmetric" meaning γ(l)=γ(Kl)\gamma(l) = \gamma(K-l), where ll denotes the sample sum, i.e., i=1KSi\sum_{i=1}^{K} S_i. To make it clearer, we updated the draft by replacing this sentence with "For γ:{0,1,,K}[0,1]\gamma: \{0,1,\dots,K\}\rightarrow [0, 1] such that γ(l)=γ(Kl),l,\gamma(l) = \gamma(K-l),\forall l,"

Minorlanguageissues.**Minor language issues. **

We have corrected those issues in the updated draft.

ResponsetoLemma3.3beingtoooptimistic.**Response to Lemma 3.3 being too optimistic. **

Lemma 3.3 indeed indicates "if and only if". This is because the statement directly follows the definition of differential privacy and applies only algebraic manipulations. We note that for the "only if" direction, the condition that we assume is that DaRRM is private for any ϵ\epsilon-private mechanisms; therefore even if favorable conditions on the mechanisms occur, our privacy cost inequality will still hold. We slightly modified and simplified part of the proof of Lemma 3.3 to emphasize both "if" and "only if" directions. See the updated proof in section C.4. Note that the same lemma statement as in Lemma 3.1 followed by its proof was indeed in section C.4 in the Appendix of the original draft, as promised in the main paper. The confusing part might be that we manually numbered all lemmas in the Appendix and there was a mismatch in the lemma numbers (although the title of each section in the Appendix has the correct lemma number). We corrected all lemma numbers in the Appendix.

评论

Thanks for your answer and the clarifications. I see that when pointed to insufficiently rigorous parts of the text, the authors can improve the text significantly. As said in my review, the lack of rigor is a general issue in the text. I have provided some detailed comments on a few sampled sections, but similar issues also exist most other sections (including, as said, the proofs in appendix). I hence suggest the authors take the time to go step by step over the complete text to make everything rigorous and clear.

评论

We have made a pass through the entire draft, including the main paper as well as all theorems/lemmas/proofs in the Appendix. We tried our best to make all theorems/lemmas/proofs as clear and rigorous as possible. Please see the updated draft. Thanks.

AC 元评审

The paper provides a framework to privately compute the majority vote of multiple classifiers using a data-dependent noise function.

Pros: Reviewers generally find the problem interesting and were impressed by the nice theoretical guarantee.

Cons: The main criticism is the quality of presentation, as commented by three reviewers. This is specifically unsatisfactory for a paper in which theoretical results are main contributions. Additional, Reviewer pyVN also concerned the design of experiments.

The rebuttal went well---authors responded to reviewers questions in details and reviewers acknowledged authors' efforts. There have been some discussions among the reviewers on whether the significance of the results outweigh the presentational issues. Eventually everyone agreed with rejection so that the authors can improve the paper to get more attention once it is published.

为何不给更高分

No reviewers seems to be able to confidently verify the correctness of the results. So, quality of presentation is indeed a valid reason to reject the paper.

为何不给更低分

N/A

最终决定

Reject