PaperHub
7.0
/10
Poster4 位审稿人
最低6最高8标准差1.0
6
8
6
8
3.0
置信度
正确性3.8
贡献度2.8
表达2.5
ICLR 2025

Optimality of Matrix Mechanism on $\ell_p^p$-metric

OpenReviewPDF
提交: 2024-09-26更新: 2025-02-25

摘要

In this paper, we introduce the $\ell_p^p$-error metric (for $p \geq 2$) when answering linear queries under the constraint of differential privacy. We characterize such an error under $(\epsilon,\delta)$-differential privacy in the natural add/remove model. Before this paper, tight characterization in the hardness of privately answering linear queries was known under $\ell_2^2$-error metric (Edmonds et al. 2020) and $\ell_p^2$-error metric for unbiased mechanisms in the substitution model (Nikolov et al. 2024). As a direct consequence of our results, we give tight bounds on answering prefix sum and parity queries under differential privacy for all constant $p$ in terms of the $\ell_p^p$ error, generalizing the bounds in Hhenzinger et al. for $p=2$.
关键词
differential privacy

评审与讨论

审稿意见
6

The paper studies worst-case lower bounds for answering linear queries under add-remove approximate differential privacy for an pp\ell_p^p error metric. In contrast, past work derived instance-specific lower bounds for unbiased mechanisms under substitution approximation differential privacy in terms of p2\ell_p^2 metrics. Its main result is a lower bound that is characterized by the factorization norm of the query matrix (Theorem 1.4). This lower bound is almost tight with an upper bound that is also provided using, as a long line of recent work has done, the matrix mechanism (Theorem D.1). The paper also specializes this result to the specific problems of prefix sum and parity by explicitly analyzing the problems' factorization norms (Theorem 1.5 and 1.6).

优点

To the best of my knowledge, the claimed novelty in the paper is accurate, and it is nice to have a more complete lower bound picture for linear queries. The paper does a reasonable job of providing high-level explanations for some of its technical challenges and solutions. It is mostly easy to read with careful explanations.

缺点

To me, the paper's primary weakness is significance. Any one of the directions of novelty referenced above -- add/remove vs substitute, pp\ell_p^p vs p2\ell_p^2, potentially biased vs unbiased -- feels pretty niche on its own, especially since we're only talking about lower bounds (as I think the authors would agree, the matrix mechanism upper bound is not a significant contribution here). As is, I don't think the paper convincingly argues why expanding known results in these directions is worth doing, beyond it just being something we could do. Since these are lower bounds, I want a stronger argument that these results somehow add "morally" to our understanding of private mechanisms -- and I'm not sure one exists (but, see questions below).

问题

Overall, I think expending more effort trying to explain why the directions of novelty add up to an "interesting" result could go a long way for this paper, but as is the qualitative improvement over known results feels weak. Some specifics:

  1. The paper repeatedly emphasizes the distinction between add-remove and substitution, but this is described in a confusing way. Throughout, the database is presented as a vector in Rn\mathbb{R}^n -- what is nn here? Is it the size of the data universe, or the number of data points? Since neighboring databases both lie in Rn\mathbb{R}^n and the paper claims to study add-remove privacy -- where I'd expect neighboring databases to literally have # data points that differ by 1 -- I expect that nn is the size of the data universe and xx is a histogram, and Appendix B.3 seems to agree. But then, I don't know why it's more natural to allow neighboring databases to have xx11\|\|x-x'\|\|_1 \leq 1, which is more like a strange "fractional" contribution model where a user can technically contribute <1<1 to many items. Edmonds-Nikolov-Ullman also appears to use the same neighboring notion as Nikolov-Tang, despite this paper's claim that it uses the same notion as this paper. Overall: I am confused by this bit.

  2. Considering general p\ell_p norms typically feels to me like generalization for its own sake. The paper makes a few attempts to argue that p2p \neq 2 is more "natural", but the arguments are vague. The error metrics are clearly different, but what do they meaningfully add to our understanding? I think this paper would be stronger if the authors would try to make as cogent as possible a case that generalizing pp actually tells us something useful.

  3. I have pretty much the same reaction to biased vs unbiased as I do to pp vs 2. I don't think anybody expected that biased mechanisms were the key to better worst-case error guarantees, so their inclusion again feels more like a technical wrinkle that necessitates different arguments than something that significantly adds to our understanding of privacy (and since it's lower bounds, I think adding to understanding matters more -- if there were new upper bounds, there would be a much simpler case that it's just a new useful thing).

评论

We thank the reviewer for the comments and suggestions.

About the privacy notion

We would like to emphasize that the main distinction between substitution model and add/remove model is that they are related to different sensitivity polytope that are not directly comparable, and thus we need a new lower bound (see also the discussion in Appendix B.3). As the reviewer pointed out, nn is the size of data universe, and the number of data points is x1\lVert x\rVert_1. We define the privacy notion using xx1\lVert x - x'\rVert_1 not to emphasize that we permit "fractional" contributions from multiple users (as the reviewer mentioned), but rather to just encompass the add/remove model. In this model, x1\lVert x\rVert_1 and x1\lVert x'\rVert_1 do not need to be equal because we allow the addition or removal of a single user. We do not intend to say that allowing fractional changes is more natural, we will make this point more clear.

"Edmonds-Nikolov-Ullman also appears to use the same neighboring notion as Nikolov-Tang"

Edmonds-Nikolov-Ullman uses the same privacy notion as ours, as it does not require x1\lVert x\rVert_1 and x1\lVert x'\rVert_1 to be equal (see also page 26, specifically the discussion between Lemma 26 and Lemma 27 in the ENU paper). Moreover, we have confirmed with the authors of NT24 in personal communication that they consider the substitution model, in which the number of users, x1\lVert x\rVert_1, is fixed.

"The error metrics are clearly different, but what do they meaningfully add to our understanding?"

We believe studying p\ell_p error contributes to our understanding in the following aspects:

  1. Studying the p\ell_p error for larger pp provides a characterization of the distribution of the error. To elaborate, let viv_i be the difference between the ii-th coordinate of the output and the ground truth, then the error we studied is just E[v1p++vnp]\mathbb{E}[|v_1|^p + \cdots + |v_n|^p]. For any additive noise mechanism that adds noise from identical distribution, by the linearity of expectation, our lower bounds provides a lower bound for the pp-th moment E[vip]\mathbb{E}[|v_i|^p] of the noise distribution. Therefore, obtaining the lower bound for various values of pp offers a more precise characterization of the "optimal" distribution, as it is well-known that the pp-th moment for p2p\geq 2 provides substantial information about the distribution more than just E[vi2]\mathbb{E}[|v_i|^2], which is implied by the squared error.

  2. We note that our lower bound on p\ell_p error is tight, as it is accompanied with a matching upper bound. So it interpolates the curve of the "correct" magnitude of the error. For example, in the private prefix sum problem, we show that the tight error is Θ~(n1/p)\tilde{\Theta}(n^{1/p}), bridging the gap between the known bounds for squared error (Θ~(n)\tilde{\Theta}(\sqrt{n})) and worst-case error (polylog(n)\text{polylog}(n)).

  3. It serves as a unifying lens to understand several prior works including [ENU20], [ALNP24] and [NT24] since both [NT24] and [ALNP24] study the similar p\ell_p error.

[ENU20] The Power of Factorization Mechanisms in Local and Central Differential Privacy

[NT24] General Gaussian Noise Mechanisms and Their Optimality for Unbiased Mean Estimation

[ALNP24] PLAN: Variance-Aware Private Mean Estimation

"I don't think anybody expected that biased mechanisms were the key to better worst-case error guarantees"

In our experience, the best possible worst-case error is usually achieved by biased mechanism, so we are not sure if we understand the reviewer's comment correctly (please correct us if not). In the context of linear query, a line of works have been working on answering the sizes of all cuts privately in a graph, which can be considered as a special case of linear query whose workload matrix is the edge-adjacency matrix in terms of all (S,T)(S,T) cut. The optimal private algorithm on this problems that gives O~(mn)\tilde{O}(\sqrt{mn}) worst-case additive error is achieved by private mirror descent ([EKKL20]), which is a biased mechanism as it computes the private gradient by multiplicative Gaussian noise. Similarly, private multiplicative weights update (MWU) usually gives the optimal worst-case error for other well-known linear queries [Vad16] and it is also a biased mechanism because of its weighing of the hypothesize (and private) dataset. Apart from linear queries, exponential mechanism is usually considered as the optimal algorithm for many private problems, while discrete exponential mechanisms are typically biased (for example, consider the randomized response mechanism).

[EKKL20] Differentially Private Release of Synthetic Graphs

[Vad16] The Complexity of Differential Privacy

评论

Overall, I'm sufficiently convinced to increase my score from weak reject to weak accept. I personally don't find this kind of paper very exciting, so I'm not going to fight for it, but I don't think it's an unreasonable fit at ICLR. Some details below:

We would like to emphasize that the main distinction between substitution model and add/remove model is that they are related to different sensitivity polytope...

OK, this makes sense, I see how the xx11\|x-x'\|_1 \leq 1 notation is a way to incorporate add-remove privacy. I'm more used to thinking of add-remove in terms of an 0\ell_0 and \ell\infty constraint than an 1\ell_1 constraint, but it seems reasonable enough.

Edmonds-Nikolov-Ullman uses the same privacy notion as ours ...

Their Section 2.2 (https://arxiv.org/pdf/1911.08339) says that adjacent databases have the same number nn of data points, and one is obtained by "replacing an element", which is substitution, not add-remove. However, the section you reference does appear to use the same xx11\|x-x'\|_1 \leq 1 notion, so maybe their earlier definition is just mismatched with the rest of the paper.

We believe studying error contributes to our understanding in the following aspects ...

I agree that the provided characterization fills a gap in the current understanding. However, my general feeling about studying p\ell_p error for 2<p<2 < p < \infty is that it's often an answer to a question that nobody is really asking, in a (maybe narrow) practical sense. For me, this is compounded by the optimal matrix mechanism being a pretty impractical algorithm -- my understanding is that convergence to this optimum is only known in polytime for some highly restricted classes of polytopes (as described in NT24) and even then the polynomial is pretty bad. So, slightly more general lower bounds tight with an unrealistic algorithm don't feel very meaningful to me. I might be less interested in this kind of result than the median ICLR attendee, of course.

I do think the added explanation given in this response would help the current submission's introduction.

In our experience, the best possible worst-case error is usually achieved by biased mechanism, so we are not sure if we understand the reviewer's comment correctly ...

I had an incorrect notion of what "unbiased" means in mind. The provided examples are helpful -- I suggest putting at least the graph cut example in the submission.

评论

We thank the reviewer for raising the score and further sharing your constructive suggestions. We will definitely incorporate them to improve the paper.

审稿意见
8

This paper studies the pp\ell_p^p-error of privately answering linear queries. The authors show a lower bound on the error in terms of a certain factorization norm (γ(p)(A)\gamma_{(p)}(A)), suggesting that the classical matrix factorization mechanism is optimal up to logarithmic factors. Based on this main result, they also provide an easy-to-use lower bound that depends on the Schatten-1 norm of the workload matrix. They then apply their results to two kinds of queries: prefix-sum and parity, and obtain nearly tight bounds for both tasks.

优点

  1. Linear query release is a fundamental problem in private data analysis. This article characterizes the error of this task with respect to pp\ell_p^p-metric, contributing to a deeper understanding of the role of differential privacy.
  2. The writing is clear. Also, the technical details are well-organized and easy to follow.

缺点

  1. It appears that the upper bound and lower bound only match in the high privacy regime. In the low privacy regime, where ε=Ω(1)\varepsilon = \Omega(1), the presented upper bound cannot match the lower bound.
  2. When pp is not a constant, the proposed bounds may not be tight for certain regimes of the parameters. For example, when n=pn = p, the upper bound in Theorem 1.5 is O(log1.5(p))O(\log^{1.5} (p)) while the lower bound is Ω(log(p))\Omega(\log (p)).

However, I think both of the above are not serious issues.

问题

I do not understand what the workload matrix for parity queries is like. Can you provide a small example?

评论

We thank the reviewer for the comments and suggestions.

"I do not understand what the workload matrix for parity queries is like. "

Depending on which parity queries are made, the workload matrix would consist of a subset of rows of a normalized n×nn \times n Hadamard matrix.

评论

Thank you for your response.

Could you incorporate a small toy example, e.g., d=2 or 3d = 2\text{ or }3, in the revision?

评论

Thank you! We agree that including a toy example would improve the clarity of the workload matrix, and we will incorporate it into the revision.

审稿意见
6

The paper investigates the asymptotic lower bounds of the expected p\ell_p errors of differentially private algorithms for answering linear queries. Specifically, given a collection of linear queries represented by a query matrix AA, it presents an asymptotic error lower bound that any differentially private algorithm will incur, expressed in terms of the matrix AA and privacy parameters.

After establishing this lower bound, the paper demonstrates that the matrix mechanism can achieve it, up to logarithmic factors. Additionally, it explores two specific linear queries: the prefix sum and the parity queries, deriving their error lower bounds as special cases of the general lower bound obtained. It also presents specific matrix mechanisms that achieve asymptotic tight upper bounds (again, up to logarithmic factors).

优点

  1. The problem addressed is of fundamental importance.
  2. The paper successfully derives asymptotic lower bounds for matrix mechanisms under p\ell_p error.
  3. It demonstrates sophisticated applications of existing techniques.

缺点

  1. The paper demonstrates that the matrix mechanism is asymptotically optimal, rather than instance optimal, for the error metric considered.
  2. The writing is generally good, but there are parts that lack consistency.

问题

  1. In Theorem 1.2, the distribution of zz does not incorporate the privacy parameter ϵ\epsilon. It seems that something is missing here.
  2. In the next paragraph, change “also obtain a characterization of err() that is tight up to log factors” to “also obtain an err() that is tight up to log factors.”
  3. Why is Theorem 1.3 included in the introduction? It appears disconnected from the surrounding discussion. For instance, do you use Theorem 1.3 to prove the lower bound of the prefix sum later in the paper?
  4. Line 113: Change v1p++vmpv_1^p + \ldots + v_m^p to v1p++vmp|v_1|^p + \ldots + |v_m|^p.
  5. In line 113, why doesn’t the approach of Nikolov & Tang explain the additive noise mechanism? Is it because the additive noise mechanism might not be unbiased?
  6. Is it common to encounter biased additive noise mechanisms? Would it be challenging to derive a lower bound for the biased case by reducing it to the unbiased one, using techniques similar to those presented in Lemma 2.3?
  7. Could you elaborate a bit more on the term “symmetry” in line 128? What does it signify in this context? Also, in line 130, clarify “the mathematical object the sequence captures…”
  8. Line 286: Change “the measure of the new distribution” to “the new distribution.”
评论

We thank the reviewer for the comments and suggestions.

"In Theorem 1.2, the distribution of zz does not incorporate the privacy parameter".

Thanks for catching this typo. We missed the privacy parameters here. The correct distribution should be zN(0,R122σ2Ik×k)z\in \mathcal{N}(0, \lVert R \rVert_{1\rightarrow 2}^2 \cdot \sigma^2 \cdot \mathbb{I}_{k\times k}) where σ2=O(log(1/δ)ε2)\sigma^2 = O(\frac{\log(1/\delta)}{\varepsilon^2}).

"Why is Theorem 1.3 included in the introduction?"

Theorem 1.3 is an easy-to-use lower bound. Since to apply Theorem 1.3, one only needs to compute the Schatten-1 norm (i.e., the sum of singular values) of the workload matrix, which is much easier than computing the γ(p)\gamma_{(p)} norm in the main theorem. Also, Theorem 1.3 is enough to derive the tight and explicit lower bound for prefix sum. We will make this point more clear.

"why doesn’t the approach of Nikolov & Tang explain the additive noise mechanism?"

For additive noise mechanisms, bounding the pp\ell_p^p error is equivalent to giving a lower bound of E[v1p++vnp]\mathbb{E}[|v_1|^p + \cdots +|v_n|^p] where {vi}(i[n])\{v_i\} (i\in [n]) are noises. Suppose the algorithm adds identically distributed noise; by the linearity of expectation, we are able to characterize the pp-th moment E[vip]\mathbb{E}[|v_i|^p] of the noise distribution. It is well-known that the pp-th moment provides substantial information about the distribution. We will make this point more clear.

"Would it be challenging to derive a lower bound for the biased case by reducing it to the unbiased one, using techniques similar to those presented in Lemma 2.3?"

We believe Lemma 2.3 is originally intended for making such reductions. We certainly do not rule out other possible ways for reduction, but we would like to emphasize that a black-box reduction from biased mechanism to unbiased mechanism in [NT24] does not give our result, as we use different privacy notion, which implies different choices of the sensitivity polytope (see also the discussions in Appendix B.3).

"Could you elaborate a bit more on the term “symmetry” in line 128? What does it signify in this context? Also, in line 130, clarify the mathematical object the sequence captures…"

By symmetry, here we mean that all the powers and roots taken on the error vector is the same. That is, it is of the form (v1p++vnp)(|v_1|_p+ \cdots + |v_n|^p) or (v1p++vnp)1/p(|v_1|_p+ \cdots + |v_n|^p)^{1/p} instead of one being pp and the other being 22.

When we consider p\ell_p norm, then it nicely interpolates between two popular norms studied in the literature, the Euclidean norm and max-norm. This is aligned with the Riesz's motivation to study the p\ell_p norm (also see footnote 1). For a given vector (or its distribution), computing its p\ell_p norm forms a sequence that converges to \ell_\infty at one end and 2\ell_2 at the other. The behavior of this sequence is better understood than the metric considered in Nikolov-Tang, which also has the same limits at the either end, but behavior in the intermediate point is not very clear.

审稿意见
8

This paper studies the so-called "factorization mechanism" for answering linear queries privately. To recap, a workload of nn linear queries over a domain of size mm can be described by a matrix A:[±1]n×mA: [\pm 1]^{n\times m}. The private query releasing algorithm asks to estimate AxAx while preventing the privacy of xx w.r.t. any adjacent xx' s.t. xx1<=1\|x-x'\|_1 <= 1.

The standard approach is to publish Ax+N(0,sigma2)Ax + N(0, sigma^2) where the Gaussian noise is calibrated to the sensitivity of AA. The factorization mechanism tries to improve over this by writing A=LRA = L R and publish L(Rx+N(0,(σ)2))L ( Rx + N(0, (\sigma')^2)) where σ\sigma' is calibrated to the sensitivity of RR only.

There is a lot of work studying the power and limitations of the factorization mechanism, trying to argue it is the "optimal" mechanism for this task. Generally, the theme along this line of research is to fix an arbitrary algorithm A\mathcal{A} (not necessarily following the paradigm above), and let B\mathcal{B} be the optimal factorization mechanism (following the approach above with an optimal L,RL,R). Then, argue that for some error metric errerr, one has err(A)err(B))err(\mathcal{A}) \preceq err(\mathcal{B})) up to insignificant factors. The optimality of the factorization mechanism has been established for 22\ell_2^2-error metric, the p2\ell^2_p metric (be a recent work [Nikolov-Tang'24].

The current paper is a close follow-up of [Nikolov-Tang'24]. It proved that the factorization mechanism is optimal under the pp\ell^p_p metric, defined as

err=E[A(x)Axpp]err = \mathbb{E}[ \| \mathcal{A}(x) - Ax \|_p^p ]

Namely, this is the pp-th moment of the ``error vector''. Some motivations are provided, for one, as pp increases from 22 to \infty, this error smoothly interpolates between the mean-squared error and the worst-case. Second, this paper argues that the error metric is more natural than p2\ell_p^2 norm considered in prior work.

The lower bound on error, implied by the proof, depends on a mysterious factorization norm and is hard to interpret. This paper gives some applications of their general lower bound to specific query families of interest. These include prefix-sum queries and sparse parity queries (or some might prefer to call it kk-way marginals)

优点

  • I like the motivation of interpolating between 2\ell_2 and \ell_\infty error via pp\ell_p^p.
  • The introduction of the paper is largely clear and easy to follow (perhaps with some background knowledge).

缺点

  • Although the gentle introduction is clear, once it goes into a more technical part, it becomes significantly harder for me to appreciate the discussion. I take Lines 254-255 "The main technical obstacle of Theorem 2.1 lies in ......" as the punchline of this paper. However, I have little idea what this means in terms of technical novelty. Given my limited time reading, I hope the authors can clarify some of my questions below.

  • Some discussions in the introduction may be improved (see some of my questions below).

问题

  • Line 111: you argued that pp\ell_p^p is more natural. I somewhat agree with this point from a notational aesthetic point. However, can you comment on why the Nikolov-Tang paper was working with an "unnatural" measure of p2\ell_p^2? This might make your claim more convincing.

  • Pargraphy beginning on Line 508: I struggled to understand why this paragraph means the Nikolov-Tang result is "instance optimal". Then I looked into their paper and found that their definition of instance optimality was somewhat nuanced, and your description was indeed accurate (but not super informative, IMO). I acknowledge it might be hard to convey the result of Nikolov-Tang concisely. But it might be worth rephrasing your conclusion a little bit.

  • Technical question: I think the current Section 2 is quite dense, and you seem to be citing Nikolov-Tang heavily. Could you maybe try to distill some of your technical punchlines and pin down the exact manipulation where you differ from prior works? I would be ok if you defer a "complete proof sketch" to appendix or so.

伦理问题详情

N/A

评论

We thank the reviewer for the comments and suggestions.

"can you comment on why the Nikolov-Tang paper was working with an "unnatural" measure of 2p\ell_2^p norm?"

From a technical perspective, [NT24] establishes a lower bound by first lower bounding the minimum variance of noise in the Euclidean geometry required to ensure privacy, where the p2\ell_p^2 lower bound is more straightforward to establish. We extend this analysis to the p\ell_p geometry. Through personal communication, the authors of [NT24] confirmed to us that they did not consider the pp\ell_p^p metric considered in this paper (also see the footnote in page 2).

"I acknowledge it might be hard to convey the result of Nikolov-Tang concisely. But it might be worth rephrasing your conclusion a little bit."

We would love to discuss the difference between instance optimality and worst-case optimality more formally in the conclusion part.

"Could you maybe try to distill some of your technical punchlines and pin down the exact manipulation where you differ from prior works?"

Due to space constraints, we had to defer the detail discussion of the difference in our techniques and previous works to Appendix B.3. For the reviewer’s convenience, we also provide a high-level summary here. Please let us know if further clarification is needed.

  1. A different privacy notion necessitates the choice of a different sensitivity polytope to exactly capture the geometry of error, and to relate to the factorization norm of the workload matrix itself.

We first note that [NT24] does not imply a lower bound for the usual setting of linear queries due to different privacy notions. We have confirmed with the authors of Nikolov-Tang that if using the same privacy notion as in our setting, a lot of results in their paper need to be modified. To elaborate, [NT24] gives a bound for mean estimation in the substitution model of privacy. Using the histogram notation, let hRXh \in \mathbb{R}^{|\mathcal{X}|} (where X=n|\mathcal{X}| = n in the notation of our paper) be the histogram of the dataset, then the substitution DP model corresponds to the neighboring notion where hh12\lVert h - h'\rVert_1 \leq 2 and xX(hxhx)=0\sum_{x \in \mathcal{X}} (h_x - h’_x) = 0.

In contrast, our privacy notion for linear queries requires hh1=O(1)\lVert h - h'\rVert_1 = O(1), representing a natural 1\ell_1 sensitivity, the same as [BDKT12], [ENU20] etc. Now answering linear queries, QQ over data X=(x1,x2,,xk)X = (x_1, x_2, \cdots, x_k), Q(X)Q(X) is equivalent to doing mean estimation in the polytope KQ={Q(x):xX}K_Q = \{Q(x):x\in \mathcal{X}\}. Even if we restrict our discussion to the substitution model, the lower bound for mean estimation from Nikolov and Tang would be in terms of Γp(KQ)\Gamma_p(K_Q). In contrast, we get a bound in terms of the factorization norms of the workload matrix AA associated with the query QQ. These two are not comparable: imagine when the workload matrix consists of vectors that are very far from the origin but are close to each other; then we have small Γp(KQ)\Gamma_p(K_Q) but a substantially larger factorization norm of the workload matrix AA (we are happy to include an example in our paper as well). Therefore, a lower bound in terms of Γp(KQ)\Gamma_p(K_Q) does not imply a lower bound in terms of the factorization norm γ(p)(A)\gamma_{(p)}(A) as ours. Essentially, this is due to the fact that the most suitable choice for "sensitivity polytope" is different for mean estimation and linear queries. This is especially important since we would like to give explicit lower bounds for explicit linear query families, and tight bounds are only possible with the right choice for "sensitivity polytope".

  1. Removing unbiasedness assumption via the Bhaskar et al. reduction.

Then, to remove the unbiasedness assumption, we chose a different path compared to [NT24] by first establishing a lower bound against data-oblivious mechanisms, where the ideas on handling the bias is summarized in Appendix B.3 (starting at line 887). We then follow the reduction of Bhaskar et al. to obtain lower bounds for general mechanisms.

Finally, a side benefit of our approach is that the width of the sensitivity polytope along the narrowest direction, denoted by κ\kappa in our paper and w0w_0 in [NT24], plays a different role in our lower bound than theirs. Unlike [NT24], our lower bound does not depend on κ\kappa and it is crucial for our explicit lower bounds for the two classes of linear queries we discuss later.

评论

Thank you for your response. That helped clarify my questions a lot. I have raised my score and encourage you to incorporate the discussions into the revision of your paper.

评论

Thank you! We appreciate your helpful comments and will certainly incorporate them into the next version of our paper.

AC 元评审

Summary of Contributions

This paper studies linear queries problem under differential privacy. The problem can be thought of as outputting an estimate of AxAx where AA is the workload matrix and xx is the histogram of the input dataset. Traditionally, most works have studied this under the 22\ell_2^2 and \ell_\infty errors. Recently, Nikolov and Tang (ITCS, 2024) studied the problem under p2\ell_p^2 error for any p2p \geq 2 and show that the so-called matrix mechanism (aka factorization mechanism) yields a nearly optimal error among all unbiased mechanisms. This paper shows a similar result but for an arguably more natural pp\ell_p^p error and without the unbiased assumption.

Strengths

  • Linear queries are one of the most important and well-studied setting in private data analysis. This paper advances our understanding on the problem.

  • pp\ell_p^p errors are natural extensions of 22\ell_2^2 and \ell_\infty errors, and are more natural than the p2\ell_p^2 error studied in (Nikolov & Tang, 2024).

  • The proofs are sophisticated and require clever use of existing techniques.

Weaknesses

  • It is not entirely clear how important pp\ell_p^p errors are since, for most previous applications, 22\ell_2^2 and \ell_\infty seem sufficient.

  • The lower bounds in this paper are for worst-case errors whereas (Nikolov & Tang, 2024) gives an "instance-optimality" result, which is a stronger type of lower bound. (This is not a strong point since all work prior to (Nikolov & Tang, 2024) also studied worst-case errors.)

Recommendation

Given the importance of linear queries and that this paper gets nearly-tight bounds for pp\ell_p^p errors, we recommend acceptance.

审稿人讨论附加意见

The reviewers asked for a few clarifications, including the adjacency notions used in the paper (add-remove DP) which is different compared to the one used in Nikolov-Tang (substitution DP) and why the techniques from the former do not apply to pp\ell_p^p. The authors provide satisfactory answers that clarify these confusions.

最终决定

Accept (Poster)