PaperHub
8.2
/10
Poster5 位审稿人
最低5最高5标准差0.0
5
5
5
5
5
3.4
置信度
创新性3.6
质量3.8
清晰度3.6
重要性3.2
NeurIPS 2025

Private Geometric Median in Nearly-Linear Time

OpenReviewPDF
提交: 2025-05-08更新: 2025-10-29
TL;DR

We give a nearly-linear time algorithm for privately computing the geometric median, addressing the main open question of Haghifam et al (NeurIPS '24).

摘要

Estimating the geometric median of a dataset is a robust counterpart to mean estimation, and is a fundamental problem in computational geometry. Recently, [HSU24] gave an $(\epsilon, \delta)$-differentially private algorithm obtaining an $\alpha$-multiplicative approximation to the geometric median objective, $\frac 1 n \sum_{i \in [n]} \|\cdot - \mathbf{x}_i\|$, given a dataset $D$ of $x_i$ for $i \in [n]$. Their algorithm requires $n \gtrsim \sqrt d \cdot \frac 1 {\alpha\epsilon}$ samples, which they prove is information-theoretically optimal. This result is surprising because its error scales with the effective radius of $D$ (i.e., of a ball capturing most points), rather than the worst-case radius. We give an improved algorithm that obtains the same approximation quality, also using $n \gtrsim \sqrt d \cdot \frac 1 {\alpha\epsilon}$ samples, but in time $\widetilde{O}(nd + \frac d {\alpha^2})$. Our runtime is nearly-linear, plus the cost of the cheapest non-private first-order method due to [CLMPS16]. To achieve our results, we use subsampling and geometric aggregation tools inspired by FriendlyCore [TCKMS22] to speed up the "warm start" component of the [HSU24] algorithm, combined with a careful custom analysis of DP-SGD's sensitivity for the geometric median objective.
关键词
Differential PrivacyGeometric MedianStochastic Convex Optimization

评审与讨论

审稿意见
5

This paper considers the problem of privately estimating the geometric median, a topic that has been investigated previously and an information-theoretically optimal sample complexity is known. In this work, the authors propose an algorithm that achieves the same optimal sample complexity while running in a nearly linear time O~(nd+d/α2)\tilde{O}(nd + d/\alpha^2) for nn input data points of dimension dd, significantly improving the previous result of [HSU24]. The workflow closely follows the framework established by [HSU24], incorporating several novel techniques that enhance runtime efficiency.

[HSU24] Mahdi Haghifam, Thomas Steinke, and Jonathan Ullman. Private geometric median. Advances in Neural Information Processing Systems, 37:46254–46293, 2024.

优缺点分析

Strengths:

  1. The estimation of the geometric median is a fundamental problem in computational geometry. This work presents a private algorithm that operates in nearly linear time, marking a substantial advancement over previous results.
  2. The desired runtime is achieved by making several improvements to different parts of [HSU24]'s framework, which are novel and technically solid.
  3. Comprehensive experiments are conducted to show the effectiveness of the proposed techniques.

Weakness:

  1. The algorithm requires prior knowledge of parameters rr and RR, which provides lower and upper bounds on the effective radius r(0.9)r^{(0.9)}.
  2. In Section 5 the authors evaluate only Algorithm 1 and 3. The presented analysis of Algorithm 2 incurs large constants that largely affect its performance in practice. Therefore, the entire algorithm is not ready for real-world deployment.

问题

  1. Is it possible to remove the requirement of knowledge rr and RR ? As mentioned in the paper, the resulting sample complexity is independent of these parameters. I wonder if this dependence in the time complexity bound can be removed as well.
  2. This work focuses on approximate DP, where a small probability δ\delta of privacy violation is allowed. Is it able to achieve the more stringent notion of pure DP?
  3. The runtime of the private algorithm is O~(nd+d/α2)\tilde{O}(nd + d/\alpha^2) while one of the non-private algorithms in [CLM+16] runs in O(d/α2)O(d/\alpha^2). Is it possible to develop a private algorithm with this time complexity?

Some (possible) typos:

  1. [Line 110] "the FriendlyCore algorithm of [HSU24]": This seems to be a wrong citation.
  2. In the proof of Lemma 5 [Line 520], the setting of μ\mu and ϵ\epsilon may only lead to a probability bound of exp(kn1n/3)\exp(-k\cdot \frac{n-1}{n} /3), which is slightly looser than the claimed δ/4T\delta / 4T.
  3. [Line 537] Lemma 6 should be referenced as Lemma 2.
  4. [Line 563] fi0.7kf_i^\star\ge 0.7 k should be corrected to fi0.75kf_i^\star \ge 0.75 k
  5. [Line 618] what is meant by "large enough CC"?
  6. [Line 619] "we run Algorithm 2 with this value of r^\hat{r}." However, Lemma 7 requires r^r(0.75)\hat{r} \ge r^{(0.75)} while Lemma 6 only guarantees r^14r(0.75)\hat{r}\ge \frac{1}{4}r^{(0.75)}. I think this should be corrected by passing 4r^4\hat{r} to Algorithm 22, and some related quantities need to be modified accordingly.
  7. [Line 623] Lemma 6 should be referenced as Lemma 7.

[CLM+16] Michael B Cohen, Yin Tat Lee, Gary Miller, Jakub Pachocki, and Aaron Sidford. Geometric median in nearly linear time. In Proceedings of the forty-eighth annual ACM symposium on Theory of Computing, pages 9–21, 2016.

局限性

yes

最终评判理由

The rebuttal has addressed my concerns. I would like to keep my positive score.

格式问题

N/A

作者回复

Thank for your encouraging words about our contribution and runtime efficiency. We will fix all typographical errors pointed out by the reviewer in the revised manuscript and address key questions below:

Knowledge of rr and RR: Thank you for the interesting question. The reason for having these constants in the algorithm are due to application of the AboveThreshold technique, a standard method in DP, in Algorithm 1 which requires a finite set of queries and we are currently not aware of extensions to this method which can handle potentially unbounded range of queries. Moreover, [BNSV15] shows that some knowledge of the parameter rr is necessary (as observed by prior work [HSU24]), although it is an interesting open question to improve our dependence.

On Pure DP: Currently, the subsampling based approach used in Algorithm 1 and 2 requires a failure probability, which is critical for the computational speed-up. Furthermore, the tools used for private stochastic convex optimization in [FKT20], which inspire the analysis of Algorithm 3 in our work also currently provide an approximate DP guarantee. Therefore, generalizing both of these techniques to handle pure DP remains an interesting direction for future work. We would like to point out that the dependence of both the runtime and the sample complexity on 1/δ1/\delta is polylogarithmic, and therefore allows for δ\delta being polynomially small in 1/n1/n.

On runtime comparison with [CLM+16]: A private algorithm cannot run in time O~(d/α2)\widetilde{O}(d/\alpha^2), given that this runtime is independent of the privacy parameter eps, which shows up polynomially in the sample complexity. We do think it is interesting to remove this additive runtime term (and obtain an O~(nd)\widetilde{O}(nd) runtime, the other runtime in [CLM+16]), although this too seems quite challenging, as existing nearly-linear time methods for DP-ERM treat the objective as a stochastic optimization problem, and thus incur the runtime overhead of (non-private) SGD.

On constants in Algorithm 2: We acknowledge in lines 289-293 that the constants in Algorithm 2 can potentially be further optimized as part of future work.

[BNSV15] M. Bun, K. Nissim, U. Stemmer, and S. Vadhan. “Differentially private release and learning of threshold functions”. In: 2015 IEEE 56th Annual Symposium on Foundations of Computer Science. IEEE. 2015, pp. 634–649.

评论

Thank you for your response.

审稿意见
5

The task is to compute geometric median of a point set in a differentially private way. The geometric median is the minimizer of ixi\sum_i ||\cdot - x_i || for Euclidean norm, and this paper is happy with approximate minimizer achieving 1+α1+\alpha multiple of the minimum.

A recent work (HSU24) has provided an algorithm solving the task with a certain sample complexity and proved that it is the optimal complexity. However, there was a significant gap in the computational efficiency between private and non-private geometric median computation in the prior work.

This work provides an algorithm similar in spirit to the one of (HSU24) that runs in a very similar time to the non-private algorithms. Per my understanding, the differences in the algorithms are subtle (but necessary for the improved runtime), but the big picture stays the same.

The algorithm runs in three phases, first, an effective radius r (i.e., the majority of points are contained in a ball of this radius) of the input is computed, then, a crude estimate of the geometric median (in a distance at most r) is computed. Finally, a custom DP-SGD analysis is performed leveraging the structure of geometric median problem.

优缺点分析

The paper is easy to read and I like the way how is it written, that it provides the intuition and sketches results before diving into the technicalities. I would only suggest to introduce the r(0.9)r^{(0.9)} notation more explicitly.

I find this to be a valuable contribution since (private) geometric median problem is quite important and the improvements in runtime are significant. The DP-SGD analysis technique used in the paper taking advantage of the fact that the gradients are unit vectors can potentially also be used elsewher.

On the other hand, the scope of the paper is very narrow and the improvements are only in the run time and that is why I do not lean to the highest score.

问题

局限性

最终评判理由

I still think the paper is good and should be accepted.

格式问题

作者回复

Thank you for the positive assessment. We will insert an explicit definition of the “r(0.9)r^{(0.9)}” notation where it first appears and polish notation throughout.

评论

Thank you for your response.

审稿意见
5

This paper proposes a DP optimization algorithm to solve the following task: given (x1,...,xn) in (R^d)^n, such that the data points have a norm of R, we aim to solve x* = argmin_x sum_i ||x - xi||. This problem is known as private geometric median. The previous work [Haghifam-Steinke-Ullman'24] proposes an algorithm to provide (1+alpha) multiplicative error with the sample complexity of sqrt(d)/(alpha*epsilon). However, the runtime of their proposed algorithms is either quadratic in n or quadratic in d. The main result of this paper is to push the time complexity to nd + d/alpha^2. Notice that d/alpha^2 is the optimal time complexity of the first-order method even non-privately. The idea behind the proposed algorithm is as follows: it has three stages.

1- In the first stage, an algorithm similar to the algorithm proposed in [Haghifam-Steinke-Ullman'24] (these algorithms are inspired by GoodRadius of [Stemmer-Nissim-Vadhan'16]). The goal is to find a radius privately around the geometric median that contains 70% of the data points. The main idea is instead of computing all the n^2 pairwise distances is sub-sampling that makes the algorithm run in linear time.

2- The second stage is inspired by the FriendlyCore of Tsfadia et al. and it consists of finding a center with the promise that there exists a radius that contains 70% of the data points. The new idea is that FriendlyCore can be sped up to run in O~(nd) time (independently of R/r) using subsampled weights.

3- Finally, the third stage, which is very interesting, is a bespoke analysis of the DP-SGD for the geometric median loss function. The idea here is to "warm-start" DP-SGD from the centerpoint given by the output of Stage 2 and the radius given by Stage 1. The main idea is to show that when gradient descent steps are contractive for an appropriate step size, we can use the reduction in [Koren-Feldman-Talwar'20] to convert the contractivity to a private algorithm.

优缺点分析

The result of this paper is interesting: it combines lots of interesting techniques to provide a very intuitive algorithm that improves the prior work's runtime.

There are a few weaknesses:

  • If we view this problem from the standpoint of DP optimization, it would be interesting to find a broader class of loss functions for which we can obtain an algorithm with multiplicative guarantee.

  • The optimal dependence of the sample complexity on R is unclear. Note that in one dimension there are algorithms that require log*(R) samples.

  • The problem of geometric median is very related to the problem of privately finding a point in the convex hull. A helpful comparison for the community is to compare the problem of private geometric median with the problem of privately finding a point in the convex hull, for instance, see the following paper: Kaplan, H., Mansour, Y., Moran, S., Stemmer, U., & Tur, N. (2025, June). On differentially private linear algebra. In Proceedings of the 57th Annual ACM Symposium on Theory of Computing (pp. 2362-2373).

Line 110 - the citation is wrong for the FriendlyCore algorithm.

问题

please see above the weakness part

局限性

yes

最终评判理由

I think this paper merits acceptance.

格式问题

no

作者回复

We thank the reviewer for their positive words about our techniques and improvement in runtime. We will fix the citation in Line 110 to replace it with [TCK+21]; great catch.

Re: broader class with multiplicative guarantees: We agree that this is a natural and important follow-up direction, and will mention this as an open problem in the revision. Some natural candidates are problems for which there exist non-private first-order methods with multiplicative guarantees, such as other GLMs (e.g., based on Huber loss) and approximate positive LPs.

Re: dependence on RR: We agree that it is unclear whether our dependence on RR is optimal, and that this is an interesting open question. We do note that many of the techniques used in prior work appear specific to the 1-dimensional setting, and that extending these tools to arbitrary dimension would likely come at a significant cost to dependences on other parameters.

Re: privately finding point in convex hull: This is an interesting point of comparison. Exactly computing the geometric median privately is certainly a more challenging problem than the private convex hull problem; on the other hand, our paper only aims to solve an approximate version. Current techniques for private convex hull appear substantially more complex than for approximate private geometric median; as noted by [KMM+25], private convex hull can be very unstable as a function of the dataset. On the other hand, because our goal is only to approximate the optimum multiplicatively (which scales at least linearly with r(α)r^{(\alpha)} for any constant alpha > 0), it cannot be affected substantially by moving a single point, so our problem is potentially much simpler. We will mention this connection in a revision, and find it interesting to further explore the relationship between these two problems.

审稿意见
5

This paper presents a novel and practically relevant algorithm for computing a differentially private geometric median in high dimensions, achieving nearly-linear runtime. The authors build upon the FriendlyCore warm-start technique and integrate a refined DP-SGD procedure with a smoothed subgradient to overcome the non-smoothness of the geometric median objective. Their proposed algorithm achieves:

  • Near-optimal sample complexity
  • Strong error guarantees comparable to prior work
  • A significant improvement in runtime from prior quadratic-time algorithms to near-linear. Empirical evaluation demonstrates a 30 times speedup on synthetic and real data benchmarks, without significant loss of accuracy, and with robust error performance across multiple trials.

优缺点分析

The geometric median problem is a fundamental problem in many domains including clustering, robust statistics and social choice. This paper makes a technically sound and valuable contribution offering a novel and efficient . The combination of a warm-start approximation with a customly analyzed DP-SGD routine is elegant and well-justified, addressing the challenge of non-smooth optimization under differential privacy.

The authors do a rigorous analysis using mathematical proofs but also give an efficient implementation and empirical evaluation of their method. The writing is clear and well-organized, with helpful figures (notably Figures 1 and 2) to illustrate empirical gains.

问题

The algorithm depends on the parameters r, R? The upper bound is more justified as there may be a natural bound on the points or one could normalize them. However, r is less justified. How can one compute it and what happens if it is not correct? Could you obtain an error with an additive term relative to r?

  • Lemma 9 seems strange. How does n >= 20 relate to (eps,delta)-dp? Seems like the condition is artificial and It would be nice to remove it.

  • I am never a fan of misusing the O tilde notation. Theorem 1 writes O tilde to hide all polylogarithmic factors (even for unrelated variables) but it should only be used to mean hiding polylogarithmic factors of the quantity inside the O tilde. It would be nice to fix the theorems appropriately rather than adding a footnote.

局限性

Yes, the authors describe the potential suboptimality of the runtime as a limitation tied to using first order methods for optimization. Alternative methods could potentially have better dependency.

格式问题

No concerns

作者回复

We thank the reviewer for their encouraging words regarding our contribution and rigorous analysis.

Need for parameters r,Rr, R : As the reviewer mentions, RR can be estimated by a norm bound on the input datapoints. Regarding the need for estimating r, as noted in the prior work [HSU24], knowledge of such an a priori lower bound is necessary by the impossibility results in [BNSV15]. We note that our dependence on rr is relatively mild (i.e., essentially a single logarithmic factor in the runtime and a single loglog factor in the sample complexity). It is an interesting open question to improve the dependence on R/rR/r, although we find it potentially challenging to do so while retaining the nearly-linear runtime.

Lemma 9 and n20n \geq 20: The n20n \geq 20 is due to the particular constants used in our application of Chernoff bounds in Lemma 8 and we will optimize it in the revised manuscript.

O~\widetilde{O} notation: We will modify O~\widetilde{O} to only include polylogarithmic factors of the quantity inside it, as our use may be confusing re: the dependence on R/rR/r. We do note that we are not aware of a uniform convention in the usage of O~\widetilde{O}, with many related works using a similar definition to our paper. Also, we remark this notation is only used in Section 1; statements in all other sections fully quantify dependences on all parameters.

[BNSV15] M. Bun, K. Nissim, U. Stemmer, and S. Vadhan. “Differentially private release and learning of threshold functions”. In: 2015 IEEE 56th Annual Symposium on Foundations of Computer Science. IEEE. 2015, pp. 634–649.

评论

Thank you for your response. Indeed, making the dependencies more explicit would be nice and make it easier for follow-up work to use and extend the resutls.

审稿意见
5

This paper presents a new efficient algorithm for computing in a differentially private way the geometric median of a set of data. Recent work has proposed two-phases (warm-up and boosting) algorithms for a multiplicative approximation of the geometric median with optimal sample complexity but running in polynomial time. The algorithm proposed in this paper achieve similar accuracy and sample complexity but runs in almost linear time (for some range of the parameters).

优缺点分析

Strengths

  • The paper address an important problem: improving the runtime of algorithms for differentially private geometric median with good average error rates.

  • The paper improves the runtime of the warm up phase by using subsampling in a

  • In the boosting phase the paper gives an analysis of DPSGD leveraging the properties of the geometric median to obtain an improved runtime.

Weaknesses

  • The proposed algorithm uses as a subroutine an adaptation of the FriendlyCore procedure from [TCK+22] but it is unclear why this departure from the original algorithm is needed.

Recent work proposed differentially private algorithms for the geometric median with average error guarantees but those algorithm have a polynomial runtime, which can be problematic in large data settings. It is thus important to have more efficient algorithms and this paper does exactly that.The proposed algorithm uses a two-phases approach, based on warm-up and boosting, as the ones in previous work but it optimizes the running time of each step.

To optimize the runtime of the warm-up, which has itself two phases, one for estimating the radius and one for estimating the centerpoints, the paper uses subsampling. In the first phase the algorithm subsample points to estimate the radius. This change is relatively straightforward. In the second phase, the proposed algorithm departs from previous work and uses as a subcomponent the FriendlyCore procedure from [TCK+22]. In orde to improve the efficiency of this procedure the algorithm uses subsample in the computation of the weights associated with each data point. This change is non-trivial and needs an interesting change in the privacy proof. However, it is unclear why this departure from the original algorithm is needed since the obtained improvement is quite limited. To optimize the runtime of the boosting phase the paper proposes an interesting ad-hoc analysis of DP-SGD specific to the geometric median. Specifically, the paper uses the properties of the geometric median to control the sensitivity of DP-SGD and obtain an optimization in nearly-linear time. Overall, the proposed algorithm shows considerable improvement thanks to non-trivial changes to the algorithms from previous work.

A couple of minor points:

  • in equation (2) the notation for the indicator function is not yet defined
  • line 110 has a wrong citation

问题

Can you clarify why you use the FriendlyCore procedure for estimating the centerpoints? What do you gain?

局限性

yes

格式问题

none

作者回复

We thank the reviewer for their kind words regarding the importance of the problem considered and the improved runtime. We will add the definition of the indicator in Eq. (2) and fix the citation in Line 110 to replace it with [TCK+21] – thank you for bringing up these comments.

We agree that there is no significant qualitative change in our runtime vs. that of [HSU24] in the centerpoint estimation phase. We stated as such in Section 1.2, but can further clarify that our motivation for including our alternative approach was to highlight the simplicity of directly adapting FriendlyCore. Indeed, our analysis of Algorithm 2 essentially directly extends [TCK+22] to tolerate inaccuracies due to subsampling. Our approach thus

(1) gives an arguably much simpler approach to centerpoint estimation (prior work [HSU24] required a custom analysis of DP-GD with geometrically decaying step sizes), and

(2) makes the observation that FriendlyCore can be sped up via subsampling, which to our knowledge has not been explicitly noted before. We do not view this section as a major technical contribution of our paper.

评论

Thank you for your response. integrating part of this response in the paper could help clarity.

最终决定

This paper presents a new, efficient algorithm for computing the geometric median of a dataset in a differentially private manner. The proposed algorithm achieves a near-linear runtime, a significant improvement over prior polynomial-time works, while maintaining similar accuracy and sample complexity. The reviewers agree that the paper addresses an important problem and that the results are significant. The improved time complexity is a strong and clear technical contribution. Given the unanimous support and the paper's strengths, my recommendation is to ACCEPT.