PaperHub
6.4
/10
Poster4 位审稿人
最低3最高5标准差0.7
4
3
4
5
3.3
置信度
创新性2.8
质量3.0
清晰度3.5
重要性2.8
NeurIPS 2025

Simple and Optimal Sublinear Algorithms for Mean Estimation

OpenReviewPDF
提交: 2025-05-10更新: 2025-10-29
TL;DR

We present algorithms with optimal sample complexity and running time for estimating a high dimensional Euclidean mean.

摘要

We study the sublinear multivariate mean estimation problem in $d$-dimensional Euclidean space. Specifically, we aim to find the mean $\mu$ of a ground point set $A$, which minimizes the sum of squared Euclidean distances of the points in $A$ to $\mu$. We first show that a multiplicative $(1+\varepsilon)$ approximation to $\mu$ can be found with probability $1-\delta$ using $O(\varepsilon^{-1}\log \delta^{-1})$ many independent uniform random samples, and provide a matching lower bound. Furthermore, we give two estimators with optimal sample complexity that can be computed in optimal running time for extracting a suitable approximate mean: 1. The coordinate-wise median of $\log \delta^{-1}$ sample means of sample size $\varepsilon^{-1}$. As a corollary, we also show improved convergence rates for this estimator for estimating means of multivariate distributions. 2. The geometric median of $\log \delta^{-1}$ sample means of sample size $\varepsilon^{-1}$. To compute a solution efficiently, we design a novel and simple gradient descent algorithm that is significantly faster for our specific setting than all other known algorithms for computing geometric medians. In addition, we propose an order statistics approach that is empirically competitive with these algorithms, has an optimal sample complexity and matches the running time up to lower order terms. We finally provide an extensive experimental evaluation among several estimators which concludes that the geometric-median-of-means-based approach is typically the most competitive in practice.
关键词
Sublinear AlgorithmsMean EstimationLearning

评审与讨论

审稿意见
4

The paper discusses sublinear mean estimation and presents three different algorithms. For each of these algorithms, running time analysis and optimality discussions are provided. Several experiments comparing proposed algorithms with existing algorithms are provided.

优缺点分析

  1. Perhaps most importantly, the new proposed algorithms are optimal in many aspects such as scalability and efficiency, which is arguably the highlight of the paper. There are some follow up questions regarding proof details which are deferred to the "Questions" section of the review, but if shown to be correct, the results are fairly interesting.
  2. Overall, the paper is well-organized and discusses background, key concepts, and contributions well. It is clear that the authors spent a lot of effort polishing the paper. Several suggestions regarding typos or formatting issues are deferred to a later section.
  3. The experiments are relatively weak. For instance, it is clear that the empirical mean consistently outperforms proposed algorithms in terms of accuracy. While the empirical mean is not included in runtime comparisons, it should be significantly faster than the proposed algorithms, if I understand correctly. Perhaps some additional experiments or discussion would be helpful to address these observations.
  4. Following point 3., I think the result of Theorem 3.2 regarding learning the mean of high-dimensional distributions is also quite interesting but there doesn't appear to be any experiments that support the dimension-free result. Might be helpful if some extra experiments or discussions can be made there.

问题

Majority of my questions are about the proofs and experiments.

Q1. In line 143, it is stated that earlier values of constants were set as a=1440,b=50a = 1440, b = 50, but neither of these appear to be discussed in detail prior to this line. Is some text missing or can some more context be shared as to why these constants were chosen?

Q2. Overall I found the proof of Lemma 2.3 to be quite vague and would prefer it to be explained in more detail, particularly because it is one of the cornerstone results of the paper. For instance, why do we get μ^iμ2>r2\|\hat{\mu}_i-\mu\|^2 > r^2 with probability at most 0.1, and where does this 0.1 come from? How does 2/81-2/81 appear in line 145-146?

Q3. In line 160 it is stated that "Lk,RkL_k, R_k have cardinality at least blogδ1/2b\log \delta^{-1}/2." What does this mean when Lk+Rk=blogδ1L_k + R_k = b\log \delta^{-1}? Following this logic, can you explain in a bit more detail how we can infer at least 1/51/5 of the samples satisfy μ^i,kμkνkμk|\hat{\mu}_{i,k}-\mu_k|\geq |\nu_k - \mu_k| from this?

Q4. Following Q3, can you clarify the logic in lines 160-161? In particularly, since Lk,RkL_k, R_k are sets, the sentence "at least for one of Lk,RkL_k,R_k, we have μ^i,kμkνkμk|\hat{\mu}_{i,k}-\mu_k|\geq |\nu_k - \mu_k|" doesn't make a lot of sense.

Q5. Following Q4., because of Q4 it's quite difficult to judge if the sentence "Depending on whether νk>μk\nu_k>\mu_k or not, at least for one of Lk,RkL_k,R_k, we have that μ^i,kμkνkμk|\hat{\mu}_{i,k}-\mu_k|\geq |\nu_k - \mu_k|." is logical. Wonder if you can help clarify it with some examples, for example say if μk=0,μ^k=2,νk=3\mu_k = 0, \hat{\mu}_k = 2, \nu_k = 3. In this case, the left hand side is 2 while the right hand side is 3 and the inequality doesn't hold.

Q6. As briefly mentioned in an earlier section of the review, it appears from the experiments that the empirical mean performs consistently better than proposed algorithms, and is likely fast as well; can you share more insights on how this may be happening?

Q7. Wonder if there are more experiments that look at how the algorithms perform under different data dimension, say, varying between 1- 1000 instead of a fixed value like 784 or 54.

Additional comments regarding formatting and typos:

  1. Line 65: "three algorithm"
  2. Line 109: "studided"
  3. Line 135: "appoximation"
  4. Line 191: "proof for the our setting"
  5. Line 313: "runing time"
  6. Line 330: "these observation"

局限性

The authors adequately addressed limitations regarding uniform sampling and data corruption.

最终评判理由

My original concerns about details of the proofs were addressed sufficiently. Overall I think the paper has good quality and it will improve with further revisions. After engaging with other reviewers, I agree about some concerns regarding novelty. Based on this I'm suggesting a score of 4.

格式问题

N/A.

作者回复

We thank the reviewer for the helpful feedback. As per NeurIPS 2025 policy, we cannot include links or images in the rebuttal or repository. We’ve therefore provided tables (in earlier responses) with the same information as the plots and appreciate your understanding given these limitations.

Regarding proofs.

Q1 Response. We apologize for the confusion. The sentence has a typo and should instead be "We set the values of the constants to be \dots ". In the answer to the next question, please find a detailed response including motivations behind the values of the constants.

Q2 Response. For the algorithms, the first step is to sample blogδ1b \log \delta^{-1} subsamples of size aε1a \varepsilon^{-1} each and compute their means. We want our final estimate to be within μ^μεOPTn\|\hat \mu - \mu \| \leq \sqrt{\frac{\varepsilon OPT}{n}} distance of the true mean, as it implies a (1+ε)(1+\varepsilon)-approximation. Define r=111εOPTnr = \frac{1}{11} \sqrt { \frac{\varepsilon OPT}{n}} and we call a mean good if μ^μr\|\hat \mu - \mu\| \leq r. Further, we define the good event E\mathcal{E} to be when at least 7/107/10 fraction of the means are good. The value 7/107/10 is chosen to an arbitrary constant strictly greater than 1/21/2.

Lemma 2.3. Event E\mathcal{E} holds with probability at least 1δ1-\delta.

The proof contains the following steps:

  1. A sample mean derived from 1440ε11440 \varepsilon^{-1} samples is good with probability 0.90.9.
  2. If we have 50logδ150 \log \delta^{-1} subsamples of 1440ε11440 \varepsilon^{-1} points each, at least 7/107/10 fraction of the mean will be good with probability 1δ1-\delta. We now give a detailed version of the proof with an explanation of the constants as well. For now, the values of constants are set as follows - a=1440,b=50a = 1440, b = 50.

Proof. From Lemma 2.1 in the paper, we know that E[μμ^2]=1SOPTn=1alogε1OPTn \mathbb{E} [ \| \mu - \hat \mu \|^2 ]= \frac{1}{|S|} \frac{OPT}{n} = \frac{1}{a \log \varepsilon^{-1}} \frac{OPT}{n}. The Markov's inequality can be used to bound the probability that a single mean is not good - P[μμ^2>r2]r2E[μμ^2]0.1\mathbb{P}[\| \mu - \hat \mu \|^2 > r^2] \leq \frac{r^2}{\mathbb{E} [ \| \mu - \hat \mu \|^2 ]} \leq 0.1. Conversely, the probability that a single mean is good is at least 0.9. We require this probability to be strictly greater that 0.70.7 in order to be able to say that the good event E\mathcal{E} holds with high probability. Specifically, the value of aa is set to 14401440 to achieve this.

Next, we use the (multiplicative) Chernoff bound to show that if we have blogδ1b \log \delta^{-1} means, then with high probability, at least 0.7 fraction of the means are good.

The Chernoff bound says P[X<(1θ)μ]exp(θ2μ/2)\mathbb{P}[X < (1 - \theta)\mu] \leq \text{exp}(-\theta^2\mu/2), where XX is the sum of independent random variables and the expected value μ=E[X]\mu = \mathbb{E}[X]. Setting μ0.9blogδ1\mu \geq 0.9 b \log \delta^{-1}, θ=2/9\theta = 2/9, we get

P[G0.7blogδ1]exp(281blogδ1).\mathbb{P}[|G| \leq 0.7 b \log \delta^{-1}] \leq \text{exp}(-\frac{2}{81} b \log \delta^{-1}).

Thus, the constant 2/81-2/81 is the result of applying the Chernoff bound by setting the parameters appropriately. We choose the value of bb so that (2b)/81<1-(2b)/81 < -1 and the probability bound achieved is at most δ\delta. Therefore, we set b=50b=50.

We will add a more detailed version of the lemma and its proof to the final version of the paper.

Q3 Response. We have LkRk=blogδ1|L_k \cup R_k| = b \log \delta^{-1} which is the total number of sample means. Recall that LkL_k consists of the sample means such that μ^kνCWM,k\hat \mu_k \leq \nu_{CWM,k}. And RkR_k is the set of sample means such that μ^kνCWM,k\hat \mu_k \geq \nu_{CWM,k}. The key detail is that there are points (such that μ^k=νCWM,k\hat \mu_k = \nu_{CWM,k}) which belong to both. Along with the definition of the coordinate-wise median, we have Lk,Rkblogδ1/2|L_k|, |R_k| \geq b \log \delta^{-1}/2. Note that if there are no points such that μ^k=νCWM,k\hat \mu_k = \nu_{CWM,k}, then Lk=Rk=blogδ1/2|L_k|= |R_k| = b \log \delta^{-1}/2.

We provide a detailed case analysis as to why at least 1/51/5 fraction of the samples are good and satisfy μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.

First, we note that when the good event E\mathcal{E} holds, at least 1/51/5 fraction of the means are good in both the sets LkL_k and RkR_k. This follows from the fact that at least 7/107/10 fraction of the means are good and the number of good means in LkL_k/RkR_k is at least 7/101/2=1/57/10 - 1/2 = 1/5.

Case I: μkνCWM,k\mu_k \leq \nu_{CWM,k}. In this case, we have for all means in RkR_k, μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.

Case II. μk>νCWM,k\mu_k > \nu_{CWM,k}. In this case, we have for all means in LkL_k, μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.

From the above statements, we have for at least one of LkL_k and RkR_k, we have that at least 1/51/5 fraction of the means are good and satisfy μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.

Q4 Response. We apologize for the typo. We want to say that the claim is true for all means in LkL_k or RkR_k. The formal statement is as follows. At least one of the following statements is true:

  1. For all μ^\hat \mu in LkL_k, we have that μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.
  2. For all μ^\hat \mu in RkR_k, we have that μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.

We will update the final version of the paper accordingly.

Q5 Response. Please refer to our answer to Q3 for the detailed proof. In the example given by the reviewer, νCWM,k=3\nu_{CWM,k} = 3, implies that there are equal number of points with μ^k<3\hat \mu_k < 3 and μ^k>3\hat \mu_k > 3. So, if there exists a sample mean such that μ^k=2<3\hat \mu_k = 2 < 3, then there will also exist at least one mean such that μ^k>3\hat \mu_k > 3. And for that mean, which is in RkR_k, we have μ^i,kμkνCWM,kμk|\hat \mu_{i,k} - \mu_k| \ge |\nu_{CWM,k} - \mu_k|.

Regarding experiments.

Q6 Response. As the reviewer correctly notes, the empirical mean is the fastest among all mean estimators, as it can be computed in closed form. Despite this computational advantage, it is not a robust estimator of the true mean—even in one dimension. That said, the empirical mean performs well when estimating the mean of subgaussian distributions, where the associated tail bounds are sharply exponential. In contrast, for heavy-tailed distributions, the best available tail bound is Chebyshev’s inequality, which is exponentially weaker. In fact, Catoni (2012) shows that there exist distributions for which Chebyshev’s inequality is tight. Nevertheless, in practice, data distributions are often not as adversarial as those constructed to demonstrate the failure of the empirical mean (e.g., Catoni (2012)). This explains why the empirical mean is frequently observed to perform well in real-world scenarios, both in terms of computational efficiency and estimation accuracy.

Q7 Response. We thank the reviewer for the suggestion. We have run two experiments: They illustrate how accuracy and runtime vary with dimensionality. Specifically, we generated a synthetic dataset of 10,000 samples drawn from multivariate Gaussian distributions with dimensions ranging from 10 to 1000. As predicted by theory, the errors of both the coordinate-wise median and the fast gradient descent estimator remain low and stable across dimensions. Interestingly, the CSS algorithm shows rapid improvement as the dimension increases, stabilizing around dimension 500. As expected, the empirical mean achieves the best accuracy since the data are drawn i.i.d. from a Gaussian distribution. In terms of runtime, we observe a roughly linear growth with dimension, in line with theory. Unfortunately, we could not insert these tables due to characters constraints, but we will add these new experimental results (and the corresponding plots) to the final version of the paper.

评论

Thank you for the detailed responses to my questions. Most of my concerns were addressed. I will update my review and final justification accordingly.

审稿意见
3

The paper considers the following question: given nn points in the dd-dimensional Euclidean space, we want to find a (1+ε)(1+\varepsilon)-approximation algorithm for finding the mean (where the objective is the 11-means cost). The goal is to randomly samply certain points and use a fast algorithm to find the approximate mean (with probability at least 1δ1-\delta). There were known lower bounds implying lower bound of Ω(1/εlog(1/δ))\Omega(1/\varepsilon \cdot \log(1/\delta)). There are almost matching upper bounds (with an extra log factor). The paper gives a simple algorithm which samples Ω(1/εlog(1/δ))\Omega(1/\varepsilon \cdot \log(1/\delta)) points and uses the well-known median of means trick to show that this gives an optimal sample complexity bound. They also give a gradient descent based algorithm which seems to work well on real data sets.

优缺点分析

Strengths:

  1. Simple algorithm achieving optimal sample complexity bound for the 1-means problem.
  2. Validation of algorithms by experiments.

Weakness:

  1. Technique is well known and has been applied to many related problems in streaming.
  2. Prior works gave almost optimal sample complexity (upto log error) for all p\ell_p norms, whereas the idea here works only for 2\ell_2 norm.
  3. Improvement over prior work is only log(1/ε)\log(1/\varepsilon).

问题

  1. What is novel in techniques here? The median of means seems to be a standard idea.
  2. What about p\ell_p norms in general? Do the ideas here extend to these more general versions?

局限性

yes.

格式问题

None

作者回复

We appreciate the reviewer’s comments and the opportunity to clarify the contributions and novelty of our work. We respond to the raised questions and comments in a unified manner.

Our primary goal is to achieve optimal runtime in addition to optimal sample complexity for robust mean estimation in streaming settings. While prior works have indeed addressed sample complexity (often for all p\ell_p norms), their algorithms are generally not optimized for runtime. In contrast, our results focus specifically on the 2\ell_2 norm because our runtime improvements rely on structural properties and lemmas (e.g., Lemmas 2.1 and 2.2) that are unique to this norm. Extending such runtime guarantees to general p\ell_p norms remains an interesting open direction.

Regarding the techniques: although median-of-means is a classical idea in robust estimation, our contributions lie in how we combine and extend known techniques to obtain new algorithmic and theoretical insights. The median-of-means estimator is designed for the one-dimensional mean estimation problem. Extending the idea to higher dimensions is non-trivial. There is no standard notion of a median for multivariate data, and it is not clear a priori what definition of multivariate median generalizes the median-of-means estimator for one-dimension.

Specifically, our contributions are:

  • We introduce a gradient-descent-based algorithm that starts from the coordinate-wise median-of-means and efficiently converges to the true mean. This process provably runs in linear time, and to the best of our knowledge, this is the first such demonstration in this setting.

  • We show that the coordinate-wise median-of-means achieves dimension-free sample complexity in a distributional mean estimation setting—an aspect that, to our knowledge, was overlooked in prior literature.

  • In addition, we propose a novel estimator based purely on order statistics, which offers a distinct alternative for mean estimation and runs in near-optimal time.

As for the claimed improvement: there is also an improvement in by the order of logδ1\log \delta^{-1} in both sample complexity and running time. While the improvement in terms of the precision ε\varepsilon may be more modest (even if it is optimal now and was not before), it can be argued that we obtain a quadratic improvement in both running time and sample complexity in terms of the failure probability. Moreover, breaking the log2(1/δ)\log^2(1/\delta) barrier has been a notable challenge in the literature, and our method addresses this without sacrificing either sample complexity or computational efficiency in other parameters.

评论

Thanks for your rebuttal. I see your point, but (i) when you say that " There is no standard notion of a median for multivariate data", I thought Theorem 3.1 is just doing this for each coordinate independently, and the proof of Theorem 3.1 is again quite standard. (ii) you say " coordinate-wise median-of-means achieves dimension-free sample complexity"...but doesn't the error depend on the dimension? Trace of Sigma matrix would be dimension dependent. So if you want to make error tiny, one needs dimension dependent sample complexity.

评论

We thank the reviewer for a quick response. (i) We would like to reiterate the fact that though the proof is standard/elementary, previous texts on mean estimation of high-dimensional distribution seem to have missed it. Please refer to survey by Lugosi and Mendelson and "Geometric median and robust estimation in banach spaces" by Minsker for further details.

(ii) Our proof does show a dimension-free bound of the sample complexity, as it does not depend on the dimension explicitly. The lower bound for the sample complexity is given by Tr(Σ)N\sqrt{\frac{\text{Tr}(\Sigma)}{N}} even for the 1-dimensional case, so that is unavoidable. Note that value of Tr(Σ)\text{Tr}(\Sigma) might not increase with the dimension dd (even though it might). Previous bounds included a factor logd\log d. For a contrast, consider two distributions with the same Tr(Σ)\text{Tr}(\Sigma), one in R1\mathbb{R}^1 and the second in R1000\mathbb{R}^{1000}. Our bounds show the same upper bound for error rate of both the distributions, while the prior results give a worse bound for the second one. Our contribution is to say that there is no additional dependence on the dimension.

评论

Thanks for your response.

审稿意见
4

The authors present several novel methods to estimate a multivariate mean of a ground data, including theoretical proofs for optimality of sampling complexity. The methods are built as variants of of an algorithm that computes several sample means and is parametrized by an aggregation function for said means. The authors prove theoretical guarantees of approximation optimality and asymptotic relative runtime optimality of those methods, and evaluate them against MNIST/F-MNIST datasets, comparing to other baseline methods.

优缺点分析

The paper is written in a very clear English and good formulations. The authors strive for theoretical soundness, and for the major part provide thorough theoretical justification for their claims.

While the methods, as the paper's title suggest, are simple, this adds more clarity to their analysis. Moreover, the fast gradient method described for one of the algorithms, that essentially replaces a gradient step with a calculation of one-dimensional median along the gradient direction, is a rather neat idea.

Perhaps because of limited space, some proofs in the main paper feel too condensed and hop over several steps, reducing the clarity. One particular place is the beginning of Lemma 2.3's proof that starts with the phrase "Recall that the values of the constants were set as a = 1440, b = 50". In all fairness, the values a and b were barely introduced; they first mentioned inline in Algorithm 1, and it's trivial how they appear from the big-O of sample sizes. However neither these particular values 1440 and 50 are mentioned before, nor is it clear how exactly they are used.

Another example of a jump in the proof, to me, is in line 164 and the next inequality. Perhaps a better way to write the inequality, while keeping it short, would be to include an part in-between, like [average across G] \ge [average across all means] \ge RHS. But this is a minor detail, and I believe the proof is correct (small correction, line 162: should the referenced lemma be 2.3?).

A bit larger concerns are related to experemental evaluation and method significance. One strength is the amount of different methods/baselines. However:

  • A sample size mm is introduced. But how does it relate to aa, bb, δ\delta or ε\varepsilon? I believe I can see that in the experiment code, but the description is missing from the paper.
  • Three fixed datasets are tested, where it is also shown that the empirical mean outperforms most methods. I strongly believe that testing on synthetic data sampled from more clean distributions would have been necessary for any conclusions, e.g. a large ground Gaussian/uniform sample. Also, not sure if accuracy should be plotted in log scale, or whether the accuracy valus in large sample sizes are not meaningfully distinct.
  • Runtime barely grows with the sample size. Unless I'm misunderstanding what sample size is, should not it be essentially linear for introduced methods? Could the issue lie in vectorized numpy implementations with bottlenecks around the calls? Wouldn't it make sense to try an aforementioned synthetic dataset with a large sample size? And if it is not supposed to be linear, then again a clarification about sample size is important.

Last small question: line 179: what kind of non-optimality is spoken about here?

问题

I believe the clarity of Lemma 2.3 should be adjusted, as per discussion before.

While this is a more theoretical paper (or even more importantly because of that), I believe an experimentation on synthetic data could be crucial here. In the simplest case, the same experiments from MNIST could be run on a sample from e.g. a Gaussian distribution, perhaps also testing limits with much larger sample sizes.

One question I have about Algorithm 2: is there any statement about algorithm's asymptotic convergence to true geometric median (in relation to T)? Unless I missed it, perhaps it is good to clarify whether it's an open question, or perhaps it is easy to construct a set of points that has a unique geometric median, but the method does not go infinitesimally close to it.

I believe addressing these issues would bring the paper to a higher quality.

局限性

Yes.

最终评判理由

I believe the additional experiments, clarification to existing figures and the addition of some missing definitions improve the quality of the paper.

Based on authors' reply and how I imagine the new version, I readjust my grading to "Borderline Accept". For a full "Accept" rating, one would have to verify the final version of the paper (which is currently not available due to review process restrictions) with updated plots to also reevaluate the significance/quality of the experimental results. The rest of the paper is solid.

格式问题

No formatting concerns.

作者回复

We would like to thank the reviewer for the valuable comments and the raised questions. Due to NeurIPS 2025 restrictions, we’re unable to share links or images in the rebuttal or repository. We’ve included equivalent information in tabular form and kindly appreciate your understanding of these constraints.

Comment. A sample size mm is introduced ... missing from the paper.

Response. The sample size mm is set to abε1logδ1ab \varepsilon^{-1} \log \delta^{-1}. To clarify, this is the result of having blogδ1b\log \delta^{-1} subsamples of size aε1a \varepsilon^{-1} each. We thank the reviewer for the comment and will add it to the paper explicitly.

Comment and Question. Three fixed datasets ... Gaussian/uniform sample; While this is ... much larger sample sizes.

Response. Please refer to the last response to reviewer otjC and tables therein.

Comment. Also, not sure if accuracy should be plotted in log scale, or whether the accuracy values in large sample sizes are not meaningfully distinct.

Response. Following the reviewer's suggestion, we plotted the accuracy in linear scale for both MNIST and the randomly generated dataset. Indeed, the accuracy values for large sample sizes essentially coincide (Tables 1 and 7).

Comment. Runtime barely grows ... sample size is important.

Response. We thank the reviewer for pointing this out. The runtime plots in the paper are shown in log-scale, which can visually compress differences and obscure linear growth, especially when the slope is small. When replotted in linear scale, the trends are indeed mildly increasing and consistent with near-linear growth, albeit with a small slope. We also repeated the experiments on the randomly generated synthetic dataset referenced above, and in that case the runtime growth is visibly linear. While we cannot upload plots, we have provided tables summarizing the above information (Tables 1 and 2 below, and 3 and 4 in reviewer otjC's last response). We believe the soft growth observed on the original dataset is largely due to implementation-level factors. In particular, our use of numpy's vectorized operations can lead to non-obvious runtime behavior, as such operations benefit from low-level optimizations (e.g., memory locality, multi-threaded backends) and often incur fixed overheads. As a result, runtime may appear sublinear over moderate input sizes, especially when memory access. Lastly, we agree that applying the methods to significantly larger datasets (as we did in the synthetic experiments) helps confirm the expected linear scaling more clearly, and we are happy to include these linear-scale plots and additional clarifications in the final version of the paper.

Table 1: MNIST — Accuracies vs. Sample Size

mmEMEM σ\sigmaFGDFGD σ\sigmaCWMCWM σ\sigmaMSSMSS σ\sigmaCSSCSS σ\sigmaWTWT σ\sigma
101.0985280.0286351.1551260.0344581.1753920.0375001.3923440.0634391.0011930.0002511.0502200.109802
151.0897860.0228531.0914810.0214141.0893860.0196511.1755790.0413411.0011930.0002901.1830680.040979
201.0509440.0170041.0710970.0161691.0794050.0153861.1503860.0337951.0011490.0002991.1667930.038011
251.0425820.0120221.0580920.0137371.0566720.0119561.1089550.0237051.0011700.0003051.1203480.024574
301.0350540.0099811.0429300.0091941.0452960.0091351.0915210.0162391.0011940.0002901.0973960.022739
1001.0095600.0021231.0118620.0028591.0122320.0023031.0416290.0111961.0012290.0002421.0420880.011341
2001.0047500.0009651.0071890.0014801.0076970.0015411.0220380.0059451.0012510.0002281.0254990.005615
5001.0019540.0004981.0029610.0006481.0027040.0005721.0106610.0022791.0012480.0002211.0126540.002951
10001.0009210.0002391.0013010.0003011.0012860.0002611.0052010.0013921.0013100.0002161.0067860.001536
20001.0004930.0001311.0007000.0001141.0007170.0001051.0028980.0006171.0012900.0001851.0037910.000942
50001.0001800.0000541.0002940.0000721.0002700.0000711.0013830.0003271.0013600.0001651.0018200.000490
100001.0000890.0000271.0001420.0000271.0001510.0000311.0006920.0001171.0013400.0001691.0009960.000299

Table 2: MNIST — Runtimes vs. Sample Size

mmFGDFGD σ\sigmaCWMCWM σ\sigmaMSSMSS σ\sigmaCSSCSS σ\sigmaWTWT σ\sigma
100.3311200.0253230.3039070.0215140.3050310.0228290.7521440.0479940.4987040.101033
150.3288240.0298530.3052480.0238600.3045510.0193280.7506690.0529150.4868010.095939
200.3211850.0223700.3133000.0264940.3047870.0216800.7539250.0492010.4692290.060214
250.3204470.0243250.3054690.0189240.3128670.0269520.7589500.0523230.4548100.045884
300.3246990.0233930.3062460.0211640.3073800.0211950.7543690.0506490.4844480.061920
1000.3322000.0297190.3055760.0225940.3042660.0215360.7532910.0479230.4715330.064969
2000.3398680.0327100.3075930.0215980.3061240.0208220.7539500.0529710.4728280.057820
5000.3412250.0329630.3111280.0206420.3099150.0217040.7578280.0394160.4769490.061728
10000.3401640.0294540.3130980.0250720.3086680.0211770.7573720.0506390.4782370.062958
20000.3476020.0296520.3203470.0267400.3178220.0223110.7605080.0509130.4985000.058959
50000.3601830.0351230.3209590.0224850.3220970.0220880.7548880.0512010.4966060.064727
100000.3698070.0326920.3365060.0231590.3452130.0299870.7599440.0473480.5268140.062030

Question. Last small question: line 179: what kind of non-optimality is spoken about here?

Response. The coordinate-wise median is not optimal with respect to the convergence rate. The rate is of the form Tr(Σ)logδ1n\sqrt{\frac{Tr(\Sigma) \log \delta^{-1}}{n}} instead of the optimal Tr(Σ)n+λmaxlogδ1n\sqrt{\frac{Tr(\Sigma)}{n}} + \sqrt{\frac{\lambda_{\max} \log \delta^{-1}}{n}}. There are other algorithms which achieve optimality in the convergence rate like the median-of-means tournaments. See the Lugosi and Mendelson survey for a detailed explanation.

Question. I believe the clarity of Lemma 2.3 ... infinitesimally close to it.

Response. Algorithm 2 does not always converge to the geometric median of the points. Take, for example, the triangle with coordinates (0,1),(0,1),(1,0)(0,1), (0,-1), (1,0). The geometric median is the point (1/3,0)(1/\sqrt 3,0). If we initialize the algorithm at the origin, the algorithm never makes any progress and stays at the origin. We wish to reiterate that our goal was only to be sufficiently close to the geometric median. As mentioned in Remark 4.4, the state of the art geometric median algorithms have a significantly higher running time than what we are aiming for. Since even the best possible algorithm converging to the geometric median will run in time Ω(ndlogε1)\Omega(nd \log \varepsilon^{-1}), a running time not achievable with current algorithms and a running time that is already superlinear in our setting, we can only hope to be sufficiently close.

评论

I thank the authors for their detailed response to the questions and discussion. In my opinion, the additional experiments (presented in the reply to Reviewer otjC) and the clarifications to parts of the paper (including the figures and some definitions) will notably improve its quality. I have adjusted my grading accordingly.

审稿意见
5

The paper suggests techniques to compute an approximation to the mean of a given n points in an Euclidean space. The approximation is (1+epsilon) multiplicative factor, using only O(log(1/delta)/epsilon) i.i.d samples. The state of the art is (1/(epsilon*delta)) is tight, but assuming that the algorithm is just "compute the mean of the sample". Instead, the authors suggest aggregation techniques and median approaches. Experiments on some datasets is also provided, as well as code.

优缺点分析

Strong: -A fundamental problem in learning and statistics.

  • Open code.
  • Provable results.

Weak:

  • The algorithm looks like a random version of a paper with a very similar goal that was appeared recently in NeuRIPS: https://arxiv.org/abs/1906.04705 Although there are not so many papers regarding this problem, the paper is not cited and it seems that the authors are unaware of the result which essentially computes the mean by multiple aggregations. I am afraid that this fact alone suffices to reject this nice paper.

  • The experiments are very poor. Why MNIST which is considered small and a "toy dataset"? Why not datasets of billions of points if the running time is sub-linear? Even random numbers would do..

问题

The idea of using the median of small samples for computing the global mean seems very related to the famous paper: https://www.tau.ac.il/~nogaa/PDFS/amsz4.pdf

Please provide some comparison, at least for the algorithm itself.

局限性

No

格式问题

No

作者回复

We would like to thank the reviewer for the useful comments, questions and suggestions. Moreover, we would like to highlight that, in accordance with the NeurIPS 2025 policy, we are not allowed to include any links in our rebuttal response, nor are we permitted to upload images—either in the rebuttal or to the already submitted repository. We are thus providing tables containing the same information as plots. We regret that you are unable to view the updated plots, and we kindly ask for your understanding and trust given these constraints.

Question. The idea ... the algorithm itself.

Response. The idea of median-of-means estimator is indeed old. Well-known references include (Nemirovski and Yudin, 1983; Jerrum et al., 1986 including the paper the reviewer has cited - Alon et al., 1996). Our work focuses on adapting the idea for higher dimensions. There are indeed many options for this including coordinate-wise median, geometric median among others. The proofs of convergence and running time we provide are novel to the best of our knowledge. Specifically, the best known algorithm for computing (approximate) geometric median (Cohen et al. 2016) leads to a much worse running time of the algorithm.

Comment. The algorithm looks like ... reject this nice paper.

Response. The paper in reference, though dealing with problems of a similar flavor, has no direct connection to the problem we aim to solve. The paper aims at computing loss-less summaries for the mean. Sublinear algorithms necessarily require some form of approximation to the mean. The difference is also highlighted in the running time: their algorithm has complexity O(nd+d4logn)O(nd + d^4 \log n). It is unlikely it could be used at any stage of the mean aggregation. But even if it could be used, its running time would be significantly worse that what we are aiming for. However, we will add a reference and explain the differences in the final version.

Comment. The experiments are very poor ... random numbers would do.

Response. Following the Reviewers otjC and 1w8V’s suggestions, we ran additional experiments with larger sample sizes. Specifically, we generated a synthetic dataset from a 200-dimensional multivariate Gaussian distribution and extended the sample size up to 5×1055 \times 10^5.

The results are in line with theory: They show that the error decreases sharply as the sample size grows and that the runtime grows linearly with sample size (Tables 3 and 4).

Unfortunately, we were unable to go beyond this sample size due to memory limitations in our environment, where the process exhausts the available RAM. We will add a note in the final version explaining this limitation. We hope that the extended results are sufficient to address the reviewers’ concerns. We have also generated two plots illustrating how accuracy and runtime vary with dimensionality: We elaborated more on this aspect in the response to Reviewer sRHN. We will add these new experimental results to the final version of the paper.

Table 3: Synthetic Dataset (d=200d = 200) — Accuracies vs. Sample Size

mmEMEM σ\sigmaFGDFGD σ\sigmaCWMCWM σ\sigmaMSSMSS σ\sigmaCSSCSS σ\sigmaWTWT σ\sigma
101.00001.17×1011.17 \times 10^{-1}1.12743.42×1023.42 \times 10^{-2}1.12763.49×1023.49 \times 10^{-2}1.25282.42×1022.42 \times 10^{-2}1.00001.11×10161.11 \times 10^{-16}1.25872.19×1022.19 \times 10^{-2}
151.00008.88×1028.88 \times 10^{-2}1.06201.22×1021.22 \times 10^{-2}1.06711.31×1021.31 \times 10^{-2}1.14271.29×1021.29 \times 10^{-2}1.00000.000.001.14421.34×1021.34 \times 10^{-2}
201.00009.42×1029.42 \times 10^{-2}1.05491.05×1021.05 \times 10^{-2}1.05899.72×1039.72 \times 10^{-3}1.11951.29×1021.29 \times 10^{-2}1.00000.000.001.12501.17×1021.17 \times 10^{-2}
251.00006.08×1026.08 \times 10^{-2}1.03995.84×1035.84 \times 10^{-3}1.03937.14×1037.14 \times 10^{-3}1.08738.39×1038.39 \times 10^{-3}1.00198.89×1048.89 \times 10^{-4}1.09158.92×1038.92 \times 10^{-3}
301.00002.72×1022.72 \times 10^{-2}1.03165.53×1035.53 \times 10^{-3}1.02994.48×1034.48 \times 10^{-3}1.06726.72×1036.72 \times 10^{-3}1.00139.67×1059.67 \times 10^{-5}1.07085.76×1035.76 \times 10^{-3}
1001.00006.47×1026.47 \times 10^{-2}1.00881.32×1031.32 \times 10^{-3}1.00911.18×1031.18 \times 10^{-3}1.03022.62×1032.62 \times 10^{-3}1.00053.69×1053.69 \times 10^{-5}1.03132.94×1032.94 \times 10^{-3}
2001.00003.85×1023.85 \times 10^{-2}1.00587.01×1047.01 \times 10^{-4}1.00576.43×1046.43 \times 10^{-4}1.01942.18×1032.18 \times 10^{-3}1.00032.54×1052.54 \times 10^{-5}1.02021.97×1031.97 \times 10^{-3}
5001.00001.66×1021.66 \times 10^{-2}1.00222.49×1042.49 \times 10^{-4}1.00222.41×1042.41 \times 10^{-4}1.00957.99×1047.99 \times 10^{-4}1.00011.03×1051.03 \times 10^{-5}1.01028.94×1048.94 \times 10^{-4}
10001.00000.000.001.00119.79×1059.79 \times 10^{-5}1.00119.97×1059.97 \times 10^{-5}1.00484.04×1044.04 \times 10^{-4}1.00016.40×1066.40 \times 10^{-6}1.00514.65×1044.65 \times 10^{-4}
20001.00005.44×1025.44 \times 10^{-2}1.00066.73×1056.73 \times 10^{-5}1.00067.67×1057.67 \times 10^{-5}1.00282.76×1042.76 \times 10^{-4}1.00013.81×1063.81 \times 10^{-6}1.00312.75×1042.75 \times 10^{-4}
50001.00009.93×1039.93 \times 10^{-3}1.00022.73×1052.73 \times 10^{-5}1.00022.63×1052.63 \times 10^{-5}1.00131.07×1041.07 \times 10^{-4}1.00012.04×1062.04 \times 10^{-6}1.00141.47×1041.47 \times 10^{-4}
100001.00001.17×10101.17 \times 10^{-10}1.00011.21×1051.21 \times 10^{-5}1.00011.04×1051.04 \times 10^{-5}1.00076.10×1056.10 \times 10^{-5}1.00011.59×1061.59 \times 10^{-6}1.00087.20×1057.20 \times 10^{-5}
1000001.00003.14×10183.14 \times 10^{-18}1.00001.43×1061.43 \times 10^{-6}1.00001.27×1061.27 \times 10^{-6}1.00016.64×1066.64 \times 10^{-6}1.00006.10×1076.10 \times 10^{-7}1.00011.00×1051.00 \times 10^{-5}
5000001.00003.14×10183.14 \times 10^{-18}1.00002.92×1072.92 \times 10^{-7}1.00002.52×1072.52 \times 10^{-7}1.00002.00×1062.00 \times 10^{-6}1.00003.98×1073.98 \times 10^{-7}1.00002.00×1062.00 \times 10^{-6}

Table 4: Synthetic Dataset (d=200d = 200) — Runtimes vs. Sample Size

mmFGDFGD σ\sigmaCWMCWM σ\sigmaMSSMSS σ\sigmaCSSCSS σ\sigmaWTWT σ\sigma
100.0231582.01×1032.01 \times 10^{-3}0.0004241.98×1041.98 \times 10^{-4}0.0004956.10×1056.10 \times 10^{-5}0.0002291.10×1051.10 \times 10^{-5}0.0004662.60×1052.60 \times 10^{-5}
150.0232872.97×1032.97 \times 10^{-3}0.0004343.57×1043.57 \times 10^{-4}0.0006633.94×1043.94 \times 10^{-4}0.0003392.44×1042.44 \times 10^{-4}0.0007015.38×1045.38 \times 10^{-4}
200.0162614.85×1034.85 \times 10^{-3}0.0002455.00×1055.00 \times 10^{-5}0.0003492.80×1052.80 \times 10^{-5}0.0001641.80×1051.80 \times 10^{-5}0.0003383.30×1053.30 \times 10^{-5}
250.0147212.97×1032.97 \times 10^{-3}0.0002736.80×1056.80 \times 10^{-5}0.0003857.80×1057.80 \times 10^{-5}0.0001678.00×1068.00 \times 10^{-6}0.0003929.60×1059.60 \times 10^{-5}
300.0151183.30×1033.30 \times 10^{-3}0.0002482.50×1052.50 \times 10^{-5}0.0003532.10×1052.10 \times 10^{-5}0.0001731.20×1051.20 \times 10^{-5}0.0003643.60×1053.60 \times 10^{-5}
1000.0170723.41×1033.41 \times 10^{-3}0.0003796.40×1056.40 \times 10^{-5}0.0004391.90×1051.90 \times 10^{-5}0.0002824.30×1054.30 \times 10^{-5}0.0006252.20×1052.20 \times 10^{-5}
2000.0192203.32×1033.32 \times 10^{-3}0.0005226.40×1056.40 \times 10^{-5}0.0006554.50×1054.50 \times 10^{-5}0.0004795.80×1055.80 \times 10^{-5}0.0010349.30×1059.30 \times 10^{-5}
5000.0216313.98×1033.98 \times 10^{-3}0.0008355.20×1055.20 \times 10^{-5}0.0010781.04×1041.04 \times 10^{-4}0.0008504.90×1054.90 \times 10^{-5}0.0020331.54×1041.54 \times 10^{-4}
10000.0222462.99×1032.99 \times 10^{-3}0.0014047.10×1057.10 \times 10^{-5}0.0018414.40×1044.40 \times 10^{-4}0.0016292.26×1042.26 \times 10^{-4}0.0036633.48×1043.48 \times 10^{-4}
20000.0252023.42×1033.42 \times 10^{-3}0.0029305.24×1045.24 \times 10^{-4}0.0032993.50×1043.50 \times 10^{-4}0.0028322.04×1042.04 \times 10^{-4}0.0078011.19×1031.19 \times 10^{-3}
50000.0470896.56×1036.56 \times 10^{-3}0.0108909.59×1049.59 \times 10^{-4}0.0081055.00×1045.00 \times 10^{-4}0.0102101.04×1031.04 \times 10^{-3}0.0184381.52×1031.52 \times 10^{-3}
100000.0421003.96×1033.96 \times 10^{-3}0.0164349.72×1049.72 \times 10^{-4}0.0167441.03×1031.03 \times 10^{-3}0.0225417.60×1047.60 \times 10^{-4}0.0377052.83×1032.83 \times 10^{-3}
1000000.2567332.73×1022.73 \times 10^{-2}0.2273102.16×1022.16 \times 10^{-2}0.2238431.91×1021.91 \times 10^{-2}0.3398042.20×1022.20 \times 10^{-2}0.4835667.13×1027.13 \times 10^{-2}
5000001.3219691.16×1011.16 \times 10^{-1}1.2782431.11×1011.11 \times 10^{-1}1.2774241.12×1011.12 \times 10^{-1}1.6956651.05×1011.05 \times 10^{-1}2.7827993.10×1013.10 \times 10^{-1}
评论

I thank the authors for adding the experimental results, and would encourage the other reviewers to accept it. Meanwhile, I noticed that there are many related results that are not cited, including papers by Jeff M. Phillips (e.g., "Near-Optimal Coresets of Kernel Density Estimate"). See many more in a survey on this problem (that is also not cited): @article{DBLP:journals/corr/abs-2111-03046, author = {Alaa Maalouf and Ibrahim Jubran and Dan Feldman}, title = {Introduction to Coresets: Approximated Mean}, journal = {CoRR}, volume = {abs/2111.03046}, year = {2021}, eprinttype = {arXiv}, bibsource = {dblp computer science bibliography, https://dblp.org} }

评论

We thank the reviewer for the references and the insightful feedback, in particular the pointer to the approximate coresets for the mean which is indeed very relevant. We will incorporate them into the final version of the paper, adding a detailed comparison between them and our work.

最终决定

This paper studies the multivariate mean estimation problem. It shows an aggregate estimator that averages mean estimates on bootstrap samples (with independent and uniform sampling) can output an (1+ε)(1+\varepsilon)-approximate mean with probability at least 1δ1 - \delta by using O(ε1logδ1)O(\varepsilon^{-1} \log \delta^{-1}) sampled points. This is an improvement of the classic O(ε1δ1)O(\varepsilon^{-1} \delta^{-1}) sample complexity and the SOTA O(ε1logδ1(logε1+logδ1))O(\varepsilon^{-1} \log \delta^{-1} (\log \varepsilon^{-1} + \log \delta^{-1})) sample complexity. The paper further presents three efficient algorithms that run the aggregate estimator in O(ε1logδ1)O(\varepsilon^{-1} \log \delta^{-1}) time. Experimental results on a real-world data set shows one of the proposed estimators achieve the best accuracy-runtime tradeoff, while the other two have competitive performance compared to the SOTA method.


The paper makes a plausible improvement on SOTA sample complexity for mean estimation, and provides efficient estimators.

What holds reviewers Vydg and sRHN back is the limited technical novelty. Reviewer Vydg also points out the improvement only applies to the 2\ell_2 norm while prior results apply to p\ell_p. Another notable weakness is the experiment section. Concerns include the limited scale of the data set (reviewers otjC & 1w8V & sRHN), insufficient comparison with empirical mean as a baseline (reviewer sRHN) and missing verification of the dimension-free property (reviewer sRHN).

Despite those concerns, most reviewers are on the positive side. I think the work is decent and the paper is well written. Its contribution may have a relatively narrowed scope, but its merits outweigh this limitation. Authors should incorporate all reviewers' feedback in revision.