Second-order Optimization under Heavy-Tailed Noise: Hessian Clipping and Sample Complexity Limits
matching upper and lower bounds for stochastic second order optimization under heavy tailed noise, high probability convergence
摘要
评审与讨论
This paper studies the problem of second-order stochastic optimization (SOSO) under heavy-tailed noise, a setting where both stochastic gradients and Hessians only have bounded p-th moments, which is common in modern machine learning tasks with outliers or data heterogeneity. The paper establishes tight minimax lower bounds on the sample complexity for any second-order optimization algorithm under heavy-tailed noise. The results demonstrate that second-order methods can still achieve provable improvements over first-order methods, even in heavy-tailed regimes. The paper proposes a normalized SGD with Hessian correction algorithm that incorporates second-order information and provably achieves the established lower bond for finding approximate stationary points. The paper also introduces Hessian clipping, an extension of gradient clipping to handle extreme noise in Hessian estimates.
优缺点分析
Strengths:
- This paper is well-written and well-structured, easy to read and follow.
- This paper provides rigorous theoretical contribution.
- The analysis is technically sound and comprehensive, covering both in-expectation and high-probability results and addressing the impact of gradient and Hessian noise.
- The originality of this paper is clear and solid. Weaknesses:
- The experimental validation is minimal, consisting only of toy synthetic experiments in low dimensions. There is no evaluation on real-world machine learning tasks to demonstrate practical effectiveness.
- The proposed methods require assumptions on hyperparameters and question setup, which may limit their practical usability.
- While theory is strong, there is no discussion of computational costs for Hessian-vector products or clipping in large-scale models.
问题
Please refer the the weaknesses section.
局限性
No
格式问题
None
Dear Reviewer tAtg,
Thank you for your thorough and thoughtful review, and for your positive evaluation of the paper’s clarity, originality, and theoretical contribution. We are especially grateful for your recognition of the strength of our analysis, including both in-expectation and high-probability results under heavy-tailed noise. Below, we respond to your comments regarding experimental validation, practical considerations, and assumptions, and we appreciate your suggestions for further strengthening the paper.
Q1: The experimental validation is minimal, consisting only of toy synthetic experiments in low dimensions. There is no evaluation on real-world machine learning tasks to demonstrate practical effectiveness.
A1: We acknowledge the importance of empirical evaluation of the algorithms on real-world machine learning tasks, and we are planning to test our algorithms on more serious tasks in the future. However, the main goal of this work is to understand the theoretical complexity limits of second-order stochastic algorithms under challenging heavy-tailed settings. Our Algorithm 1 was already tested before in reinforcement learning literature [Salehkaleybar et al., 2022, Fatkhullin et al., 2023] with promising results. We are hopeful that our theoretical enhancement with gradient and Hessian clipping (Algorithm 2) will further improve the practical performance, and we leave this for future work.
We also find controlled synthetic experiments insightful and important, especially in the initial phase of algorithmic development. We conduct the following additional ablation studies, which provide insights on theoretically derived complexities and the sensitivity to clipping threshold tuning.
We solved a simple quadratic problem, , in dimension , with synthetic noise sampled from a symmetrized Pareto distribution. We considered both a fixed heavy-tailed regime with tail index and a variable tail index ranging from to . We start each algorithm from the same point sampled uniformly at random from the standard normal distribution. Our empirical analysis consists of three experimental settings:
1. Effect of Clipping Level on Iteration Complexity: We ran Algorithm 2 (Clip NSGDHess) with until it reached the target accuracy of . The stepsize and momentum parameters were set to and . Clipping levels were varied from to using a multiplicative step of 10. The results are shown below:
| Iterations |
Table 1: Intermediate clipping () yields the lowest iteration complexity. Extremely small or large values lead to slower convergence. The proposed algorithm can gracefully tolerate very small values of clipping parameter , but using values larger than leads to very slow convergence. This observation motivates the need for gradient and hessian clipping.
2.Comparison with Clip NSGDM under Varying Tail Index: We compared Clip NSGDHess with Clip NSGDM (as proposed in [2]). For Clip NSGDM, we used the stepsize and momentum according to theory. For Clip NSGDHess, both parameters were set to . We fixed , , and iterations.
| Tail index, | Iteration Complexity (Clip NSGDM) | Iteration Complexity (Clip NSGDHess) |
|---|---|---|
Table 2: Clip NSGDHess consistently outperforms Clip NSGDM, with the gap increasing for smaller , i.e., heavier-tailed noise. Visually the picture looks similar to theoretical complexities in Figure 1 of our original submission.
[2] Cutkosky, Ashok, and Harsh Mehta. "High-probability bounds for non-convex stochastic optimization with heavy tails." Advances in Neural Information Processing Systems 34 (2021): 4883-4895.
3.Sensitivity to Gradient Clipping for Fixed Hessian Clipping: We fixed the Hessian clipping level and varied the gradient clipping level :
- For : we set from to with multiplicative step ;
- For : we set from to with multiplicative step ;
- For : we set from to with multiplicative step .
In all cases, the algorithm was run until . We consistently observed that both extremely small and large values of led to higher iteration complexity. The optimal values were:
- for ;
- for ;
- for .
This pattern confirms that intermediate clipping levels are crucial for performance.
Q2: The proposed methods require assumptions on hyperparameters and question setup, which may limit their practical usability.
A2: This is a valid concern, and we can offer a strategy that avoids any tuning requirement in Algorithm 1. We can set step-size , momentum . Following the proof strategy similar to (Hübler et al, 2024), we can establish convergence of our algorithm with this step-size strategy. Similarly to the observations in (Hübler et al, 2024), our convergence rate will become slower: instead of (and the constant behind ) will be larger. Nevertheless, this strategy avoids any parameter tuning and ensures non-asymptotic convergence. In case of Algorithm 2, the situation is more complex and we do not have a good strategy for setting clipping thresholds when problem parameters are unknown. Since the main focus of our paper is understanding the sample complexity upper and lower bounds when all parameters are known, we omit the discussion of these details.
Florian Hübler, Ilyas Fatkhullin, Niao He. From Gradient Clipping to Normalization for Heavy Tailed SGD. AISTATS 2025.
Q3: While theory is strong, there is no discussion of computational costs for Hessian-vector products or clipping in large-scale models.
A3: Thank you for your comment! The computational costs for Hessian-vector products is almost the same as the costs for gradients, refer to e.g., to (Pearlmutter, 1994). The idea is that let us imagine we have a backpropagation procedure for computing the gradient of . After this procedure we get . To compute Hessian-vector product we need to construct a function , and compute the gradient of the new function with respect to via backpropagation Therefore, we obtain the desired Hessian-vector . Thus, the computational cost is .
Pearlmutter, Barak A. "Fast exact multiplication by the Hessian." Neural computation 6, no. 1 (1994): 147-160.
Thank you for your detailed response to my review. I appreciate you taking the time to address the points raised.
This paper studies nonconvex stochastic optimization for smooth functions where both stochastic gradients and Hessians have -th bounded central moment for . The authors establish minmax lower bounds on the sample complexity of second order stochastic methods for the considered problem class and zero-respecting algorithms. A near optimal algorithm, i.e., Normalized SGD with Hessian correction, is proposed and shown to achieve in-expectation upper bounds that almost matches with the lower bound. In addition, the, a clipped variant of the previous algorithm, i.e., Normalized SGD with Hessian correction and clipping, is shown to achieve nearly optimal high-probability convergence.
优缺点分析
Strengths:
- This paper for the first time proposes SOSO methods to address heavy-tailed noise.
- The minmax lower bound is shown to be tight in several special cases.
- Near-optimal algorithms are designed to match the lower bounds in the in-expectation, and high-probability sense with logarithmic failure probability dependence, respectively.
Weakness:
- There is a lack of compelling numerical or analytical examples showcasing the advantages of SOSO over first-order methods.
问题
- In line 160, is it -th order?
- In definition 1, can the authors provide some examples on algorithms that belong to the defined algorithm family?
局限性
N.A.
最终评判理由
I have read the reviews from other reviewers and author's rebuttal. I decided to keep my original score.
格式问题
N.A.
Dear Reviewer GxUC,
Thank you very much for your detailed and encouraging review. We appreciate your positive assessment of the paper’s originality, theoretical rigor, and clarity. We are glad that you found the contributions—including the tight minimax lower bound and the near-optimal SOSO algorithms under heavy-tailed noise—novel and significant. Below, we address your thoughtful comments regarding numerical validation, clarify technical points (such as the algorithm family in Definition 1), and correct minor inconsistencies in the manuscript.
Q1: In line 160, is it -th order?
A1: Yes, this is a typo, thanks for catching this.
Q2: In definition 1, can the authors provide some examples on algorithms that belong to the defined algorithm family?
A2: Examples of algorithms satisfying definition 1 include many common algorithm. For example, SGD (with and without momentum), stochastic BFGS and L-BFGS, Stochastic Newton’s method (with and without cubic regularization), first and second-order trust-region method. To support our claim we refer to Carmon et al., 2019 (page 80 in the published version), see also Arjevani et al., 2020. We will add a remark about this important question in our revision.
Y. Carmon, J. C. Duchi, O. Hinder, and A. Sidford. Lower bounds for finding stationary points I. Mathematical Programming, May 2019.
Yossi Arjevani, Yair Carmon, John C. Duchi, Dylan J. Foster, Ayush Sekhari, Karthik Sridharan. Second-Order Information in Non-Convex Stochastic Optimization: Power and Limitations. COLT 2020.
I thank the author for answering my questions. It is appreciated.
In this submission, the authors present a theoretical analysis of second-order optimization in the presence of heavy-tailed noise, addressing the common issue that such methods often underperform and lack guarantees in this setting. They consider scenarios where stochastic gradients and Hessians have only bounded first- and second-order moments, and derive tight lower bounds on the sample complexity for any second-order method. To complement their analysis, the authors propose a normalized stochastic gradient descent algorithm that leverages second-order information and matches the established lower bounds. They also introduce a novel gradient and Hessian clipping technique, along with high-probability upper bounds to support its effectiveness.
优缺点分析
This paper makes several high-quality theoretical contributions, summarized as follows:
-
It establishes a lower bound on the sample complexity for any second-order stochastic optimization algorithm under the p-BCM noise model, showing that using higher-order derivatives does not necessarily improve the convergence rate under the same noise assumptions.
-
It proposes a stochastic gradient descent method that leverages second-order Hessian information and proves that its sample complexity matches the previously established lower bound.
-
It introduces a gradient and Hessian clipping technique and demonstrates that this method can achieve near-optimal sample complexity with high confidence.
These theoretical advances contribute significantly to the optimization community’s understanding of second-order stochastic methods in high-noise settings.
The main weakness of this submission is the lack of sufficient empirical results. The paper should include more numerical experiments evaluating the proposed Normalized SGD with Hessian correction algorithm and the gradient and Hessian clipping methods. Presenting empirical performance would help verify whether these novel algorithms align with the theoretical predictions.
For instance, since the normalized SGD with Hessian correction combined with clipping requires a clipping threshold parameter , the authors should explore the impact of varying values on the algorithm’s performance. This would help assess whether the clipping technique enhances robustness. Incorporating such numerical experiments would significantly strengthen the paper and, if the empirical results support the theory, greatly increase its publication potential.
问题
-
From assumptions one to four, this analysis seems to be independent of the convexity of the objective function. All results are applicable in the non-convex setting. If we add the assumption of convexity or even strong convexity, how will these convexity condition improve the lower bound and the sample complexity of the proposed algorithms? All assumptions from this submission are relevant to the smoothness or Lipshiz continuity of both gradients and Hessian. If we add some lower bounds for the smallest eigenvalue of the Hessian, how will the the number of oracle queries required to find an ε-stationary point be reduced? Will this be some future research direction?
-
I'm not sure whether the proposed Normalized stochastic gradient decent with Hessian correction (with gradient clipping and Hessian clipping) belong to the class of second-order optimization algorithms. In fact this proposed algorithm utilize the second-order Hessian matrix information, but in general second-order optimization algorithm correspond to Newton type methods such as Newton's method, quasi-Newton method (BFGS, DFP ,SR1) or L-BFGS with iterations like: where is some matrix that approximates the exact Hessian matrix . This Hessian approximation matrix is usually constructed using some first-order gradient informations. However, the proposed variant SGD method has no such iteration updating forms. Hence, strictly speaking, the proposed Normalized SGD with Hessian correction belong to the first-order optimization algorithms which utilized some second-order Hessian queries. This method don't belong to the formal "second-order" optimization class.
-
From Theorem 2 and Theorem 3, both the step size from the normalized SGD and the clipping threshold from the clipping technique requires the estimate the initial sub-optimality . How to estimate this initial sub-optimality in practice? Theorem 2 and 3 presented that and need to be set by values of this initial sub-optimality to reach the corresponding sample complexity. Hence, the proposed methods are impractical since it is in general very difficult to estimate the initial sub-optimality accurately.
局限性
The authors have presented all the limitations of this work in the last section of the paper.
最终评判理由
The rebuttal results of numerical experiments solve my concerns. I raised the score from 3 to 4.
格式问题
No Paper Formatting Concerns
Dear Reviewer oE5T,
Thank you for your detailed and thoughtful review. We sincerely appreciate your recognition of our theoretical contributions, including the lower bound under the p-BCM noise model, the matching upper bound via our second-order algorithm, and the development of the gradient and Hessian clipping technique that achieves near-optimal sample complexity with high probability.
We would like to address your main concern regarding the empirical evaluation:
Q1: The main weakness of this submission is the lack of sufficient empirical results. The paper should include more numerical experiments evaluating the proposed Normalized SGD with Hessian correction algorithm and the gradient and Hessian clipping methods. Presenting empirical performance would help verify whether these novel algorithms align with the theoretical predictions.
A1: During the rebuttal period, we conducted several numerical experiments that are not included in the current version of the paper but will be added in the final revision. Due to NeurIPS policy, we are unable to include figures; however, we describe the results below. We solved a simple quadratic problem, , in dimension , with synthetic noise sampled from a symmetrized Pareto distribution. We considered both a fixed heavy-tailed regime with tail index and a variable tail index ranging from to . We start each algorithm from the same point sampled uniformly at random from the standard normal distribution. Our empirical analysis consists of three experimental settings:
1. Effect of Clipping Level on Iteration Complexity: We ran Algorithm 2 (Clip NSGDHess) with until it reached the target accuracy of . The stepsize and momentum parameters were set to and . Clipping levels were varied from to using a multiplicative step of 10. The results are shown below:
| Iterations |
Table 1: Intermediate clipping () yields the lowest iteration complexity. Extremely small or large values lead to slower convergence. The proposed algorithm can gracefully tolerate very small values of clipping parameter , but using values larger than leads to very slow convergence. This observation motivates the need for gradient and hessian clipping.
2.Comparison with Clip NSGDM under Varying Tail Index: We compared Clip NSGDHess with Clip NSGDM (as proposed in [2]). For Clip NSGDM, we used the stepsize and momentum according to theory. For Clip NSGDHess, both parameters were set to . We fixed , , and iterations.
| Tail index, | Iteration Complexity (Clip NSGDM) | Iteration Complexity (Clip NSGDHess) |
|---|---|---|
Table 2: Clip NSGDHess consistently outperforms Clip NSGDM, with the gap increasing for smaller , i.e., heavier-tailed noise. Visually the picture looks similar to theoretical complexities in Figure 1 of our original submission.
[2] Cutkosky, Ashok, and Harsh Mehta. "High-probability bounds for non-convex stochastic optimization with heavy tails." Advances in Neural Information Processing Systems 34 (2021): 4883-4895.
Q2: From assumptions one to four, this analysis seems to be independent of the convexity of the objective function. All results are applicable in the non-convex setting. If we add the assumption of convexity or even strong convexity, how will these convexity condition improve the lower bound and the sample complexity of the proposed algorithms?
A2: In the convex case, under bounded variance () the upper and lower bounds were derived, e.g., in [1]. It turns out that when the function is convex, the use of second-order information does not improve the stochastic gradient complexity . Therefore, we do not expect further improvement for other for our method as well. However, the situation in our non-convex setting is different and the sample complexity can be reduced uniformly for all using the second-order information.
[1] Artem Agafonov, Dmitry Kamzolov, Alexander Gasnikov, Ali Kavis, Kimon Antonakopoulos, Volkan Cevher, Martin Takáč. Advancing the lower bounds: An accelerated, stochastic, second-order method with optimal adaptation to inexactness. ICLR 2024.
Q3: All assumptions from this submission are relevant to the smoothness or Lipshiz continuity of both gradients and Hessian. If we add some lower bounds for the smallest eigenvalue of the Hessian, how will the the number of oracle queries required to find an -stationary point be reduced? Will this be some future research direction?
A3: Since we assume -Lipschitz continuous gradient (smoothness) and the function is twice-differentiable, we already have a lower bound for the smallest eigenvalue of the Hessian equal to . Therefore, such assumption does not help improve the complexity in terms of the . However, one can consider a more refined setting where the smallest eigenvalue is and improve the sample complexity in terms of dependence on . We leave such refinement for future work.
Q4: I'm not sure whether the proposed Normalized stochastic gradient decent with Hessian correction (with gradient clipping and Hessian clipping) belong to the class of second-order optimization algorithms. In fact this proposed algorithm utilize the second-order Hessian matrix information, but in general second-order optimization algorithm correspond to Newton type methods such as Newton's method, quasi-Newton method (BFGS, DFP ,SR1) or L-BFGS with iterations like: where is some matrix that approximates the exact Hessian matrix . This Hessian approximation matrix is usually constructed using some first-order gradient informations. However, the proposed variant SGD method has no such iteration updating forms. Hence, strictly speaking, the proposed Normalized SGD with Hessian correction belong to the first-order optimization algorithms which utilized some second-order Hessian queries. This method don't belong to the formal "second-order" optimization class.
A4: Thank you for your thoughtful comment. We respectfully clarify that the use of second-order information is the defining characteristic of second-order optimization methods, and such methods are not necessarily restricted to the Newton-type update form . For instance, cubic Newton methods and trust-region methods are widely recognized as second-order algorithms, yet they do not conform to this canonical iteration. Similarly, our proposed Normalized SGD with Hessian correction explicitly utilizes Hessian information to adapt the updates, thereby leveraging second-order curvature information. Therefore, while our method differs structurally from Newton and quasi-Newton schemes, it naturally belongs to the broader class of second-order optimization methods.
Q5: From Theorem 2 and Theorem 3, both the step size from the normalized SGD and the clipping threshold from the clipping technique requires the estimate the initial sub-optimality . How to estimate this initial sub-optimality in practice? Theorem 2 and 3 presented that and need to be set by values of this initial sub-optimality \delta to reach the corresponding sample complexity. Hence, the proposed methods are impractical since it is in general very difficult to estimate the initial sub-optimality accurately.
A5: We can propose two strategies. If a lower bound on is available (e.g., it is zero) we can replace with , where can be estimated, e.g., using a mini-batch. In general any upper bound on suffices. Another valid strategy that will work for Algorithm 1 is to set step-size , momentum if is known or , momentum if is unknown. Following the proof strategy similar to (Hübler et al, 2024), we can establish convergence of our algorithm. If is known the convergence rate of the algorithm will have the same dependence on and as Theorem 2, i.e., and if is unknown, the method will still converge but slower, i.e., .
Florian Hübler, Ilyas Fatkhullin, Niao He. From Gradient Clipping to Normalization for Heavy Tailed SGD. AISTATS 2025.
This paper studies second-order optimization algorithms with heavy-tailed noise. A lower bound on the sample complexity for second-order algorithms is established along with an algorithm to match it in terms of first moment error. A second algorithm with clipping with high probability guarantees for the estimation error is proposed and shown to achieve near-optimal sample complexity.
优缺点分析
This paper has interesting contributions which, as far as I am concerned, are indeed novel. The assumptions, contributions and setting of the paper are clearly identified early on, making it easy to read. I was not able to check the proofs in detail.
Regardless, I would encourage the authors to proof-read their paper as I found a few typos while reading it. (e.g. E[ || \nabla f(x,\xi) || ]^p \leq B when talking about bounds on pth moment).
One weakness of this submission is the lack of numerical studies to illustrate the theory developed in the paper. While, the authors provide two plots in page 3, very little discussion is provided regarding the setup of their experiments and how these plots are related to the theorems in their paper. I would encourage the authors to include a simple toy problem illustrating their sample complexity bound. Figure 1 shows a plot of the sample complexity, but not details on this experiment are provided.
In terms of clarity, I also think that there is some work needed. For example, the authors should explicit say which algorithm (i.e. first equation of section B) they focus on in the very beginning of the paper to engage the reader and avoid confusion. Also, the authors reference Table 1 in the main text, but do not mention this table is in the supplementary material, which might create confusion.
I think a statement on how the choice of g_0 changes the sample complexity of algorithm 2 should be included in the main text. Fine details do not need to be included, but the fact that it only changes a constant on the sample complexity and not the dependency on \epsilon.
问题
-
Is \xi assumed to be i.i.d.? If not, is the expectation taken at steady-state (if \xi is Markovian for example)?The authors need to explicitly state their assumptions on the noise.
-
Is the BCM assumption that much weaker than the p-moment bound assumption of [Zhang et al., 2020]? Why is the difference between these assumptions that important?
-What is the difference between O(.) and \Omega(.)? Why the need for two different notations? Also, Thm 1 uses \Theta(.), what is that?
局限性
The assumptions of this paper are clearly identified and the authors provide a discussion on thee limitations of their work.
最终评判理由
The authors' rebuttal addressed my concerns. I retain my initial score, which recommended acceptance.
格式问题
I have no concerns on this matter
Dear Reviewer JJgW,
Thank you for your review and for your positive evaluation of the novelty and clarity of our contributions. We are glad to hear that you found the setting, assumptions, and theoretical results clearly presented. We also appreciate your constructive suggestions regarding clarity, notation, and experimental presentation, and we address each of your comments and questions in detail below.
Q1: I would encourage the authors to include a simple toy problem illustrating their sample complexity bound.
A1: We want to mention that Figure 1 is not an experiment. It is an illustration of theoretically optimal complexity bounds (upper and lower bounds are matching) for SOSO and FOSO in order to support understanding. This plot implies that for the entire range of tail indices , methods with access to second-order information have reduced sample complexity than methods that only have access to first-order information. Figure 2 is a toy experiment, which motivates us to study second-order methods and algorithms with clipping in depth. The most important information related to this experiment is provided in the caption to Figure 2. During the rebuttal phase, we conduct the following additional ablation studies on the same toy problem, which provide insights on theoretically derived complexities and the sensitivity to clipping threshold tuning.
We solved a simple quadratic problem, , in dimension , with synthetic noise sampled from a symmetrized Pareto distribution. We considered both a fixed heavy-tailed regime with tail index and a variable tail index ranging from to . We start each algorithm from the same point sampled uniformly at random from the standard normal distribution. Our empirical analysis consists of three experimental settings:
1. Effect of Clipping Level on Iteration Complexity: We ran Algorithm 2 (Clip NSGDHess) with until it reached the target accuracy of . The stepsize and momentum parameters were set to and . Clipping levels were varied from to using a multiplicative step of 10. The results are shown below:
| Iterations |
Table 1: Intermediate clipping () yields the lowest iteration complexity. Extremely small or large values lead to slower convergence. The proposed algorithm can gracefully tolerate very small values of clipping parameter , but using values larger than leads to very slow convergence. This observation motivates the need for gradient and hessian clipping.
2.Comparison with Clip NSGDM under Varying Tail Index: We compared Clip NSGDHess with Clip NSGDM (as proposed in [2]). For Clip NSGDM, we used the stepsize and momentum according to theory. For Clip NSGDHess, both parameters were set to . We fixed , , and iterations.
| Tail index, | Iteration Complexity (Clip NSGDM) | Iteration Complexity (Clip NSGDHess) |
|---|---|---|
Table 2: Clip NSGDHess consistently outperforms Clip NSGDM, with the gap increasing for smaller , i.e., heavier-tailed noise. Visually the picture looks similar to theoretical complexities in Figure 1 of our original submission.
[2] Cutkosky, Ashok, and Harsh Mehta. "High-probability bounds for non-convex stochastic optimization with heavy tails." Advances in Neural Information Processing Systems 34 (2021): 4883-4895.
3.Sensitivity to Gradient Clipping for Fixed Hessian Clipping: We fixed the Hessian clipping level and varied the gradient clipping level :
- For : we set from to with multiplicative step ;
- For : we set from to with multiplicative step ;
- For : we set from to with multiplicative step .
In all cases, the algorithm was run until . We consistently observed that both extremely small and large values of led to higher iteration complexity. The optimal values were:
- for ;
- for ;
- for .
This pattern confirms that intermediate clipping levels are crucial for performance.
Q2: I think a statement on how the choice of g_0 changes the sample complexity of algorithm 2 should be included in the main text. Fine details do not need to be included, but the fact that it only changes a constant on the sample complexity and not the dependency on \epsilon.
A2: Thank you for the suggestion! We will add an additional discussion about it in the main text of the camera-ready version.
Q3: Is \xi assumed to be i.i.d.? If not, is the expectation taken at steady-state (if \xi is Markovian for example)? The authors need to explicitly state their assumptions on the noise.
A3: On each iteration we sample independently from the same distribution, so the sequence is assumed to be i.i.d. We will clarify this in the revision.
Q4: Is the BCM assumption that much weaker than the p-moment bound assumption of [Zhang et al., 2020]? Why is the difference between these assumptions that important?
A4: Our -BCM assumption is weaker than -th bounded moment. To see this, it is sufficient to show it for . By bias-variance decomposition On the other hand, -th bounded moment assumption can be restrictive as it implies bounded gradients even in case as can be seen from the above decomposition. The class of Lipschitz smooth functions with unbounded gradients is much richer than the one with bounded gradients. In particular, it includes an important sub-class of quadratic functions.
Q5: What is the difference between O(.) and \Omega(.)? Why the need for two different notations? Also, Thm 1 uses \Theta(.), what is that?
A5: Thank you for your question, we will add this in the notations section.
- if there exist and such that for all the following inequality holds, ;
- if there exist and such that for all the following inequality holds, ;
- if there exist and and such that for all the following inequality holds, ;
For more details, we refer to the book: Cormen, Thomas H., Charles E. Leiserson, Ronald L. Rivest, and Clifford Stein. Introduction to algorithms. MIT press, 2022.
I thank the authors for their detailed responses. I encourage them to include the numerical results discussed in their rebuttal in the final submission. I am maintaining my current score.
The paper is an analysis of second-order stochastic optimization under the assumptions of heavy tailed stochastic gradients and hessians: it proposes a normalized SGD algorithms that achieves the established lower bounds on the sample complexity, and a novel gradient and Hessian clipping strategy to cope with large deviation instabilities, also achieving nearly-optimal performances. The main novelty of the paper consists in a sound and rigorous theoretical analysis of this relevant set-up, which has been appreciated for its clarity and originality. I recommend however the authors to integrate the text with the numerical experiments discussed with the Reviewers, as one of the weaknesses pointed out has been in fact the minimal experimental validation. Nevertheless, the relevance of the topic — essential for the construction of more realistic theoretical setup and more effective algorithmic solutions — makes the paper a valid and impactful contribution, and therefore I lean towards its acceptance.