Nonlinearly Preconditioned Gradient Methods: Momentum and Stochastic Analysis
摘要
评审与讨论
This paper considers preconditioned gradient methods for nonconvex optimization problems under the generalized smoothness condition. It first introduces preconditioned gradient descent with momentum when the full gradient is available. In addition, the authors propose a stochastic variant of the preconditioned gradient descent method and study its convergence guarantee.
优缺点分析
Strengths:
- Extension of the preconditioned gradient descent to the stochastic setting is an interesting topic to explore.
- The authors provide a rigorous theoretical analysis of the proposed methods.
- This paper presents empirical evidence to demonstrate the effectiveness of the proposed methods.
Weakness:
-
The authors do not provide a comparison of the convergence rates of the proposed methods with existing approaches under the generalized smoothness setting. The theoretical advantages of the proposed methods over existing approaches are not clearly articulated.
-
In Theorem 3.1, the R.H.S. of (15) has a non-vanishing term of , which implies that the proposed stochastic preconditioned gradient descent cannot have exact convergence to the stationary point.
-
In Theorem 3.4, the choice of batch size is too large. To obtain -stationary point, one needs to choose a batch size , which is impractical for large-scale machine learning applications.
问题
- With an appropriate choice of the reference function , anisotropic smoothness is equivalent to -smoothness. The -smoothness framework already encompasses a wide range of machine learning applications. Therefore, a key concern is the significance of further generalizing from -smoothness to anisotropic smoothness in the context of machine learning. Specifically, I would like to know which machine learning applications fall within the class of anisotropically smooth functions but not within the class of -smooth functions.
局限性
The theoretical guarantees for the stochastic variants of preconditioned gradient descent appear to be limited, as highlighted in Weaknesses (2) and (3).
最终评判理由
The author addressed my concerns. I am satisfied with their response.
格式问题
No formatting concern.
We appreciate the reviewer’s time and effort in providing feedback.
In the following we address the Weaknesses stated by the reviewer.
The authors do not provide a comparison of the convergence rates of the proposed methods with existing approaches under the generalized smoothness setting. The theoretical advantages of the proposed methods over existing approaches are not clearly articulated.
We would like to stress that the notion of smoothness that we work with in this paper is different from that of other papers in the generalized smoothness framework. Therefore it is not straightforward to compare the obtained convergence rates except for specific cost functions. Nevertheless, a major advantage of our work is that it provides a unified analysis for a family of algorithms, thus eliminating the need to produce new results for every different setting.
In Theorem 3.1, the R.H.S. of (15) has a non-vanishing term of , which implies that the proposed stochastic preconditioned gradient descent cannot have exact convergence to the stationary point.
Indeed, in Theorem 3.1 we show convergence up to under a new noise assumption that is less restrictive than bounded variance and when the stochastic estimator is potentially biased. We believe that this is an interesting standalone result (such as [3, Section 4.1]), but it also helps provide some insight on our main result, Theorem 3.4 (see also our reply to reviewer 28hb).
In Theorem 3.4, the choice of batch size is too large. To obtain -stationary point, one needs to choose a batch size , which is impractical for large-scale machine learning applications.
Theorem 3.4 requires large batch sizes to obtain convergence up to , which to the best of our knowledge is standard for all clipping-type methods under (generalized) smoothness that do not utilize some variance reduction technique or momentum (see [1, Section 3.1], [2, Section 2], [3, Section 4.1]). Using such a technique to remove the need for a large batch size, possibly under a weaker noise assumption, is an important research direction that we believe deserves separate work.
In what follows we answer the Question asked by the reviewer.
1. With an appropriate choice of the reference function , anisotropic smoothness is equivalent to -smoothness. The -smoothness framework already encompasses a wide range of machine learning applications. Therefore, a key concern is the significance of further generalizing from -smoothness to anisotropic smoothness in the context of machine learning. Specifically, I would like to know which machine learning applications fall within the class of anisotropically smooth functions but not within the class of -smooth functions.
We would like to stress that although a function that is -smooth is also anisotropically smooth relative to a suitably chosen reference function, the corresponding descent inequalities are in general not equivalent. In fact for this specific , anisotropic smoothness is less restrictive than -smoothness, as we detail in our answer to reviewer VEKm. Moreover, to the best of our knowledge, although -smoothness has been shown empirically to better describe some machine learning applications, proving these empirical findings remains an open question.
Regarding the significance of anisotropic smoothness as a notion of generalized smoothness, we would like to stress that as we state in the introduction (lines 42-46) our goal is not only to obtain the most general descent inequality but also to fit the best one to the cost function at hand. From a majorization-minimization perspective we want to find a tight model of the cost function in order to obtain faster algorithms that do not require additional tuning. Therefore, similarly to how -smooth functions also belong to the class of -smooth functions (for ), but the standard approach to minimizing them is standard GD with stepsize , there exist functions that are both anisotropically smooth and -smooth, but the anisotropic descent inequality describes a tighter model and thus the best approach is the nonlinearly preconditioned gradient method. We next provide some simple examples of 1d functions that are provably anisotropically and -smooth, to better motivate this.
- The function is both anisotropically and -smooth, but the method described in Equation (1) of the paper for converges in just one iteration.
- The function is a simple 1d nonconvex function that is anisotropically smooth relative to with and thus we can use HGD to minimize it. It is also -smooth with constants , where we can freely set . We can thus use the algorithm described in [4, Equation (3.2)] to minimize this function. We compare the two methods and present the results in the following table, where the number denotes the required iterations to achieve accuracy .
| Initial point | HGD | [4, Equation (3.2)] , | [4, Equation (3.2)] , |
|---|---|---|---|
| 10 | 25 | 55 | |
| 21 | 83 | 84 | |
| 32 | 155 | 120 |
These simple examples showcase the existence of functions that are better described by the anisotropic descent inequality.
We believe that we have addressed all the reviewer's concerns and weaknesses, and we kindly ask the reviewer to reevaluate their rating.
References:
[1] Hübler, Florian, Ilyas Fatkhullin, and Niao He. "From gradient clipping to normalization for heavy tailed sgd." arXiv preprint arXiv:2410.13849 (2024).
[2] Kornilov, Nikita, et al. "Sign Operator for Coping with Heavy-Tailed Noise in Non-Convex Optimization: High Probability Bounds Under -Smoothness." arXiv preprint arXiv:2502.07923 (2025).
[3] Yang, Junchi, et al. "Two sides of one coin: the limits of untuned SGD and the power of adaptive methods." Advances in Neural Information Processing Systems 36 (2023): 74257-74288.
I have some extra questions:
-
Regarding the response to my question, what is the reference of [4]?
-
You mention that From a majorization-minimization perspective we want to find a tight model of the cost function in order to obtain faster algorithms that do not require additional tuning. Is there any theoretical justification to show that your algorithm is faster than other algorithms?
-
You mention that Therefore it is not straightforward to compare the obtained convergence rates except for specific cost functions. Could you provide some examples of the specific functions, and compare the obtained convergence rates with existing methods?
Dear reviewer tUyH
Regarding the response to my question, what is the reference of [4]?
See below for the reference of [4]. We apologize for the omission.
You mention that From a majorization-minimization perspective we want to find a tight model of the cost function in order to obtain faster algorithms that do not require additional tuning. Is there any theoretical justification to show that your algorithm is faster than other algorithms?
We would like to stress that our analysis incorporates a generalized PL condition under which we obtain global linear rates with constants that do not depend on the initialization. This implies that our algorithm is in fact provably faster than existing methods for various functions, where the methods for have sublinear rates or at most linear rates with constants depending on the initialization (see for example [5]). Moreover for various examples, our obtained rates are better than those in the related literature, see also our reply to the next question.
You mention that Therefore it is not straightforward to compare the obtained convergence rates except for specific cost functions. Could you provide some examples of the specific functions, and compare the obtained convergence rates with existing methods?
In the following we answer the remaining questions. To begin with, since anisotropic smoothness and -smoothness form different classes of functions, the cost functions have to be specified in order to compare the obtained convergence guarantees. More concretely, consider the reference function . Then, the convergence guarantees for the stationarity measure can be translated to the standard one, and we obtain standard sublinear rates albeit with different constants than those of the -literature. We next provide some standard examples from the -literature, where our methods are provably faster.
- (which can be found in [6, Theorem 1]). This function is 1-anisotropically smooth w.r.t. and satisfies the generalized PL inequality with constant 1. This implies the convergence of the base method in Equation (1) in one iteration and of our method Algorithm 1 with a linear rate from Theorem 2.3. The same can be said for w.r.t. . Moreover, is also anisotropically smooth w.r.t. with constant and satisfies the generalized PL inequality, thus implying a global linear convergence rate. Since and are strongly convex functions, the analysis of [5] also applies but leads to linear convergence constants that depend on the initialization.
- Consider any 1d polynomial function that is -smooth, i.e. (found e.g. in [7, Lemma 2]). Then it is anisotropically smooth w.r.t. , with constant . Using simple algebra, we find that our Theorem 2.1 describes the following optimal (choosing ) rate: . For , this is a better rate than the best known one for -smoothness [4, Theorem 3.1].
We can thus see that anisotropic smoothness is not only more general than -smoothness (see also our reply to reviewer VEKm) but also leads to faster algorithms.
References
[4] Vankov, Daniil, et al. "Optimizing -Smooth Functions by Gradient Methods." arXiv preprint arXiv:2410.10800 (2024).
[5] Lobanov, Aleksandr, et al. "Linear Convergence Rate in Convex Setup is Possible! Gradient Descent Method Variants under -Smoothness." arXiv preprint arXiv:2412.17050 (2024).
[6] Chen, Ziyi, et al. "Generalized-smooth nonconvex optimization is as efficient as smooth nonconvex optimization." ICML (2023).
[7] Zhang, Jingzhao, et al. "Why gradient clipping accelerates training: A theoretical justification for adaptivity." arXiv preprint arXiv:1905.11881 (2019).
Thanks for your response! It has addressed my concerns, so I have raised my rating.
This paper presents a theoretically rigorous and practically motivated study of nonlinearly preconditioned gradient methods (NPGMs), specifically extending them with momentum (m-NPGM) and developing a stochastic variant. The paper analyzes convergence under anisotropic smoothness, a generalization of Lipschitz smoothness, and establishes novel results under generalized PL conditions and (L0, L1)-smoothness. It also provides empirical comparisons on deep learning and matrix factorization tasks.
优缺点分析
Strength: 1) Theory: This work introduces a novel heavy-ball extension of nonlinear preconditioning methods with provable convergence guarantees in the general nonconvex setting under a generalized PL condition. This is the first result regarding the convergence of the method with momentum under anisotropic smoothness conditions. 2) Assumption and setting: the analysis is established under anisotropic smoothness and preconditioned Lipschitz conditions, which cover many applications. In particular, the analysis is a non-trivial extension of the previous work due to the particular structure of the anisotropic descent inequality and the lack of global upper bounds for the gradient difference term. Therefore, the techniques developed here have the potential to be applied to develop and analyze other algorithms in this setting.
Weakness: The presentation of the content lacks details. All the main results are presented right after the assumptions/conditions, without any elaboration on the proof sketch and technical novelty. The paper discusses a variety of settings, but could focus on one or two of them to further elaborate the technical details. The experimental section could be strengthened with ablation on choice of reference function and effect of momentum coefficient on convergence, etc.
问题
see above
局限性
see above
最终评判理由
I want to thank the authors for providing the response. I am satisfied with the response and have raised the rating to 4.
格式问题
no
We appreciate the reviewer’s time and effort in providing feedback.
In the following we address the Weaknesses stated by the reviewer.
The presentation of the content lacks details. All the main results are presented right after the assumptions/conditions, without any elaboration on the proof sketch and technical novelty. The paper discusses a variety of settings, but could focus on one or two of them to further elaborate the technical details.
We appreciate any feedback on the presentation of our work. However, in this case the comment feels unjustified, as we motivate below.
One of our main theorems is Theorem 2.1 and the three paragraphs that follow it are dedicated to discussing both the difficulties of the setting that we study and its limitations. Then, we present the generalized PL condition and briefly explain it in one paragraph, but we provide additional details on it and further study its relation to the classical PL condition in Appendix D. We do not discuss the difficulty of the setting after Theorem 2.3, since this is similar to Theorem 2.1.
We then move on to establishing the preconditioned Lipschitz continuity assumption, which we state after an introductory paragraph. Following the statement of the assumption we show that it holds for Lipschitz smooth functions and even for -smooth ones, thus showcasing its generality and applicablity. Then, we state our convergence theorem under this new assumption. We do not provide a discussion after this statement due to spacing constraints, but our proofs are presented in full detail in the appendix.
Regarding Section 3, we present it in a way such that the main convergence result Theorem 3.4 follows naturally. The logic of the presentation is as follows: we first show in Theorem 3.1 the convergence of the method up to under a new noise assumption. Then, in Proposition 3.2 we show that this noise assumption is actually implied by the standard bounded variance assumption and in Example 3.3 we provide an example where the bounded variance assumption fails and our noise assumption holds. Following that we provide the convergence result Theorem 3.4 where we assume bounded variance and that we have access to a minibatch oracle. Finally, we discuss the limitations and advantages of our approach.
In general, we agree that in some places of the manuscript we could provide more details and the page limit constraint is the main reason we have not done so. Nevertheless, we do not believe that this is a valid reason to reject the paper.
The experimental section could be strengthened with ablation on choice of reference function and effect of momentum coefficient on convergence, etc.
Regarding the choice of preconditioner in our experiments, please see our comment on weakness 3 stated by reviewer A43W. As for the ablation study on the momentum parameter, we consider this an interesting research direction, as the current paper primarily focuses on theoretical analysis.
We believe that we have addressed all the reviewer's concerns and weaknesses in our paper, and we kindly ask the reviewer to reevaluate their rating.
Dear reviewer 28hb
Could you let us know if your concerns have been addressed and whether you have any further questions?
This paper investigates smooth noncovnex optimization problem using anisotropic gradient descent (GD) method & its extensions. They combine the notion of momentum with nonlinear preconditioned GD & propose a Heavy-Ball (HB)-type method with a convergence rate of for a strongly convex reference function & being anisotropically smooth relative to (a notion that was proposed by previous works). Later, they extend their convergence result to the scenario when satisfies a generalized PL inequality and is 2-subhomogenous. Additionally, they study their method for functions that satisfy a preconditioned-smoothness assumption which contains -smoothness as a special case. A convergence analysis of the stochastic preconditioned GD method under a more general noise assumptions revealed that the convergence is not guaranteed without variance reduction techniques (like increasing batch sizes). Numerical experiments revealed competetive behaviour of the proposed method.
优缺点分析
Strengths
1- The problem is well-motivated. Although, the gradient clipping method is more explored and studied in the literature, its general form (1) is less attended. This is the general motivation in this work. (Lines 42-54)
2- The contribution is good. To further study the general form of nonlinear preconditioned GD, they propose an HB-type method with its convergence analysis under the assumptions of anisotropic smooth relative to (lines 155-156 theorem 2.1), a generalized PL inequality where is 2-subhomogenous (lines 197-201 theorem 2.3), and a preconditioned Lipschitz continuity (lines 225-228 theorem 2.6) . A stochastic extension is also studied (for the first time as the authors claim) under a slightly more general noise assumption (lines 241-243 theorem 3.1). Additionally, they show that if the mini-batch size increase (classical approach to reduce the variance estimation error), the true convergence is possible in theorem 3.4.
Weaknesses
1- The writing could improve, sometimes the reader might feel lost or diconnected. Adding a bit more coherency might help with this (this point exists throughout the whole paper. Examples are Paragraph on line 22 is incoherent with the previous paragraph, paragraph on line 203 hard to read sentence on line 207).
2- ADAM is a momentum-based algorithm, however it was not compared in the Numerical simulations (Figure 1, right).
问题
First, thank you for your efforts.
1- In line 117 it is mentioned that this work mainly focuses on strongly convex reference functions. Then they propose some examples. What are other examples of strongly convex reference functions?
2- Did you ensure that all the methods in your simulations are compaired fairly in terms of the choice of the step-sizes? I'm mainly referring to the neural network simulations where the step-size was chosen through sweeping on .
3- Why ADAM was not compared in the training on CIFAR10 ResNet18 experiment with momentum?
4- Do you think it would be possible to extend these results to the constrained optimization?
Thanks!
局限性
Yes
最终评判理由
After reading all the comments, I decide to raise my score to 5 as I see clear positive responses from other reviewers who responsed. The authors have clearly addressed my concerns regarding their simulation results. Due to comparative novelty aspects existing in NeurIPS, I do not raise my score to 6 as this work is developing on the same line of prior works like reference [30]. However, still they bring valuable results as mentioned in the "Strengths" section of my initial comment.
格式问题
No formatting issues
We thank the reviewer for their positive evaluation of our paper, and we appreciate their recognition of the significance of our work.
In the following we address the Weaknesses stated by the reviewer.
1. The writing could improve, sometimes the reader might feel lost or diconnected. Adding a bit more coherency might help with this (this point exists throughout the whole paper. Examples are Paragraph on line 22 is incoherent with the previous paragraph, paragraph on line 203 hard to read sentence on line 207).
We acknowledge that our writing could improve at places in the manuscript. We will refine the writing by simplifying the language and providing additional details.
2. ADAM is a momentum-based algorithm, however it was not compared in the Numerical simulations (Figure 1, right).
We did not include Adam in Figure 1 (right) because this experiment is the same (Resnet18 on the Cifar10 dataset) as the one shown in Figure 1 (center), where Adam is already included. We aimed to avoid overcrowding the image on the right.
In what follows we answer the Questions asked by the reviewer.
1. In line 117 it is mentioned that this work mainly focuses on strongly convex reference functions. Then they propose some examples. What are other examples of strongly convex reference functions?
There are many interesting strongly convex reference functions. In order to avoid repetition we only present the 1d kernel functions that generate , along with the corresponding preconditioner , in the following table:
We remark that the second and third reference functions lead to a simplified version of Adagrad and the gradient clipping algorithm respectively through [1, Examples 1.5 & 1.7].
2. Did you ensure that all the methods in your simulations are compaired fairly in terms of the choice of the step-sizes? I'm mainly referring to the neural network simulations where the step-size was chosen through sweeping on .
In order to clear possible confusion we remark that except for the experiment with momentum, in all the other ΝΝ experiments we tuned the methods via an adaptive gridsearch. We did not use the same gridsearch for the experiment with momentum to lower the computational cost, since SGDm achieves state-of-the-art performance with these parameters.
3. Why ADAM was not compared in the training on CIFAR10 ResNet18 experiment with momentum?
Please see our comment on the second weakness.
4. Do you think it would be possible to extend these results to the constrained optimization?
Indeed, we believe that an extension to the constrained case is certainly possible and is interesting future work. We remark that another advantage of the anisotropic smoothness framework that we study is that it has a natural proximal extension for additive nonsmooth problems [2].
References:
[1] Oikonomidis, Konstantinos, et al. "Nonlinearly Preconditioned Gradient Methods under Generalized Smoothness." ICML (2025).
[2] Laude, Emanuel, and Panagiotis Patrinos. "Anisotropic proximal gradient." Mathematical Programming (2025): 1-45.
I thank the authors for their detailed explanation and clarifications.
Just one more clarification: The caption of Figure 1 states that the simulation in the middle is for methods without using momentum and it includes Adam, (I suppose you discarded the momentum parameter in your implementation), then the figure on the right is for the methods with momentum, but it does not include Adam. Do you mean that momentum does not play a crucial role for the Adam? Or the figure in the middle is Adam with momentum?
Dear reviewer QeMc
The middle figure includes Adam with the standard momentum parameters . See line 34 of the file neural_networks/main.py in the supplementary material:
return problem(get_device(), lambda params: torch.optim.Adam(params, lr)).train(epochs=epochs)
To avoid confusion, we will move Adam to the rightmost figure.
We apologize for the misunderstanding.
Thank you for your response.
Yes, I believe that would be improving the presentation there. I might raise my score in the next phase.
This paper proposes a nonlinearly preconditioned heavy-ball-type algorithm and provides its convergence guarantees in the general nonconvex setting and under a generalized PL condition. It also presents a stochastic variant and analyzes its convergence behavior under both new and standard noise assumptions. Finally, it tests the proposed methods and shows their good performance on various tasks, including neural network training and matrix factorization.
优缺点分析
Strengths
-
This paper extends the nonlinearly preconditioned gradient descent method by incorporating a momentum term and studies the convergence behavior under different conditions.
-
This paper also proposes a stochastic variant of the base method and analyzes it under different noise assumptions.
-
The numerical experiments consider two different applications and show the good performance of the proposed methods.
Weaknesses
-
Some concepts are used without any explanations, such as -convexity. It may be better to give a brief introduction in the appendix.
-
It is better to compare the standard convergence analysis of gradient clipping under the -smooth condition and the one under this anisotropic smoothness. For example, does the paper obtain a stronger result in the stochastic setting? Does the analysis become simpler? Are the assumptions weaker? I think the current manuscript lacks a discussion on this aspect.
问题
- It is better to explain sigmoid preconditioners in the main text.
- Is there any relationship between the so-called Hyperbolic Gradient Descent with the gradient descent in hyperbolic space [1]?
- Is it better to use instead of ?
- To obtain the next iterate, Equation (2) is minimized over what?
- Define in Line 103
- Define in Equation (3)
- It is better to give the formulation of the stochastic optimization problem at the beginning of Section 3.
- What about the numerical performances of algorithms induced by other reference functions?
- Is it possible to include numerical experiments on PL-type functions?
[1] Wilson, B., & Leimeister, M. (2018). Gradient descent in hyperbolic space. arXiv preprint arXiv:1805.08207.
局限性
Yes.
最终评判理由
I believe the authors have addressed all my concerns. They are also committed to incorporating several discussions into the revised manuscript.
格式问题
I have not found any major formatting issues in this paper.
We appreciate the reviewer’s time and effort in providing feedback and their acknowledgement of our papers' novelty, quality and clarity.
In the following we address the Weaknesses stated by the reviewer.
1. Some concepts are used without any explanations, such as -convexity. It may be better to give a brief introduction in the appendix.
We would like to thank the reviewer for their suggestion. Indeed, we will provide additional details on the important concept of -convexity in a revised version of the manuscript.
2. It is better to compare the standard convergence analysis of gradient clipping under the -smooth condition and the one under this anisotropic smoothness. For example, does the paper obtain a stronger result in the stochastic setting? Does the analysis become simpler? Are the assumptions weaker? I think the current manuscript lacks a discussion on this aspect.
We would like to stress that although there exist connections between the two notions of smoothness, they are not equivalent and thus directly comparing the analysis of the methods is not straightforward.
Are the assumptions weaker?
From [1, Proposition 2.9] we know that anisotropic smoothness encompasses a class of functions that is wider than that of -smooth functions. In other words, anisotropic smoothness is a weaker assumption than -smoothness. To better see this, note that for , the second-order sufficient condition for anisotropic smoothness from [1, Definition 2.4] takes the following form:
while for -smoothness it is . Therefore, we can see that this second-order sufficient condition 1) does not require lower bounds on the eigenvalues of and 2) can be less restrictive than the one for -smoothness stated above. Some simple examples of functions that are anisotropically smooth but not include , , (by choosing as the function itself). It is also important to note that our preconditioned Lipschitz continuity Assumption 2.4 is at least as general as -smoothness and can be even less restrictive when the Jacobian of is symmetric, following the discussion on lines 204-212 of the manuscript.
We remark that another advantage of the anisotropic smoothness framework is that it can better describe the functions that are between Lipschitz smoothness and -smoothness in that it generates tighter upper bounds. For further details please see our reply to reviewer tUyH.
Does the analysis become simpler?
Our analysis covers a whole family of algorithms (parameterized by the reference function ) under a very general condition and can thus be considered simpler, since it does not require further effort for every specific instance of the algorithm and the corresponding descent inequality, as is standard in the related literature. Moreover, it is illuminating in that it links the choice of the preconditioner (and thus algorithm) with the function class of the cost function.
Does the paper obtain a stronger result in the stochastic setting?
In the stochastic setting we obtain standard convergence guarantees for this type of methods (without some variance reduction technique) [3, Section 4.1], but under a different condition.
We aim to add a discussion on this subject in a revised version of the manuscript.
In what follows we answer the Questions asked by the reviewer.
- We thank the reviewer for their suggestion, we will include some discussion on the sigmoid shape of the preconditioners in the main text.
- We are not aware of any relationship between HGD and gradient descent in hyperbolic space. However, the general update described in (1) is connected to works from the geometry of optimal transport literature [4, 5].
- Thank you for the suggestion, we will replace with .
- The main iterate is generated by minimizing the r.h.s. of the inequality (2) over . We will specify this in the manuscript.
- We will include the suggested changes in a revision.
- We will include the suggested changes in a revision.
- We will include the suggested changes in a revision.
- In our paper we focus mostly on the algorithms generated by , since their behavior is between that of plain GD and gradient clipping (see also the response to reviewer A43W). We did not provide experiments with reference functions that have bounded domains, since they perform some form of gradient clipping for which there already exists a rich literature.
- We believe that in order for such experiments to be meaningful, the algorithms should be applied with the correct stepsize obtained by the theory. For that reason we have left such a project for future work. Nevertheless a toy experiment is presented in our reply to reviewer tUyH, where the function satisfies the generalized PL inequality along the sequence of iterates.
We believe that we have addressed all the reviewer's concerns and weaknesses, and we kindly ask the reviewer to reevaluate their rating.
References:
[1] Oikonomidis, Konstantinos, et al. "Nonlinearly Preconditioned Gradient Methods under Generalized Smoothness." ICML (2025).
[2] Laude, Emanuel, and Panagiotis Patrinos. "Anisotropic proximal gradient." Mathematical Programming (2025): 1-45.
[3] Lu, Haihao, Robert M. Freund, and Yurii Nesterov. "Relatively smooth convex optimization by first-order methods, and applications." SIAM Journal on Optimization 28.1 (2018): 333-354.
[4] Gangbo, Wilfrid, and Robert J. McCann. "The geometry of optimal transportation." (1996): 113-161.
[5] Figalli, Alessio, Young-Heon Kim, and Robert J. McCann. "When is multidimensional screening a convex program?." Journal of Economic Theory 146.2 (2011): 454-478.
I am satisfied with the authors' responses and decide to raise my rating to 4.
The paper proposes nonlinearly pre-conditioned gradient methods (NPGMs) that combine the dual-space/anisotropic-descent framework with heavy-ball momentum and with a stochastic variant. It establishes global convergence bounds for the momentum version under anisotropic smoothness and shows a linear rate when a generalized PL condition holds. For the stochastic setting, the authors derive non-asymptotic rates under a novel noise assumption that subsumes bounded variance and prove linear convergence under generalized PL. Finally, they include experiments on MNIST/CIFAR-10 classification and matrix-factorization tasks.
优缺点分析
Strengths
- The paper is well structured and has a nice flow.
- They provide rigorous proofs for all stated theorems, and the assumptions are clearly presented. In particular, they both deterministic and stochastic analyses, including mini-batch dependence.
- The authors give the first convergence proof for heavy-ball momentum inside anisotropic preconditioning setting.
Weaknesses
- In the statement of Theorem 2.1 the authors make the restriction , limiting practical momentum choices (usually ), however, they note this as an open question.
- The stochastic results (Section 3) cover only the base (no-momentum) method. Following the work from Section 2, one would expect results for the momentum-based method. After all, this is what practitioners usually deploy.
- Empirical accuracy gains over tuned SGD/Adam are modest on CIFAR-10; error bars overlap after 200 epochs.
问题
See weaknesses
局限性
Yes
最终评判理由
The authors have made an adequate job addressing my concerns. Therefore I maintain my initial rating of 4.
格式问题
None
We thank the reviewer for their positive evaluation of our paper.
In the following we address the Weaknesses stated by the reviewer.
1. In the statement of Theorem 2.1 the authors make the restriction , limiting practical momentum choices (usually ), however, they note this as an open question.
Indeed, Theorem 2.1 covers momentum parameters , a limitation that we acknowledge in our manuscript. Nevertheless, in Theorem 2.6 we provide convergence guarantees of the method for under a condition that we show is more general than standard Lipschitz smoothness and is at least as general as -smoothness. This condition can also be less restrictive than -smoothness, when the Jacobian of is symmetric, following the discussion in lines 204-212. In that sense our result Theorem 2.6 is state-of-the-art, especially considering that it covers the whole family of algorithms described in Algorithm 1 and it does not require considering specific instances (e.g. normalized gradient) as is standard in the literature. Therefore, Theorem 2.1 describes an even broader result, since it holds under an even less restrictive condition, with the caveat that .
2. The stochastic results (Section 3) cover only the base (no-momentum) method. Following the work from Section 2, one would expect results for the momentum-based method. After all, this is what practitioners usually deploy.
In this paper we have indeed not studied the proposed algorithm with momentum in the stochastic setting. We believe that this is an important research direction that deserves separate work, especially when considering that our paper is the first that studies the momentum extension and the stochastic version of the base method in the general nonlinearly preconditioned gradient framework.
3. Empirical accuracy gains over tuned SGD/Adam are modest on CIFAR-10; error bars overlap after 200 epochs.
In our experiments we have focused on the algorithms generated by and (called HGD) which we find interesting because they behave in a way that is between standard GD and gradient clipping. This intermediate behavior is especially clear in one dimension, where it can be illustrated by plotting the preconditioner . In that sense, the aim of our experimental section is to show that it keeps the best of both worlds in that it performs similarly to or even better than state-of-the-art methods in a variety of problems including NN training, while at the same time having stronger theoretical guarantees than GD. Therefore, HGD is a robust solver that can tackle a variety of problems, supported by theoretical guarantees that extend beyond the usual Lipschitz smoothness assumptions.
I thank the authors for their response. I will maintain my positive score.
This paper studies preconditioned gradient methods for smooth nonconvex optimization problems. This work is a follow up to the work of Oikonomidis, Quan, Laude, and Patrinos [30] who introduces the idea of preconditioned gradient methods which generalize gradient clipping techniques. Moreover, the assumptions which they prove convergence generalizes (L_0,L_1)-smooth functions.
In this paper, they analyze (i) a heavy ball variant of the method (algorithm 1) and (ii) a stochastic version of the method (which as theorem 3.4 specifies requires very large minibatches such that the gradient noise is less than 1/K where K is the number of iterations). Their numerical results are okay with no meaningful gains training CIFAR10/MNIST (figure 1) but significant gains demonstrated on matrix factorization with the Movielens 100K dataset (figure 2). One weakness of these empirical results is that there is limited modern relevance of gains on this task.
Comments:
Please explain clearly how heavy tail distributions are captured by your assumption that E[\phi(\grad \phi*(\grad f(x)))] <= \sigma^2.
“We set the learning rate by performing a parameter sweep over {1, 0.1, 0.01}.” Use a finer grid.