Provable Benefit of Adaptivity in Adam
We show that Adam converges under non-uniform smoothness.
摘要
评审与讨论
This paper provides the convergence analysis of Adam under ()-smooth condition.
优点
-
Beyond the -smooth condition, the authors analyze the convergence of Adam under )-smooth condition. Their analysis yields a specific convergence rate (to small error).
-
Moroever, the authors also provide a function on which Adam works but SGD fails.
缺点
-
Several assumptions in this paper differ significantly from standard analysis. Notably, in Assumption 3.2, the constant should be dependent on the parameter , and when . According to the main result in Theorem 4.1, the RHS is on the order of , which falis for large , a critical regime in stocastic optimization.
-
While the authors provide an example that Adam works but SGD fails, there may exist more examples under which Adam fails but SGD works. The results in this paper do not provide insights into the conditions under which Adam is likely to fail, as well as the efficiency of Adam in deep learning.
-
The title suggests a focus on the benefits of "Adaptivity," but this paper lacks a thorough explanation regarding these advantages.
问题
Please see "Weaknesses".
We express our appreciation to the reviewer for your constructive feedback. We will thoroughly revise our paper based on these valuable suggestions. However, due to the brevity of the rebuttal period, we are unable to complete this process in time and have decided to withdraw the paper.
In this paper, the authors studied the theoretical properties of Adam optimizer. For the existing theoretical analysis, they rely on the bounded smoothness assumption (L-smooth condition), which is not practical in real-world machine learning models. Therefore, the authors proposed the very first convergence result of Adam without assuming the L-smoothness. The L-smoothness is replaced by a (L_0, L_1)-smooth condition, where the norm of Hessian matrix can be upper bounded by a linear combination of constant and gradient norm.
优点
- The writing of this paper is very clear. The theoretical proofs are sound and solid. Although I didn't go through all the proof details in the appendix, I believe all these proofs are theoretically correct.
- The reference list of this paper is also very complete.
- The biggest strength is that the authors empirically verified the (L_0, L_1)-smooth condition under the case of LSTM and Transformers, which are both widely used model architecture.
缺点
- The authors should pay more attention on the details of paper writing. For example, in the last line of Page 2, there is a reference error (Appendix ??).
问题
For me, there are no additional questions and suggestions. I think this paper worth an "accept" because of its theoretical soundness.
We express our appreciation to the reviewer for your constructive feedback. We will thoroughly revise our paper based on these valuable suggestions. However, due to the brevity of the rebuttal period, we are unable to complete this process in time and have decided to withdraw the paper.
This paper studies convergence of the Adam algorithm. Previous work shows that Adam can converge to a bounded region around a stationary point under L-smooth condition. However, L-smoothness does not hold in practical settings (e.g., training of DNNs), so the main focus in this paper is to weaken the assumption to -smoothness. The authors show that Adam can converge to a bounded region even under this weaker assumption. Moreover, they also demonstrate the Adam's benefit over SGD by comparing the upper bound of Adam's convergence rate with the lower bound of the SGD's.
优点
- They prove the convergence of Adam under the -smoothness condition, which is novel.
- Benefits of Adam over SGD is explained via the derived convergence bound.
缺点
- Theoretical contribution is a little minor. Basically, the result is an extension of [1]. Analysis under the -smooth assumption is a novel point, but technical novelty of their proof is limited.
- They do not take the bias correction technique of Adam into consideration, which is widely used in practice. I think the result will not be so different even when the bias correction is incorporated, but they should mention it at least.
- They mention that the first term on the right hand side of Eq. (4) is . This means that the convergence of Adam deteriorates when is chosen to be close to , which seems to contradict to practical observations. They should add discussion on it.
[1] Zhang, Yushun, et al. "Adam can converge without any modification on update rules." Advances in Neural Information Processing Systems 35 (2022): 28386-28399.
问题
- As I pointed out in the Weakness section, the first term on the right hand side of Eq. (4) is , which means that the convergence of Adam deteriorates when is chosen to be close to , which seems to contradict to practical observations. How can the gap between theory and practice be explained?
We express our appreciation to the reviewer for your constructive feedback. We will thoroughly revise our paper based on these valuable suggestions. However, due to the brevity of the rebuttal period, we are unable to complete this process in time and have decided to withdraw the paper.
This manuscript provides the convergence analysis for randomly shuffled Adam under the -smooth condition and stochastic gradient noise with affine variance. The authors also refine the existing lower bounds of SGD by following existing work [Zhang et al., 2019a].
优点
The paper studied an interesting and important problem. Understanding Adam in deep learning is very important.
缺点
- One of the main theorems (Theorem 4.1) is very weak.
(i) It says that unless , the convergence rate of Adam is worse than SGD. One has to choose very large to converge to -stationary points. As the second term in RHS of formula (4) is of order , and one has to choose to make the second term of (4) on the RHS to be small, then has to be , and hence has to be to make the first term of RHS of (4) to be less than . This is much worse than the complexity of SGD. In this setting, the SGD complexity indeed depends on the magnitude of gradient on a sublevel set, but this quantity can be bounded by smoothness constants and independent of . The dependency of SGD is still . Therefore this contradicts the main claim of this paper about the "provable benefit of adaptivity in Adam".
(ii) The Theorem 4.1 requires gradient calls to find -stationary points, because the RR-Adam algorithm described in Algorithm 1 denotes as as epochs, not iterations. The hidden dependency makes this paper extremely weak. For example, there is a simple baseline that significantly dominates RR-Adam analysis in this paper: one can run gradient descent with clipping as in [Zhang et al. 2019a] in a deterministic case (e.g., every epoch), then the number of gradient oracle calls is at most ) (Theorem 3 in Zhang et al. 2019a).
(iii) The provable benefit of Adam is proved in [Li et al. 2023] and they prove the gradient complexity of Adam in the setting of stochastic optimization, which I believe is more general than the finite-sum setting as considered in this manuscript and they achieved a much better complexity results.
-
The presentation of the proof of Theorem 4.1 is very difficult to read. There are too many "constants" (e.g., from to on Page 18) which depend on each other. I suggest the authors to highlight which term is dominating and what are their orders.
-
The lower bound results are standard extensions of existing lower bounds in [Zhang et al. 2019a, Crawshaw et al. 2022]. As I saw from the proof, the hard instances are the same and the proof idea is also closely following these prior works. The only main difference is concatenating two 1-dimensional instances from previous works to a single 2-dimensional instance. I think this is a very marginal contribution.
Overall, although this paper studied an interesting problem, the upper bound results are too weak and dominated by previous works, and the lower bound results come from straightforward extensions from previous works. Therefore I recommend rejecting this paper.
问题
Can the authors address my questions? I am happy to increase my score if my concerns can be addressed.
伦理问题详情
N/A
We express our appreciation to the reviewer for your constructive feedback. We will thoroughly revise our paper based on these valuable suggestions. However, due to the brevity of the rebuttal period, we are unable to complete this process in time and have decided to withdraw the paper. We would like to address a point of contention regarding the comparison of our work with that of [Li et al., 2023], which we believe to be both unfair and inappropriate.
-
Firstly, our draft was submitted to flagship conferences nearly a year before [Li et al., 2023] made their submission to arXiv, and our work was even cited by them. We believe that drawing a comparison under such circumstances is not correct.
-
Secondly, the analysis presented in our paper holds true when . Here, represents a term added to to improve numerical stability and can be as minimal as . Contrarily, Li et al., [2023] operates under the assumption that , which creates a vastly different framework. This assumption allows the adaptive learning rate to be instantly upper-bounded, significantly simplifying the analysis. Conversely, in our proof, deriving the upper bound of presents a central and complex challenge. The bounds provided by [Li et al., 2023] show a polynomial dependence on , which could potentially result in an extremely loose bound. We acknowledge and appreciate the contribution of [Li et al., 2023] in providing an rate in the stochastic setting. However, due to the stark differences in setup, as discussed above, comparing our work with theirs is not even meaningful.