Adaptive Bilevel Optimization
摘要
评审与讨论
The paper studies deterministic Bilevel optimization problems of the form
$
&\min_{x} f(x; y^*(x)),\text{ where }y^*(x)=\arg\min_{y}g(x;y)
$
An algorithm is proposed which adapts to an unknown smoothness constant, assuming a known strong-convexity constant. The algorithm guarantees in the convex bounded loss setting and in the non-convex unconstrained setting.
优点
- The proposed algorithm does appear to adequately achieve adaptivity to the smoothness constant, as promised.
- The math is easy to follow and appears to be correct.
- The exposition is generally clear and free of typos. The experiment seems like a reasonable application and demonstration of the algorithm and includes several baselines for reference.
缺点
- The paper seems to focus on an easier deterministic optimization problem setting than its contemporaries such as Huang et al (2022), which develop similar algorithms for the stochastic Bilevel Optimization setting. In this sense the results may be rather limited in scope.
- The novelty of the algorithms is unclear to me: the approaches used seem similar to works such as Levy (2017) and Orabona (2023), which also show adaptivity to smoothness using normalized gradients. My main concern is that the results presented here are a straight-forward application of existing techniques to a new problem setting. This would be fine of course, but would of limited novelty.
Minor Gripes:
- The experimental results would be easier to visually parse if you keep the same color mapping between experiments (e.g. ABO is black in one plot but green in another).
- The conclusion mentions developing an adaptive step-size schedule, which feels like a contradiction of terminology: schedules are fixed in advance, like Cosine annealing or a linear decay schedule, whereas stepsize adaptation implies adapting the stepsize on-the-fly, as done in the algorithms presented here
References:
- Huang, Feihu, et al. "Enhanced bilevel optimization via bregman distance." Advances in Neural Information Processing Systems 35 (2022): 28928-28939.
- Levy, Kfir. "Online to offline conversions, universality and adaptive minibatch sizes." Advances in Neural Information Processing Systems 30 (2017).
- Orabona, Francesco. "Normalized Gradients for All." arXiv preprint arXiv:2308.05621 (2023).
问题
- Could you elaborate on what new techniques were introduced that wouldn't be possible using existing results? Is there a reason that the convergence of the inner-loop, for instance, need to be re-proven and can't just call out to an existing result? The algorithm doesn't appear to be doing anything particularly surprising that I haven't encountered already, e.g. particularly in Levy (2017) or Orabona 2023
This paper studies and important problem of designing adaptive (parameter agnostic) algorithms for bilevel optimization. The main result of the work states that in deterministic setting, their algorithm converges with the rate O(1/T) in convex and non-convex cases. The key feature of the proposed algorithm is that it only requires the knowledge of the strong convexity parameter of the lower level problem and does not require smoothness constants. From technical side, the authors incorporate Adagrad-Norm type step-sizes in both inner and outer loop iterations.
优点
-
The work proposed a nested adaptive algorithm and proved O(1/T) convergence rate in the deterministic setting without using the smoothness parameters of the problem.
-
Experiments are conducted to validate the efficiency of the proposed method.
缺点
-
Knowledge of strong convexity parameter is needed (and crucial in the analysis), which is a big weakness of the paper from theory side. In the deterministic setting (considered in the paper), this issue should not be too difficult to handle, e.g., using line-search or other adaptive methods. In minimization and min-max setting, such problem has been overcome even in the stochastic case, see, e.g., [1,2,3].
-
The authors do not make explicit the dependence on the constants and only report convergence in terms of . This should be fixed before the paper gets accepted. It could happen that the dependence on problem parameters can be e.g., exponential, making it essentially useless to derive a polynomial dependence on .
-
It is very confusing that Theorem 1 requires the bounded domain for convex case and unbounded domain (with Euclidean setting) for non-convex. A good theory should be able to handle both constraint/unconstrained, Euclidean/non-Euclidean cases at the same time.
-
There are several technical problems in the statements of theorems and some gaps in the proofs.
- Theorem 2 does not state any additional assumptions, while it becomes clear from Appendix F that a strong additional assumption (called reciprocity condition by authors) is needed. It seems that such assumption essentially means that the distance generating function is smooth. This is very limiting for mirror descent framework.
- Page 21 is not necessarily a contradiction.
- Same page 21. In the proof, the diameter of appears. However, it can be infinite in the unconstrained case. Thus, the next Theorem E.1. is wrong.
-
There a many incomplete, grammatically incorrect or repeating sentences in the draft. For example, in contributions section, first sentences in the first and third contributions missing a verb. In section 6, the sentence "It is known that..." is repeated.
-
In Assumption 2, what does it mean "relative to "?
-
Use different markers in the plots, lines are indistinguishable when printed in gray-scale. What is the advantage of using Shannon entropy compared to the Euclidean distance?
[1] Yair Carmon, Oliver Hinder. Making SGD Parameter-Free. COLT 2022.
[2] J. Yang, X. Li, N. He. Nest your adaptive algorithm for parameter-agnostic nonconvex minimax optimization. NeurIPS 2022.
[3] X Li, J Yang, N He. TiAda: A Time-scale Adaptive Algorithm for Nonconvex Minimax Optimization. ICLR 2023.
问题
-
What is the motivation for using the Fenchel coupling in the step-sizes (12)? Why not using more common distance measures, e.g., symmetrized Bregman divergence, e.g., . What is the connection with Fenchel coupling?
-
Is the set in Definition 1 assumed to be closed? If yes, why consider its closure? If no, how to project?
-
I am curious why the authors decide to use Adagrad type step-sizes. Is it possible to simply use a line-search for this problem?
-
Why the update rule (7) is called "normalized". Usually the normalized gradient step has the property, usually normalization with a power is called Normalized GD.
-
On page 19, after "Focusing on the last term of the (RHS) we have:", why the second inequality holds?
Typos:
- Lemma C.1. 2. should be .
- Lemma C.4. x^+ should be y^+.
- On page 21, F(x^*, x_k) must be a typo, since x_k is in the primal space.
- On page 23, "we will make the following ... inequality".
This paper proposes a new adaptive optimization algorithm based on mirror descent for a class of possibly nonconvex smooth bilevel problems with strongly-convex lower level. It provides the convergence analysis for the proposed algorithm and prove that it obtains a convergence rate . Meanwhile, it provides some experimental results to verify the efficiency of the proposed algorithm.
优点
This paper proposes a new adaptive optimization algorithm based on mirror descent for a class of possibly nonconvex smooth bilevel problems with strongly-convex lower level. It provides the convergence analysis for the proposed algorithm and prove that it obtains a convergence rate . Meanwhile, it provides some experimental results to verify the efficiency of the proposed algorithm.
缺点
Although this paper assert that the proposed adaptive algorithm does not rely on the Lipschitz constants, it strictly depends on the unknown strong convexity parameter of function and strong convexity parameter of function . Meanwhile, it requires to compute Hessian matrix and its inverse. This algorithm is a double-loop algorithm, and number of iteration in inner loop increases as number of iteration in outer loop increases. The proposed algorithm only considers the deterministic bilevel problems but does not consider the stochastic problems. Clearly, the proposed algorithm can not be efficiency in solving large-scale problems. Recently, there exist many efficient single-loop Hessian-free or full first-order gradient algorithms. Meanwhile, the convergence analysis of the proposed algorithm basically follows the existing convergence analysis. In summary, the novelty of this paper is limited. Thus, the level of this paper does not reach the level of ICLR.
问题
-
In the proposed algorithm, the function is regularizer. For easily reading, please give a specific example for function .
-
In the experiments, how to choose this regularizer ?
-
In the proposed algorithm, given . If is not bounded, the proposed algorithm is meaningless. Thus, the algorithm implicitly assume that is bounded, which is a stricter assumption than the existing bilevel algorithms.
-
In the convergence analysis, the authors only provide the convergence results of the proposed algorithm on the convex function and the bounded . When the function is convex and the set is not bounded, the authors can also provide its convergence results?
-
In the convergence analysis, the authors only provide the convergence results of the proposed algorithm on the non-convex function and . When the function is non-convex and the set , the authors can also provide its convergence results?
This work studied bilevel optimization with adaptive learning rate. Here adaptivity means that the developed algorithm is free from smoothness Lipschitz constants for both upper- and lower-level objective functions and learning rates. Compared to existing approaches that require heavily tuning for the theoretical guarantee and practical implementation, the proposed approach in this work can achieve similar convergence rate without such tuning efforts. The proposed method uses the idea from mirror descent. Motivated from Levy 2017, it uses a gradient norm accumulation idea to approximate the stepsizes. The outer-level stepsize is also designed to be adaptive based on an idea similarly to gradient mapping. In theory, they focus on the deterministic case, and obtain the same bounds as existing results but with less tuning efforts.
优点
-
This work studied a very interesting problem in bilevel optimization with less tuned parameters or without the requirement on knowledge of smoothness parameters. Given more complicated bilevel structure, getting less parameter tuning can be of interest in theory and in practice.
-
The paper is well written. The story is clear, and the algorithm is easy to follow with each adaptive design carefully explained. Proof sketch is also provided for readers to check and understand the main idea.
-
The idea of using the gradient mapping accumulation into the Fenchel coupling geometry to stabilize the convergence seems to be effective.
缺点
-
The developed algorithm is not entirely adaptive. The stepsize requires the information of the strong-convexity parameter of the inner-level function . This means that in experiments, you still need to tune the stepsizes. If I have a misunderstanding, please let me know.
-
The algorithms are limited to the deterministic setting. Is this idea able to extend to stochastic case.
-
The hypergradient requires an accurate computation of Hessian matrix inversion. So, a more practical case is to approximate the Hessian-inverse-vector product by solving a linear system or using the Neumann Series expansion. In this case, can the adaptive designs and analysis still work?
-
There are some works on auto-tuning stepsizes for (stochastic) bilevel optimization. For example, [1] uses some idea from stochastic line search to auto-tune the learning rates without knowing the smoothness parameters. Their approach is also appliable in stochastic case. I feel that the proposed method should be also compared to bilevel methods with adaptive learning rates, which are missing in the comparison.
[1] Fan, Chen, Gaspard Choné-Ducasse, Mark Schmidt, and Christos Thrampoulidis. "BiSLS/SPS: Auto-tune Step Sizes for Stable Bi-level Optimization." arXiv preprint arXiv:2305.18666 (2023).
- In the experiments, the comparison algorithms (including the proposed one) work in the stochastic case with data sampling. For example, SABA, BSA, VRBO are all stochastic algorithms. However, the theory is given in the deterministic case. Thus, it is unclear whether the proposed adaptive designs should be further adjusted in the stochastic setting? This is important because the adaptive designs such as the introduced scaled norms in line 6, 8, 9 may be biased in the stochastic setting,
问题
Overall, I think this is an interesting and very important topic, and the authors provide a simple and encouraging method. However, given there are several concerns, I am on the slightly negative side, but I am ok to increase my score after the rebuttal. See the weakness part for questions.
This paper proposed an adaptive algorithm for solving bilevel optimization problems with strongly convex inner function. If outer function is convex, the algorithm achieves a convergence rate of in terms of the outer objective function. If the outer objective is non-convex, the algorithm achieves an best-iterate guarantee for the squared norm of the gradient of the outer objective function.
优点
Originality: The authors proposed a adaptive optimization algorithm based on mirror descent for solving bilevel optimization problem without knowing the Lipschitz constants.
Quality: Compared with approximation bilevel method, their approach is more robust and practical based on the numerical experiments.
Clarity: The overall structure and presentation of the paper is clear and easy to follow.
缺点
-
The authors stated it is the first adaptive method for solving bilevel optimization problems. But I saw a paper [a] about adaptive methods for stochastic bilevel optimization problem. The statement is not true.
-
Although the algorithm utilized the adaptive method without knowing the Lipschitz constants, it did not achieve the state-of-the-art convergence rate. The state-of-the-art convergence rate for convex-strongly-convex case is in [b]. Is it possible to improve the proposed algorithm to that rate?
-
All the numerical results are plotted in terms of iterations. It would be better if the authors provide some results in terms of running time.
References:
[a]. BiAdam: Fast Adaptive Bilevel Optimization Methods . Feihu Huang, Junyi Li, Shangqian Gao.
[b]. Lower Bounds and Accelerated Algorithms for Bilevel Optimization. Kaiyi Ji, Yingbin Liang.
问题
All my questions are listed in the Weaknesses section.
There are many issue raised by the review team, but the authors did not response to any review's comment.
为何不给更高分
There are many issue raised by the review team, but the authors did not response to any review's comment.
为何不给更低分
N/A
Reject