Provable Faster Zeroth-order Method for Bilevel Optimization with Optimal Dependency on Error and Dimension
We improve the best-known complexity of fully zeroth-order bilevel optimization.
摘要
评审与讨论
This paper considers bilevel optimization. In particular, the authors aim at solving the stochastic bilevel optimization via zeroth-order methods. Motivated by the fact that a recent work has high dimensional dependency in the complexity bound, the authors propose VRZSBO, and prove that it requires less oracle complexity than the existing work, in the context of zeroth-order bilevel optimization algorithms. Numerical simulations are provided to validate the performance of their algorithm under different settings and their better convergence as compared to existing algorithms.
优点
-
The authors provide a better upper bound for the oracle complexity of zeroth-order bilevel optimization, without the need of additional assumptions.
-
The technical claims are in general sound.
缺点
-
Incomplete related work. Some important bilevel optimization algorithms are missing, even if they are not based on zeroth-order methods. It would be good if the authors could include a more comprehensive table containing existing algorithms with comparisons of assumptions, complexities, and types of oracles needed. See, for example, the related works in table 1 in [1].
-
Presentation.
- the authors claim one contribution is that they do not have additional assumptions like 3rd Lipschitz. However, this is not well reflected in Table 1.
- It seems that Section 1.1 can be moved to other sections for better presentation — for those who are familiar with bilevel optimization this section seems unnecessary and for those who are unfamiliar with bilevel optimization this section is full of technical details and hard to follow.
- It would be better if the authors could provide a well-organized Section 3 — the details of analysis described in 3.5 - 3.7 can be moved to the appendix to make the main context more readable. For example, proof sketch in section 3.6 is fairly standard and unnecessary to be included in the main body of the paper.
-
There are many important baselines missing in the experiments.
-
Overfull box in lines: 257 - 262, 409, 453, etc.
Reference
[1] BILEVEL OPTIMIZATION UNDER UNBOUNDED SMOOTHNESS: A NEW ALGORITHM AND CONVERGENCE ANALYSIS
问题
-
Can the authors explain the motivation of studying zeroth-order bilevel algorithms? The experiments can not reflect if VRZSBO can still be a better choice in large-scale settings.
-
Can the authors explain why (d_1 + d_2) dependency is optimal? For example, is there a reference for the lower bound of the dimension dependency.
-
The oracle cost mentioned in Remark 2 is in expectation (see (14) - (16)). However, the cost in (Aghasi and Ghadimi, 2024) is not in expectation, as the estimator therein does not require an additional sampling step (i.e., (14) - (16)). I am wondering if the authors could provide more context on this unfair comparison.
We sincerely appreciate the reviewers' valuable suggestions regarding the writing of our paper, which have been very helpful in improving our work. Following your recommendations, we have revised the structure of the paper, included comparisons with more algorithms in the tables, and added additional experiment.
Q1: Can the authors explain the motivation of studying zeroth-order bilevel algorithms? The experiments can not reflect if VRZSBO can still be a better choice in large-scale settings
1.Bilevel is a highly general framework that can model many real-world problems. Min-max problems and single-level optimization problems are both special cases of bilevel problems. Zeroth-order optimization has already been extensively studied in single-level and min-max problems, attracting significant interest. Therefore, studying zeroth-order algorithms under the bilevel framework to obtain more general conclusions compared to the aforementioned cases is highly appealing.
2.Moreover, reformulating some zeroth-order optimization tasks using the bilevel framework can potentially achieve better results. For example, as shown in [1] and [2], combining adversarial attacks (a common application of zeroth-order algorithms) with meta-learning (a common bilevel application) led to improved performance over existing methods.
Given that [3] demonstrates the use of zeroth-order algorithms for fine-tuning large language models (LLMs), achieving up to 12× memory reduction and up to 2× GPU-hour reduction, and [4] highlights the close connection between the bilevel framework and LLM , zeroth-order bilevel algorithms hold significant potential for large-scale applications. Our VRZSBO algorithm, with its strong theoretical guarantees and low complexity, can serve as an important baseline for future research in zeroth-order bilevel optimization.
Q2 Can the authors explain why dependency is optimal? For example, is there a reference for the lower bound of the dimension dependency.
Thank you for raising this question. We answer it in the General Response .
Q3: the cost in (Aghasi and Ghadimi, 2024) is not in expectation while our oracle cost is in expectation
This problem is actually results in the difference between PAGE[5] and Spider[6].To achieve a single-loop algorithm, PAGE selects a large batch size with a probability of , while SPIDER uses a large batch size once every iterations. This results in both methods having the same complexity, but PAGE's results are presented in an expected form. This reflects a trade-off between the single-loop structure and randomness. We do not consider this a significant issue, as in single-level problems, comparing the complexities of SPIDER and PAGE is not regarded as unfair.
Reference:
[1]Boosting Black-Box Adversarial Attacks with Meta Learning
[2]Simulating Unknown Target Models for Query-Efficient Black-box Attacks
[3]Fine-Tuning Language Models with Just Forward Passes
[4]LLMS ARE HIGHLY-CONSTRAINED BIOPHYSICAL SE QUENCE OPTIMIZERS
[5]PAGE: A Simple and Optimal Probabilistic Gradient Estimator for Nonconvex Optimization
[6]SPIDER: Near-Optimal Non-Convex Optimization via Stochastic Path Integrated Differential Estimator
Using finite-difference approximations and the PAGE method for nonconvex optimization, this work introduces a novel zeroth-order algorithm, VRZSBO, for solving bilevel optimization problems with a strongly convex lower-level problem. The authors show that VRZSBO converges to an -stationary point with a function query complexity of in the finite-sum setting and in the expectation setting, where and denote the dimensions of the upper- and lower-level variables, respectively. Experimental results on hyper-representation tasks using linear and two-layer network embedding models further validate the algorithm’s effectiveness.
优点
The paper is easy to follow. Zeroth-order algorithms for bilevel optimization remain underexplored, even in the nonconvex-strongly-convex setting, making the introduction of the novel VRZSBO algorithm a valuable algorithmic contribution to this area.
The proposed VRZSBO algorithm achieves a significant improvement over the best-known complexity of ZDSBA (Aghasi and Ghadimi, 2024), reducing the query complexity from to in the expectation setting. Additionally, this work provides a complexity analysis for the finite-sum setting.
缺点
It is well-known that the Hessian-vector product can be implemented through finite-difference approximations of the gradient along a given direction. Given this, the novelty of the proposed results seems limited compared to existing methods.
Moreover, as noted in Chapter 9 of Numerical Optimization by Jorge Nocedal and Stephen J. Wright, finite-difference approaches are often impractical due to potential inaccuracies in function evaluations (e.g., rounding errors), which prevent the distance of finite differences from being arbitrarily small.
The experiments could be more comprehensive. Incorporating updated first-order algorithms, such as F2SA (Kwon et al., ICML 2023), would provide a more thorough comparison. Furthermore, including commonly used bilevel optimization tasks, such as hyperparameter selection and data hyper-cleaning, would further validate VRZSBO’s effectiveness, showcasing its reliability across various bilevel optimization problems.
I would be willing to increase the score based on the rebuttal.
问题
Apart from the questions raised in the Weaknesses section, some additional questions are as follows:
Why does the proposed algorithm achieve an optimal result with respect to dependence on dimension? A bit more discussion is needed.
In Assumption 2, why is there no stochastic assumption on ?
Based on the proofs of Theorems 1 and 2, the iteration count in these theorems depends on , which may in turn depend on (in the finite-sum case) and (in the expectation case). Therefore, to derive the complexity analysis results (i.e., the total oracle cost), does this require certain conditions on the initial points? Additionally, does it require an assumption of a lower bound for ? A bit more discussion on these points would be helpful.
In Theorems 1 and 2, the zeroth-order tuning parameter does not appear—why is this the case? Is there a relationship between and , potentially stemming from Corollary 1 and Lemma 3 through an intermediate parameter ? Additionally, in Figure 1, ablation studies examine the effects of different choices of and on the zeroth-order estimator. Were these experiments conducted with consideration of the relationship between and ?
The theoretical results in this work suggest that smaller zeroth-order tuning parameters and yield better performance, yet the experimental results in Figure 1 do not support this conclusion. Could the authors explain this discrepancy?
In the experiments, why was there no comparison with ZDSBA from Aghasi and Ghadimi (2024)?
In the convergence analysis, for example, in Theorem 1, it is stated that . This implies that the Lyapunov function is decreasing. However, how is this possible when is nonconvex? It seems that the sign before may be incorrect.
In Theorem 1, why is there no in the finite-sum case? Is there an assumption that ?
Page 1, Line 045: In , should be .
Page 2, Lines 102-103: Verify the form of .
Page 3, Line 145: Replace “, , ” with “, , ”.
Page 5: Line 228: The referenced equation (2) may not be appropriate.
Page 5: Lines 230-231: does not include an inverse matrix operator.
Page 5, Lemma 1: It would be clearer to replace y^ with since is a function.
Page 5, Definition 1: In , replace with , and in , replace with .
Page 6, Lemma 3: Clarify the meaning of in the definition of .
Page 7, Line 327: “Section 3.1” should be changed to “Sections 3.2–3.3”.
Page 7, Equations (14), (16): Keep the notations consistent with those in (10)–(11).
Page 7, Line 369: Should replace according to Lemma 9?
Page 7, Line 372: Clarify what represents in the definition of .
Page 8, Lines 383-387: Should replace based on Lemmas 6–7?
Page 8, Line 398: Replace with .
Page 8, Lemmas 4-5: Define \ell_{Z^} and \ell_{Y^}.
Page 10, Figure 1: Correct typos, particularly in the caption.
Q: Based on the proofs of Theorems 1 and 2, the iteration count T in these theorems depends on , which may in turn depend on n (in the finite-sum case) and ϵ (in the expectation case). Therefore, to derive the complexity analysis results (i.e., the total oracle cost), does this require certain conditions on the initial points? Additionally, does it require an assumption of a lower bound for Φ(x)? A bit more discussion on these points would be helpful.
Good question. recall that
$ \mathbb{E}[\psi _0]:&=\Phi(x _t)+\frac{18(L ^{2}+2 r _{z} ^{2}\rho ^{2})\eta _x}{\eta _y\mu}(|| y _{0}- y ^{\star} _{0}|| ^{2})+\frac{36\eta _x L ^{2} \rho ^{2}}{\eta _z\mu} (|| z _{0}- z ^{\star} _{0}|| ^{2})\\ &+\mathbb{E}[ \frac{3\eta _x}{2p}||v _{x, 0}-\hat{h} _{x, 0}|| ^{2}+ \frac{72(L ^{2}+2 r _{z} ^{2}\rho ^{2})\eta _x}{p\mu ^{2}}|| v _{y, 0} -\hat{h} _{y, 0} || ^{2}+ \frac{144 L ^{2} \rho ^{2}\eta _x}{p\mu ^{2}}|| v _{z, 0} -\hat{h} _{z, 0} || ^{2}], $For the variance terms , by computeing , , with batchsize (finite-sum case) and (the expectation case) when , the variance terms will decrease to 0(finite-sum case)and (expectation case), which makes independt of and , we will add this initialization step to our algorithm. The remaining terms , and , to the best of our knowledge, appear in almost all bilevel optimization literature. Since and are solutions to strongly convex problems, we can run algorithm times to decrease and to nearly 0 .This trick is called "warm start", which is effective in certain experiments. And of course we require an assumption of a lower bound for , which is standard in bilevel optimization, we will add this assumption to improve clarity.
Q: In Theorems 1 and 2, the zeroth-order tuning parameter h does not appear—why is this the case? Is there a relationship between v and h, potentially stemming from Corollary 1 and Lemma 3 through an intermediate parameter δ? Additionally, in Figure 1, ablation studies examine the effects of different choices of h and v on the zeroth-order estimator. Were these experiments conducted with consideration of the relationship between v and h?
Your understanding is correct. In Theorems 1 and 2, does not explicitly appear because its effect is incorporated into through the intermediate parameter , as shown in Corollary 1 and Lemma 3, we will include the value of in the statements of Theorems 1 and 2 to improve clarity. In the experiments, we fix v to the best-performing value, then test h , and then swap by fixing h to the best-performing value and testing v .
Q: The theoretical results in this work suggest that smaller zeroth-order tuning parameters v and h yield better performance, yet the experimental results in Figure 1 do not support this conclusion. Could the authors explain this discrepancy?
Good question. We would like to clarify that, theoretically, the value of v indeed improves with smaller values. However, for h, there is an optimal value: it should not be too large or too small, as shown in the proof of Lemma D.6. In our experiments, the optimal value of h corresponds to a middle-range value, which is consistent with our theoretical analysis. As for v , the experimental trend does not fully align with our theoretical expectations, fortunately, we observed that the best performance in the experiments still corresponds to the smallest value of v. Based on our experience, selecting a sufficiently small v (e.g., 0.001 or 0.00001) remains effective.
Reference:
[1]On the Complexity of First-Order Methods in Stochastic Bilevel Optimization
[2]Using Complex Variables to Estimate Derivatives of Real Functions
I thank the authors for their detailed responses and have adjusted my score accordingly.
Thank you for reconsidering our work and for the updated score. We appreciate your valuable feedback and the time you invested in reviewing our paper.
We sincerely thank you for taking the time to provide such detailed and thoughtful feedback. Your suggestions are helpful for improving our paper. We have supplemented the experiments and corrected the typos based on your comments in the revised version, and we will now address the another questions you raised.
W: finite-difference approaches are often impractical due to potential inaccuracies in function evaluations (e.g., rounding errors), which prevent the distance of finite differences from being arbitrarily small.
1.To the best of our knowledge, achieving the best theoretical upper bound for a Hessian-free method appears challenging without employing finite differences. First-order methods based on penalty functions can at best achieve a complexity of ([1], ICML 2024), whereas finite-difference-based method achieves a significantly improved complexity of . 2.Although we did not employ this approach in our paper, we would like to introduce an intersting method that can effectively address the issue of inaccuracies in function evaluations, provided that complex number computations are acceptable[1]. Using a Taylor’s series expansion ( denote the th standard basis vector,v is a smoothing parameter, z is a vector):
f(x+\mathrm{i} ve_{i})=f(x)+\mathrm{i} v \langle e_{i}, \nabla f(x)\rangle+\mathcal{O}(v ^2) \and
we can estimate the gradient and hessian-vector product as : and . This estimation method is not subject to rounding errors because it doesnt involve subtraction,serves as a good complement to our approach, particularly when the estimation might be affected by rounding errors.
Q: Why does the proposed algorithm achieve an optimal result with respect to dependence on dimension? A bit more discussion is needed.
Thank you for raising this question. We answer it in the General Response.
Q: In Assumption 2, why is there no stochastic assumption on ?
Sorry about this typo. We have corrected it.
This paper explores zeroth-order algorithms for addressing stochastic bilevel optimization problems. Expanding on earlier work that introduced a zeroth-order bilevel method using Gaussian smoothing, the authors enhance the dimensional dependency in sample complexity from to . They also utilize variance reduction techniques to optimize the epsilon dependency in sample complexity from to , achieving optimal bounds. The new algorithm undergoes theoretical analysis, demonstrating convergence to a stationary point with optimal complexity in both expectation and finite sum setting. Additionally, experiments are conducted to validate the algorithm's effectiveness.
优点
The paper presents several key strengths that enhance its contribution to the field of stochastic bilevel optimization. Notably, it significantly improves the dimensional dependency in sample complexity from to and employs variance reduction techniques to optimize epsilon dependency, achieving optimal bounds. The thorough literature review effectively situates the work within existing research, underscoring its relevance. Additionally, the clear and structured presentation of the methodology supports reproducibility, while the empirical validation through experiments demonstrates the algorithm's practical effectiveness.
缺点
Lack of Novelty: The contributions of this paper build upon established concepts, integrating ideas from quadratic auxiliary functions, PAGE-type variance reduction, and zeroth-order gradient estimation. While these improvements are valuable, they may not be unexpected, in my humble opinion. For instance, ZDSBA built on previous work as a double-looped algorithm utilizing Hessian inverse approximation, so it is understandable that the incorporation of auxiliary quadratic functions would enhance dimension dependency. Additionally, the introduction of various variance reduction techniques is likely to improve epsilon dependency. Although I recognize the meaningful nature of these contributions, I feel that the paper does not fully meet the acceptance criteria for ICLR. Therefore, I will reserve my opinion on acceptance until after the rebuttal stage.
问题
Gaussian Smoothing vs. Coordinate-wise Smoothing: Based on my previous experiences, I have not observed a significant difference in the theoretical bounds between these two smoothing techniques. Consequently, I believe the improvements presented in the paper may not be attributed to coordinate-wise smoothing. The paper does not address the rationale behind choosing coordinate-wise smoothing over Gaussian smoothing. Could the authors please clarify the reasoning for this choice?
Thank you for taking the time to review our paper. We appreciate the opportunity to clarify our contributions and provide a detailed comparison between our method and ZDSBA.
Q: Gaussian Smoothing vs. Coordinate-wise Smoothing
To simplify notation, let . The improved efficiency of our method with respect to dimensionality stems from two key factors: (1) Our single-loop structure eliminates the inner loop in [1], saving complexity from solving inner problems. (2) We use a finite-difference approximation for Hessian-vector or Jacobian-vector products, requiring only two gradient evaluations, with oracle cost. This is significantly more efficient than the Gaussian smoothing method in [1, Proposition 2.5], where approximating the Hessian-vector product incurs a variance of , leading to complexity for hypergradient estimation and outer loop iterations. Our method reduces this cost by . Combining the two above, our complexity is faster than that of [1].
- Gradient Estimation: Coordinate-wise smoothing requires oracle calls for accurate gradient estimation, while Gaussian smoothing uses calls but introduces a variance of , ultimately requiring oracle calls to cancel this variance. Hence, the overall complexity for both remains times that of standard first-order methods.
- Hessian-Vector/Jacobian-Vector Estimation: Finite-difference approximation via coordinate-wise smoothing requires oracle calls for accurate estimation, while [1, Proposition 2.5] requires calls but introduces a variance of due to term , leading to complexity in the outer loop. Our method reduces this by . Conclusion: Combining the improvement in Hessian-vector/Jacobian-vector estimation and the speedup from the single-loop structure, our method achieves an overall improvement compared to ZDSBA [1]. Additionally, our coordinate-wise results suggest the potential for Gaussian smoothing to achieve samples with variance, as both smoothing approaches perform similarly in gradient estimation, we have incorporated this comparison into the latest version of the paper.
W: Lack of Novelty Thank you for your comment. We have already addressed a similar question raised by Reviewer L5fV, and our response can be found in their section. We kindly refer you to that response for a detailed explanation, moreover, we would like to emphasize that our differences with SPABA also extend to comparisons with other methods that achieve near-optimal complexity [2],[3]. We believe that our answer should address your concern as well.
Reference: [1]Fully Zeroth-Order Bilevel Programming via Gaussian Smoothing
[2]Achieving O (ϵ−1.5) Complexity in Hessian/Jacobian-free Stochastic Bilevel Optimization
[3]A Lower Bound and a Near-Optimal Algorithm for Bilevel Empirical Risk Minimization
In this paper, the authors propose a zeroth-order method for solving stochastic bilevel optimization problems, where the lower-level problem is strongly convex and only function values are accessible for both levels. Their method can be regarded as a zeroth-order approximation of the SPABA method proposed by (Chu et al., 2024), where the gradients and Hessian-vector products are replaced by their zeroth-order estimators. Specifically, the authors employ a coordinate-wise zeroth-order estimator, approximating each coordinate of the gradient by finite differences. They analyze their proposed algorithm under both general stochastic and finite-sum settings, demonstrating that the function value query complexity scales linearly with the problem’s dimension and has optimal dependence on the target accuracy.
优点
In the context of zeroth-order algorithms for bilevel optimization, the proposed algorithm improves dimensional dependence compared to prior work and introduces a simpler single-loop update format.
缺点
- My major concern is the use of the coordinate-wise zeroth-order estimator in Definition 1. As noted by the authors, this estimator requires function value queries in total, which may be as computationally expensive as obtaining the exact gradient by backpropagation. While it could be argued that only function values are accessible in certain applications, this does not hold for most common bilevel optimization applicationss, including the hyper-representation problem used in the experiments.
- In addition, the novelty of the paper appears somewhat limited. If my understanding is correct, the algorithmic framework closely follows the SPABA method by (Chu et al., 2024) such as the use of the PAGE variance reduction technique, and the main modidfication is to substitute the exact gradients and Hessian-vector products by their zeroth-order estimators. However, since the coordinate-wise zeroth-order estimator is employed, the gradient estimator error can theoretically be made arbitrarily small by selecting a sufficiently small (see Lemma 2). Indeed, the authors choose in Theorems 1 and 2, which means that the gradient error is also on the order of . Consequently, the convergence analysis may not significantly differ from that of the original SPABA method.
Tianshu Chu, Dachuan Xu, Wei Yao, and Jin Zhang. SPABA: A single-loop and probabilistic stochastic bilevel algorithm achieving optimal sample complexity. ICML 2024
问题
- The author claims that the dependence on the dimension is optimal. While this appears reasonable, a rigorous lower bound is necessary to support this claim.
- The experiments should include the SPABA method as a baseline, as it is the most relevant algorithm for comparison with the proposed method. Additionally, comparisons across different algorithms should be based on runtime, given the potentially significant differences in per-iteration costs.
We appreciate the reviewer’s thoughtful comments and feedback. In response to the concerns raised, we address the following points:
W: Concerning the use of the coordinate-wise zeroth-order estimator in Definition 1.
We understand the reviewer’s concern regarding the practical applicability of the coordinate-wise zeroth-order estimator. In response to this, we offer the following points:
1.We chose the coordinate-wise zeroth-order estimator in order to achieve stronger theoretical results.As discussed in our response to Reviewer Nm7C, the use of the coordinate-wise zeroth-order estimator enables us to achieve a complexity of . This result serves as an important baseline, as it matches the complexity of the most well-known zeroth-order min-max methods, which are often used in practice.
-
Early Stage of Zeroth-Order Bilevel Optimization Research We remind the reviewers that the first fully zeroth-order bilevel optimization algorithm was only recently proposed, so research in this area is still in its early stages. Given this, the improvement in our work from to is a significant advancement. Our theoretical analysis provides valuable insights that future researchers can use to design methods tailored to specific problems. For example, a hypergradient estimator requiring only function evaluations or potentially even achieving faster speeds than first-order methods in practical tasks.
-
Moreover, reformulating some zeroth-order optimization tasks using the bilevel framework can potentially achieve better results. For example, as shown in [10] and [11], combining adversarial attacks (a common application of zeroth-order algorithms) with meta-learning (a common bilevel application) led to improved performance over existing methods.
W: Novelty of the paper
To address your concerns about novelty of the paper, we clarify them and improve discussion of challenges as follows:
1. Analysis of zeroth-order estimator We would like to emphasize that, although in our analysis choosing a sufficiently small can indeed make the gradient error very small, the same cannot be done for the parameter used in Hessian-vector estimation. This is because, as shown in the proof of Lemma 2, there exists an optimal value for : it should neither be too large nor too small. Our experiments on the optimal choice of also support this point. We believe that our results can effectively help avoid performance degradation caused by excessively small , which might otherwise occur if one blindly minimizes without considering its optimal value.
2.Even beyond the zeroth-order estimators, our approach to convergence analysis differs significantly from that of the SPABA method. Both methods are based on the quadratic auxiliary function framework proposed in [3], which typically divides the convergence analysis into three key components: (1) Descent Analysis of the Inner Problem, (2) Variance Reduction Analysis, and (3) Convergence Analysis of . Our analysis framework is novel and non-trivial .Below, we compare our approach to SPABA in each of these aspects.
(1) Descent Analysis of the Inner Problem We now compar the difference between their descent in inner variable z (Lemma F.3) and ours (Lemma D.6).Although our analysis claims to leverage the strong convexity property, our proofs actually rely only on the Quadratic Growth (QG) condition for the objective function of . In the unconstrained case, this condition is equivalent to the Polyak-Łojasiewicz (PL) condition [6]. In contrast, the proofs in SPABA require a property we refer to as co-coercivity, which applies specifically to strongly convex problems. Unfortunately, functions satisfying only the PL condition do not exhibit the co-coercivity property. This distinction makes our analysis more general and potentially of broader interest, especially given the recent exploration of nonconvex-PL (NC-PL) bilevel problems [7],[8],[9].
(2)Variance Reduction Analysis We now compar the difference between their variance descent in inner variable z (Lemma F.4) :
$ &\mathbb{E}\left[\left||v _{k+1} ^z-D _z\left(x _{k+1}, y _{k+1}, z _{k+1}\right)\right|| ^2\right] \\ &\leq \left(1-p+\frac{4(1-p)}{b}\left(L _2 ^g\right) ^2 \gamma _k ^2\right) \mathbb{E}\left[\left||v _k ^z-D _z\left(x _k, y _k, z _k\right)\right|| ^2\right] \\ & +\left(2\left(L ^f\right) ^2+4 R ^2\left(L _2 ^g\right) ^2\right) \frac{(1-p)}{b} \alpha _k ^2 \mathbb{E}\left[\left||v _k ^x\right|| ^2\right] \\ & +\left(2\left(L ^f\right) ^2+4 R ^2\left(L _2 ^g\right) ^2\right) \frac{(1-p)}{b} \beta _k ^2 \mathbb{E}\left[\left||v _k ^y-D _y\left(x _k, y _k, z _k\right)\right|| ^2\right] \\ & +\left(2\left(L ^f\right) ^2+4 R ^2\left(L _2 ^g\right) ^2\right) \frac{(1-p)}{b} \beta _k ^2\left(L _1 ^g\right) ^2 \mathbb{E}\left[\left||y _k-y ^*\left(x _k\right)\right|| ^2\right] \\ & +\frac{4(1-p)}{b}\left(L _2 ^g\right) ^2 \gamma _k ^2 L _z ^2 \mathbb{E}\left[\left||z _k-z ^*\left(x _k\right)\right|| ^2\right] \\ & +\frac{4(1-p)}{b}\left(L _2 ^g\right) ^2 \gamma _k ^2 L _z ^2 \mathbb{E}\left[\left||y _k-y ^*\left(x _k\right)\right|| ^2\right] $while ours Lemma D.7 are:
$ &\mathbb{E} \left|| v _{z, t+1} -\hat{h} _{z, t+1} \right|| ^{2}\\ &\le (1-p)\mathbb{E}[\left||v _{z, t}- \hat{h} _{z, t}\right|| ^{2}]\\ &+\frac{2(1-p)}{b}((8 r _{z} ^{2}+3 L ^{2})\left|| c _{t+1}-c _{t} \right|| ^{2}\\ &+8L ^{2}\left|| z _{t+1}-z _{t} \right|| ^{2}+16 r _{z} ^{2} \rho \delta+6\delta ^{2}). $From a formal perspective, it is evident that our approach to error estimation differs significantly from SPABA, as the error terms we choose to analyze and utilize are substantially distinct.
(3) Convergence Analysis of . Although the potential function we use is similar in form to that of SPABA, our choice of coefficients is superior. For example, on page 40 of their paper, to ensure has a negative coefficient, the step size must satisfy . On page 41, to ensure has a negative coefficient,the step size must satisfy ,a step size satisfying both conditions may not exist when , which is highly likely to occur,however,in our potential function, the use of different weighting ratios can address this issue. Moreover, based on the reasons mentioned above, there is a significant difference between our approach and other works with near-optimal complexity, such as [4] and [5]. Therefore, we believe that our theoretical analysis is not merely a simple imitation of any single paper.
Q: The dependency on the dimension.
Thank you for raising this question. We answer it in the General Response.
Q: The experiments should include the SPABA method as a baseline, as it is the most relevant algorithm for comparison with the proposed method. Additionally, comparisons across different algorithms should be based on runtime, given the potentially significant differences in per-iteration costs. We have added a new set of experiments as per your request, including SPABA as a baseline for comparison and reporting the runtime statistics.
Reference:
[1]Zeroth-Order Alternating Gradient Descent Ascent Algorithms for a Class of Nonconvex-Nonconcave Minimax Problems
[2]Fully Zeroth-Order Bilevel Programming via Gaussian Smoothing
[3]A framework for bilevel optimization that enables stochastic and global variance reduction algorithms
[4]Achieving O(ϵ−1.5) Complexity in Hessian/Jacobian-free Stochastic Bilevel Optimization
[5]A Lower Bound and a Near-Optimal Algorithm for Bilevel Empirical Risk Minimization
[6]Linear Convergence of Gradient and Proximal-Gradient Methods Under the Polyak-Łojasiewicz Condition
[7]Optimal Hessian/Jacobian-Free Nonconvex-PL Bilevel Optimization
[8]On Penalty Methods for Nonconvex Bilevel Optimization and First-Order Stochastic Approximation
[9]On Finding Small Hyper-Gradients in Bilevel Optimization: Hardness Results and Improved Analysis
[10]Boosting Black-Box Adversarial Attacks with Meta Learning
[11]Simulating Unknown Target Models for Query-Efficient Black-box Attacks
I appreciate the authors' detailed response. However, I still have reservations regarding the use of the coordinate-wise zeroth-order estimator. This coordinate-wise finite-difference approach makes it not much different from a standard gradient oracle, in terms of both the computational cost and convergence analysis. While I understand that the authors aim to establish a baseline for zeroth-order methods in bilevel optimization, it is well-known that gradients can be well approximated using function value queries, and thus a first-order method can be simulated using a zeroth-order oracle. In some sense, the contribution of this paper appears to be formalizing this observation for bilevel optimization problems.
In addition, while there are some technical differences in the convergence analysis between this paper and SPABA, ultimately they appear to achieve similar convergence rates. It remains unclear whether any of these differences are essential for the algorithm to work properly with a coordinate-wise zeroth-order estimator.
As a side note, I noticed that the two plots in Figure 3 appear to be identical. Intuitively, one would expect a larger dimension to result in a longer runtime. Could the authors double-check if this is correct?
Thank you for your thoughtful feedback. We appreciate the opportunity to address your concerns.
Regarding the coordinate-wise zeroth-order estimator
We reaffirm that our use of the coordinate-wise zeroth-order estimator aims to establish better theoretical convergence results. While we acknowledge that the coordinate-wise approach is well-known, we argue that our specific application of this estimator to approximate the hypergradient in bilevel optimization introduces unique insights. For instance, as demonstrated in Lemma D.6 of our paper, the optimal choice of the parameter (step size for finite differences) is well-known: it cannot be too large or too small to ensure the correctness of the hypergradient estimation. This delicate balance is critical in the bilevel setting and is not a well-known result, adding a layer of novelty to our analysis. Additionally, our framework showcases how zeroth-order estimators can effectively handle the nested structure of bilevel problems without access to explicit gradients.
On convergence analysis and comparisons with SPABA
Regarding the technical differences in the convergence analysis between our paper and SPABA, we would like to clarify that, as discussed in our general response, the dependence of the convergence rates on in our work is already very close to optimal. Therefore, it is not surprising that our dependence on aligns with SPABA’s. However, the differences in the convergence analysis between our work and SPABA are indeed meaningful and carry practical implications, not only for zeroth-order methods but also for the analysis of first-order methods, as we have outlined in our previous responses.
Regarding the identical plots in Figure 3
We sincerely apologize for not updating the figure titles. The two plots in Figure 3 show the test loss and train loss, respectively. The experimental setup did not involve any change in the dimensionality of the problem. We will update the figure titles in the revised version.
We kindly ask you to reconsider the score in light of the clarifications and improvements made in the revised version. We are confident that our work provides meaningful contributions to the field of bilevel optimization, particularly in the context of zeroth-order methods.
We would like to thank all the reviewers for their thoughtful feedback and for raising important questions about our work. Several of these questions share common themes, and we have provided a general response below to address these shared concerns comprehensively.
G: the optimality of and , and lowerbound. After we carefully check the lower bounds under our assumptions,we indeed find that our claim on "optimal" is not rigorous. We have decided to clarify the statements related to "optimal" to "best-known", and add a discussion on the dependencies on and in our revised version, based on the following facts:
- Our dimension dependency matches the best-known result in simpler minmax problems [1, 2], and the dependency aligns with the best-known complexity for first-order NC-SC bilevel problems [3, 4, 5], making our results "best-known."
- Since [6] establishes a lower bound for single-level smooth convex problems, which are simpler than NC-SC bilevel problems (e.g., by setting ), a dependency is inevitable.
- Our assumptions align with those used to derive the lower bound in [4] for NC-SC finite-sum bilevel problems, confirming that the dependency is inevitable in the finite-sum case .
- The lower bound for single-level problems[8] in the expectation case is constructed under the mean-square-smooth assumption, which is slightly stronger than the typical assumption of smoothness about smoothness of for all . Thus, strictly speaking, is not necessarily optimal in for the expectation case.
Based on existing lower bounds, the dependency on is inevitable. However, the optimality of the dependency remains unclear. It would be particularly interesting if the exponent of could be reduced to less than 1 for lower level strongly convex optimization, as this would imply a theoretical speedup of zeroth-order algorithms compared to first-order algorithms (assuming the gradient computation cost is ).
[1]Zeroth-order algorithms for nonconvex–strongly-concave minimax problems with improved complexities
[2]Zeroth-Order Alternating Gradient Descent Ascent Algorithms for a Class of Nonconvex-Nonconcave Minimax Problems
[3]Achieving O (ϵ−1.5) Complexity in Hessian/Jacobian-free Stochastic Bilevel Optimization
[4]A Lower Bound and a Near-Optimal Algorithm for Bilevel Empirical Risk Minimization
[5]SPABA: A Single-Loop and Probabilistic Stochastic Bilevel Algorithm Achieving Optimal Sample Complexity
[6]Optimal Rates for Zero-Order Convex Optimization: The Power of Two Function Evaluations
[7]On the Complexity of First-Order Methods in Stochastic Bilevel Optimization
[8]Lower bounds for non-convex stochastic optimization
This paper studies zeroth-order method for bilevel optimization. In particular, the authors proposed a single-loop accelerated zeroth-order algorithm for bilevel optimization. The reviewers believe that the proposed algorithm closely follows some existing works, and the novelty is limited. Moreover, it requires significant amount of work to improve the presentation and readability of the paper.
审稿人讨论附加意见
Further discussed the novelty and numerical experiments. The reviewers were not convinced.
Reject