Stochastic Regret Guarantees for Online Zeroth- and First-Order Bilevel Optimization
We propose a novel search direction for stochastic online bilevel optimization, enabling first- and zeroth-order algorithms to achieve sublinear regret without window smoothing.
摘要
评审与讨论
This article presents two online bilevel optimization algorithms SOGD and ZO-SOGD. Among them, SOGD achieves the sota local regret bound without using window-smoothed functions. ZO-SOGO provides a fully-zero-order approach and achieves the hypergradient estimation only with function values. The authors present the theoretical analysis of the regret as well as the convergence analysis. Both two algorithms perform well in a series of numerical experiments.
优缺点分析
Strength:
- Both algorithms are sing-loop.
- SOGO achieves the sota regret bound without extra computation of the window-smoothed function. ZO-SOGO provides a fully-zero-order approach for online BO, which is especially useful in the training of large models.
- The authors present a detailed proof of the regret bound, which have been quickly checked by the reviewer and no errors have been found.
- The experiment result successfully validate the theoretical finds.
Weakness:
- The authors fail to present the comparison of different algorithms under wall-clocked time in the experiment, which may illustrate the practical value of the proposed algorithms. Also, additional experiment should be taken to validate how step size affects the convergence.
- The author lack a detailed comparsion for the computation complexity of different algorithms.
问题
- Whether the authors can present the comparison of different algorithms under wall-clocked time?
- Whether the author lack a detailed comparsion for the computation complexity of different algorithms?
- How does the step size affect the algorithm performance practically? Please provide some experimental results.
- Can the authors provide a brief proof sketch? It may help readers understand the proof.
局限性
Yes.
最终评判理由
This submission is a theoretically valuable paper for discussion with rich theoretical results and enough experiments. Moreover, all the concerns have solved during rebuttal and discussion period. Thus, the reviewer recommends to the acceptance.
格式问题
NA
We thank Reviewer R7y3 for the constructive comments. We address each point below.
Question 1: Can authors present wall-clock time comparison? Weakness 1: Missing wall-clock time comparison and step size analysis
Response: Thank you for your review and suggestions. We would like to clarify that the runtime plots presented in both Figures 2 and 3 already show wall-clock time measurements in seconds. Please refer to the left subplot in Figures 2 and 3, where we report time elapsed.
Specifically,
-
In Figure~2 (left panel), we present the actual wall-clock time of our ZO-SOGD method compared to the baseline methods in the online adversarial attack experiment. The recorded times range from approximately to seconds at steps.
-
In Figure 3 (left panel), we present wall-clock time comparisons for our SOGD method against OAGD and SOBOW in the parametric loss tuning task. The results show a clear efficiency gain, with SOGD completing under 25 seconds at , compared to SOBOW, which takes roughly 220 seconds due to its reliance on conjugate gradient (CG) iterations.
All wall-clock times were measured using the same hardware platform across all algorithms to ensure fair and consistent comparisons.
In the revised version, we will make this point more explicit by stating "wall-clock time" clearly in the figure captions and experimental section to avoid any ambiguity.
Question 2: Detailed computational complexity comparison? Weakness 2: Lack of detailed computational complexity comparison
Response: Thank you for your valuable comments. We will provide the total computational cost estimates for each method.
Total complexity:
- SOGD:
- ZO-SOGD:
- SOBOW/SOBBO: due to conjugate gradient (CG) steps
As an example, for SOGD, the regret bound
implies that
Let . Then , which is sufficient to ensure that the average regret , leading to a total cost of .
For SOGD, we have gradient computations, Hessian-vector products, and Jacobian-vector products. Since the algorithm operates over iterations, the total number of operations is . Note that to achieve -stationarity, we need iterations. Thus, the computational complexity is .
We note that the total cost for other baselines (SOBOW/SOBBO) is derived using their window size and the factor for conjugate gradient (CG) steps. The OAGD baseline incurs higher cost due to exact system solves at each iteration.
We will include the above discussion in the main paper in the revised version.
Question 3: How does step size affect algorithm performance practically?
Response: Thank you for your valuable feedback.
We apologize that these details were not prominently featured in the main paper. We would like to clarify that a comprehensive hyperparameter sensitivity study—including step sizes, momentum parameters, and other key settings—has been included in Appendix E of our submission within the Supplementary Material: zip file. This appendix provides detailed ablation studies showing how different parameter choices affect algorithm performance across various problem settings, along with practical guidelines for parameter selection. If the reviewer requires any specific clarification regarding these studies, we are happy to provide further details.
In summary, we conducted extensive tuning for ZO-SOGD in the adversarial attack scenario. This includes inner/outer step sizes (, ), smoothing parameters (, ), and momentum terms (, ). While ZO-SOGD is somewhat sensitive to these choices, it consistently performs well across a wide range.
- Best performance was achieved with , ;
- Smoothing parameters , yielded stable behavior;
- Momentum led to the strongest attacks (accuracy ).
We will make Appendix E more visible in the main text for clarity. Please let us know if further details are needed.
Question 4: Can authors provide a brief proof sketch?
Response: We thank the reviewer for the constructive comment.
Detailed proof sketches are provided in Appendix C (lines 516–529) for our first-order method (SOGD) and in Appendix D (lines 797–811) for the zeroth-order method (ZO-SOGD). For convenience, we summarize the key ideas below using simplified explanations and light math.
SOGD Proof Sketch:
The analysis starts with Lemma C.2, which bounds the error between the momentum-based estimate and the true inner gradient . Using this, Lemma C.4 shows convergence of toward the best-response solution , i.e., minimizing . Lemma C.5 controls the error in versus the ideal second-order term involving and . Lemma C.8 then bounds the error . Lemma C.9 focuses on the hypergradient estimator and its closeness to the true gradient . Finally, Lemma C.11 establishes a projection bound that aggregates the above errors to complete the regret analysis.
ZO-SOGD Proof Sketch:
For the zeroth-order case, Lemma D.9 quantifies the error between the zeroth-order estimate and the smoothed gradient . Lemma D.11 shows that the iterates converge to a smoothed best response . The auxiliary vector estimates are handled in Lemma D.15, bounding the difference from the smoothed version of , while Lemma D.17 ensures the convergence of to . Lemmas D.19–D.22 complete the proof by addressing hypergradient approximation and projection error bounds.
We will include this summary in the main paper in the revised version, if space permits.
Thanks for the comment. All the problems are solved and the reviewer with raise the final rating.
This paper studies online bilevel optimization using both first-order and zeroth-order oracles. It achieves stochastic sublinear regret without relying on window smoothing, leveraging novel search directions. In the zeroth-order setting, hypergradient estimation is performed using only function values. The effectiveness of the proposed algorithm is demonstrated through numerical experiments.
优缺点分析
Strengths:
-
Unlike prior approaches, the proposed algorithms do not rely on window smoothing, thereby avoiding the significant computational overhead associated with evaluating multiple past gradients at each iteration. This makes the method more scalable and better suited for real-time or large-scale applications.
-
In previous works, hypergradient estimation typically requires multiple iterations of the inner objective. In contrast, this paper achieves hypergradient estimation using a single subproblem iteration.
-
In the zeroth-order setting, the algorithm efficiently approximates computationally intensive second-order terms (Hessian and Jacobian) using only a function value oracle.
Weaknesses :
-
Although a variance reduction-based approach is used in Equation 6a, the paper lacks a clear discussion of the motivation behind this choice and its specific contributions to the overall method. A more explicit explanation of how variance reduction influences the results would strengthen the presentation.
-
The paper would benefit from a clearer articulation of its limitations and potential directions for future work.
问题
-
What is the purpose of Lemma 2.1? It appears to be a direct application of the exponential moving average (EMA) and may not require a separate lemma.
-
In theorem 2.6, what can be said about the optimality of the regret bound with respect to the time horizon ? Which specific factor is responsible for the improvement over OSO?
-
Is the dimensional dependence in the zeroth order setting optimal ? In theorem 3.2, Line 200, the dimensional dependence seems to be but in Remark 3.3, this is reported as . Which statement is correct?
局限性
The discussion of limitation is rather superficial. For example, there are many more natural questions that are left open after reading this paper. Are the regret bounds optimal? If not, what are the lower bounds for regret in each problem setting?
最终评判理由
My initial score was positive. The authors addressed my concerns and I raised the score.
格式问题
none
We thank Reviewer yzQD for the detailed questions. We address each point below:
Weakness 1: Although a variance reduction-based approach is used in Equation 6a, the paper lacks a clear discussion of the motivation behind this choice and its specific contributions to the overall method. A more explicit explanation of how variance reduction influences the results would strengthen the presentation.
Question 1: What is the purpose of Lemma 2.1? It appears to be a direct application of the exponential moving average (EMA) and may not require a separate lemma.
Response: We appreciate this constructive comment. Lemma 2.1 plays a pivotal role in enabling online learning without relying on window smoothing. As discussed below Lemma 2.1 (Lines 104--107), for specific choices of and , the time-smoothed gradient forms a momentum-type search direction. Our key insight is that, instead of using time-smoothed gradients directly as a search direction (as in previous online bilevel optimization (OBO) methods such as OAGD, SOBOW, and SOBBO), we can design momentum-type updates to achieve local regret bounds without smoothing.
However, our initial theoretical analysis revealed that achieving regret bounds via momentum updates—without gradient smoothing—would require assuming bounded inner-level gradients, a restrictive condition under our strong convexity assumption (Assumption 2.2). To address this, we introduce Equation (6a), which replaces time-averaged gradients with a variance reduction momentum-based update that avoids such assumptions. Specifically, the first two terms () implement a momentum update and eliminate time smoothing, while the remaining term controls noise amplification in the nested bilevel structure and avoids additional assumptions on bounded gradients.
Furthermore, we show that under a suitable choice of , this variance reduction mechanism in Equation (6a) enables us to achieve regret rates without window smoothing. This represents a significant improvement over the typical regret rates obtained by standard online single-level and bilevel methods. We will expand this discussion in Section 2 to better motivate the specific form used.
Question 2: In Theorem 2.6, what can be said about the optimality of the regret bound with respect to the time horizon ? Which specific factor is responsible for the improvement over OSO?
Response: Thank you for your comment. As we do not provide matching lower bounds, we do not claim these bounds are optimal. In fact, establishing lower bounds for bilevel regret remains an open problem in both first-order and zeroth-order settings. To our knowledge, all prior works—including those listed in Table 1 (e.g., OAGD, SOBOW, SOBBO)—focus exclusively on upper bounds. For further discussion, please see our response under Limitations.
Below, we discuss the regret bound with respect to the time horizon for both OBO and online single-level optimization (OSO).
As discussed in Remark 2.7 and shown in Theorems 3.1 and 4.2 of OSO-type methods [Hallak et al., 2021], in OSO, dynamic regret bounds include a regularity term (e.g., , representing online time variations) that quantifies the temporal change in the cost functions. The average regret bound is meaningful only when this regularity term is sublinear. Please see also [Besbes et al., 2015], [Hazan, 2017], [Aydore et al., 2019], [Zhang et al., 2020], and [Guan et al., 2024].
Our method's regularity term is analogous to those in prior work, but the key difference is that our dependence on and yields strictly better scaling than previous methods when grows slowly. For example, in Theorems 3.1 and 4.2 of [Hallak et al., 2021], the stochastic component appears as or , along with regularity terms. Hence, their average regret bound scales as under the assumption that the regularity term is sublinear. However, under similar assumptions, our average regret bound scales as , which shows significant improvement in dependence on .
Similarly, in the OBO setting (see Table 1), all existing methods incorporate regularity terms such as , , or in their regret bounds. Specifically, all existing methods (e.g., OAGD, SOBOW, SOBBO) yield regret bounds like or with ; please see [Ataee Tarzanagh et al., 2024], [Lin et al., 2024], and [Bohne et al., 2024].
However, our average regret bound features a term that is multiplied by a factor strictly better in than previous methods, which often incur an rate due to window averaging (please refer to Table 1).
Overall, our method is able to achieve regret bounds that improve over both single-level and bilevel settings.
Question 3: Is the dimensional dependence in the zeroth order setting optimal? In Theorem 3.2, Line 200, the dimensional dependence seems to be , but in Remark 3.3, this is reported as . Which statement is correct?
Response: Thank you for catching this important discrepancy. In Remark 3.3, we wrote to highlight the dependence on both the upper-level and lower-level dimensions. However, we simplified the expression and omitted the exponent for brevity. The correct dimensional dependence appears in Theorem 3.2, where the regret bound consists of two dominant terms: the first term scales as , and the second as .
This higher-order dependence is expected in the zeroth-order bilevel setting, where gradients, Jacobians, and Hessians must be estimated using only function evaluations.
We will revise Remark 3.3 to clarify that although reflects the qualitative dimensional dependence, the exact bound includes powers stated in Theorem 3.2.
Weakness 2: Paper lacks clear articulation of limitations and future work
Limitations: The discussion of limitations is rather superficial. For example, there are many more natural questions that are left open after reading this paper. Are the regret bounds optimal? If not, what are the lower bounds for regret in each problem setting?
Response: Thank you for raising this important point. We agree that the current discussion of limitations can be expanded. While our work establishes the first sublinear regret guarantees for stochastic OBO without window smoothing, we do not claim these bounds are optimal. Indeed, deriving lower bounds for bilevel regret remains an open and challenging problem, in both first-order and zeroth-order settings. To date, all existing works have focused exclusively on upper bounds (please refer to Table 1).
We will revise the limitations section to explicitly state that the optimality of our regret bounds is not known, and that lower-bound analysis is an important direction for future work.
Beyond this, another limitation is that the work focuses on nonconvex online bilevel learning, which is more challenging than the convex case. Extending our results to cover convex and strongly convex bilevel problems, particularly in the zeroth-order setting, would be a valuable direction for future research. While first-order convex and strongly convex guarantees exist for some prior OBO methods [Tarzanagh et al., 2022], no such guarantees currently exist for zeroth-order bilevel optimization. This represents an open and important area for further investigation.
We will add a discussion of these potential improvements and outline lower bound questions for future work.
I want to thank the authors for addressing my concerns. I read the rebuttal and I have no further questions.
This paper studies the online bilevel optimization problem with stochastic and zero-order feedback. It proposes a new performance measure for OBO that avoids the need for window-smoothed objectives, as used in prior work. The paper then presents algorithms for both the stochastic and zero-order feedback settings, providing theoretical guarantees under the proposed performance metric. Experiments on black-box attacks and loss tunning are conducted to demonstrate the effectiveness of the proposed methods.
优缺点分析
Strengths:
- The proposed performance measure eliminates the need for smoothing used in previous works, making it more suitable for scenarios where the objective function changes over time.
- Experiments are conducted to empirically validate the effectiveness of the proposed algorithms.
Weaknesses:
-
Clarity of the performance measure: The notation for and is somewhat confusing to me, as it appears these notations are used to denote both expected and individual functions in (1), leading to ambiguity in understanding the performance measure. If represents the expected function, it is unclear to me over what randomness the expectation in equation (2a) is taken. On the other hand, if denotes the individual realization, it seems that BL-Regret cannot reach zero even with optimal predictions at each iteration. Additionally, the BL-Regret depends on , which is a parameter of the algorithm itself. It is not clear what is an appropriate choice of for evaluating different methods, or how to fairly compare regret across methods using different schedules.
-
Theorem 2.6 and Theorem 3.2: The regret bounds presented appear superlinear in the worst case, as can be as large as . In such scenarios, the guarantees become worse than the trivial regret that one could achieve simply by choosing arbitrary predictions within a compact set .
-
Claims about convergence rates: In line 161, the authors claim the proposed method surpasses the rate for online stochastic optimization (OSO). However, this argument is unclear because the bound still depends on , which in the worst case can grow as , resulting in a convergence rate worse than . The same concern applies to the claim made at line 212.
-
Experiments: Although the paper focuses on OBO with time-varying and , the two experimental cases studied appear essentially to be offline learning problems, where it is unclear why modeling time-varying objectives is necessary. Even if the validation datasets and training datasets vary, in the offline setting it would be more natural to assume they are drawn from the same fixed distribution rather than a time-varying distribution.
问题
Questions:
- Could you provide additional justification for the proposed performance measure?
- Could you offer a more detailed discussion on the tightness of the proposed regret bounds?
- In the experiments, what is the benefit of applying an algorithm designed for time-varying objectives instead of treating the data as stochastic samples from a fixed distribution?
局限性
NA
最终评判理由
One of my concerns about the paper lies in the theoretical guarantee, specifically the dynamic regret bound (Theorem 2.6). This bound implies that the dynamic regret is sublinear only when and . In the worst case, the bound becomes superlinear at a rate of . By contrast, the algorithms in Table 1 achieve at worst dynamic regret bound (though the performance measure is slightly different). Since no lower bound is provided, it is unclear whether this gap is due to the inherent hardness of the problem or the suboptimality of the proposed algorithm.
That said, I acknowledge that the paper introduces an interesting approach by avoiding window smoothing, and my concerns regarding the experimental results have been adequately addressed. Therefore, I am willing to raise my score to 4.
In the revision, I encourage the authors to provide a more comprehensive discussion of their results. For example, in Remark 2.7, it would be helpful to clearly specify the conditions under which the regret is sublinear, identify the scenarios in which the proposed bound is tighter than existing approaches, and discuss its limitations in the worst-case setting.
格式问题
NA
We thank Reviewer LBrC for the detailed feedback.
Q1: Additional justification for proposed performance measure? Weakness 1- Clarity of the performance measure: - The notation for and ... iteration.
Response: Thank you for your comment. We use and to denote individual stochastic realizations, while the expected functions are defined as: and ,as introduced in Equation (1).
The key insight is that represents the output of a stochastic bilevel optimization algorithm at iteration (detailed in Section 2). As such, is a random variable that depends on all stochastic information observed by the algorithm up to time , including the random samples used in the estimation of (implicit) gradients and Hessians in earlier iterations.
Our bilevel regret is defined in Equation (2a) as , where the gradient of the expected function is evaluated at the random point . Although is an expected function (not a random sample), its gradient becomes a random quantity due to its dependence on , which is itself random.
The outer expectation in our regret is over all randomness in the algorithm up to time , including sampled gradients and Hessians (see Algorithms 1 and 2). Regret thus measures the expected stationarity gap, with bounds given in Theorems 2.6 and 3.2.
Q1: Additional justification for proposed performance measure? Weakness 1- Clarity of the performance ... BL-Regret depends on , ... schedules.
Response: Thank you for your comment. This dependency is standard in the analysis of online algorithms involving projected gradient descent. We refer the reviewer to Definitions 2.2 and 2.5 in the seminal work of [Hazan, 2017], where the regret for projected gradient descent is defined as:
with updates
where
Their Proposition 2.3 shows projected gradient is a continuous mapping, so Brouwer’s fixed point theorem guarantees at least one fixed point where projected gradient (and thus average regret) vanishes. Definition 2.5 also shows single-level regret depends on stepsize, as does our bilevel regret on . Our analysis extends these results to the bilevel case. It is important to contextualize our regret relative to prior OBO work. Previous OBO methods in Table 1 (e.g. OAGD and SOBOW) address the unconstrained case (), where the projection operator becomes the identity. Thus, our bilevel regret definition (Eq. 2b) simplifies to
which is independent of and matches the regret definitions used by OAGD and SOBOW. Our work extends OBO to the constrained setting and establishes an improved bound. The stepsize in Theorem 2.6 (Eq. 15) ensures convergence and is standard in non-convex online optimization.
For fair experimental comparison (Section 4):
(1) In unconstrained settings, all methods use plain gradient descent, so regret is independent of ;
(2) In constrained settings, all use the same projection and tune .
Q2: Detailed discussion on tightness of regret bounds? Weakness 2- Theorem 2.6 and Theorem 3.2: The regret bounds presented appear superlinear ... a compact set .
Response: Thank you for your comment. It is widely recognized that it is impossible to achieve dynamic regret in general. For example, [Zhang et al. (2018)], on Page 3, Section 2.2 states: "While it is impossible to achieve a sublinear dynamic regret in general, we can bound the dynamic regret in terms of certain regularity of the comparator sequence or the function sequence."
The condition is standard in OBO literature [Tarzanagh et al., (2022)] and online single-level optimization [Besbes et al., (2015)].
Even with , our bound scales as compared to window-smoothed methods requiring , which give regret; albeit neither regret is usable as they are not sublinear.
When for some , the number of objective functions differing from previous ones can scale linearly with , making good performance impossible—information from to is uninformative for . consists of and . We assume for , and provide examples where in many online settings.
Example 1:
Consider , where
, .
Then,
By ,
Then,
As , approaches a constant, i.e., grows slower than .
Moreover, if , then
Since , we have .
Example 2: Let , , and define
where and .
Then and by bounding term by term one finds
In summary, when the variations decay faster than , we obtain . If the decay is like , then .
Weakness 3- Claims about convergence rates:.... same concern applies to the claim made at line 212.
Response: We appreciate the reviewer’s comment.
As discussed in response to Weakness 2 and Remark 2.7, and as shown in Theorems 3.1 and 4.2 of [Hallak et al., 2021], dynamic regret bounds in single-level online optimization include a regularity term (e.g., or ) that quantifies the temporal variation of the cost functions. The rate is meaningful only when this regularity term is sublinear; please also see [Aydore et al., 2019] and [Zhang et al., 2020]
Specifically, our method's regularity term is analogous to those terms in prior work, but the key difference is that our dependence on and yields strictly better dependence than that of previous methods when grows slowly. For example, in Theorems 3.1 and 4.2 of [Hallak et al., 2021], the stochastic component appears as alongside regularity terms. Hence, their bound also depends on a regularity term, and it is commonly assumed that this term is sublinear in order to achieve a sublinear regret.
Similarly, in the bilevel setting ( please refer to Table 1), all existing methods incorporate regularity terms such as , , or in the regret bounds. These methods (e.g., OAGD, SOBOW, SOBBO) yield regret bounds like or , with ; please see [Ataee Tarzanagh et al., 2024] and [Lin et al., 2024]. In contrast, our approach features a term that is multiplied by a factor strictly better in than previous methods, which often incur a due to window averaging; please refer to Table 1.
Q3: Benefit of time-varying ... from fixed distribution? Weakness 4- Experiments: Although the paper .. .a time-varying distribution.
Response: We respectfully disagree with the characterization of our experiments as "essentially offline learning problems." Both experimental settings are genuinely online with time-varying objectives. For example, in the adversarial attack setting (Section 4), the optimal attack strategy changes abruptly between stages (see Eq. (28)), reflecting dynamic defenses in real-world scenarios. These are not i.i.d. samples, but represent structural shifts requiring online adaptation of hyperparameters . This setting is well-established in the online learning literature, where distribution shifts are a central motivation for online algorithms [Hall and Willett, 2015].
Our method is also well-suited for online meta-learning (OML), where agents must quickly adapt to sequential tasks; please see Section 4.3 in [Ataee Tarzanagh et al., 2024] for discussion on OBO for OML. For additional applications, including traffic flow prediction and wireless network control, please refer to Appendix A in [Lin et al., 2024].
Thank you for the detailed explanation of the notations and the clarifications regarding the experiments. However, I remain unclear about the arguments concerning the regret guarantees. While I agree with the authors that sublinear dynamic regret is generally unachievable, the results presented in the paper appear to be superlinear in the worst case, specifically of the order $O(T^{\frac{4}{3}})$. It seems that the proposed method only achieves sublinear regret under the condition $\Delta_T = o(T^{\frac{2}{3}})$. In the case of single-level optimization (Besbes et al. 2015), the worst-case regret is $O(T)$, which implies that it does not exceed the trivial bound derived from the boundedness of the loss function. Regarding the comparison with Tarzanagh et al. (2022), the definition of regret differs from that used in this paper. Nevertheless, their bound of $O(T/W + H_{1,T} + H_{2,T})$ seems still result in $O(T)$ regret in the worst case when $W = \sqrt{T}$.
Response: We appreciate the reviewer for providing additional feedback. Below, we provide further discussion regarding the worst-case superlinear regret bound.
We agree with the reviewer that under worst-case regularity, such as , the dynamic regret bound of our method is at most . As previously discussed with Reviewer yzQD, we do not claim any lower bounds, and the regret may be loose in specific worst-case scenarios involving the regularity . Superlinear regret is often unavoidable when the environment exhibits linear variation, and many existing online algorithms can similarly incur superlinear dynamic regret in such non-stationary settings.
We first review seminal works on online gradient descent [Zinkevich (2003)] and online mirror descent [Hall and Willett (2015)] that can lead to superlinear regret bounds under worst-case regularity shifts. We then highlight recent results showing that such bounds are not uncommon in online learning algorithms operating in non-stationary environments—particularly when assumptions similar to ours on the rate of change are not imposed. Next, we compare these results to the regret bounds in [Tarzanagh et al. (2022)] and [Besbes et al. (2015)]. Finally, we discuss adaptive algorithms in the single-level setting that have the potential to be extended to the bilevel setting to improve regret dependence on and regularity measures such as .
Superlinear regret in worst-case regularity scenarios under non-stationary environments. In the seminal work of [Zinkevich (2003)] for online gradient descent and its extension to online mirror descent by Theorem 2 of [Hall and Willett (2015)], the regret is , where is the path-length of the comparator sequence . This bound becomes superlinear when , resulting in regret. For example, consider a sequence of loss functions where the optimal solutions drift linearly: for some unit vector with . In this case
This scenario naturally occurs when tracking a moving target with constant velocity or in trending non-stationary environments.
We now discuss further examples showing that worst-case superlinear regret bounds are not uncommon in online learning algorithms operating in non-stationary environments, particularly when assumptions similar to ours on the rate of change are not imposed.
-
Yi et al. (2021) in Table 1 show that their Algorithm 1 achieves regret bound in the convex setting. When , this leads to a regret of .
-
Guo et al. (2022) in Theorem 3 and Remark 3 show that their online learning algorithm achieves regret in the convex setting. When , this gives a regret of .
-
Theorem 3.2 in [Zhang et al. (2020)] shows that the regret includes a term in the convex setting, where denotes the maximum squared variation of the objective between consecutive time steps. When , this leads to a regret of , which is superlinear. Theorem 4.2 shows that the accumulated gradient of the smoothed functions has regret bounded by , where is the sum of expected function variations over time. When , the regret becomes , again resulting in a superlinear bound.
-
Theorem 1 in [Kalhan et al. (2021)] shows the regret bound , where is the function variation and is the gradient variation. When , this leads to a regret of .
-
Theorem 1 in [Chen et al. (2017)] shows that the regret bound is with step size where is the path-length of the comparator sequence. When , this leads to a regret of . In Corollary 1, they set to get .
Comparison with OAGD [Tarzanagh et al. (2022)] and SOBBO [Bohne et al. (2024)]. As the reviewer pointed out, our regret definition differs from that of Tarzanagh et al. (2022), as ours is projection-based and window-free in the stochastic setting, which requires more technical effort to achieve a sublinear regret bound. Besides many algorithmic novelties of our algorithm such as single step, recursive momentum update, window-free, and derivative-free algorithm, we compare their regret bound below.
-
In deterministic bilevel optimization, Tarzanagh et al. (2022) show that with and the path-length regularity , the dynamic regret is . Importantly, regardless of how benign the environment is (i.e., how small are), their regret bound remains . In contrast, our method adapts to the complexity of the environment. While our worst-case bound is , in benign settings—such as when the regularity —we can achieve sublinear regret. Moreover, as the regularities decrease further, our regret bound continues to improve.
-
In stochastic bilevel optimization, our regret remains sublinear under suitable conditions on regularities. Importantly, in the stochastic bilevel optimization setting, our method achieves an (average regret) dependence on the noise variance, which improves dependency on over the bound for stochastic OBO [Bohne et al. (2024)] and surpasses the rate in prior work; please refer to Remark 2.7.
Comparison with [Besbes et al. (2015)]. We agree with the reviewer that Besbes's work has better dependency on function variation . We note that our regret bound in the nonconvex case may not be directly comparable to the results of Besbes et al. (2015) in convex settings. Nevertheless, the key to achieving favorable dependency in the regret bound is to use adaptive function variation-based step size and batch size. For example, in their Theorem 3, the authors select an adaptive batch size and step size as follows:
where is the batch size (as defined in their work) and is the step size at iteration , both depending on the cumulative function variation . This adaptive parameterization is essential for attaining the improved regret bound . However, the quantity is typically not accessible in practice. Therefore, we adopt a simple step size for some constant , without incorporating any regularity-dependent terms.
Adaptive algorithms for improved regret. Finally, we note that a similar adaptive step size and batch size strategy in Eqn (117) could also improve our dependence on regularity measures such as function variation. However, this type of information—such as function variation or path length—is typically not available in practice. Therefore, it is crucial to develop an adaptive algorithm that can achieve the optimal rate without prior knowledge of or , following the expert-tracking algorithm of [Zhang et al. (2018)], which improves to . Exploring how to extend such adaptive techniques to improve regret bounds in bilevel learning remains an open and promising research direction.
The paper presents a unified first-order (SOGD) and zeroth-order (ZO-SOGD) framework for online bilevel optimization (OBO) that achieves sub-linear stochastic bilevel local regret without any window smoothing. Key ingredients are 1) momentum-type projected search direction that simultaneously updates leader, follower, and auxiliary variables in a single loop, 2) oracle-efficient hyper-gradient estimation that needs only one inner linear-system step per round, and 3) a fully function-value–only variant that estimates gradients, Hessians, and Jacobians via two-point finite differences.
Theoretical contributions include tight (first-order) and (zeroth-order) regret bounds under standard non-convex/strongly-convex assumptions. Empirical studies on online adversarial attacks and parametric loss tuning under distribution shift confirm the framework’s practical gains over state-of-the-art OBO and single-level baselines.
优缺点分析
Strengths:
- Removing the window-smoothed regret requirement is a clear conceptual advance.
- Proofs carefully track stochastic variance, path-length, and gradient-drift terms and tighten the noise dependence to which is sharper than the factor in SOBBO/SOBOW.
- Experiments show ZO-SOGD consistently outperforms strong black-box attack optimizers (ZO-O-Adam, SignSGD, etc.) in success rate and perturbation norm while maintaining reasonable run-time, and SOGD matches or exceeds OAGD / SOBOW on adaptive loss tuning with lower latency. Weaknesses: The code for the experiments is not provided in the supplementary material. It must be required for reproducibility.
问题
-
Could you make a GitHub or submit your codes and experiments for the sake of reproducibility of your result?
-
Could you elaborate on concrete scenarios where the proposed bilevel local regret better captures performance than traditional window-smoothed metrics? Have you identified tasks where the two measures give conflicting evaluations?
-
The theoretical analysis uses order- step sizes. In the experiments I suppose you tune them heuristically, did you observe a clear regime where the theory-suggested schedule outperforms alternative decays (e.g., )?
局限性
Yes.
最终评判理由
The author’s reply resolves my concerns; I’m keeping my current rating. However, the proofs read quite densely and could be clarified with revisions. Compared with other bilevel optimization papers I’ve read, I had more difficulty following these arguments.
格式问题
None.
We sincerely thank Reviewer ETap for the valuable and constructive feedback. We address each point below.
Question 1: Could you make a GitHub or submit your codes and experiments for the sake of reproducibility of your result?
Response: Thank you for raising this important point. We fully agree on the importance of reproducibility. To that end, we contacted the Area Chairs to inquire whether we could share an anonymous code link during the rebuttal phase and received the following guidance:
"Because of known concerns on identity leakage, we are prohibited using any links in the rebuttal, including, but not limited to anonymous or non-anonymous URL links, or updating your submitted GitHub repository."
Accordingly, while we are currently unable to share the code during the rebuttal phase, we will release the complete codebase upon acceptance to facilitate full reproducibility of our results.
Question 2: Could you elaborate on concrete scenarios where the proposed bilevel local regret better captures performance than traditional window-smoothed metrics? Have you identified tasks where the two measures give conflicting evaluations?
Response: We thank the reviewer for this insightful question. While the bilevel local regret values themselves may not always show dramatic differences between window-smoothed and non-smoothed approaches, the key distinction lies in the underlying algorithmic mechanisms and their real-world performance. Our contribution is not just the regret metric, but the first and zeroth order novel recursive momentum-type search directions (Eq. 6) that fundamentally differs from window-smoothed gradient methods used in prior work.
Traditional OBO methods like OAGD, SOBOW, and SOBBO all rely on window-smoothed search directions of the form:
which inherently averages past gradients. This averaging introduces lag and can lead to conflicting gradient directions when objectives change rapidly. In contrast, our recursive momentum approach
maintains a single adaptive direction that immediately responds to changes while preserving useful historical information through the momentum term.
Experiment: Periodic Objective Oscillations: Following your valuable suggestion, we evaluate a synthetic scenario where the leader's objective exhibits oscillations with frequency parameter :
Table 1: Gradient-based Regret Under Different Change Frequencies
| Method | Time (s) | Our Regret | Smoothed Regret |
|---|---|---|---|
| For (Slow changes) | |||
| OAGD () | 45.2 | 498.0 | 485.3 |
| SOBOW () | 68.4 | 542.6 | 521.8 |
| SOGD () | 22.6 | 596.9 | 596.9 |
| For (Rapid changes) | |||
| OAGD () | 52.3 | 1842.7 | 127.5 |
| SOBOW () | 75.8 | 1205.3 | 186.2 |
| SOGD () | 24.9 | 758.4 | 758.4 |
In rapidly changing environments (), OBO [Tarzanagh et al., (2022)] with smoothing regret () shows extremely low window-smoothed regret (127.5) because opposing gradients cancel out when averaged, suggesting near-optimal performance. However, its local regret (1842.7) reveals it's actually performing poorly—the algorithm is stuck between conflicting directions.
This demonstrates why window-smoothed metrics can be fundamentally misleading: they evaluate performance against averaged gradients that may not represent any actual objective the algorithm needs to track. Our local regret correctly captures that SOGD (758.4) significantly outperforms window-based methods in dynamic environments.
Question 3: The theoretical analysis uses step sizes. In the experiments I suppose you tune them heuristically, did you observe a clear regime where the theory-suggested schedule outperforms alternative decays (e.g., )?
Response: We appreciate the reviewer's careful observation. From Theorem 2.6, we use and . For the zeroth-order case in Theorem 3.2, we have , with similar scaling for the other parameters. These schedules are derived to balance the bias-variance trade-off in our stochastic bilevel setting and to achieve the regret bound.
While it is difficult to validate this theory in our real-world highly nonconvex experiments--and we heuristically tuned the parameters--we can demonstrate its advantages in synthetic settings. To validate our theoretical step size schedule, we conducted experiments using the periodic objective oscillation scenario from our response to Question 2.
Table 2: Step Size Schedule Comparison on Periodic Objectives
| Step Size Schedule | Our Regret () | Our Regret () |
|---|---|---|
| (Fixed) | 396.6 | 892.4 |
| 325.5 | 850.6 | |
| (Theory) | 596.9 | 758.4 |
The faster decay proves too aggressive for our bilevel setting, especially in rapidly changing environments. This schedule reduces the step size too quickly, preventing the algorithm from adapting to sudden changes in the bilevel landscape.
The advantage of our theoretical schedule becomes particularly pronounced in environments with rapid changes. When the oscillation frequency is high (), the performance gap widens dramatically: the schedule achieves a regret of 758.4, while the schedule yields 850.6. Conversely, in smoother settings with , the performance difference narrows (596.9 vs. 325.6), as rapid adaptation is less critical.
We emphasize that our approach significantly reduces computation time and memory usage, as we do not use window smoothing. Please refer to Table 1.
My concerns have been taken care of by them!
This paper proposes a new online bilevel optimization method. It introduces a novel search direction that does not rely on deterministic window-smoothed regret minimization as existing approaches. It then shows that both first- and zeroth-order (ZO) stochastic OBO algorithms leveraging this direction achieve sublinear stochastic bilevel regret without window smoothing.
All reviewers agree that eliminating the dependence on window smoothing is a significant contribution to online bilevel optimization. In addition, the experiments confirm the effectiveness of the proposed method. A minor concern after the rebuttal phase is that the paper lacks a more comprehensive discussion of its regret bound, particularly with regard to its limitations in worst-case scenarios. It is highly recommended that this issue be addressed in the camera-ready version. Overall, this paper makes important contributions to online bilevel optimization and is recommended for acceptance.