Online Bandit Nonlinear Control with Dynamic Batch Length and Adaptive Learning Rate
We propose a novel bandit-type algorithm for the online nonstochastic control problem using a dynamic batch length and an adaptive learning rate, operating under a much weaker controller stability assumption than exponentially stabilizing notions.
摘要
评审与讨论
This paper addresses the problem of online bandit nonlinear control, proposing a new algorithm called Dynamic Batch length and Adaptive learning Rate (DBAR). The algorithm in this paper is designed to handle a nonlinear dynamical system with a mix of stabilizing and destabilizing controllers, offering stability guarantees and regret bounds.
优点
- The paper provides rigorous theoretical analysis, including stability proofs and regret bounds, which demonstrate the effectiveness of the DBAR algorithm under weaker assumptions. The comparison with existing methods is clear and it strengthens the contribution.
- The introduction of the DBAR algorithm relaxes the requirement for exponentially stabilizing controllers, allowing a broader range of controllers to stabilize the system. It also achieves better result for known and unknown .
缺点
- While the theoretical analysis is thorough, the experiments are somewhat limited. The paper focuses on low-dimensional systems (a 2D linear system and a simple nonlinear system), and it would be beneficial to demonstrate the algorithm's performance in more complex, high-dimensional scenarios to validate the practical applicability of DBAR.
- The writing of this paper is not very clear, especially in Section 2 and Section 3. I suggest using more references, formulas, and illustrations to explain the rationale behind the assumptions and the intuition behind the algorithm design, rather than relying solely on extensive text descriptions.
问题
- Could you include more experimental results on complex systems?
- Could you explain the necessity of Assumption 2.5? You may refer to previous literature to support your explanation.
- Could you clarify the proof ideas and intuition behind Theorems 4.5 and 4.6? The current explanation is quite brief, and it's difficult to grasp the core proof strategy as well as how the earlier algorithm design relates to this proof. I suggest expanding this section and moving some of the lemma proofs, such as Lemma 4.7, to the appendix.
Question 3: Thanks for the recommendation. We will defer Lemma to the appendix and add the following explanation.
To clarify, our proof strategy starts from dividing the expected total cost into mix loss and mixability gap as in [4,5]. This technique is common in hedge setting, which is an instance of multi-armed bandit problems. Since the earlier algorithm relies on the "constant learning rate", they divide the expected cost into the "" and the "" (notice that the denominator of the mix loss and the term inside the exponential function are both , the constant learning rate). However, to deal with the absence of exponentially stabilizing controllers, we need to adopt an adaptive learning rate, thus the proof strategy should be modified to analyze , where a new is an adaptively selected learning rate at batch . Both the modified mix loss and the mixability gap lead to an additional term that is in and . Both terms are on the order of (Lemma ), so they do not increase the regret compared to the earlier work [3].
Note that our adaptive learning rate is determined by . Accordingly, while we obtain additional terms linear in due to adaptive learning rate, we benefit from having the term instead of in a constant learning rate case. Since increases (an adaptive learning rate decreases) when the state norm is in an unstable region, the term decreases to the extent that alleviates the multiplicative exponential term (see in Table , "Dynamic Batching") that arises when using a dynamic batch length. In summary, our final regret has polynomial terms more in regret compared to the previous work, but instead alleviates multiplicative exponential term.
[1] Sontag, "Input to state stability: Basic concepts and results", Nonlinear and Optimal Control Theory, 2008.
[2] Khalil, "Nonlinear Systems", Pearson Education, 2015.
[3] Li et al., "Online switching control with stability and regret guarantees", L4DC, 2023.
[4] van Ervan et al., "Adaptive hedge", NeurIPS, 2011.
[5] de Rooij et al., "Follow the leader if you can, hedge if you must", JMLR, 2014.
Many thanks for reading our rebuttal.
Thank you for your response! I will maintain my positive score and recommend accepting this paper.
Thanks for the feedback again! Please let us know if you have any further questions.
We are very grateful to the reviewer for a thorough evaluation of our work and providing valuable feedback. We will improve the paper based on the provided comments.
We have improved the paper and uploaded a new PDF file with two important revisions:
-
Experiments: We have replaced the experiments in Example with high-dimensional experiments (100 states) on leader-follower systems. The leader is represented by a previous ball-beam (nonlinear) system, and the followers leverage the leader's state to stabilize themselves. Please check the experimental setup in Example (see pages 9-10).
-
Illustration: In a new Figure , we have illustrated a concept of the controller pool for the unknown nonlinear system, and how general an asymptotically stabilizing notion is. We hope that this figure helps the reader to understand the core concepts (see page 4, Figure ).
Responses to the reviewer:
Weakness 1: Thanks for the comment. We have now provided the experiments on high-dimensional systems in Example .
Weakness 2: Thank you for the comment. We will improve our presentation and strive to better illustrate our results. As a first step, we have presented Figure in Section to illustrate a concept of stabilizing notions and the controller pool for unknown system.
Question 1: Yes, we have now provided 100-dimensional systems.
Question 2: The assumptions such as Lipschitz continuous is quite mild. To understand Assumption , we provide the following example (also see the pictorial illustration in Figure 1(b)). Given an unknown system, we can consider different regimes to model the possibilities of the unknown parameters of the system and then design a controller for each possibility. For example, consider a robot operating in an uncertain environment where there are different scenarios that could happen in the environment and we model it by a vector "" for simplicity. We consider different intervals such as or , etc. for "" and then design a controller for each scenario. Now, we have a pool of controllers and we know at least one of them should work but do not know which one since we do not know the exact parameter of the system. If the intervals chosen for each "" and "" are too wide, a stabilizing controller may not exist and then Assumption will be violated. This assumption is always satisfied if we have a rich set of controllers in the given pool, as long as the system is stabilizable (if the system is not stabilizable, then its behavior cannot be controlled no matter what action we take). The core part of Assumption is the existence of a controller satisfying Definition and , which means that we only require that at least one input-to-state stable (ISS) and incrementally stable (IS) controller "exists" in the candidate pool. The ISS and IS principles are a widely understood concept [1, 2] for handling disturbances in nonlinear systems, particularly in applications such as automotive and robotics systems, as it ensures that a vehicle or robot can perform safely despite variations in exogenous inputs like road conditions or obstacles. Specifically, [3] assumed the existence of "exponentially" ISS and IS stabilizing controllers; the stabilizing behavior of such controllers should be exponentially fast, which may be quite restrictive for nonlinear systems to have such controllers. On the other hand, in our work, we only assume the existence of "asymptotically" ISS and IS stabilizing controllers whose stabilizing behavior can be arbitrarily slow. As mentioned in the paper (Lines 69-79), the difference between exponential stability and asymptotic stability is tantamount to the difference between convexity and strong convexity. As witnessed in many machine learning problems, strong convexity is a very restrictive assumption while convexity is more realistic. Therefore, the existence of an asymptotic ISS and IS controller is not a particularly restrictive assumption.
Continued in the next rebuttal (references are also provided).
In this paper, the authors study the online bandit nonlinear control, which aims to learn the best stabilizing controller from a pool of stabilizing and destabilizing controllers of unknown types for a given nonlinear dynamical system. They develop an algorithm, named Dynamic Batch length and Adaptive learning Rate (DBAR), and study its stability and regret. DBAR achieves better regret bounds than Exp3 under even weaker assumptions.
优点
-
The problem of online bandit nonlinear control is well-motivated.
-
This paper weakens the assumptions in previous works and derive a stronger regret bound. The proof looks correct.
-
There are simulations to support the theoretical findings.
缺点
Main concerns:
-
It seems to me that the results heavily depend on the assumption that has finite limitation (Thm 4.2) or (Thm 4.6), this assumption does not seem to be very mild ( has finite limitation is far from the conclusion in Lem 4.3). What will the result be like without such assumptions? Or is there any further justifications of such assumption?
-
In the experiments, the claim 'While the simulations are on low-dimensional systems for illustration purposes, similar observations can be made for high-dimensional systems.' is made. Is this a conjecture, or is there any experimental evidence for this? It seems not hard to do experiments with higher dimensionality.
Some questions and suggestions for writing:
-
The notation first appears in the table, while it is not defined until Def 2.6.
-
The previous papers depend on 'exponential stability'. What is its definition and how is it stronger compared to 'asymptotic stability'?
-
I did not quite understand the motivation of Lemma 4.3. This is too weak to support the assumption .
-
Listing the steps to solve for in Theorem 4.6 could help justify the results.
问题
Please refer to the weakness part.
We are very grateful to the reviewer for a thorough evaluation of our work and providing valuable feedback. We will improve the paper based on the provided comments.
We have improved the paper and uploaded a new PDF file with two important revisions:
-
Experiments: We have replaced the experiments in Example with high-dimensional experiments (100 states) on leader-follower systems. The leader is represented by a previous ball-beam (nonlinear) system, and the followers leverage the leader's state to stabilize themselves. Please check the experimental setup in Example (see pages 9-10).
-
Illustration: In a new Figure , we have illustrated a concept of the controller pool for the unknown nonlinear system, and how general an asymptotically stabilizing notion is. We hope that this figure helps the reader to understand the core concepts (see page 4, Figure ).
Responses to the reviewer:
Main concern 1 (): Thank you for the comment. There may have been a confusion here since only some of our results depend on . In our paper, there are two types of theorems: those that depend on and those that do not. The ones depending on is Thm and the second part of Thm . Thm implies finite-gain stability when , and the second part of Thm implies if . As the reviewer has noticed, these assumptions are restrictive but still extends beyond the exponentially stabilizing notions (geometrically series) and include some asymptotically stabilizing notions. Such theorems were given to fairly compare the existing works between ours. For example, in Table , we presented the result when to compare the regret between ours and the papers with exponentially stabilizing notions.
On the other hand, Thm and the first part of Thm do not depend on . Whatever comes in, we achieve an asymptotic system stability (Thm ), and concurrently, a sublinear regret bound is attained (see "we achieve a sublinear regret bound" in Thm ). In this case, we obtain a higher regret than , but the regret bound is still sublinear. To the best of our knowledge, this is the first result in the literature that a sublinear regret is achieved even exponentially stabilizing controllers do not exist. To illustrate how challenging our setting is, let us further present a one-dimensional system, where the current system state is (newly presented in Figure ). The goal is to achieve a state near , and we would like to detect this stability by observing whether one arrives at a state less than , where is an arbitrarily small positive number. Exponentially stabilizing controllers guarantee to detect the stability in time. However, with an asymptotically stabilizing controller, if the controller is designed to keep the system state unchanged for an arbitrarily long time and then collapse the state towards afterward, one cannot detect the stability before time regardless of how small is. In such a case, even though the controller ultimately achieves the goal, it may take a lot of time to learn whether a closed-loop system would be stable or not. Our algorithm achieves a sublinear regret even if only those troublesome controllers exist and simple (exponentially stabilizing) controllers do not exist at all. This achievement is due to our design of a dynamic batch length and an adaptive learning rate.
To help reader understand which part needs restriction, we will add the relevant explanations in the paper.
Main concern 2 (experiment): We have now provided the experiments on high-dimensional systems in Example .
Continued in the next rebuttal.
Question 1: Thanks for the comment. In the new PDF, we added the explanation in the caption of Table 1: " is the set of destabilizing controllers and denotes its cardinality."
Question 2: As mentioned in the response to Main concern 1, consider a single state (also see Figure ). Exponential stability means there are positive constants and such that , which means that the system should go to an equilibrium exponentially fast (the equilibrium is zero in this case). Asymptotic stability says should eventually go to the equilibrium but there is no restriction on how fast it should go. It is just saying . In many systems in real world, say autonomous systems, we cannot make them go to a desired behavior exponentially fast. This is a very strong notion but is commonly used in control theory since it makes the mathematical analysis much simpler. Our paper adopts the realistic setting that only requires asymptotically stabilizing controllers.
Question 3: Lemma is not to support the assumption . This lemma is to prove Thms and regardless of the assumption. Please notice that Thm does not require .
Question 4: Technically, one can select any for the algorithm to work. It turns out that only affect the final regret only up to a constant. The simplest selection would be . Then the polynomial batch length would start from and subsequently have for , where is a batch number. To sum up, one can think of the batch length as growing proportionally to the square root of . We will add some exposition to the paper to address this comment.
Many thanks for reading our rebuttal.
To address the reviewer's first main concern about , we have slightly revised the introduction part, and added the relevant explanations in the paper. The revisions are blue-highlighted.
Please note that
-
A dynamic batch length allows us to achieve both asymptotic system stability and a sublinear regret. However, we suffer from a multiplicative exponential term in return.
-
In Table 1, when using an adaptive learning rate, we can see that is improved to , and this is when as we mentioned in the previous rebuttal. However, even when is not restricted, is improved to (see Corollary D.10), so it is true that multiplicative exponential term can be alleviated in all cases.
Thus, our algorithm DBAR, the combination of these two components, effectively stabilizes the potential explosion of the system and enjoys the improved regret.
Based on your constructive comments, we have added the following explanations in the updated manuscript:
-
In Lines 88-90 (in the introduction), we specified that a dynamic batch length (designed to grow unboundedly, but the growth amount eventually saturates) contributes to both "asymptotic system stability" and "a sublinear regret".
-
In Lines 95-97 (in the introduction) and Lines 311-312, we specified that an adaptive learning rate contributes to alleviating a multiplicative exponential term in all cases. Moreover, we also indicated that the regret is achievable for a specific class of stabilizing controllers.
-
In Lines 256-260, we again emphasized that a dynamic batch length contributes to both "asymptotic system stability" and "a sublinear regret", and this was possible due to the design of . For example, if we pick a geometric sequence (1,2,4,8, 16,...) as a dynamic batch length, both stability and a sublinear regret are not achievable.
Many thanks.
Thanks for your detailed response. My main concerns are addressed and I have increased my score.
Thank you very much! Please let us know if you have any further questions.
This paper studies the problem of online bandit non-linear control where both transition dynamics and cost functions are non-linear and adversarial, and we only receive bandit feedback on either of them. Removing the assumption on exponentially stabilizing controllers, this paper ensured regret with the help of only asymptotically stabilizing controllers.
Technically, this paper uses geometricly increasing batch lengthes to remove the requirement of exponentially stabilizing controllers, and further make the dependency to via an adaptive learning rate tuning scheme.
优点
- The removal of exponentially stabilizing policies looks pretty good.
- The resulting regret is satisfactory.
- The algorithms are stated and motivated clearly.
- Results are supplemented with numerical evaluations.
缺点
From the main text, it looks like the two main techniques, geometrically-increasing batch lengthes and geometrically-decaying learning rates, are the main technical contributions of this paper. But these two techniques seems both extensively studied in other online learning problems. So --
- How are they different from those used in other online learning problems? Say, do you specifically adapt them to this control theory problem?
- Do they bring special difficulties under this specific setup of nonlinear control? Say, do they break some previous analysis ideas in online nonlinear control?
- Why previous works didn't use them?
[EDIT. After rebuttal, it turns out that batch lengths and learning rates are both not simply geometric, but follow some more meticulous designs.]
Minor:
- To create a reference inside parentheses, please try using the command
\citep{}instead of manually typing(\citet{}). - Section 2 is notation heavy. I suggest the authors to add subsections / paragraphs to group the definitions.
- The results part (especially Section 4.2) contains a huge amount of formal statements with almost no informal descriptions of them; it is definitely not enjoyable to read. I suggest the authors to replace them with informal versions (and also add a paragraph of intuitive description before/after each of them) and defer the formal ones into the appendix.
问题
See Weaknesses.
We are very grateful to the reviewer for a thorough evaluation of our work and providing valuable feedback. We will improve the paper based on the provided comments.
We have improved the paper and uploaded a new PDF file with two important revisions:
-
Experiments: We have replaced the experiments in Example with high-dimensional experiments (100 states) on leader-follower systems. The leader is represented by a previous ball-beam (nonlinear) system, and the followers leverage the leader's state to stabilize themselves. Please check the experimental setup in Example (see pages 9-10).
-
Illustration: In a new Figure , we have illustrated a concept of the controller pool for the unknown nonlinear system, and how general an asymptotically stabilizing notion is. We hope that this figure helps the reader to understand the core concepts (see page 4, Figure ).
Responses to the reviewer:
Weakness (Geometrical series and adaptation to control problem): Thanks for the comment. Our first technique is polynomially increasing batch length. We carefully designed the batch length to concurrently achieve asymptotic system stability and sublinear regret. It turns out that if we design the batch length to geometrically increase, we cannot achieve sublinear regret. Thus, the increase amount saturates as time goes by, which renders our algorithm more practical. The second technique is adaptive learning rate based on the state norm. It does not geometrically decay; we only decrease the learning rate if the state norm is large (state is unstable), and subsequently increase the learning rate if the state returns to a stable area.
In online learning, the agent may progressively lengthen the batch length as the agent learns the system. Such a strategy is often used in incremental learning or curriculum learning, where the data availability or task complexity may increase over time. For example, [1] used geometrically increasing (doubling) batch length as they try to learn longer Atari games and escape local optima learned from short games. However, such a setting is different from ours, since we focus on a "single trajectory setting" in the control problem. Often, in a single trajectory setting, geometrically increasing batch length will lead to learning instabilty due to higher variance in learning a good policy. Thus, instead of using geometrically increasing length, we have carefully designed a polynomially increasing length to both guarantee stability and enough exploration.
For the learning rate, in the context of online learning, it is well-known that reducing the learning rate will lead to early exploration and later stabilization. The work [2] adapts this advantage of decreasing learning rate to control problem and presents an algorithm based on non-increasing learning rate, but the paper only proves the results for constant learning rate. However, our claim is that decreasing the learning rate, regardless of the current state, does not significantly improve the control performance. We provide theoretical guarantees on our scheme based on the stability of state norm, and thus the rate is not necessarily non-increasing.
Weakness (special difficulties and comparison with prior works): The previous work [2] assumed the existence of exponentially stabilizing controllers. The stabilizing behavior of such controllers should be exponentially fast, which may be quite restrictive for nonlinear systems to have such controllers. In that case, they do not need to use dynamic batch length since it does not take a lot of time to learn whether a closed-loop system would be stable or not. In our challenging setting where those simple controllers may not exist (stabilizng behavior can be arbitrarily slow in our case), a previously used "fixed batch length" fails to identify the system stability. Thus, we need a dynamic batch length to deal with a removal of exponentially stabilizing controllers. The use of a dynamic batch length in turn necessitates the use of an adaptive learning rate. For more details on the distinction between exponential and asymptotic notions, see Appendix A.
Minor 1: Thanks for the recommendation. We updated the paper accordingly.
Minor 2: To help readers understand Section 2 (assumptions and definitions), we provided the glossary on Appendix B. We will guide the readers to refer to Appendix B while reading Section . Based on your comment, we have also presented Figure to illustrate the concept of an asymptotically stabilizing notion in the given controller pool.
Minor 3: Thanks for the comment. We will soon add informal statements accordingly and defer some of formal theorems to the appendix.
[1] Fuks et al., "An Evolution Strategy with Progressive Episode Lengths for Playing Games", IJCAI, 2019.
[2] Li et al., "Online switching control with stability and regret guarantees", L4DC, 2023.
Many thanks for reading our rebuttal.
Dear Authors,
Thank you for your response! Yes, I misread your Assumption 3.1 as . You mentioned that "It turns out that if we design the batch length to geometrically increase, we cannot achieve sublinear regret." -- can you kindly point me to the corresponding discussions in the text or give a quick justification here?
I also admit that I interpreted your learning rate tuning too easily. Can you give a quick explanation on why "It indicates that the learning rate decreases in unstable states and increases back to the initial value when the state norm returns to a stable region" and also include it after Line 302 in your revision?
I agree that both of them seem to be new. I'd be happy to increase my rating if you can let me understand what issues these two techniques are trying to tackle, and how they tackled these issues.
Best, Reviewer 7GNK
Thanks for the comments! We are really happy to hear back.
First of all, for a dynamic batch length, Assumption 3.1 indicates that , which means the ratio of two consecutive batch lengths should approach 1 as time goes by (this includes polynomially increasing batch). The necessity of this assumption arises in the equation (14) (see Lines 993-999 in the Appendix), where we studied the asymptotic system stability. If we have , then the last part of (14) would be instead of for some , and this quantity is not guaranteed to converge to 0 since the number of Breaks goes to infinity as the number of batches goes to infinity. In other words, unless we have as our Assumption 3.1. Thus, the asymptotic system stability is violated. Not only that, the equation (14) is a core part of the sublinear regret proofs as in Line 1481 (Algorithm 1), Line 1637 (Algorithm 2), Line 1833 (Algorithm 3). Thus, if we have the ratio greater than 1, we cannot have a sublinear regret. Picking the ratio of increasing batch length to be 1 is perhaps the only way to obtain both asymptotic system stability and a sublinear regret, and we particularly picked polynomially increasing batch length among such batch length designs to achieve regret for the case .
We really thank the reviewer's comment and we included a brief discussion with blue highlight (Lines 256-260) under Assumption 3.1 in the updated manuscript.
Second, for an adaptive learning rate, Line 27 in Algorithm 1 says that , where is an initial learning rate. and are determined in Lines 11-20 in the algorithm. To summarize, if the norm of the first state in the next batch is large ( large), we adjust to be large ( is increased by 1, and is increased according to the norm). Subsequently, if is small enough (smaller than ), we let be , meaning that we have exactly . This implies that starting with an initial learning rate, if the state norm increases, a learning rate may decrease accordingly; however, if the state norm is stabilized, a learning rate increases back to an initial learning rate. We still can obtain a sublinear regret without this technique, but this allows us to alleviate the multiplicative exponential term in the regret bound ( is improved to ).
As you recommended, we included the following brief sentence with blue highlight in Line 302 in the updated manuscript: "Since increases when the state norm is large, and resets to zero for sufficiently small state norm, the corresponding learning rate decreases in unstable states and increases back to the initial value when the state norm returns to a stable region."
To conclude, our dynamic batch length lets us achieve both asymptotic system stability and a sublinear regret, and an adaptive learning rate further improves this sublinear regret to in the case when . Thus, our algorithm DBAR, the combination of these two components, effectively stabilizes the potential explosion of the system and enjoys the improved regret.
Many thanks!
Thank you for your explanations and revisions. I have updated my review accordingly.
Thank you. Please let us know if you have any further questions!
The paper addresses the online bandit nonlinear control problem, where the objective is to learn an optimal controller for a nonlinear dynamical system amid unknown stabilizing and destabilizing controllers. The authors introduce the DBAR (Dynamic Batch length and Adaptive learning Rate) algorithm, which adapts batch length and learning rate to improve control stability and minimize regret without requiring exponentially stabilizing controllers. Unlike existing approaches that need stronger stability assumptions, DBAR works with a weaker, asymptotic stability notion. This flexibility allows DBAR to achieve stability and a regret bound even when the pool of stabilizing controllers is limited or includes only asymptotically stabilizing ones.
Theoretical contributions include proving asymptotic and finite-gain stability of DBAR and bounding its regret. The authors also compare DBAR's performance with existing algorithms, illustrating through simulations that DBAR provides improved stability and reduced regret in both linear and nonlinear systems under adversarial disturbances.
优点
Originality: The paper introduces a new algorithm, DBAR, combining a dynamic batch length and an adaptive learning rate, which expands on existing Exp3 approaches for bandit learning. This design theoretically broadens the applicability of online control by relaxing the requirement for exponentially stabilizing controllers, a novel aspect in nonlinear control with adversarial disturbances.
Quality: The paper presents a well-structured algorithm and offers some theoretical analysis, including stability and regret bounds, which demonstrate rigor in method development. The authors provide proofs for their main results and discuss conditions under which stability is achieved, showcasing technical precision.
Clarity: The problem setup, assumptions, and main results are generally clear, with key terms defined (e.g., asymptotic stability and finite-gain stability) and a well-organized structure. The authors also make efforts to contextualize the algorithm in comparison to prior work, which aids readability for readers with a background in online control or bandit learning.
缺点
Originality: While DBAR's approach to adapting the learning rate and batch size is new, the actual contribution may be seen as incremental. The improvements over existing algorithms are mainly in modifying specific assumptions rather than introducing a breakthrough method, and it is unclear if these modifications substantively advance the field.
Quality: The theoretical guarantees, though valuable, rest on assumptions that may be difficult to satisfy in real-world applications, such as requiring knowledge of an ISS controller in advance. Furthermore, empirical validation is limited to relatively simple experiments, which may not fully demonstrate the algorithm’s effectiveness or scalability in more complex settings.
Significance: The practical impact of DBAR appears limited, as the algorithm’s assumptions—particularly on controller stability—may not align well with typical scenarios in nonlinear control. The significance for broader ML audiences is thus moderate, as DBAR seems primarily useful in niche settings where exact stability knowledge is available.
Significance: The paper's experimental evaluation does not robustly validate the theoretical claims, as it relies on lower-dimensional systems that may not capture the challenges posed by high-dimensional, complex environments. This limits the confidence in DBAR's utility beyond the examples shown, reducing its impact.
问题
-
How does the dynamically increasing batch length affect the algorithm's computational cost, particularly in real-time applications? It would be helpful to discuss any trade-offs between stability assurance and computational efficiency.
-
The current experimental setup involves relatively simple, low-dimensional systems. Can the authors provide results on more challenging benchmarks to better demonstrate DBAR’s performance and stability claims?
Weakness 2 (Quality): There may have been a misunderstanding. Indeed, our assumption does not require prior knowledge of an ISS controller. We do not need to know which controller is ISS or not; we only require that at least one ISS controller "exists" in the candidate pool. The ISS principle is a widely understood concept for handling disturbances in nonlinear systems, particularly in applications such as automotive and robotics systems, as it ensures that a vehicle or robot can perform safely despite variations in exogenous inputs like road conditions or obstacles. Moreover, as mentioned above, we only assume the existence of "asymptotically" ISS stabilizing controllers whose stabilizing behavior can be arbitrarily slow. Therefore, the existence of an ISS controller is not a particularly restrictive assumption. To support the scalability, we have presented new experiments in Example .
Weakness 3 (Significance): Controller stability and ISS stability is the same concept in our context. We are not sure about the reviewer's comment. Perhaps, some terminologies in the paper have caused a confusion. We believe that the typical scenario in nonlinear control is to design a controller that takes the states to an equilibrium, and this is about the design of stabilizing controllers. We have addressed the same problem in our paper. We do not assume stability knowledge. Given an unknown system, we can consider different regimes to model the possibilities of the unknown parameters of the system and then design a controller for each possibility. For example (see a new Figure for the pictorial illustration for readers), consider a robot operating in an uncertain environment where there are different scenarios that could happen in the environment and we model it by a vector "" for simplicity. We consider different intervals such as or , etc. for "" and then design a controller for each scenario. Now, we have a pool of controllers and we know at least one of them should work but do not know which one since we are not aware of the exact parameter of the system. Our method learns a correct controller from the pool. Our stability notion is the same as the one routinely considered in the control theory area as well as the new area of machine learning for control systems.
Weakness 4 (Significance): Thank you for your comment. We have a mathematical proof showing that the proposed idea works. We have provided new high-dimensional experiments in Example .
Question 1: The magnitude of learning rate and batch length does not affect the algorithm's complexity. For the learning rate, it is simply used to calculate Line 28 of Algorithm regardless of its magnitude. For the batch length, the only (potential) intensive part is Line 21 of Algorithm . Given the batch and its length , it simply adds all of the bandit feedback over , so the complexity of the line is . Since the sum of all batch lengths is the algorithm time , the total complexity would be and this is certainly a linear-time algorithm.
Question 2: We have now provided 100-dimensional systems in Example .
Many thanks for reading our rebuttal.
Dear Reviewer BMap,
Since the discussion period will end in a couple of days, we would greatly appreciate it if you could check our responses and let us know whether there are further concerns or issues that we can address. We hope that the reviewer will kindly reconsider their score if our responses are satisfactory. Many thanks.
Best regards,
Submission1215Authors
We are very grateful to the reviewer for a thorough evaluation of our work and providing valuable feedback. We will improve the paper based on the provided comments.
We have improved the paper and uploaded a new PDF file with two important revisions:
-
Experiments: We have replaced the experiments in Example with high-dimensional experiments (100 states) on leader-follower systems. The leader is represented by a previous ball-beam (nonlinear) system, and the followers leverage the leader's state to stabilize themselves. Please check the experimental setup in Example (see pages 9-10).
-
Illustration: Based on your constructive comment, in a new Figure , we have illustrated a concept of the controller pool for the unknown nonlinear system, and how general an asymptotically stabilizing notion is. We hope that this figure helps ML audiences understand the core concepts of our setup in nonlinear control. (see page 4, Figure ).
Responses to the reviewer:
Weakness 1 (Originality): We appreciate the reviewer's comment. As the reviewer stated in the strengths, our main contribution is on relaxing the requirement for exponentially stability controller. We carefully designed the polynomially increasing batch length and adaptive learning rate based on the state norm to achieve desirable properties even when only asymptotically stable controllers exist. We would like to emphasize that we did not use an existing method as is and our results all depend on our method of adjusting the batch length and learning rate in the learning process.
The existing methods only deal with exponentially stable controllers while we focus on a much broader class of asymptotically stable controllers. To illustrate how significant the difference between the two types of controllers is, let us further present a one-dimensional system, where the current system state is (newly presented in Figure ). The goal is to achieve a state near , and we would like to detect this stability by observing whether one arrives at a state less than , where is an arbitrarily small positive number. Exponentially stabilizing controllers guarantee to detect the stability in time. However, with an asymptotically stabilizing controller, if the controller is designed to keep the system state unchanged for an arbitrarily long time and then collapse the state towards afterward, one cannot detect the stability before time regardless of how small is. In such a case, even though the controller ultimately achieves the goal, it may take a lot of time to learn whether a closed-loop system would be stable or not. Our algorithm works even if only those troublesome controllers exist and simple (exponentially stabilizing) controllers do not exist at all. For more details, please see Appendix A. Furthermore, to provide a second perspective on this issue, as mentioned in the paper (Lines 69-79), the difference between exponential stability and asymptotic stability is tantamount to the difference between convexity and strong convexity. As witnessed in many machine learning problems, strong convexity is a very restrictive assumption while convexity is more realistic.
Meanwhile, our numerical experiment also shows that the relaxed controller stability assumption is crucial. Notice that Li et al.'s work assumes at least one exponentially stabilizing controller in the pool (see Table 1). In Figure , Li et al.'s work ("fixed , fixed ") stabilizes the linear system (the state norm reaches near zero), putting aside the time needed for the stabilization. This is because an exponentially stabilizing controller always exists for the linear system. However, in the nonlinear system case where an exponentially stabilizing controller may not exist (see Figure ), Li et al.'s algorithm incurs the system to explode, while our DBAR ("dynamic , adaptive ") successfully stabilizes the system and achieves a good regret. By comparing Li et al's algorithm in Figures and , we observe that allowing a broader class of controllers can greatly affect the algorithm's performance, and our DBAR works well even when exponentially stabilizing controllers do not exist.
Continued in the next rebuttal.
Dear reviewers: Since the discussion period would end in a week, could you please read our rebuttal and let us know if you have any further concerns or comments? We provided extensive answers to your previous comments and hope to discuss them with you during this discussion period. We appreciate your time and effort.
Best regards,
Submission1215Authors
Summary: This paper investigates online nonlinear control using bandit feedback. Its primary contribution lies in introducing dynamic batch length and learning rate techniques to relax the commonly used exponentially stabilizing controller assumption. Instead, it adopts a weaker assumption termed the asymptotic stabilizing assumption, achieving a better regret bound. Experimental results support the authors' claims.
Strengths:
- The use of dynamic batch size and adaptive step size to mitigate stability issues is a compelling approach.
- The motivation for this work is clear and well-articulated.
- The experiments, while conducted on simulations, seem sound and adequately support the proposed methodology.
Weaknesses:
- The paper’s presentation needs significant improvement and currently falls short of publication standards. Both my own reading and other reviewers' comments highlight the excessive reliance on dense notations, making the paper less accessible to readers unfamiliar with the field.
- The significance of the asymptotic stability assumption is insufficiently emphasized, leaving the contribution somewhat unclear in the context of existing literature. The draft lacks clarity in positioning this work within the broader research landscape.
Decision: I recommend rejecting this paper in its current form due to the weaknesses outlined above.
审稿人讨论附加意见
During the discussion phase, the authors expanded on the motivation behind this work and added additional simulation experiments. However, I believe that the current writing quality still falls short of the standard expected for an ICLR publication.
Reject