Sleeping Reinforcement Learning
摘要
评审与讨论
This paper considers a tabular episodic reinforcement learning setting where the set of available actions is not fixed but varying over episodes, states and time steps. The paper studies two different ways the available actions are revealed to the learner: per-episode (available actions are revealed at the beginning of each episode) and per-stage (available actions are revealed only at each time step within an episode). In the per-episode regime, the paper proposes an algorithm that works for both adversarial and stochastic cases, and proves that its sleeping regret is not larger (in big-O notation) than standard RL. In the per-stage regime, the paper proves both lower and upper bounds in this regime for two different types of action distributions - one where the action availability is independent of past states and actions and one where the action availability is dependent on the previous state and action.
给作者的问题
-
Could the authors outline any significant technical challenges for the stochastic per-stage regime where the type of the distribution of the action availability is unknown? In particular, why doesn't your current estimation method for works for ? The distribution does not depend on the episode, so couldn't we simply estimate the conditional probability that an action might be available given the current state, previous state, previous action and previous availability?
-
Going back to the exponential dependency on : the lower bound construction in Theorem 4.2 requires the availability to change based on previous availability . Do you think the exponential dependency on A can be removed if the availability depends only on and ?
论据与证据
This is a theory paper. All theorems are clearly stated and accompanied by their proofs in the appendix.
方法与评估标准
The approach is sound and the results are significant. There are two different notions of regret, one for the per-episode regime (Definition 3.1) and one for the per-stage regime (Definition 4.2). These notions of regret make sense.
理论论述
For the upper bounds, I only skimmed their proofs in the appendix. Since the algorithms are based on the standard approach of optimism, all the upper bound proofs seem fine.
For the lower bounds, I did carefully check the proofs. In particular, I checked the proof of Theorem 4.2, which is the most important lower bound in this paper. I found the proof correct.
实验设计与分析
The paper has no experiments.
补充材料
I checked the lower bound proofs in the paper, in particular Appendix E.2 (Proof of Theorem 4.2). The proof is correct.
与现有文献的关系
The paper extends the existing literature on sleeping bandits to reinforcement learning. Sleeping reinforcement learning is much harder than sleeping bandits, so the contributions of the paper are significant. The paper also make novel technical contributions in deriving optimistic algorithms for sleeping reinforcement learning. I find it interesting that such an optimism-based approach works (in the per-episode regime) even for adversarial action availability, which was not the case for sleeping adversarial bandits (e.g. Nguyen and Mehta, AISTATS 2024).
In the independent and stochastic per-stage regime, the paper proposes an approach that estimates the distribution of action availability. This approach is similar to existing approaches in sleeping bandits with stochastic availabilities (e.g. Saha et al, ICML 2020).
The most surprising result to me is that on both independent and Markovian stochastic per-stage regimes, the dependency on the number of actions is exponential. This was not the case in sleeping bandits.
遗漏的重要参考文献
I am not aware of any missing important references.
其他优缺点
No other comments.
其他意见或建议
No other comments.
We thank the Reviewer for the time spent reviewing our work, and for having appreciated the significativeness of our work. We also thank the Reviewer for the comments on the results which we will use to expand the discussion on the results exploiting the additional page. Below, our answer to the Reviewer's questions.
Could the authors outline any significant technical challenges for the stochastic per-stage regime where the type of the distribution of the action availability is unknown? In particular, why doesn't your current estimation method for works for ? The distribution does not depend on the episode, so couldn't we simply estimate the conditional probability that an action might be available given the current state, previous state, previous action and previous availability?
We can adapt the design of the estimator for to handle by incorporating conditional probabilities, as the Reviewer correctly suggests. However, we conjecture that the resulting regret bound would remain of the same order (with exponential dependence on ) as in our current approach (Theorem 4.1), which relies on the augmented MDP approach. That said, we believe our method allows us to avoid several non-strictly necessary calculations. We will add a comment on that in the paper.
Going back to the exponential dependency on : the lower bound construction in Theorem 4.2 requires the availability to change based on previous availability . Do you think the exponential dependency on A can be removed if the availability depends only on and ?
Yes, this dependency can be removed in the described scenario and the regret bound we would obtain should be very similar to the independent availability scenario, as the additional complexity in the Markovian case is generated, as the Reviewer properly noticed by looking at the lower bound construction, by the challenge in estimating the transition probabilities over the action sets. We will add a comment on that using the additional page. Thank you for pointing it out.
This paper studies a new paradigm called Sleeping Reinforcement Learning, where the available action set varies during the interaction with the environment. The authors study several settings, including the per-episode disclosure, in which the available action sets are revealed at the beginning of each episode, and the per-stage disclosure, in which the available actions are disclosed only at each decision stage. The authors provide algorithms, upper and lower bounds for the problem.
给作者的问题
It is possible to also provide a lower bound for the setting of independent per-stage disclosure? I believe this may help better understand the problem.
论据与证据
Yes.
方法与评估标准
Yes.
理论论述
No.
实验设计与分析
No.
补充材料
No.
与现有文献的关系
The work is related to the works on sleeping bandits, where the action sets vary in time. To the best of my knowledge, this is the first theoretical work on sleeping RL.
遗漏的重要参考文献
No.
其他优缺点
Strengths:
- This is the first work to theoretically study the important sleeping RL setting. It is well motivated, the authors provide many examples to show the importance of this setting.
- The paper is well-written and easy-to-follow.
- The authors conduct a complete study in this topic, providing algorithms with regret upper bounds, and proving novel lower bounds.
Weaknesses:
- It may be better to also provide a lower bound for the setting of independent per-stage disclosure.
- In Lines 130-132, , , are used before definition.
其他意见或建议
There is a typo in Line 195.
We thank the Reviewer for the time spent reviewing our paper and for having appreciated the motivation, clarity and novelty of our work. Below, our answer to the Reviewer's comments.
It may be better to also provide a lower bound for the setting of independent per-stage disclosure.
It is possible to also provide a lower bound for the setting of independent per-stage disclosure? I believe this may help better understand the problem.
We thank the Reviewer for giving us the opportunity to elaborate on this. After careful consideration, we argue that a tight lower bound for the independent per-stage disclosure case is the same as that of standard MDPs, i.e., (see Domingues et al., 2021). Clearly, the lower bound for standard MDPs is a lower bound for the independent per-stage disclosure case, since the latter includes a smaller class of problems in which all actions are always available, i.e., for every . Concerning the tightness, we need to carefully compare it with the upper bound of Theorem 5.1. First of all, the lower bound (Domingues et al., 2021) is for the expected regret (as customary in the literature), while our upper bound for the per-stage disclosure case of Theorem 5.1 is for the regret in high probability. Thus, we need to convert the latter to the expected regret in order to perform the comparison. To do so, we start from the first regret bound of Theorem 5.1 (disregarding constants and logarithmic terms, but not terms depending on ):
Notice that by disregarding the dependence on and for sufficiently large , we obtain the second regret bound of Theorem 5.1 (still in high probability). To get the upper bound for the expected regret, we make the choice of , whenever , obtaining (since always):
The latter is of order whenever . We will adjust the presentation of the result in the final version of the paper, clarifying the difference between expected regret and regret with high probability bounds, and, consequently, the discussion on the result and the future works. Thank you for rising this point.
In Lines 130-132, , , are used before definition.
We agree with the Reviewer that (at least) an informal definition is needed also in this part. It was an oversight, we fixed it. Thank you.
Typo in Line 195.
Thanks, we fixed it.
The paper introduces Sleeping Reinforcement Learning (SleRL), a new reinforcement learning paradigm where the set of available actions varies over time due to external constraints or stochastic processes. Two settings are considered: per-episode disclosure where available actions for all states are revealed at the start of each episode and per-stage disclosure where available actions are revealed only at the current step. They show an upper bound and a lower bound for the regret of a modification of UCBVI in each setting.
给作者的问题
- Is it possible to add some numerical results to verify your results?
- Is it possible to reduce the bound gaps, i.e. the exponential term and , between Theorem 4.1 and 4.2?
- Is it possible to reduce the lower bound condition for with a factor in Theorem 5.1?
update after rebuttal: I raise the score from 2 to 3. Please refer to the comment below for reason.
论据与证据
The regret bounds are supported by proofs.
方法与评估标准
The algorithm is a modification of a classical algorithm called UCBVI, and it makes sense for this problem.
理论论述
I have roughly checked the proof of Theorem 3.1 and it looks correct.
实验设计与分析
No experimental designs.
补充材料
No.
与现有文献的关系
This work is related to Multi-Armed Bandits under the name of “Sleeping” MABs and Reinforcement Learning with constrained action spaces.
遗漏的重要参考文献
This paper has discussed essential references.
其他优缺点
Strengths:
- This paper is a pure theoretical paper and offers detailed analysis about the upper bound and lower bound for the algorithm.
- This paper uses a novel construction (Figure 3) to show that an exponential dependence on the number of actions is unavoidable in the regret and proves a lower bound (Theorem 4.2).
Weaknesses:
- This paper lacks numerical analysis to test their theoretical results.
- The gap between upper bound and lower bound may be too loose for Markovian Per-stage Disclosure.
其他意见或建议
- In the definition of three regrets, is there a typo? I guess and should be the same thing, the number of episodes.
- It would be better to show some numerical results to verify the bounds.
We thank the Reviewer for the time spent reviewing our work. Below, our answer to the Reviewer's questions and concerns.
Is it possible to add some numerical results to verify your results?
To numerically validate the algorithm and show the impact of action availability on performance, we consider a modification of the well-known Frozen Lake environment (https://gymnasium.farama.org/environments/toy_text/frozen_lake/). In the original version, the agent must traverse a frozen lake (i.e., a grid) from a start to a goal position, avoiding holes in the surface (i.e., unavailable positions in the grid). We modify it by letting such holes open and close stochastically (i.e., the unavailability of each position is sampled at each time step from a Bernoulli with parameter ). We assume the start and goal to be fixed in the top-left and bottom-right corners of the grid, respectively.
The agent can move up, down, left, right, or stay. If the agent selects an action that would move it to an unavailable state, it stays in the current position. The reward function is defined as follows: the agent receives reward for state-action pairs that have the goal as next state, and otherwise. The episode stops at the end of the time horizon (not when we reach the goal state). Given that all positions (except the start, the goal, and the position occupied by the agent) have the same probability of being frozen at any time step, it is easy to see that the optimal policy tries to move along the diagonal that connects the start to the goal. To compute the value of the optimal policy, we perform a Montecarlo simulation for each evaluated setting. We compare our S-UCBVI (Algorithm 7) with standard UCBVI (Azar et al., 2017), observing that the latter interacts with the environment as shown in Figure 1. We compare them in terms of reward, to highlight that both algorithms converge, yet UCBVI converges to a suboptimal objective.
We evaluate grids of size , with , varying the stochasticity parameter as . We consider a time horizon of for each episode and episodes.
The Reviewer can find the plots of the experiments here:
https://drive.google.com/file/d/19WlZpKUHoSoYxx4PgT3jNLqyfmrm9twu/view?usp=sharing
and the code here:
https://drive.google.com/file/d/1NBoDjr9P_eaUcO1Na4nyzE71q4zgAKwR/view?usp=sharing
As expected, when , we observe no performance difference. Instead, as the environment stochasticity increases, we observe a greater gap in the performances of the two algorithms, with S-UCBVI (our algorithm) achieving the optimal value (obtained via Montecarlo simulations as described above), represented as an horizontal line, and UCBVI not reaching such optimum. This is due to the fact that UCBVI cannot observe action availabilities, effectively playing in an environment with a lower optimum.
Is it possible to reduce the bound gaps [...] between Theorem 4.1 and 4.2?
The lower bound of Theorem 4.2 that shows an exponential dependence in the cardinality of the action set represents a statistical barrier in the learning for Sleeping MDPs with Markovian per-stage availability. For this reason, as common in the literature (e.g., MDPs with adversarial transitions [1]), the goal is no longer matching such an exponential lower bound, but rather finding structures (e.g., independent action availability) that overcome such barriers. The result of Theorem 4.1, indeed, has to be interpreted as just an exemplification of the fact that, when accepting such an exponential dependence, the Sleeping MDPs with Markovian per-stage availability can be addressed through the augmented MDP method. We will clarify this in the paper.
[1] Tian, Y., Wang, Y., Yu, T. and Sra, S. Online Learning in Unknown Markov Games. ICML 2021
In the definition of three regrets, is there a typo? I guess and should be the same thing, the number of episodes.
We checked and the notation is correct and compliant with (Azar et al., 2017) where is the total number of interactions (see Section 2), is the number of episodes and is the horizon of the single episode.
Is it possible to reduce the lower bound condition for with a factor in Theorem 5.1?
Just like in (Azar et al., 2017), for the UCBVI algorithm and its analysis, it is challenging to remove such a dependence on the horizon from the minimum condition. This is exacerbated, in our case compared to (Azar et al., 2017), since we consider stage-dependent transitions (doubling the exponent of ). Nevertheless, there exist works [2] that succeed in mitigating such a condition at the price of more complex (and less effective in practice) algorithms. Importing such techniques to the Sleeping MDP setting can be an interesting future work.
[2] Zhang, Z., Chen, Y., Lee, J. D., and Du, S. S. Settling the sample complexity of online reinforcement learning. COLT 2024.
I appreciate your responses for my questions. The numerical experiment shows the advantage of your algorithm over UCBVI, and your responses answer my questions about the regret bound. I have adjusted my score accordingly.
This paper introduces Sleeping Reinforcement Learning, in which available actions vary over time. It provides a theoretical analysis for this problem, along with algorithms with regret guarantees and matching lower bounds. While minor concerns were raised regarding exponential dependencies and missing lower bounds in specific cases, the rebuttal and added experiments addressed these concerns. Reviewers have converged on the significance of the contribution.