Solving robust MDPs as a sequence of static RL problems
摘要
评审与讨论
This work proposes a new robust RL methods for MDP with uncertain transitions called IWOCS (Incremental Worst-Case Search). This method incrementally builds a discrete, non-sa-rectangular uncertainty set and a sequence of candidate robust policies. IWOCS can be utilized for classic RL tasks with discrete state/action spaces and Deep RL domains with continuous task settings. The experiments on windy gridworld and Mujoco tasks demonstrate its effectiveness.
优点
- Improving robustness of RL against transition uncertainty is meaningful for downstream tasks, such as RL applications on robotics.
- The extensive experiments on classic and deep RL settings demonstrate the effectiveness of the new method, especially in tasks with continuous state space and action spaces.
缺点
- In appendix. I, the author shows the training times for SAC and IWOCS in Mujoco control tasks. As shown in Table 6, IWOCS requires about twice time cost for training compared to SAC, leading to additional burdens in real applications. According to this phenomenon, please provide the following results to analyze and improve the new method:
- Provide a more detailed breakdown of where the additional computation time is spent in IWOCS (e.g. worst-case search, policy optimization, etc.). This would help identify specific areas for potential optimization.
- Provide potential methods or approaches to reduce the computational cost of IWOCS.
- Include training times for other baseline robust RL methods to provide more context for evaluating IWOCS's computational efficiency relative to the state-of-the-art.
- This work focuses on the RL task with uncertain transition, such as robotics control with dynamic friction settings. This issue is also researched in Meta-RL domains, such as Meta RL methods with task inference. Such methods can adapt to new environment transitions with limited interactions. Please explain the relationship or difference of this work and such meta RL methods. A brief introduction of Meta-RL with uncertain transition should be included in the related work, with highlighting key similarities and differences between IWOCS and meta-RL approaches.
- Some typos in the manuscript, such as line 085: denote -> denotes
问题
- In this work, IWOCS only considers a unique starting (Line 067). However, in some tasks such as continuous control tasks, we need to tackle the distribution . For example, as shown in Algorithm 2, we need to compute and during training. How to compute these values for uncertain ? Discuss how IWOCS might be extended to handle a distribution of starting states, such as in IWOCS+SAC algorithm for Mujoco control tasks.
- As shown in table 2&3, IWOCS* obtains higher performance than IWOCS, which is normal because IWOCS* utilize grid-search rather than CMA-ES. Please provide a detailed comparison of computation times for IWOCS and IWOCS* across different uncertainty set dimensions. Discuss the trade-offs between performance and computation time when using grid-search vs CMA-ES, and provide guidelines for when to use IWOCS vs IWOCS* based on problem characteristics.
- Can the idea proposed in this work be utilized in other robust RL problems, such as robust RL under observation perturbations? Please discuss any modifications or potential challenges when applying IWOCS to observation perturbation problems.
This work addresses the robust static MDP problem. It is worth noting that when optimizing over stationary policies, the optimal robust static MDP model is equivalent to the dynamic variant. The authors propose an iterative approach that incrementally identifies worst-case transition models. Specifically, given a finite set of uncertain transitions, the method iterates over the uncertainty set to find the optimal policy at each iteration until the condition is satisfied. The authors further demonstrate a deep RL variant of their algorithm across a variety of standard deep RL benchmarks.
优点
(1) The paper is well-written, showcasing the authors' deep understanding of prior work and foundational concepts in the robust RL community.
(2) The proposed methodology is clearly and comprehensively explained.
(3) The authors have thoroughly evaluated the performance of their algorithm across a wide range of deep RL control benchmarks, effectively validating its robustness and applicability.
缺点
The paper has two main weaknesses:
(1) The primary weakness is the lack of theoretical contribution, as most theoretical results presented are well-known. The application of established theories such as the lack of duality gap and the equivalence between static and dynamic stationary models, to the IWOCS setting seems trivial and adds limited novelty. A more in-depth discussion of how IWOCS provides advantages (i.e. scalability analysis, sample complexity) over existing robust MDP algorithms would strengthen the paper.
(2) The motivation of the paper appears somewhat contradictory. On one hand, it is stated that the static model is more challenging to solve than the dynamic model and is valuable for decision-makers seeking robustness to static transition model uncertainty. On the other hand, the authors note that, under mild assumptions, the static and dynamic models are equivalent. This positioning makes it seem as though the approach achieves similar outcomes to robust value iteration (RVI) but with added complexity. The paper could benefit from more detailed analysis of IWOCS against other robust algorithms beyond short discussion on "rationale to why IWOCS might be a time-efficient algorithm in large scale robust RL problems".
Minor Issue: Algorithm 1 missing close bracket in the if statement:
问题
-
In line 199, the authors state, “We consider a (small) discrete set of MDPs,” yet it remains unclear why IWOCS is necessary in this context if can be computed directly when the model uncertainty set of MDPs is small. Specifically, in Algorithm 1, it would be helpful to clarify the complexity of the "Find worst " step compared to a direct computation of .
-
In line 252, the authors state that " can be chosen with another criterion than the worst-case without loosing property 1 and 2". However, for this to hold under alternative criteria, it may be necessary to assume equality when swapping the maximization over policies into other criteria. This assumption may not hold for criteria beyond average and extreme values, potentially limiting the generalizability of this claim.
-
In numerical experiment, the proposed method demonstrate a better performance than other robust RL algorithms. However, this algorithm aim to converge to the the optimizer for RVI, it is unclear to my why it would converge to a better rather than identical optimizer. It is because we worked in deep learning setting and the robust algorithm has no guarantee to converge to optimal? In tabular domain, would the proposed method and RVI have the same optimizer?
This paper proposes solving a robust MDP using a worst-case model search. In particular, it focuses on a static model, where the transition function does not change during trajectory. This setting ensures that the policy computed is less conservative than a policy that solves for a dynamic model. However, solving problems with a static model is harder than problems with dynamic models.
To mitigate this issue, the paper assumes the agent optimizes a stationary policy. Under such assumption, the paper shows that it is sufficient to find optimize for the worst-case transition function: .
In this case, given the uncertainty set, the proposed approach interleaves the optimization of the policy and the update of the transition dynamics, effectively searching for the worst-case transition function.
The paper provides empirical evidence that the method can converge to robust policies in a tabular setting as well as in a function approximation setting, adapting the SAC method to handle such a sequence of problems.
优点
- The quality of the paper is high. The arguments are presented in a concise and effective way. The description of the algorithm for the tabular case is precise.
- The writing is clear, and the mathematical notation is precise.
- Strong empirical results against the baseline algorithms under consideration.
缺点
-
The main weakness of the paper is its originality. This is limited since prior work proposes a similar approach to estimate the worst-case MDP to compute a robust policy based on a single transition function (Gadot et al. 2024).
-
The significance of the proposed approach also seems limited when we put it in perspective with the work from Gadot et al. (2024), since this method requires a specialized implementation of the SAC algorithm, while the method from Gadot et al. (2024) does not requires any changes on the underlying RL algorithm. Furthermore, their method already shows the scalability of deep RL algorithms to robust settings, which puts the empirical contribution of this paper in check. While the paper provides a considerable overview of the related work literature, it misses a key reference, as discussed above. The paper needs a clear discussion of such an approach and, if appropriate, an empirical comparison with it. Lastly, the presentation of the worst-case identification on the deep IWOCS is difficult to parse. Furthermore, the provided supplemental material does not provide any instructions about how to use the code. A pseudo-code detailing this procedure and instructions to reproduce the experiment would significantly increase the clarity and reproducibility of the paper.
-
Gadot, U., Wang, K., Kumar, N., Levy, K. Y., and Mannor, S. (2024). Bring your own (Non-robust) algorithm to solve robust MDPs by estimating the worst kernel. ICML, 14408–14432. https://proceedings.mlr.press/v235/gadot24a.html
问题
Could you please elaborate on the differences between the proposed method and the approach from Gadot et al. (2024)?
The paper targets reinforcement learning for MDPs whose transition probabilities are known up to an uncertainty set. Hence, it aims to robustify against these uncertainties and justify the proposed approach empirically.
优点
- The general problem targeted by the paper is relevant and timely.
- The amount of empirical results is appropriate.
缺点
- It is hard to talk about soundness of the results because there is no precise problem statement.
- As such, it is impossible to prove a meaningful result, which is critical when dealing with uncertainties and their impact on performance and safety.
- The literature within which the paper places itself misses a large body of more relevant work on RL and IRL (which solves a series of RL problems as a subroutine) for robust/uncertain MDPs. This body of work originates from probabilistic formal methods and investigates how existing approaches for planning in robust/uncertain MDPs can be utilized in RL. Examples include the following. https://www.prismmodelchecker.org/papers/neurips22.pdf, https://arxiv.org/html/2312.06344v1, https://arxiv.org/html/2312.06344v1, https://ojs.aaai.org/index.php/ICAPS/article/view/19785
- The method is, at best, ad hoc. The implementation leaves many choices to the user.
问题
- The contrast with dynamic uncertainties is an odd choice. That is a generalization of the model considered in the paper. Why would one claim that the setting in the paper is more challenging? Especially without even giving a precise description of what is more challenging.
The paper addresses the robust MDP problem in the context of uncertain transition dynamics, assumed to lie within a rectangular uncertainty set. Building on the non-duality-gap property established in previous work, the authors propose a discretization approach to tackle the adversarial problem. Specifically, they sample transition functions from the uncertainty set and iteratively solve the non-robust problem using the sampled transition dynamics. A deep version is also introduced to handle continuous state-action spaces. Numerical results are provided on various MuJoCo benchmarking tasks, comparing the proposed methods with other robust MDP approaches.
优点
- Robust MDP is an important area of study.
- The proposed algorithm appears simple to implement.
- The Deep IWOC represents a meaningful improvement over existing robust MDP methods, which mainly focus on discrete state-action spaces.
缺点
-
The primary sampling-based approach raises significant concerns. Typically, solving a max (or min) problem through sampling requires an exponential number of points to achieve the desired accuracy, making this approach generally intractable from a theoretical standpoint.
-
There is no discussion on the number of transition samples needed to obtain a robust solution.
-
The sampling-based approach appears overly simplistic and could be seen as a straightforward baseline for robust MDP. Combined with the lack of theoretical analysis (e.g., theoretical analyses on the impact of the number of samples on the convergence), the contributions feel incremental.
-
The proposed algorithm does not seem scalable. The authors should consider comparisons with some model-free algorithms and test on larger benchmarking tasks.
问题
- How does the sampling-based approach converge as the number of samples increases? Can the authors provide more analyses (theoretical and experimental) to support the sampling-based approach?
- Would this approach work with non-rectangular uncertainty sets?
- How does your method compared to model-free algorithms (e.g, Q-learning, SAC, PPO, TRPO ...)?
- How does your method perform with other environments other than Mujoco (Atari, DM control)?
- Could a simple dual-gradient method solve the minimax problem?
伦理问题详情
I do not see any critical ethical concerns
Dear authors,
If possible, please provide a rebuttal to kick-start the discussion about the paper.
Thanks, AC
The paper develops Incremental Worst Case Search (IWOCS), a method designed to solve the robust RL problem (i.e. to find a single policy that works in any of a set of environments).
The main strengths of the paper are: (1) the problem is important; (2) the proposed algorithm is simple; (3) experiments are extensive.
I don't think that the paper is ready for the following reasons.
- Weak theory / no sample complexity guarantee.
- Presentation not very clear.
- Novelty issues relative to the Gadot paper.
For these reasons, and because the authors failed engage in a discussion with reviewers, I recommend rejection.
审稿人讨论附加意见
All reviewers agree the paper has significant weaknesses (and the AC agrees).
Reject