PaperHub
5.5
/10
Rejected6 位审稿人
最低3最高8标准差1.5
6
3
5
8
5
6
2.8
置信度
正确性2.8
贡献度2.8
表达2.5
ICLR 2025

Minimax Optimal Regret Bound for Reinforcement Learning with Trajectory Feedback

OpenReviewPDF
提交: 2024-09-28更新: 2025-02-05
TL;DR

We prove a nearly optimal regret bound for reinforcement learning with trajectory feedback.

摘要

We study the reinforcement learning (RL) problem with trajectory feedback. The trajectory feedback based reinforcement learning problem, where the learner can only observe the accumulative noised reward along the trajectory, is particularly suitable for the practical scenarios where the agent suffers extensively from querying the reward in each single step. For a finite-horizon Markov Decision Process (MDP) with $S$ states, $A$ actions and a horizon length of $H$, we develop an algorithm that enjoys an optimal regret of $\tilde{O}\left(\sqrt{SAH^3K}\right)$ in $K$ episodes for sufficiently large $K$. To achieve this, our technical contributions are two-fold: (1) we incorporate reinforcement learning with linear bandits problem to construct a tighter confidence region for the reward function; (2) we construct a reference transition model to better guide the exploration process.
关键词
Reinforcement learning theoryregret analysistrajectory feedback

评审与讨论

审稿意见
6

This paper studies reinforcement learning (RL) with trajectory feedback. The learner can only observe the accumulative noisy reward once a trajectory is terminated. The state transition model is unknown to the learner. The paper develops an online learning algorithm and shows that its regret bound is optimal.

优点

  1. The paper is well-written and easy to follow.
  2. RL with trajectory feedback is an interesting problem.
  3. This is a solid theory paper. The analysis is rigorous and sound.

缺点

  1. There is no experiment.
  2. In the upper bound of Theorem 7, the last three terms dominate. In contrast, the abstract and the introduction claim that the upper bound is determined by the first term. The authors should clarify under what conditions, if any, the first term dominates. If the first term is not asymptotically dominant, the authors should explain why they only focus on the first term in the abstract and the introduction.
  3. The introduction claims that the developed algorithm for RL with trajectory feedback achieves the same asymptotically optimal regret bound as the standard RL. The authors should explain why trajectory feedback does not lead to a worse regret bound and what properties of their algorithm allow them to overcome the information disadvantage of only receiving trajectory feedback.
  4. Section 3 should clarify that how the expected trajectory reward is a linear function of the state-action visitation frequencies.
  5. Some mathematical derivations are not intuitive. The authors can add explanations about what the mathematical properties mean and how they are derived. Here are some examples. 5.1 The second key observation on page 5. 5.2 Inequality (3). 5.3 The equation in (8).
  6. There are a few typos: The third term of the upper bound in Theorem 7. P1 in Line 4 of Algorithm 2. D2 in Line 6 of Algorithm 2.

问题

  1. In the upper bound of Theorem 7, the last three terms dominate. In contrast, the abstract and the introduction claim that the upper bound is determined by the first term. The authors should clarify under what conditions, if any, the first term dominates. If the first term is not asymptotically dominant, the authors should explain why they only focus on the first term in the abstract and the introduction.
  2. The introduction claims that the developed algorithm for RL with trajectory feedback achieves the same asymptotically optimal regret bound as the standard RL. The authors should explain why trajectory feedback does not lead to a worse regret bound and what properties of their algorithm allow them to overcome the information disadvantage of only receiving trajectory feedback.
评论

For your questions about (1) the lower order term and (2) the intuition why trajectory-feedback does not hurt the regret performance, please refer to the common response.

About linearity of trajectory reward in Section 3: Thanks for the suggestion. We have provided a proof in the revision, which we also present here. E[rτ]=E[h=1Hrh]=s,a,hE[I[(sh,a,h)=(s,a)]]rh(s,a)=dr\mathbb{E}[r_{\tau}] = \mathbb{E}[\sum_{h=1}^H r_h] = \sum_{s,a,h}\mathbb{E}[ \mathbb{I}[(s_h,a,_h)=(s,a) ] ]r_h(s,a) = d^{\top}r.

About mathematical derivations in Section 5: Thanks for the suggestion. We explain as follows. *

About the second observation in page 5: dPπ(s,a,h):=Prπ,P[(sh,ah)=(s,a)]=τPrπ,P[,τ]I[(sh(τ),ah(τ))=(s,a)]d_{P}^{\pi}(s,a,h):=Pr_{\pi,P}[(s_h,a_h)=(s,a)] = \sum_{\tau}Pr_{\pi,P}[,\tau] \cdot I[ (s_h(\tau),a_h(\tau))=(s,a) ], where (sh(τ),ah(τ))(s_{h}(\tau),a_h(\tau)) is the hh-th state-action pair in τ\tau.

About (3): This equation is based on confidence regions for linear bandits.

About (8): the equation is by definition that Wπ(r,p)=Eπ,p[h=1Hrh]=s,a,hdpπ(s,a,h)rh(s,a)=(dpπ)rW^{\pi}(r,p) =E_{\pi,p}[\sum_{h=1}^H r_h] = \sum_{s,a,h}d_{p}^{\pi}(s,a,h)\cdot r_h(s,a) = (d_{p}^{\pi})^{\top}r.

评论

Thanks for the response. I would like to maintain my rating 6 as the paper does not conduct experiments to show the performance of the developed algorithm.

审稿意见
3

This paper proposes an algorithm that achieves a minimax optimal regret bound for RL with trajectory feedback. Specifically, the proposed algorithm achieves a regret bound of O~(SAH3K)\tilde{O}(\sqrt{SAH^3K}) for an episodic MDP.

优点

  • The paper tackles a well-motivated problem, proposing an asymptotically optimal regret bound algorithm for scenarios with trajectory-level feedback.

  • Up to Section 4, the authors clearly present their motivations, objectives and proof sketch, making it easier for readers to grasp the core concepts of the paper.

缺点

The paper has numerous typos and complex, undefined notations, giving the impression of a hastily composed and unpolished draft. The structure and expressions appear to be inspired by [1], yet there are instances of notations that are either not used or are undefined in the context of this paper. A thorough revision to polish the paper would improve its clarity.

In particular, Section 5 lacks logical flow, as many concepts are listed without a thorough theoretical explanation, making it difficult for readers to follow, especially given the insufficient description of each pseudocode. Considering page limitations, substituting theoretical proofs with a more accessible explanation of the pseudocode in the main paper could enhance readability.

The proposed algorithm also appears to have impractical elements. See Questions.

Typos

  • Line 46: "ofte" should be "often"

  • Algorithm 2: D2\mathcal{D}_2 is not defined

  • Algorithm 3, Line 3: The font of “plan” needs correction

  • Line 388: "we" should be capitalized to "We"

  • Line 483: S11S^1 1 should be corrected to S11S^{11}

  • Algorithm 5: λπ\lambda_\pi is not defined.

[1] Zhang, Zihan, et al. "Near-optimal regret bounds for multi-batch reinforcement learning." Advances in Neural Information Processing Systems 35 (2022): 24586-24596.

问题

  • On line 310, optimization is mentioned for the double-state reward function r(s,s,h,h)r(s,s',h,h'). Could the authors provide more detail on this?

  • In Algorithm 4, Line 7, does πˉ\bar{\pi} refer to what is stated in Eq (6)? Is it possible to obtain πˉ\bar{\pi} in a tractable manner?

评论

About writings: Thanks for the careful review. We have fixed the typos accordingly in the revision. In Algorithm 2, D2\mathcal{D}_2 is redundant so we remove D2\mathcal{D}_2.

About λπ\lambda_{\pi} in Algorithm 5: In Algorithm 5, we use λπ\lambda_{\pi} to denote the probability of π\pi following the probability vector λ\lambda in Line 12.

About structure of Section 5: We have presented the high-level idea in Section 4. We borrow the ideas to construct a reference model from [1], and then conduct exploration with the help of the reference model. We also present the usage of each algorithm and the corresponding lemmas in Section 5.1-2.

About the double-state reward in Line 310: In Line 4 Algorithm 4, the optimization problem is maxπPrπ[τ]minϕΛ1ϕ,1\max_{\pi}\Pr_{\pi}[\tau] \min\\{ \phi^{\top}\Lambda^{-1}\phi , 1\\}. We simplify the target as maxπPrπ[τ]ϕΛ1ϕ\max_{\pi}Pr_{\pi}[\tau] \phi^{\top}\Lambda^{-1}\phi, which is actually (s,a,h,s,a,h)I[(sh,ah)=(s,a),(sh,ah)=(s,a)]Ws,a,h,s,a,h\sum_{(s,a,h,s',a',h')} \mathbb{I}[(s_h,a_h)=(s,a), (s_{h'},a_{h'}) =(s',a') ]W_{s,a,h,s',a',h'} where W=Λ1W = \Lambda^{-1}. We name it as the double-state reward function because the classical reward function could be viewed as a single-state reward function, where the reward only depends on one state-action pair. To optimize a double-reward function, the learner makes decisions according to its trajectory, which means its policy is non-Markovian. To our best of knowledge, there is no existing algorithms which could solve this problem efficiently, even if assuming the knowledge of the transition model.

About π~\tilde{\pi} in Algorithm 4, Line 7: π~\tilde{\pi} could be regarded as an approximate solution to Eq(6). As discussed above, we do not know how to solve the optimization problem in line 4 algorithm 4 efficiently, which is the major difficult to compute π~\tilde{\pi}.

审稿意见
5

The authors present a regret bound for RL with trajectory feedback. The lower bounds improves upon existing results reducing the dependency in the state space and are minimax optimal. To achieve the bounds, the authors use better confidence bounds than previous work and a more refined exploration scheme.

优点

A strong theoretical result improving upon SOTA for an important problem

Refined techniques for a well studied problem

I appreciate the authors' attempt to provide an informal explanation of the proofs. It is not trivial.

缺点

I am curious to have a formal understanding of the term "for sufficiently large K" when presenting THM1. This is important because it can render the result.

The presentation of the algorithm is convoluted. Especially. it is not clear what is Design in Algorithm 5. In general, I think this can be grossly simplified.

The last term in THM 7 is kind of disappointing as it renders the result less exciting.

I believe the examples in the introduction were discussed by others before. A proper reference is needed.

There are some typos in the text that should be fixed: e.g., "ofte" (l46) and "fro" (l250)

问题

Q1: Please elucidate the initial K issue.

Q2: The paper seems incremental wrt Efroni et al (2021). What is the main contribution wrt Efroni et al? Is it a mix and match of a model and know techniques? If not, are the new techniques applicable elsewhere?

Q3: Can you say something about the constant c>0 in l262? Does it depend on the problem itself?

Q4: L282-285: as far as I understand the statement "it holds that \pi^* \in \Pi_{\ell+1}" is a high probability even rather than an almost sure statement. So, if I don't understand what is going on. More specifically, for any finite T or K_\ell all elements in Eq (9) and after are random variables, but they seem to be treated as fixed because everything is taken in the limit. But it cannot be taken as such (at least not without a proof). I am probably missing something here, but this seems like a loose usage of \tilde{O} notations. Please explain.

Q5: How is K_0 set in Algorithm 1?

Q6: It seems to me that the computation needed for Algorithm 5, line 9 would be quite difficult. Is this correct? Can you say something about the complexity of the different algorithms?

Q7: What do you suspect is the real dependence of the low order terms in THM7?

I would be happy to raise my score if the questions above are answered properly, especially Q4.

评论

About burn-in time: According to the final regret bound, the burn-in time would be S24A16H28S^{24}A^{16}H^{28}, which is impratically large. The main issue is from the initial step (line 3 in Algorithm 1) to learn the reference model, and it is a challenging problem to reduce this term (e.g., reduce the order of HH to tight).

About Design in Algorithm 5: The Design step (Line 11-13, Algorithm 5) stems from experimental design, where the target is to find a distribution over a set of datapoints to maximize some certain potential function.

About the examples in the introduction: Similar examples were indeed discussed by prior work on RL with trajectory feedback (e.g., Efroni et al (2021)) and we will add references in the next revision accordingly.

About comparison with (Efroni et al., 2021): In result, the main contribution is a near-tight asymptotic regret bound. In technique, we formulate the RL problem as a linear bandit problem with feature as the trajectory, while (Efroni et al., 2021) used the occupancy distribution as features. The major difference is that, the number of all possible trajectories is (SA)H(SA)^H, while the number of deterministic policies is ASHA^{SH}. This is way we can save an S\sqrt{S} factor compared to the best result in Efroni et al., 2021.

About c>0c>0 in line 262: It does not depend on the problem. In fact, the exact inequality would be: with proability 1δ1-\delta, Λ13TΛπˉ3log(d/δ)I\Lambda\geq \frac{1}{3}T\Lambda_{\bar{\pi}} -3\log(d/\delta)\mathbf{I}.

About line 282-285: We present (9) and (10) to explain the high-level idea. There is possibility that some of the concentration inequalities might fail, and the learner would suffer a linear regret KHKH. But this probability is very small, so we ignore this probability in explaining the intuitions. On the other hand, the exact definition of Π+1\Pi_{\ell+1} is presented in Line 9 Algorithm 5, where we eliminated policies with value smaller than the best LCB.

About K0K_0: We set K0=100000S4.5A1.5H8.5log(SAHKδ)K_0 = 100000S^{4.5}A^{1.5}H^{8.5}\log(\frac{SAHK}{\delta}) (see Appendix.A).

About computation cost in Algorithm 5 line 9: You are correct about this. In general this step is inefficient. But it is easy to test whether an element is in Π\Pi_{\ell} since maxπWπ(u,p^)\max_{\pi}W^{\pi}(u,\hat{p}) could be computed with optimistic backward planning, which means we can optimize some functions with good properties overΠ\Pi_{\ell} (e.g., value function). However, the target function τPrπ,p[τ]minϕτΛ1ϕτ,1\sum_{\tau}Pr_{\pi,p}[\tau]\min\\{ \phi_{\tau}\Lambda^{-1}\phi_{\tau} ,1\\} in Line 4 Algorithm 4 is more complicated, where we have to play enumeration to find the solution. This is the main difficulty in making the algorithm efficient in computation.

审稿意见
8

The paper studies a RL problem with delayed rewards, which are sampled after letting a trajectory roll out for a certain time horizon. This is a practical scenario that appears in several applications and that, to date, doesn't have a detailed and comprehensive solution. The paper shows how optimal regrets can be obtained that are similar to those of traditional RL algorithms.

优点

The paper studies a novel and interesting variant of the standard RL problem, where rewards are generated with a delay rather than at each time instants. The provided bounds are technically sound and improve upon the existing literature.

缺点

  1. Rewards are assumed to be a linear function of the trajectory (sequence of states and actions). This may not be practical in real-life. This is not practical even in the case discussed in the introduction - medical care.

  2. Learning the optimal policy, with the assumption of linear rewards, is now a (linear) regression problem. This is the main part from where the similarity between regret bounds for traditional RL and the proposed method arises. Can you comment on more complicated or realistic bandit models?

  3. Using a parameterized policy that can be systematically updated after the end of each trajectory, instead of policy elimination, might result in tighter regret bounds. Is this the case?

问题

Do the methods/ideas extend to cases where rewards are nonlinear functions of the trajectories?

The connections to linear bandits is interesting, and it would be beneficial to expand upon this connection (formulation and examples). The literature in this area is also rather extensive, and a more comprehensive review/comparison would be helpful.

Some numerical/illustrative examples could help the reader appreciate the theoretical results.

评论

About more complicated or realistic bandit models: In this work, we introduce linear bandit problem due to the natural formulation of reinforcement learning. As for extensions, it would be an interesting problem to replace the tabular MDP with more complicated MDPs, e.g., (generalized) linear MDP. We think advanced bandit models are related to these problems.

About parameterized policy without elimination: As far as we can see, it is unclear that parameterized policies could help improve the regret bound. On the other hand, we make delayed updates to keep statistical independence, so it might be improper to make frequently updates.

About the case where ewards are nonlinear functions of the trajectories: The general idea is to formulate the problem as a bandit problem with structure. There is possibly more interesting formulation because the arm set in this problem is the set of trajectories, which also has a good structure.

About connections to linear bandits: Thanks for the suggestion and we will improve the presentation accordingly.

审稿意见
5

This paper considers the episodic reinforcement learning with trajectory feedback. In this paper, the authors proposed an online learning algorithm that achieves optimal regret which scales as sqrt(SAH^3K). Unlike existing methods directly applied linear bandit, the proposed algorithm does not suffer from the regret with higher order on S.

优点

This paper provides rigorous theoretically analysis for the proposed approach. Instead of building confidence region as linear bandit, the proposed algorithm maintain a policy set within a constant range, which can be done through data sampling in batches. The theoretical motivation behind this idea is very clearly presented in the paper with rigorous proofs. Comparing to existing work, the proposed algorithm achieve optimal regret bound for RL with trajectory feedback only. Given the trajectory feedback is more general for real application, the proposed work has a potential to be applied to real problem with better performance guarantees.

缺点

The paper focuses on proposing algorithm that has tighter regret bound. While the rigorous proofs are appreciated, a more comprehensive comparison would help with understanding of the importance of the difference. It will become clearer with some toy experiment showing how the linear bandit performance in the episodic MDP environment when state/action is huge and time horizon is long. And since the proposed algorithm still require exponential time cost, it is also needed to compare the complexity in the paper.

问题

Beside what I wrote in the previous section, I would also like to understand more about Theorem 1 and 7. The statement of 'a RL problem with trajectory feedback has the same asymptotically optimal regret as standard RL' is a very intriguing result. I would like to read more explanation on the interpretation of Theorem 7 in the papaer.

评论

Please find our answers to the questions about inefficiency and why trajectory-feedback RL is as easy as standard RL in the common response.

审稿意见
6

This paper addresses reinforcement learning with trajectory feedback, where the reward in a Markov Decision Process is not observed throughout an episode but only as a cumulative reward at the episode’s end. The authors propose an algorithm that achieves a regret bound with statistical behavior comparable to the state of the art in standard reinforcement learning literature.

优点

The paper proposes an algorithm that achieves a better asymptotic regret bound than those previously established for reinforcement learning with trajectory feedback. The authors employ a novel method in their proofs, which appears robust.

缺点

First, I must admit that I am not well-versed in the literature on regret bounds in reinforcement learning with bandits.

To me, the framework of this paper appears quite limited. Moreover, the asymptotic regret bounds seem similar to those achieved in more classical frameworks, giving the impression that the work involves only minor modifications to standard proofs. However, I may be mistaken, as I am not deeply familiar with this literature.

问题

Here are few questions :

  1. Could the authors elaborate on the statement in lines 261–262: "Therefore, by running [...] for some constant c>0c>0"? A more detailed explanation of this step would be helpful.
  2. The definition of the set in (10) seems unclear to me. First, how do the authors compute maxWπ(R^,P)\max W^{\pi'}(\hat{R},P)? Additionally, how is the quantity on the right-hand side of the selection criterion, i.e., the practical value of O~\tilde{O}?
  3. Can the method employed in this paper be adapted to consider cases where KK is not that large compared to SS, AA and HH, similar to the approach in [Zhang et al.]?
评论

About the constant cc in lines 261-262: By this statement, we mean that by running π\pi for TT rounds to get trajectories τi\tau_i with feature ϕτi\phi_{\tau_i} for 1iT1\leq i \leq T, the information matrix Λ=i=1Tϕτ,iϕτ,i\Lambda = \sum_{i=1}^{T}\phi_{\tau,i}\phi_{\tau,i}^{\top} would be an upper bound of 12TEπ[ϕτϕτ]\frac{1}{2} T E_{\pi}[\phi_{\tau}\phi_{\tau}^{\top} ] with high probability (allowing some logarithmic error terms).

About equation (10): Our goal was to present some high-level intuition. In the algorithm, we do not compute Π+1\Pi_{\ell+1} with (10). Instead, we eliminate policies by comparing the value of some policy with the best LCB (lower confidence bound) over all policies (see Line 9 in Algorithm 4). By presenting equation (10), we aim to show that, for any policy πΠ\pi\in \Pi_{\ell} the gap between its UCB and its LCB is at most O~(SAH3/K)\tilde{O}(\sqrt{SAH^3/K_{\ell}}). To compute maxπWπ(R^,P)\max_{\pi}W^{\pi}(\hat{R},P), we can run simple backward planning since here we assume the knowledge of PP. In the case PP is unknown, we compute maxπWπ(R^,P^)\max_{\pi}W^{\pi}(\hat{R},\hat{P}) with optimistic backward planning, where P^\hat{P} is the empirical transition model.

About small KK: As far as we see, the answer is negative. Our algorithm depends on the knowledge of a reference transition model which requires a large sample complexity (in SS, AA and HH) using current techniques.

AC 元评审

This work contributes a new algorithm for online RL with a special form of information feedback. However, the manuscript itself is full of typos, the importance of the results is not demonstrated experimentally, and there do not appear to be any highly innovative components in the planning and policy learning steps. For this reason, the paper does not meet the bar for acceptance at this time.

审稿人讨论附加意见

NA

最终决定

Reject