PaperHub
6.1
/10
Poster4 位审稿人
最低3最高4标准差0.4
3
3
4
3
ICML 2025

Minimax Optimal Regret Bound for Reinforcement Learning with Trajectory Feedback

OpenReviewPDF
提交: 2025-01-23更新: 2025-07-24
TL;DR

We prove a near-optimal regret bound of $\tilde{O}(\sqrt{SAH^3K})$ for finite-horizon MDP with trajectory feedback.

摘要

In this work, we study reinforcement learning (RL) with trajectory feedback. Compared to the standard RL setting, in RL with trajectory feedback, the agent only observes the accumulative reward along the trajectory, and therefore, this model is particularly suitable for scenarios where querying the reward in each single step incurs prohibitive cost. For a finite-horizon Markov Decision Process (MDP) with $S$ states, $A$ actions and a horizon length of $H$, we develop an algorithm that enjoys an asymptotically nearly optimal regret of $\tilde{O}\left(\sqrt{SAH^3K}\right)$ in $K$ episodes. To achieve this result, our new technical ingredients include (i) constructing a tighter confidence region for the reward function by incorporating the RL with trajectory feedback setting with techniques in linear bandits and (ii) constructing a reference transition model to better guide the exploration process.
关键词
reinforcement learningonline learningtrajectory feedback

评审与讨论

审稿意见
3

This paper studies the setting of tabular MDP, with trajectory feedback, i.e., every round after executing a policy the learner obtains the entire trajectory with the total reward along the trajectory. The per-step reward is not given to the learner. This paper provided an algorithm for this setting and also proved that the algorithm achieves O~(SAH3K)\tilde{O}(\sqrt{SAH^3K}) regret upper bound, which also matches the lower bound, asymptotically.

给作者的问题

I have the following questions:

  1. Regarding the "burning in" term in the regret, e.g. the term that does not scale with K, what is the best you can obtain?
  2. Is it possible to generalize the results into linear MDP, or MDP with function approximation, since the tabular MDP is most restrictive?
  3. The algorithm designed in this paper first learns the transition model PP, then learn the reward model rr. Is it possible to learn these two simultaneously?

论据与证据

Yes

方法与评估标准

Yes

理论论述

Yes, I checked all proofs. They all look correct to me.

实验设计与分析

No experiment is provided in this paper.

补充材料

I went through all the supplementary materials, including the proof.

与现有文献的关系

The closest literature to this paper is (Efroni et al., 2021), which proposes the setting of learning MDP with trajectory rewards. In the previous paper, they show the regret upper bounded by O(S2A2H4K)O(\sqrt{S^2A^2H^4K}), which is suboptimal. This paper improve this rate to O(SAH3K)O(\sqrt{SAH^3K}).

遗漏的重要参考文献

No. As far as I know, this paper includes sall related references.

其他优缺点

Strength:

  1. This paper is well-written, except few typos I pointed out below.
  2. The paper provides a tight characterization to the regret.

Weakness:

  1. This paper studied the setting of tabular MDP, which is quite restrictive. Additionally, this paper merely tightens the bounds provided in Efroni et al. 21, which is incremental.
  2. I concern about the novelty of this paper. The results are good results, and they match the lower bound, which is also good, but by looking at the proving techniques, it feels that everything is well developed in previous literature and nothing feels special to me.
  3. The upper bound of the algorithm requires a large "burning in" phase, which I believe, still having improvement space.

其他意见或建议

Typos: (1) in Eq. (1), (Y^t - \phi_{\tau^t}^T r) to (Y^t - \phi_{\tau^t}^T r)^2 (2) On the right side of Line 175, d^\pi_P to d^\pi_P(\tau) (3) In Line 722 (Line 6 of Algorithm Raw-Exploration), missing symbol "\leftarrow"

Suggestions:

  1. The author mentioned that the preference-based learning can be a possible application to the RL with trajectory feedback. It would be better write down explicitly how does results in this paper implies about preference-based learning.
作者回复

We are grateful for the reviewer's detailed comments and constructive suggestions. Below we present our response.

Regarding significance: We would like to stress that even for RL in the standard setting, minimax optimal regret bounds have not been obtained until recent few years, and the burn-in terms were not fully optimized until very recently. Therefore, as the first work that achieves minimax optimal regret bound for RL with trajectory, we believe our work is non-trivial in terms of technique.

Regarding technical novelty: While the algorithmic framework depends on previous work, the key technical observation is that the number of possible trajectories O((SAH)O( (SA^H) is smaller to that of possible policies O(ASH)O( A^{SH}). This is why we can remove the extra S\sqrt{S} factor compared to previous work Efroni et al., 21.

Regarding the large burn-in terms: Indeed our current regret bound has large burn-in terms, and we would leave improving those terms as a future direction. The current term is S24A4H32S^{24}A^4H^{32}. The main reason of such a large burn-in term is due to learning the reference model. This term can be significantly reduced under the assumption of an effective reference model.

Regarding other comments:

About the typos: Thanks for pointing out the typos. We will fix the typos accordingly in the next revision. In the second typo, we use dPπd_{P}^{\pi} to denote the feature vector with respect to π\pi, which does not depend on the trajectory τ\tau.

About the preference-based learning: Preference-based RL (PbRL) is another RL paradigm to deal with the lack of reward signals. We will discuss the similarities and differences between the two settings in the next revision.

Regarding the questions:

Abut the exact burn-in term: Following the current analysis, the final regret bound is O~(SAH3K+S24A4H32)\tilde{O}(\sqrt{SAH^3K}+S^{24}A^4H^{32}).

About the extensions to other settings, e.g., linear MDP and MDP with function approximation: At the moment, we are not aware of any possible extensions to linear MDP or MDP with function approximation. We believe these problem settings should enjoy some advantages of low-rank structure so that we can construct a feature space with lower dimensions. For example, in linear MDP, we can take (ϕs1,a1,ϕs2,a2,...,ϕsH,aH) (\phi_{s_1,a_1}, \phi_{s_2,a_2},..., \phi_{s_H,a_H}) as a dHdH-dimensional unified feature to learn the reward kernel (θ1,θ2,...,θH) (\theta_{1}, \theta_2,..., \theta_H). However, the exploration and policy design over this feature space is more complicated compared to the counterparts in the tabular case. It requires more efforts to solve the experimental design problem over these complicated structures.

About learning the reward and transition kernel simultaneously: Thanks for the interesting question. Although we cannot establish this definitively, we conjecture the answer is negative in the worst case. The problem could be reformulated as finding a mixed policy πˉ\bar{\pi} to meet two following conditions simultaneously (we set ϕ(τ)\phi(\tau) to be ϕτ\phi_{\tau} for display):

minπˉΔΠmaxπΠτPrπ,P[τ]ϕ(τ)Λ(πˉ)1ϕ(τ)=O(SAH)\min_{\bar{\pi}\in \Delta^{\Pi}}\max_{\pi\in \Pi}\sum_{\tau}Pr_{\pi,P}[\tau]\phi^{\top}(\tau)\Lambda(\bar{\pi})^{-1}\phi(\tau)=O(SAH)

minπˉΔΠmaxπΠs,a,hdPπ(s,a,h)dPπˉ(s,a,h)=O(SAH).\min_{\bar{\pi}\in \Delta^{\Pi}}\max_{\pi\in \Pi}\sum_{s,a,h} \frac{d_{P}^{\pi}(s,a,h)}{d_{P}^{\bar{\pi}}(s,a,h)}=O(SAH).

For the first problem, the target is to maximize log(det(Λ(πˉ)))\log(\det(\Lambda(\bar{\pi}))) where we recall that Λ(πˉ)=τPrπ,P[τ]ϕ(τ)ϕ(τ)\Lambda(\bar{\pi}) = \sum_{\tau}Pr_{\pi,P}[\tau]\phi(\tau)\phi^{\top}(\tau). For the second problem, the target is to maximize log(Πs,a,hdPπˉ(s,a,h))=log(det(diag(Λ(πˉ)))\log(\Pi_{s,a,h}d_{P}^{\bar{\pi}}(s,a,h))=\log(\det(diag(\Lambda(\bar{\pi}))), where diag(Λ)diag(\Lambda) denotes the matrix by setting all non-diagonal elements of Λ\Lambda to 0. As a result, the two optimization problem are substantially different when h2h\geq 2. So we conjecture that the answer to your question is probably negative in the worst case.

审稿意见
3

This paper investigates reinforcement learning with trajectory feedback, where the agent receives only the cumulative reward for an entire trajectory rather than individual state-action rewards, while still observing all visited state-action pairs. The authors establish the first asymptotically nearly optimal regret bound of O(SAH3K)O(\sqrt{S A H^3 K}) for this setting, matching the asymptotically optimal regret bound in standard RL. To achieve this, they construct a tighter confidence region for rewards by leveraging the structure of the linear bandit instance associated with RL with trajectory feedback.

给作者的问题

Why the authors choose to use an elimination-based online batch learning process rather than using the new confidence region directly with a linear bandit algorithm? I may be overlooking a key step and would appreciate further clarification.

论据与证据

All theoretical claims are followed by proofs in the appendix or the main paper.

方法与评估标准

As a theoretical paper, the proposed algorithm seems well suited for the problem.

理论论述

I've reviewed some of the proofs, and while most appear correct, I would appreciate further clarification on the proof of Lemma B.1. This lemma seems to play a key role in improving the dependence of the regret bound on SS . However, the definition of λ(πˉ,π)\lambda(\bar{\pi}, \pi) is unclear. How do the authors define the probability that πˉ\bar{\pi} "distributes" over π\pi? Additionally, the interpretation of the partial derivative of FF with respect to this quantity is not well explained. These concepts may be standard in this type of analysis, but it would be helpful to reintroduce them rigorously, especially for readers in RL who may not be familiar this.

实验设计与分析

not applicable.

补充材料

I checked some of the proofs, in particular for Lemma B.1 see the comment above on theoretical claims.

与现有文献的关系

Efroni et al. (2021) established a regret bound of O~(S2A2H4K)\tilde{O}(\sqrt{S^2 A^2 H^4 K}) for RL with trajectory feedback, which holds for all K>0K > 0, not just asymptotically. The authors suggest that refining the analysis could improve this bound to O~(S2AH3K)\tilde{O}(\sqrt{S^2 A H^3 K}), but reducing the dependence on the number of states requires a fundamentally new approach. This paper presents the first algorithm that asymptotically achieves the lower bound of O~(SAH3K)\tilde{O}(\sqrt{S A H^3 K}) for RL with trajectory feedback, matching the optimal rate in standard RL, though the bound holds only in the asymptotic regime (for large KK, not all K>0K > 0). Like Efroni et al. (2021), the algorithm leverages the connection between RL with trajectory feedback and linear bandits but introduces a novel confidence region around rewards. Additionally, it employs a policy elimination method to learn the transition kernel, similar to the approach of Zhang et al. (2022b).

遗漏的重要参考文献

To the best of my knowledge, the related work is clearly presented.

其他优缺点

Strenghts:

  • The paper is generally well-written.

  • The technique of building the confidence region around the reward estimation by considering trajectories, rather than defining it for each deterministic policy and applying a union bound, enables an improved dependence on the number of states in the regret bound and appears to be a novel approach.

  • This work is the first to achieve an asymptotically near-optimal regret bound for this problem.

Weaknesses:

  • Some parts of the paper require further clarification, such as the proof of Lemma B.1 and the precise application of the classical Kiefer-Wolfowitz theorem in this setting (as claimed on page 5, line 224).

  • The proposed algorithm is not computationally efficient, whereas some existing approaches, such as UCBVI-TS from Efroni et al. (2021), are computationally efficient but yield worse regret bounds (O(H7S4A3K)O(H^7 S^4 A^3 K)).

  • The regret bound established in this work holds asymptotically in KK and does not apply for all K>0K > 0, unlike the bound in Efroni et al. (2021).

其他意见或建议

  • The phrase on page 6, lines 282-284, may be lacking some connectors, which makes the formulation unclear: ex.: "to the L1 norm"

  • Page 4 line 208: YtY_t should be YtY^t?

作者回复

We are grateful for the reviewer's detailed reviews and valuable suggestions. Below we present our response.

Regarding Lemma B.1: In Lemma B.1, πˉ\bar{\pi} represents a mixed policy. That is, by following πˉ\bar{\pi}, the learner takes each deterministic policy πΠ\pi\in \Pi with a certain probability as λ(πˉ,π)\lambda(\bar{\pi},\pi). As a result, F(πˉ)F(\bar{\pi}) could be regarded as a multi-variable function with respect to the probability vector [λ(πˉ,π)]πΠ[\lambda(\bar{\pi},\pi)]_{\pi\in \Pi}. We will revise the proof accordingly in the next version.

Regarding clarification of KW theorem: In the proof, we do not directly use KW theorem. Instead, we generalize the KW theorem to a distributed version in Lemma B.1. In the next revision, we are planning to present the original version of the KW theorem together with our generalized version to make the difference clear.

Regarding computational inefficiency: We admit that the current algorithm is computationally inefficient, and indeed, devising algorithm with minimax optimal regret bound and polynomial running time for RL with trajectory feedback is an intriguing open problem. However, we notice that obtaining statistically (nearly) optimal algorithms is usually the first step for completely settling problems in machine learning theory, and we believe our techniques would be useful for later developments in the theory of RL with trajectory feedback.

Regarding large burn-in terms: Indeed our current regret bound has large burn-in terms, and we would leave improving those terms as a future direction. On the other hand, we would like stress that even for RL in the standard setting, minimax optimal regret bounds have not been obtained until recent few years, and the burn-in terms were not fully optimized until very recently. Therefore, as the first work that achieves minimax optimal regret bound for RL with trajectory, we believe it is reasonable to have large burn-in terms in our bound. Moreover, we believe our new techniques are crucial for fully understanding the regret bound of RL with trajectory feedback.

Regarding other comments and suggestions:

About lines 282-284: We will rewrite this sentence as "Instead of approximating PP by pp under the L1L_1-norm, it is required the trajectory distribution under PP could be covered by that under pp, up to a constant ratio."

About line 208: Thanks for pointing out the typo. We will fix accordingly.

Regarding the elimination-based online batch learning: In this work, the fundamental problem is still reinforcement learning. More precisely, even if assuming that the reward is known, we are required to learn the transition kernel. This step cannot be trivially implemented using a linear bandit algorithm. As a result, we choose to use a linear bandit algorithm to learn the reward function and an RL algorithm to learn the transition kernel. Another two reasons to apply the eliminated-based batch learning are that: (1) Designing an algorithm that can simultaneously learn both the reward function and transition kernel presents significant challenges. As a result, we have to learn the reward and transition kernel separately; (2) By batch learning, we can efficiently reduce the statistical dependencies between different batches, which makes the analysis less complicated.

审稿人评论

I thank the authors for addressing my concerns and for committing to clarify some parts of the paper (such as Lemma B.1 and the use of KW theorem). I will keep my positive score.

审稿意见
4

This work considers the problem of online learning in a tabular finite horizon MDP with stochastic rewards and aggregate bandit feedback, where agent observes only the sum of rewards she collected after each episode.

An algorithm based on policy elimination is proposed, which builds on the linear bandits perspective to the problem. This approach, while not computationally efficient, achieves nearly minimax optimal regret in all problem parameters (assuming the number of episodes KK is sufficiently large).

One of the main observations used to design the algorithm, is that the number of possible trajectories in the MDP ((SA)H(SA)^H) is much smaller than the number of "arms", i.e., deterministic policies (ASHA^{SH}). Thus, building on confidence regions based on trajectories rather than policies, leads to dependence on S\sqrt S rather than SS that would be the result of a vanilla linear bandits algorithm, even in the known dynamics case.

update after rebuttal

I chose to increase my rating to accept. After further consideration, and taking into account that none of my concerns were major, I think there is more than enough in this work to merit acceptance.

给作者的问题

I don't have any important questions.

论据与证据

Yes

方法与评估标准

N/A

理论论述

I went over proofs of Lemmas 5.2 and 5.3, and the proof overview of Theorem 5.6. I didn't find any significant issues.

实验设计与分析

N/A

补充材料

The parts relevant to the proofs I checked.

与现有文献的关系

The problem we initially proposed by Efroni et al. (2021). Later, Cohen et al. (2021) study the adversarial setting, and Cassel et al. (2024) study the stochastic Linear MDP setting.

All the previous approaches provide O~(K)\widetilde O(\sqrt K) regret in the the tabular stochastic setting, but with suboptimal dependence on the rest of the problem parameters S,A,HS,A, H. However, they are computationally efficient.

The approach proposed in this work gives optimal dependence on all parameters (assuming KK large enough), but is not computationally efficient.

遗漏的重要参考文献

None that I know of.

其他优缺点

Strengths

  • The approach yields a (near) minimax optimal regret bound for the problem.
  • The technical overview and presentation of the algorithm are well written and relatively clear.

Weaknesses

  • The algorithm is not computationally efficient, and the burn-in period (requirement on KK for minimax optimality) is quite demanding.

Ultimately, establishing the optimal dependence on problem parameters is secondary in importance, even more so for a non-computationally efficient approach. Further, this is essentially the first paper that proposes a non-computationally efficient approach to the problem. That said, it is clear the problem is far from trivial and it seems this work offers some valuable insights. The algorithmic approach is pretty elegant and intuitive, at least at a high level.

其他意见或建议

Algorithms

  • Algorithm 5 line 3 computes c(s,a,h)c(s,a,h) (the occupancy of πˉ\bar\pi wrt pp?) but it does not seem to be used anywhere.
  • Algorithm 2 does not pass a δ\delta (lines 3,6) to Raw-Exploration (algorithm 6)
  • I suggest you add a "Conditioned on parameter settings in Section A,..." to the relevant Lemma statements.
  • In the proof overview of Theorem 5.6, should y~=2ϵ0\tilde y=2\epsilon_0 be y~=2ϵ\tilde y=2\epsilon_\ell?

Lemma 5.2

  • In the proof, P~P^2\tilde P \to \hat P_2 (or change the Lemma statement to be about P~\tilde P)
  • Eq. 28 is not synced with Lemma 5.3 (e.g., log2(T/δ)log(T)log(1/δ)log^2(T/\delta) \nleq \log (T) \log(1/\delta))
  • Also in Eq. 28, the last inequality seems to follow from a condition on KK that is not mentioned in the statement of the Lemma.

Lemma 5.3

  • “By Lemma D.2 … it holds that RRR \in \mathcal R” is this the R\mathcal R defined in line 13 of Algorithm 4, when invoked with (p,T,Π)(p, T, \Pi) of the current Lemma? Better to be explicit about this to help the reader. Also, R\mathcal R denotes your reward distribution, which is unrelated.
  • In Eq. 30, same comment as for Lemma 5.2, log2(Z/δ)log(Z)log(1/δ)\log^2(Z/\delta) \nleq \log (Z) \log(1/\delta) for general ZZ
  • There seems to be a sum over τ\tau missing in Eqs. 31 and 32

Misc

  • "To circumvent issues mentioned above, practitioners often rely on heuristics (e.g., reward shaping (Ng et al., 1999) or reward hacking (Amodei et al., 2016))." - Amodei et al. propose methods to avoid reward hacking. Reward hacking refers to the scenario where the agent exploits an inaccurate reward signal.
  • "The optimal Q-function and V-function are given by ..." Why use sup\sup and not max\max for QQ^*?
  • "we write the inner product xyx^\top y as xyxy for simplicity" - I would suggest to revise this decision, it generates place for confusion with scalar multiplication.
  • "the regret stemming from learning P~\tilde P can be bounded by..." It this point it is not clear what is the reference transition kernel so the sentence does not convey much information.
作者回复

We sincerely appreciate the reviewer's thorough evaluation and constructive feedback. Below we present our response.

Regarding the computational issue:

We admit that the current algorithm is computationally inefficient, and indeed, devising algorithm with minimax optimal regret bound and polynomial running time for RL with trajectory feedback is an intriguing open problem. However, we notice that obtaining statistically (nearly) optimal algorithms is usually the first step for completely settling problems in machine learning theory, and we believe our techniques would be useful for later developments in the theory of RL with trajectory feedback.

Regarding the burn-in terms:

Indeed our current regret bound has large burn-in terms, and we would leave improving those terms as a future direction. On the other hand, we would like stress that even for RL in the standard setting, minimax optimal regret bounds have not been obtained until recent few years, and the burn-in terms were not fully optimized until very recently. Therefore, as the first work that achieves minimax optimal regret bound for RL with trajectory, we believe it is reasonable to have large burn-in terms in our bound. Moreover, we believe our new techniques are crucial for fully understanding the regret bound of RL with trajectory feedback.

Regarding other comments and suggestions:

About c(s,a,h)c(s,a,h): We are sorry for this mistake. c(s,a,h)c(s,a,h) is only used in the analysis for Algorithm 5. We will move the definition of c(s,a,h)c(s,a,h) to Appendix D.4 in the next revision.

About δ\delta in Algorithm 2: We will use δ\delta as a common parameter across all algorithms and delete δ\delta from the input of Algorithm 6.

About parameters in Appendix.A: Thanks for the suggestion. We will refine the lemma statements to improve clarity.

About the proof overview of Theorem 5.6: Thank you for pointing out the typo. It should be y~=2ϵ\tilde{y}=2\epsilon_{\ell}.

About P~\tilde{P} in Lemma 5.2: Thank you for pointing out the typo. We will replace P~P^2\tilde{P}\to \hat{P}_2 accordingly.

About Eq.(28): Here we use the fact that log(A)log(B)=O(log2(AB))\log(A)\log(B)=O(\log^2(AB)) for A,B1A,B\geq 1. We have revised the inequality and reset the value of ϵ0\epsilon_0 as ϵ0=90000log3(SAHKδ)(SAH2K14+S4AH6K12)\epsilon_0=90000\log^3(\frac{SAHK}{\delta})\left( \frac{SAH^2}{K^{\frac{1}{4}}}+\frac{S^4AH^6}{K^{\frac{1}{2}}} \right). There should be an extra HH factor in the right-hand side of Eq.(28) to make the inequality holds. The revised version of Eq.28 is as follows. maxπΠdetWπ(R^,P)Wπ(R,P)Hlog(SAH)log(16/δ)(b1+325SAHlog(K)log(8SAH/δ)Kˉ1)1000log2(SAHKδ)(SAH2K14+4SAH2b1).\max_{\pi\in \Pi_{\mathrm{det}}}| W^{\pi}(\hat{R},P) - W^{\pi}(R,P)|\leq H\sqrt{\log(SAH)\log(16/\delta)}\left(b_1+ 325\sqrt{\frac{SAH\log(K)\log(8SAH/\delta)}{\bar{K}_1}} \right)\leq 1000\log^2\left(\frac{SAHK}{\delta}\right)\cdot \left(\frac{SAH^2}{K^{\frac{1}{4}}} + 4SAH^2b_1 \right). We remark that this HH factor is in the second order term so the final regret bound is still asymptotically optimal.

About R\mathcal{R}: We are sorry for the abuse of notations. You are correct that R\mathcal{R} is the confidence region defined in line 13 of Algorithm 4. The proof of Lemma D.2 is under the proof of Lemma 5.2, which studies the properties of Algorithm 4. We will make this clear in the next revision.We will also replace R\mathcal{R} with R\mathsf{R} when introducing the reward distribution.

About Eq.(30), Eq.(31) and Eq.(32): Thanks for pointing out the typos. We will fix accordingly.

About the term "reward hacking": We will remove the term "reward hacking" from this paragraph.

About sup\sup: We will replace sup\sup with max\max.

About xyxy and xyx^{\top}y: We will fix the notation about sup\sup and xyx^{\top}y. We only use xyxy as a shorthand of xyx^{\top}y in Appendix D.4 to reduce the complexity of the notations.

About "the regret stemming from learning P~\tilde{P} can be bounded by ...": We will mention that "P~\tilde{P} serves as an efficient tool to help design the exploration policy" before this sentence.

审稿意见
3

This work studies Reinforcement Learning with only Trajectory Feedback, where the agent does not observe the reward for each individual step separately. Under this setting, the author proposes a novel algorithm based on the arm-elimination method over all possible deterministic policies and achieves a near-optimal regret bound, which matches the lower bound for the setting with single-step rewards up to logarithmic factors.

给作者的问题

  1. As the author mentions in the regret guarantee, the upper bound matches the lower bound for a more powerful learner with single-step rewards, suggesting that the main challenge in learning an MDP comes from estimating the transition dynamics. Given this, it seems unreasonable that the algorithm in Ref-Model only dedicates a small fraction of rounds to estimating the transition matrix and then primarily focuses on reward estimation using the estimated transition model for the majority of rounds. More explanation is needed to justify why the key challenge can be effectively addressed in the first K\sqrt{K} rounds.

  2. In Lemma 5.2, the Ref-Model is claimed to achieve an approximate transition probability function with a small error ϵ0\epsilon_0. According to the regret analysis in this first stage, ϵ0\epsilon_0 is approximately 1/K1/\sqrt{K}. However, in general cases, it typically requires O(1/ϵ2)O(1/\epsilon^2) rounds to obtain an ϵ\epsilon-optimal estimator. Thus, it seems unreasonable that the algorithm can achieve a 1/K1/\sqrt{K}-optimal approximation with only K\sqrt{K} rounds. Further clarification is needed on how this estimation is justified.

论据与证据

The author provides a clear claim of the result in the theorems and includes a proof sketch to outline the key steps of the theoretical analysis.

方法与评估标准

The main contribution of this work focuses on the theoretical analysis of regret and does not have experiment.

理论论述

The author provides a clear proof sketch for the case where the transition probability is already known. In this setting, the author utilizes the observation that the number of trajectories is significantly smaller than the number of deterministic policies and represents the reward of each policy as a linear combination of the rewards for each trajectory. By leveraging this method, the author reduces the problem to a linear bandit problem, achieving better performance with a lower-dimensional representation.

However, the correctness of the analysis when the transition probability is unknown is not entirely clear, and I have several concerns:

  1. As the author mentions in the regret guarantee, the upper bound matches the lower bound for a more powerful learner with single-step rewards, suggesting that the main challenge in learning an MDP comes from estimating the transition dynamics. Given this, it seems unreasonable that the algorithm in Ref-Model only dedicates a small fraction of rounds to estimating the transition matrix and then primarily focuses on reward estimation using the estimated transition model for the majority of rounds. More explanation is needed to justify why the key challenge can be effectively addressed in the first K\sqrt{K} rounds.

  2. In Lemma 5.2, the Ref-Model is claimed to achieve an approximate transition probability function with a small error ϵ0\epsilon_0. According to the regret analysis in this first stage, ϵ0\epsilon_0 is approximately 1/K1/\sqrt{K}. However, in general cases, it typically requires O(1/ϵ2)O(1/\epsilon^2) rounds to obtain an ϵ\epsilon-optimal estimator. Thus, it seems unreasonable that the algorithm can achieve a 1/K1/\sqrt{K}-optimal approximation with only K\sqrt{K} rounds. Further clarification is needed on how this estimation is justified.

实验设计与分析

The main contribution of this work focuses on the theoretical analysis of regret and does not have experiment.

补充材料

No, due to time limitations, I only reviewed the main paper and did not check the supplementary material.

与现有文献的关系

This work mainly focuses on reinforcement learning with trajectory feedback; however, the proposed algorithm is highly computationally inefficient, making it primarily relevant for the theoretical analysis of reinforcement learning rather than practical applications.

遗漏的重要参考文献

This paper provides a comprehensive discussion of related work in linear bandit and trajectory-based reinforcement learning method.

其他优缺点

  1. The proposed algorithm is highly computationally inefficient, making it primarily relevant for the theoretical analysis of reinforcement learning rather than practical applications.

  2. For the non-dominant term, the dependency on S,A,HS, A, H is extremely large, making the regret guarantee non-trivial and only meaningful for an excessively large number of rounds KK.

其他意见或建议

  1. The notation of KK and TT is consing. As the author mention in the introduction, there may exists a gap of the episode length HH between the number of episode and tnumber of stage HH. However, the author use the both notation in algorithm, e.g., Algorithm 1 take a oracle to Traj-Learning with parameter KK and when introduce the Traj-Learning algoritmh, with a notation of TT. better to union.

  2. For the non-dominant term in Theorem 5, there is still a dependency on the number of KK. It would be better to explicitly separate the effect of KK in the non-dominant term to provide a clearer understanding of its impact on the regret bound.

作者回复

We are grateful for the reviewer's detailed assessment and helpful suggestions. Below we present our response.

Regarding your concerns about correctness:

  1. In the first K0=O(K)K_0=O(\sqrt{K}) episodes (other factors ignored), the main target is to identify the infrequent state-action-state triples with probability O(σ0)=O(K1/2)O(\sigma_0)=O(K^{-1/2}). For the left triples, we want to compute an (3,σ0)(3,\sigma_0)-approximate estimation for the transition probability Ps,a,hP_{s,a,h}. Note that this approximate transition matrix is only for designing the exploration policy, not for computing the near-optimal policy. Hence, computing such an approximate transition matrix requires only O(K)O(\sqrt{K}) episodes.

  2. We set ϵ0=O(K1/4)\epsilon_0 = O(K^{-1/4}) ignoring other factors (please refer to Appendix.A). You are correct that it requires 1/ϵ021/\epsilon_0^2 episodes to learn an ϵ\epsilon-optimal policy.
    Indeed, we compute this ϵ0\epsilon_0 optimal policy to improve the efficiency of exploration in the following episodes. In the main algorithm, we use K0=O(K)K_0=O(\sqrt{K}) episodes to identify the triples with probability O(K1/2)O(K^{-1/2}). If we apply direct exploration, the regret in this stage might be K0HK_0H, which violates the minimax bound O(SAH3K)O(\sqrt{SAH^3K}) due to the dependencies on S,AS,A and HH. Instead, we first compute the set of ϵ0\epsilon_0 optimal policies, and then conduct exploration within this policy set. In this way, the regret due to exploration is bounded by O(K1H+K0ϵ)=O(SAH3K)O(K_1 H + K_0 \epsilon) = O(\sqrt{SAH^3K}).

Regarding the computational inefficiency:

We admit that the current algorithm is computationally inefficient, and indeed, devising algorithm with minimax optimal regret bound and polynomial running time for RL with trajectory feedback is an intriguing open problem. However, we notice that obtaining statistically (nearly) optimal algorithms is usually the first step for completely settling problems in machine learning theory, and we believe our techniques would be useful for later developments in the theory of RL with trajectory feedback.

Regarding the large dependency on S,A,HS,A,H:

Indeed our current regret bound has large burn-in terms, and we would leave improving those terms as a future direction. On the other hand, we would like stress that even for RL in the standard setting, minimax optimal regret bounds have not been obtained until recent few years, and the burn-in terms were not fully optimized until very recently. Therefore, as the first work that achieves minimax optimal regret bound for RL with trajectory, we believe it is reasonable to have large burn-in terms in our bound. Moreover, we believe our new techniques are crucial for fully understanding the regret bound of RL with trajectory feedback.

Regarding other comments or suggestions:

We are sorry for the typos. The TT notation in Section 5 denotes number of episodes, not number of steps. We will use Kˇ\check{K} to replace TT throughout Section 5 and the corresponding analysis in the next revision. We will also separate KK from other factors. The final regret bound would be O~(SAH3K+S24A4H32)\tilde{O}(\sqrt{SAH^3K}+S^{24}A^4 H^{32}).

审稿人评论

Thanks for the rebuttal and it addresses my concern regarding the algorithm. As the authors mentioned, the approximate transition matrix is only used for designing the exploration policy. There is a subsequent exploration stage in Algorithm 5, and the update policy is based on newly collected data (Algorithm 5, Line 7).

Additionally, in Algorithm 1, all Traj-Learning is based on the reference model computed during the first stage. It is interesting that if we replace it with the estimated transition probability function from Algorithm 5 during the iterative structure, would that help reduce the burn-in terms?

Overall, I will maintain my positive score.

作者评论

Thanks for the follow-up question.

Regarding improving the burn-in terms using new estimated transition matrix: We think the answer is not. In our current analysis, a (3,σ0)(3,\sigma_{0})-approximate transition matrix suffices for designing the exploration policy. A better approximation of the transition matrix does not improve the efficiency of policy design.

最终决定

We thank the authors for their submission. All in all, the contributions were clear, the paper is well-written and the proofs appear to be sound. The reviews were reserved about the algorithm not being computationally efficient, and excessively large regret incurred during the burn-in period. Perhaps these can be studied in future work.