Provably Efficient UCB-type Algorithms For Learning Predictive State Representations
We developed a provably efficient upper confidence bound algorithm for online and offline low rank decision-making problem, which is modeled by predictive state representation.
摘要
评审与讨论
This work establishes the first UCB-type algorithms for PSR problems. The proposed algorithm is computationally tractable compared with previous methods and obtains sample complexity which matches the previous best upper bound w.r.t. the rank of PSR and accuracy level. The counterpart of the algorithm and analysis in the offline case has also been established.
优点
- Motivation: Given previous methods studying PSR problems that are not computationally tractable or sample from posteriors maintained over a large model set, studying a computationally tractable method for PSR is well-motivated.
- Novelty: This work proposes the first UCB bonus to enable efficient exploration of PSR problems, the core design of which is a new MLE guarantee. The newly proposed MLE guarantee and UCB bonus might be of independent interest.
缺点
- The proof of Theorem 1 relies on some sort of MLE guarantee, a confidence bound constructed by the empirical features and relating the empirical bonus to the ground-truth bonus. Such an approach shares similar spirits as previous works studying MDP with low-rank structures (say [1]). I would like to suggest the authors give more comparisons on this.
- The result in Theorem 1 seems to critically depend on the newly proposed MLE guarantee in this work. It might be better if discussions about how this MLE guarantee is established and to what extent it differs from previous ones are given.
- By “computationally efficient”, it in most cases means that the algorithm enjoys polynomial computation efficiency. In some parts of this work, the authors claim that the proposed algorithms are computationally efficient. In my opinion, the proposed algorithm is indeed computationally tractable or oracle-efficient but might not be “computation efficient”, since the polynomial efficiency of the proposed algorithm is not guaranteed (correct me if I am wrong).
[1] Uehara et al. Representation Learning for Online and Offline RL in Low-rank MDPs. ICLR, 2022.
问题
Please see the weakness part above.
Q3: By “computationally efficient”, it in most cases means that the algorithm enjoys polynomial computation efficiency. In some parts of this work, the authors claim that the proposed algorithms are computationally efficient. In my opinion, the proposed algorithm is indeed computationally tractable or oracle-efficient but might not be “computation efficient”, since the polynomial efficiency of the proposed algorithm is not guaranteed (correct me if I am wrong).
A3: We apologize for the inaccurate description of the computational efficiency of our algorithm. The proposed algorithm is indeed oracle-efficient. We have changed the term "computationally efficient" to "computationally tractable" in the revision.
We thank the reviewer again for the helpful comments and suggestions for our work. If our response resolves your concerns to a satisfactory level, we kindly ask the reviewer to consider raising the rating of our work. Certainly, we are more than happy to address any further questions that you may have.
Thanks for the detailed responses. I have no further questions and I would like to maintain my current evaluation for this work.
We thank the reviewer very much to check our response and for your effort into this review process.
We thank the reviewer for the helpful comments, and would like to provide the following responses. The paper has been revised accordingly to reflect these responses, where the changes are highlighted in the blue color. We hope that these discussions can clarify the reviewer's concerns, and we will be more than happy to answer any further questions you may have.
Q1: The proof of Theorem 1 relies on some sort of MLE guarantee, a confidence bound constructed by the empirical features and relating the empirical bonus to the ground-truth bonus. Such an approach shares similar spirits as previous works studying MDP with low-rank structures (say [1]). I would like to suggest the authors give more comparisons on this.
[1] Uehara et al. Representation Learning for Online and Offline RL in Low-rank MDPs. ICLR, 2022.
A1: While both [1] and this work utilize MLE oracle and UCB-based design, our method and analysis differs significantly from [1] as follows.
First, we adopt a constrained MLE oracle, which provides a more stable estimation (page 6) compared with vanilla MLE used in [1]. Here, the stable estimation implies that the estimated model is guaranteed to be close to the true model not only in expectation, but also on empirical samples. Specifically, we establish the estimated model guarantee in terms of the empirical conditional probability distributions, i.e. This stability indicates that is close to the true feature , which allows us to use empirical features to design bonus function . This cannot be achieved by vanilla MLE estimation in PSRs.
Second, due to non-Markovian property of PSR, the step-back technique, which is frequently used to relate the empirical bonus to the ground-truth bonus in low-rank MDPs, cannot be employed in our setting. Instead, we develop a new technique in Lemma 13, which controls the difference between two bonus functions using different features.
Third, we exploit the physical meaning of the features in PSR described in Sec 4.1 for efficient exploration. Specifically, each coordinate of a feature represents the transition probability from a history to a core test. This unique property of PSR further allows us to successfully design a UCB-based approach and addresses the challenge brought by PSR regarding how to relate the empirical bonus to the true bonus (see Lemma 4).
Q2: It might be better if discussions about how this MLE guarantee is established and to what extent it differs from previous ones are given.
A2: Our MLE guarantee is established by analyzing the conditional probability for those sampled history . Due to the constraint, we can find a good cover of the parameter space in terms of the conditional probability distribution distance (see Proposition 5), i.e. . Equipped with the optimistic cover, we then apply Chernoff bound and union bound to obtain the final guarantee.
The key difference of our proposed constrained MLE oracle from previous ones is that the constraint provides a more stable MLE solution. Specifically, we establish the estimated model guarantee in terms of the empirical conditional probability distributions, i.e. This stability indicates that is close to the true feature , which allows us to use empirical features for planning and makes our algorithm computationally tractable.
For a clearer comparison, we note that previous works either does not have "conditional distribution" guarantee [2], or only have guarantee on the "expected" distance [1] instead of on the empirical samples.
[2] Liu, et al., Optimistic MLE—A Generic Model-based Algorithm for Partially Observable Sequential Decision Making, 2022.
This paper studies learning predictive state representations (PSRs) using UCB-type approaches. The proposed PSR-UCB algorithm is based on an upper bound for the total variance distance between estimated and true models. The algorithm requires supervised learning oracles based on which it is computationally efficient and has a last iterate guarantee. An offline variant PSR-LCB is also proposed.
优点
- This paper is in general well-written and smooth to follow, with the exception of some proofs in the supplementary material.
- The originality of the paper is high as it proposes a novel UCB-based algorithm for PSRs. Related works are covered in detail.
- The theoretical proofs seem to be rigorous.
- The proposed UCB-based algorithm enjoys strong theoretical guarantees and might be of interest to researchers.
缺点
My main concerns with this paper are as follows.
- First, the theoretical results crucially depend on the MLE oracle, which is called times during the implementation. However, the paper falls short in discussing the difficulty of solving the MLE estimation. At least, it would be better if, for some special class of 's, the authors could show how the estimation could be implemented and what's computationally costly.
- This paper is purely theoretical and lacks experiments even on synthetic data. I would appreciate simulation results validating the sample complexity presented in the theory.
问题
I've detailed my primary concerns and suggestions above. However, I have a couple of specific questions regarding certain aspects:
-
On Page 6, just below equation (2), could the authors elaborate on why learning $\bar{\psi}^(\tau_h^{k, h+1}) \mathbb{P}_{\theta^}^{\pi^{k-1}}(\tau_h^{k, h+1}) $ is very small?
-
In Algorithm 1, the UCB-approach seems to serve as a pure exploration mechanism aimed at enhancing the estimation of . Yet, in Algorithm 2, the upper confidence bound for is directly employed to construct a lower confidence bound for , which the output policy then optimizes. Could you shed light on what drives this distinction between the designs of the two algorithms? Or, is there an underlying equivalence between them?
Q4: In Algorithm 1, the UCB-approach seems to serve as a pure exploration mechanism aimed at enhancing the estimation of . Yet, in Algorithm 2, the upper confidence bound for is directly employed to construct a lower confidence bound for , which the output policy then optimizes. Could you shed light on what drives this distinction between the designs of the two algorithms? Or, is there an underlying equivalence between them?
A4: We note that the ultimate goal of both algorithms is to output a near-optimal policy.
In Algorithm 1, we choose a reward-free design for online exploration, since it can provide stronger performance guarantees including a near-accurate estimated model. The bonus value also serves as an upper confidence bound (UCB) for . By deploying the policy that maximizes the UCB of , we are collecting data where the estimated model has the most uncertainty. Thus, the accuracy of the estimated model can be enhanced.
In Algorithm 2, it is hard to estimate the model accurately since the offline dataset has partial coverage of the target policy, or the optimal policy. Following the commonly adopted pessimism principle in offline RL [1], we directly construct a lower confidence bound for in light of the inequality and an UCB of .
In summary, both Alg.1 and Alg. 2 utilize the UCB of , which is . But due to different purposes, Alg. 1 maximizes to enhance the accuracy of the estimated model under reward-free design, and Alg. 2 maximizes the LCB under the pessimism principle to handle partial coverage.
[1] Uehara et a., Representation Learning for Online and Offline RL in Low-rank MDPs.
We thank the reviewer again for your feedback and suggestions. Hopefully, our responses have addressed your main concerns. If so, we kindly ask that you consider raising the score of your evaluation. If you still have remaining questions, please let us know, and we will be very happy to provide further clarification.
I appreciate the authors' detailed responses to my questions. I'll keep my current rating as there are still no simulation results.
We thank the reviewer for checking our response. We understand the reviewer's decision. We are working on the numerical experiments, which take some time. Since most of the studies along this line of research on theoretical POMDP don't have numerical experiments, we need to code from scratch. We are doing our best.
We thank the reviewer for the helpful comments, and would like to provide the following responses. The paper has been revised accordingly to reflect these responses, where the changes are highlighted in the blue color. We hope that these discussions can clarify the reviewer's concerns, and we will be more than happy to answer any further questions you may have.
Q1: Solving MLE oracle. It would be better if, for some special class of 's, the authors could show how the estimation could be implemented and what's computationally costly.
A1: We note that Eq (2) can be solved as an unconstrained optimization problem in practice as we explain below.
In the constraint, the threshold can be set as small as , where is an arbitrary positive constant (see Theorem 1). Clearly, such a threshold is approximately zero (note that is typically chosen to be a small constant, and all variables in the denominator are large in practical systems). This means that as long as we guarantee is nonzero, the constraint is satisfied. Hence, in practice, we can simply parameterize to have nonzero elements to satisfy the constraint, which is needed anyway to prevent infinite value of in the objective function of Eq (2). Then Eq (2) will become a standard unconstrained optimization to maximize the likelihood value (MLE problem). The computational cost is thus equivalent to the cost of performing supervised learning with loss function , which can be solved via methods such as stochastic gradient descent.
Q2: This paper is purely theoretical and lacks experiments even on synthetic data. I would appreciate simulation results validating the sample complexity presented in the theory.
A2: Thanks for the suggestion. We are actively exploring some synthetic settings to validate our algorithm and theoretical results. We hope to produce some results soon.
Q3: Could the authors elaborate on why learning might be harmful when is very small?
A3: We apologize for the confusion here. This statement should be interpreted as since is not small with high probability, it is harmful to have an estimated model where is very small. In Proposition 7 (page 19), we prove that must be greater than with high probability. In other words, it is highly unlikely that is very small. Therefore, under this high probability event, if a model assigns a small probability on , then this model is unlikely to be close to the true model. In other words, using for future planning is harmful because it differs from the true model too much.
This paper adds the UCB (or LCB) term to online or offline POMDP. The modification compared with (Liu et al., 2022) is that 1) a hard threshold p_\min adding to the selection of and 2) the UCB term used in exploration.
优点
- This paper is well-written and easy to follow
- The online algorithm can provide a `last-iteration' bound, and the bonus term can be efficiently calculated.
缺点
- Compare with (Liu et al., 2023), why during the exploration, the action trajectory needs to be sampled from , why we need the last term ? In addition, why we do not need to guarantee here?
- The sample complexity of the online algorithm is confusing, it looks more like a 'reward-free' paradigm, in which we are concerned about the sample complexity instead of the regret. When we talk about the 'last iteration regret', we should assume this regret bound works for all and the algorithm will not change regarding different , which is not the case in this paper: we need to predefine a and get the sample complexity bound. Thus I think this result is kind of strange.
- I wonder what will the algorithm look like if we do not only use but use the optimistic value function in the online algorithm, should the result be the same with (Liu et al., 2023)?
- Why the constrain related to p_\min is easy to hold? Is there any theoretical justification? In other words, what will happen if that constrain does not hold for any ?
问题
Please see weakness
We thank the reviewer for the helpful comments, and would like to provide the following responses. The paper has been revised accordingly to reflect these responses, where the changes are highlighted in the blue color. We hope that these discussions can clarify the reviewer's concerns, and we will be more than happy to answer any further questions you may have.
Q1 (a): The action trajectory needs to be sampled from , why we need the last term ?
A1 (a): We emphasize that the key difference between our work and [3] is that we try to estimate the feature of the empirical sample , while [3] analyzes the estimation error of in expectation, i.e., , which can be upper bounded by using importance sampling. Therefore, it suffices to sample to obtain an accurate estimate of and ensure the estimation error shrinks. Such exploration is exactly what the algorithm has done at step , i.e. uniformly sampling actions from at step . Therefore, [3] does not need to include additional for exploration.
However, our estimation error (Eq. (b) in Lemma 2, page 22) cannot be controlled in a similar way as [3]. This is the reason that we need to explore explicitly. The importance of uniformly selecting actions from has been explained in the second paragraph on page 6. We restate it as follows. First, the actions in assist us to learn the prediction feature since the -th coordinate of equals . Selecting improves the probability of observing , which will provide more information to estimate . Second, the action sequence helps us to estimate because the -th coordinate of represents . Similarly, selecting might provide more data to estimate . Thus, by uniformly selecting actions from , we are able to collect the most informative samples for estimating the true model .
In summary, the novel design that uses empirical features to quantify the uncertainty level requires a new exploration strategy, i.e. uniform sampling from . In addition, from the analysis for in Lemma 2 on page 22, the final result depends on the size of the set of exploration action sequences. Thus, this new strategy does not affect the overall sample complexity, since
[1] Guo et al., Provably Efficient Representation Learning with Tractable Planning in Low-Rank POMDP.
[2] Uehara et a., Representation Learning for Online and Offline RL in Low-rank MDPs.
[3] Liu, et al., Optimistic MLE—A Generic Model-based Algorithm for Partially Observable Sequential Decision Making, 2022.
[4] Zhan, et al., PAC Reinforcement Learning for Predictive State Representations
Q3: I wonder what will the algorithm look like if we do not only use but use the optimistic value function in the online algorithm, should the result be the same with (Liu et al., 2023)?
A3: Great point! Using the optimistic value function in the online algorithm will also allow us to find a near-optimal policy at the last iteration. However, the drawback is that it no longer guarantees that the estimated model is accurate. This is because the optimistic policy now depends the reward as well as the uncertainty level, which has less incentive to explore the trajectories generating low rewards. As a result, the accuracy of estimated model over those trajectories cannot be guaranteed.
In terms of the sample complexity, in the large regime when , our result achieves better order on than the result in Liu et al. (2023), but has an additional factor . In the low rank regime where , our result has better order on .
Q4: Why the constrain related to is easy to hold? Is there any theoretical justification? In other words, what will happen if that constraint does not hold for any ?
A4: We prove that with high probability, the true model always satisfies this constraint (See Proposition 7 on page 19). Moreover, a simple model that makes trajectories uniformly distributed also satisfies this constraint. Specifically, let . Then clearly satisfies the constraint.
In practice, we can set for any absolute constant . Clearly, such a threshold is approximately zero (note that is typically chosen to be a small constant, and all variables in the denominators are large in practical systems). This means that as long as we guarantee is nonzero, the constraint is satisfied. Hence, in practice, we can simply parameterize to have nonzero elements to satisfy the constraint, which is needed anyway to prevent infinite value of in the objective function of Eq (2). Then Eq (2) will become a standard unconstrained optimization to maximize the likelihood value (MLE problem). Such an unconstrained problem can be solved efficiently through standard gradient descent based approaches.
We thank the reviewer again for your feedback and suggestions. We also would like to bring to your attention that your individual ratings on Soundness (3 good), Presentation (4 excellent) and Contribution (3 good) don't seem to be quite consistent with your final rating 5 of our paper. If our response resolves your concerns, could you please kindly consider raising your rating of the paper, which will also make it aligned with your individual ratings? Of course, if you have any further questions, we would be very happy to address them.
Q1 (b): In addition, why we do not need to guarantee here?
**A1 (b):**The reason that we do not need to guarantee is because of the bonus-based design and the stable MLE oracle. First, we note that for reward-known design, there is no need to guarantee even for confidence-set-based design (Eqn. (6) in [4]). For reward-free design in model-based algorithms, a key requirement is to output a model that is sufficiently accurate. In bonus-based algorithms [1,2], we typically just select an arbitrary model from , where the only property we need is that the model is close to the MLE solution (see Eqn. (2)), and the estimation error of the model is controlled by the empirical bonus term as
.
However, in confidence-set-based algorithms [3], the estimated model is optimized within at each iteration, and there is no ``computable'' control on the accuracy of the models within a confidence set, since the accuracy of those models in depends on ground-truth bonus function, which is unknown. Under this situation, the only guarantee they have is that the summation of the estimation errors of any sequence of models selected from grows sublinearly in . Mathematically, holds for any and . Therefore, to guarantee the accuracy of the final output model , the analysis in [3] requires , so that the accuracy of the model in the final confidence set enjoys all the the guarantees in previous iterations, i.e. .
Q2: The sample complexity of the online algorithm is confusing, it looks more like a 'reward-free' paradigm, in which we are concerned about the sample complexity instead of the regret. When we talk about the 'last iteration regret', we should assume this regret bound works for all and the algorithm will not change regarding different , which is not the case in this paper: we need to predefine a and get the sample complexity bound. Thus I think this result is kind of strange.
A2: We would like to clarify that our performance metric is the sample complexity, not the regret.
The last-iterate guarantee refers to the sample complexity needed to guarantee the performance of the final output policy and estimated model of our algorithm.
In the current version of our algorithm, we use a predefined to control the maximum number of episodes run by the algorithm, which is just for ease of analysis. We can actually remove from line 2 in Alg. 1, since the algorithm is guaranteed to stop when line 10 is satisfied. For this case, only serves as a hyperparameter used to tune other parameters, such as and in Theorem 1, so that the algorithm is guaranteed to stop within episodes.
We emphasize that assuming a prefixed is standard for RL algorithms when sample complexity is considered [1,2,3,4]. As a separate note, none of algorithm in [2,3,4] can relax the assumption of a prefixed , while our algorithm can afford such flexibility.
[1] Guo et al., Provably Efficient Representation Learning with Tractable Planning in Low-Rank POMDP.
[2] Uehara et a., Representation Learning for Online and Offline RL in Low-rank MDPs.
[3] Liu, et al., Optimistic MLE—A Generic Model-based Algorithm for Partially Observable Sequential Decision Making, 2022.
[4] Zhan, et al., PAC Reinforcement Learning for Predictive State Representations.
Dear Reviewer dtbr,
As the author-reviewer discussion period will end soon, we will appreciate it if you could check our response to your review comments. This way, if you have further questions and comments, we can still reply before the author-reviewer discussion period ends.
We also would like to bring to your attention that your individual ratings on Soundness (3 good), Presentation (4 excellent) and Contribution (3 good) don't seem to be quite consistent with your final rating 5 of our paper. If our response resolves your concerns, we kindly ask you to consider raising the rating of our work. Thank you very much for your time and efforts!
This paper presents an UCB-based algorithm to learn parameters in POMDP using predictive state representations (PSRs) to simplify the number of potential historical and future trajectories. PSRs assume a smaller subset of possible trajectories with a low-rank basis. Therefore, PSRs can efficiently reduce the dimensionality required to represent all the historical or future trajectories.
Using PSRs, the authors show how to use maximum likelihood method to find the most likely PSR parameters, and then use an additional bonus function to quantify the uncertainty of the learned PSR parameters. The algorithm leads to no-regret guarantees with a sample complexity bound induced from the regret analysis. The UCB construction is quite different from other UCB algorithms that I have seen in the literature of multi-armed bandits.
优点
Solid theoretical analysis: I checked the main paper but didn’t check the appendix. The main paper and proof sketches are theoretically sound to me.
UCB algorithm for learning parameters in POMDP (reduction in learning space): careful and detailed analysis of a complex problem of learning parameters of POMDP from historical trajectories. Previous online learning in sequential problems usually learns the tabular-from parameters directly with easily constructed UCBs. This paper analyzes UCBs in predictive state representations (PSRs), which can reduce the number of potential historical trajectories in POMDP to maintain UCBs.
New analysis: the analysis is indeed new to me. I appreciate the novelty in theoretical contributions.
缺点
Computation: the computation cost is still expensive with a dependency on how good the PSR is (the size of ). Could you elaborate more on the overall computation complexity and point out what is the bottleneck?
Reward-free claim: the theoretical results and the analysis are reward-free by design. The major learning problem is on learning the transition probability using PSR. The reward function is bounded within [0,1] so I am not surprised that the algorithm and results are reward-free.
Overly complicated notations: I understand this is a difficult problem with many notations needed for the analysis. But the presentation is very difficult to follow and its clarity can be significantly improved. I believe there is a better way to present and help readers understand the analysis better.
Sample complexity: I thought a major advantage of PSRs is to discard the dependency on the number of all possible history with size . However, this term still appears in your sample complexity with a dependency on . The dependency on the action and observation space is hidden in where . Could you explain how your algorithm improves and compares with the online learning algorithms without using PSRs?
Offline learning: the offline learning part directly follows the same proof as the online learning part. Could you elaborate if there is any new insights brought by the offline learning analysis?
问题
Questions: please answer my questions mentioned in the weaknesses section. I am happy to raise the score if some of my questions are clarified.
Typos or confusions:
- Equation 1: what does mean? Your definition of measures the probability of choosing a sequence of actions conditioned on a sequence of observations. only specifies the future trajectory after time but does not specify the history prior to . Please clarify.
- After finishing reading the paper, I still don’t see what does the “physical meaning” mean in Section 4.1.
- Line 4 in Algorithm 1: notations of and are confusing. You have a different use of superscript on and defined previously. The following discussion in Page 6 is helpful though.
- Equation 3: , abused notation which collides with the definition of the bonus function with a different involved. Similarly, there is a typo in the following sentence: where should be unless I misunderstood.
Q5: Could you elaborate if there is any new insights brought by the offline learning analysis?
A5: One fundamental difference between online and offline learning is that, during online learning, the learner is able to decide the exploration policies adaptively, while in the offline setting, the behavior policy adopted to collect the dataset cannot be controlled by the learner. As a result, it leads to two major differences in the offline learning analysis.
First, offline learning requires some coverage assumption on the behavior policy adopted to collect the dataset in order to ensure that the information contained in the dataset is sufficient to recover a near-optimal policy. Specifically, our analysis relies on the newly proposed coverage coefficient (Sec. 5.2) for PSRs. Unlike the analysis for online learning, where we use elliptical potential lemma to show the closeness between the output policy and the optimal policy, this coverage coefficient helps us to establish the relationship between the optimal policy and the behavior policy (See Appendix D.4). Second, the action sequence chosen by the behavior policy may also be different from the online exploration policy. Specifically, online exploration policies have the freedom to uniformly explore core action sequence, while does not necessarily do so. To handle such difference, importance sampling is used in the offline learning analysis to control the estimation error by (See Appendix D.2).
Those differences in the analysis highlight the importance of coverage assumption and core actions sequence in offline learning for PSRs. As offline PSR learning is still an active research area, an interesting direction is to weaken the coverage assumption such that it is more suitable for the dataset in real applications.
We then address the typos and confusions as follows.
Q6: Equation 1: what does mean? Your definition of measures the probability of choosing a sequence of actions conditioned on a sequence of observations. only specifies the future trajectory after time but does not specify the history prior to . Please clarify.
A6: We apologize for the confusion. The maximization in Eqn. (1) should also be over all possible . In other words, , where the optimization over is over all possible distributions on future action sequences . We have clarified this in the revision.
Q7: After finishing reading the paper, I still don’t see what does the “physical meaning” mean in Section 4.1.
A7: By physical meaning, we are referring to the sentence each coordinate of a prediction feature of represents the probability of visiting a core test conditioned on and taking the corresponding core action sequence right after the term ''physical meaning'' in Sec 4.1. We have revised this statement to make it clearer in the revision.
Q8: Line 4 in Algorithm 1: notations of and are confusing.
A8: The superscripts represent the index of episodes during online exploration. The superscripts indicate the state sequence and the action sequence, respectively. We have updated the presentation of Sec. 4.1 (change the order of algorithm description and the pseudo-code) to avoid potential confusion.
Q9: Equation 3: , abused notation which collides with the definition of the bonus function with a different involved. Similarly, there is a typo in the following sentence: .
A9: We apologize for the typo. Note that each element in has the form . Thus, means that we are extracting all the history before the step in the dataset , i.e. . We have corrected it as in the revision. To avoid the notation collision, we now use to make it different from that used in the bonus function.
We thank the reviewer again for your feedback and suggestions. Hopefully, our responses have addressed your main concerns. If so, we kindly ask that you consider raising the score of your evaluation. If you still have remaining questions, please let us know, and we will be very happy to provide further clarification.
I would like to thank the authors for clarifying my questions and for revising the paper. I also agree with other reviewers that there is no simulation or evaluation and it is certainly a weakness of the paper, but I would still like to increase my score based on the theoretical contribution of the paper.
We thank the reviewer very much for checking our response and kindly increasing the rating. Many thanks for your effort into this review process.
We thank the reviewer for the helpful comments, and would like to provide the following responses. The paper has been revised accordingly to reflect these responses, where the changes are highlighted in the blue color. We hope that these discussions can clarify the reviewer's concerns, and we will be more than happy to answer any further questions you may have.
Q1: The computation cost is still expensive with a dependency on how good the PSR is (the size of ). Could you elaborate more on the overall computation complexity and point out what is the bottleneck?
A1: The computational cost of our algorithm can be summarized as follows. Our algorithm requires calling MLE computation oracle once in every iteration (line 7 in Alg. 1). Calculating the bonus term (line 8 in Alg. 1) takes flops in iteration with the Sherman-Morrison formula. Finally, a quasipolynomial time complexity has been shown in [1] for planning (line 9 in Alg. 1). Therefore, the overall computation complexity is if MLE oracles are available. The computational bottleneck is therefore determined by the MLE oracle and the planning procedure.
The above MLE oracle and the planning procedure are what our algorithm shares in common with the existing confidence-set-based POMDP approaches. Beyond those, the existing confidence-set-based approaches also rely on an optimistic planning procedure that iterates over a large function class , which may require combinatorial search, leading to additional exponential computational complexity. In contrast, our bonus-based approach avoids such optimistic planning procedure, and thus is more advantageous for practical implementation.
[1] Golowich, et al., Planning in Observable POMDPs in Quasipolynomial Time.
[2] Liu, et al., Optimistic MLE—A Generic Model-based Algorithm for Partially Observable Sequential Decision Making, 2022.
Q2: I am not surprised that the algorithm and results are reward-free.
A2: Our algorithm and results are indeed reward-free by design. The main advantage of such a reward-free design over reward-based design is that it guarantees to output a near-accurate model for model-based algorithms. We can also modify Line 9 in Alg. 1 to let , so that the policy depends on as well as the reward information . We can show that such a design can also identify a near-optimal policy under the same sample complexity: The current proof stays almost the same, except that we remove the claim . However, it lacks the guarantee for the estimated model. This is because the optimistic policy now depends on the reward as well as the uncertainty level, which has less incentive to explore the trajectories generating low rewards. As a result, the accuracy of estimated model over those trajectories cannot be guaranteed.
Q3: Complicated notations.
A3: Thanks for the suggestion. We will try our best to simplify the notations and make our presentation as clear as possible in the revision.
Q4: I thought a major advantage of PSRs is to discard the dependency on the number of all possible history with size . However, this term still appears in your sample complexity with a dependency on . The dependency on the action and observation space is hidden in where . Could you explain how your algorithm improves and compares with the online learning algorithms without using PSRs?
A4: We would like to clarify that the term only appears in the logarithm factor, as suggested by , and . In fact, when we plug in , we have (See Proposition 3). Therefore, compared with previous PSR work [2] and other online learning algorithms without using PSRs [3], our algorithm improves the computation tractability using bonus-based design (see detailed explanation in the second paragraph of A1). Meanwhile, our sample complexity depends polynomially on , and not on , which is the same as previous works [1,2,3].
[1] Golowich, et al., Planning in Observable POMDPs in Quasipolynomial Time.
[2] Liu, et al., Optimistic MLE—A Generic Model-based Algorithm for Partially Observable Sequential Decision Making, 2022.
[3] Liu, et al., When Is Partially Observable Reinforcement Learning Not Scary?
In this paper, the authors introduced UCB exploration for predictive state representation. This is theory-oriented paper. With MLE oracle and the planning oracle, the authors provided computation and sample complexity for general sequential decision-making problem, which includes Markov decision processes (MDPs) and partially observable MDPs (POMDPs).
为何不给更高分
There are several issues should be addressed:
1, Polynomial computational-complexity relies on computation oracle, which should be clearly stated as "oracle-efficient".
2, Comparing to Optimistic-MLE, the UCB planning oracle is required additionally.
3, Comparing to Golowich, et al, in which no oracle is required, the current claim on computation complexity is unfair.
Considering this, the necessity of UCB upon optimistic MLE should be carefully discussed: the proposed algorithm requires both optimistic MLE oracle and UCB planner oracle, while optimistic MLE only requires MLE estimator.
为何不给更低分
The major confusing part of the paper lies in 1) , the motivation of using UCB upon optimistic MLE, and 2) the claim in "computation-efficient", other than that, the paper analysis is solid.
Accept (poster)