A Computationally Efficient Algorithm for Infinite-Horizon Average-Reward Linear MDPs
First computationally efficient algorithm for infinite-horizon average-reward linear MDPs.
摘要
评审与讨论
This paper tackles the open problem of proving a rate optimal regret bound in average reward infinite horizon linear mdp with a computationally efficient algorithm. The problem it is solved successfully, using a double loop structure inspired by hong et al. 2025 and a deviation controlled mechanism to make sure that the iterates produced by least square value iteration with different clipping thresholds could be bound by the difference of the clipping thresholds.
After rebuttal
My opinion remains positive. The assumption on bounded features and weights is clear now but I think it would be better if the authors make clear that finding features with the desired boundness properties requires computing a MVEE.
Please make sure to add the comparison with the failed attempt in Hong et al. 2025v1. I think this is very important for the placement of the paper in the literature.
给作者的问题
How does the Bellman optimality equation in Assumption A relates to structural assumptions often imposed in Average Reward infinite horizon problems such as Communicating/Weakly Communicating MDPs ? I think this should be discussed better in the manuscript before stating Assumption A.
Why in Assumption B is reasonable to assume that the reward weights are bounded by . I think that in case of very badly conditioned features the reward weights might need to be larger than to realize the reward function. Let imagine to have two features and , now to write the reward function the weight norm should be as large as . However, can be made arbitrary small so the quantity can become arbitrary larger that .
Moreover, the authors say that these assumptions are without loss of generality and refer to Wei at al. 2021 to see why this is the case.
Unfortunately Wei et al 2021 is only for the tabular setting so it does not contain information about this assumption that involves linear MDPs.
论据与证据
Claims are convincing
方法与评估标准
Rigorous proofs.
理论论述
I checked the correctness of all the theoretical statements. I found just a minor error at the end of page 15 in the equation at line 820 should be replaced with , otherwise the authors could not conclude that the first ferm equals .
实验设计与分析
see above
补充材料
No supplementary material available
与现有文献的关系
Yes, related to the previous attempt towards solving this problem. This is quite well done but I have a request for the author. please find it below.
To my knowledge, there was a previous attempt in the first version of the paper uploaded in arxiv that claimed to achieve the same results but without the double loop structure and therefore with a better time complexity. Unfortunately, the authors of Hong et al. 2025 claimed in their v2 that the previous result contains a mistake.
I think that the mistake in Hong et al 2025 is in the maximum over the past Q functions, this algorithmic technique makes the covering number of the state value functions space to be exponential in . This is indeed avoided by the current submission that replaces the maximum over past value functions with the double loop structure.
In my opinion, it would be helpful for the community if in the current submission, you could add a paragraph explaining the differences with Hong et al. 2025 v1 and in particular explaining why the current approach fixes the issue with the previous approach.
遗漏的重要参考文献
None
其他优缺点
Well written paper
其他意见或建议
Make clear that Algorithm 1 and 2 need to know both and , especially it seems very difficult to obtain an algorithm that works without knowing the value of in advance given the double loop structure. The author should make this limitation clearer.
Thank you very much for your thorough review and suggestions for improving the paper. Here are our responses.
Minor error in proof. Thank you very much for catching this. We will replace with for the equality to hold.
Relation to previous work. Thank you for the suggestion. We fully agree that detailed comparison with Hong 2025 v1 will benefit the community. A short explanation is that Hong 2025v1 that uses "max-pooling" of Q functions, which had an error, happened to allow computationally efficient clipping without additional algorithmic modification. Hong 2025v2, which fixes the error using double loop structure could no longer use the computational efficient clipping. Our submission introduces an algorithmic modification that allows for computationally efficient clipping under the double loop structure. We will mention the attempt made by Hong 2025v1 in Section 2.4, and add detailed comparison in the Appendix.
Knowledge of T, H. We will make it clear to the reader that is required by adding in the list of input parameters in the algorithm box, and mentioning it in the analysis when we set . We will add a remark following the Theorem statement that using doubling-trick, we can get away with the knowledge of at the cost of incurring additional factor of in the regret. As for the knowledge of , we will remind the reader after the Optimism lemma (Lemma 3.4) and after the main theorem statement that knowledge of an upper bound of span is required. As mentioned in our response to Reviewer CyKa, we will also mention that there is a recent work that relaxes this assumption for the tabular setting, and that we leave relaxing this assumption to future work.
Discussion on Bellman optimality equation. Thank you for the suggestion. We will add discussion on the relation to Communicating/Weakly Communicating MDPs and Ergodic assumption. Bellman optimality equation assumption is strictly weaker than Communicating/Weakly Communicating MDP which is strictly weaker than ergodic assumption. We will mention this fact with brief reasoning and citation.
On boundedness of reward vector. Note that the paper by Wei et al. 2021 "Learning infinite-horizon average-reward mdps with linear function approximation." discusses this in detail in Appendix A. (Based on your comment, we think you were referring to Wei et al. 2020. "Model-free reinforcement learning in infinite-horizon average-reward markov decision processes"). Their reasoning is as follows. Given any feature mapping with and a reward function in the form such that , there always exists an invertible matrix such that and , which enables us to work with and instead that satisfy and . They argue in their proof that such a transformation is the transformation that transforms the MVEE of the feature mappings into a unit ball. When citing their work, we will add which section of their paper to refer to (Appendix A), and briefly explain their argument.
Thanks a lot for your response ! I indeed did a mistake and I was referring to Wei et al. 2020 instead of Wei et al. 2021. That point is clear now but I think it would be better if the authors make clear that finding features with the desired boundness properties requires computing a MVEE.
Please make sure to add the comparison with the failed attempt in Hong et al. 2025v1. I think this is very important for the placement of the paper in the literature.
Thank you for the suggestion. We will mention the MVEE method for the transformation for ensuring boundedness, and mention that the computation of MVEE is required.
As for comparison with 2025v1. We do agree that full comparison with 2025v1, not just with 2025v2/3, is important. We will devote a section in appendix for this. Unfortunately, the comparison is too technical to describe fully in the rebuttal due to space and format constraints. Here is a brief summary:
- [modified notation a bit for better comparison with our submission] Single-loop algorithm (computationally inefficient) in 2025v1 has a value iteration structure that leads to a term in the regret bound. To match the subscript for bounding this term, 2025v1 introduces "max-pooling" value functions such that is increasing in . To avoid covering issue, they keep track of the index such that is increasing in . Also, they need to restart the algorithm periodically, otherwise the value functions will always be at all time steps. They employ the "doubling trick" that restarts algorithm when determinant of covariance doubles. However, they erroneously claim that holds. For this to make sense, should be in the current episode, enforcing to be chosen by max-pooling from the second time step in the episode and onward. However, taking max-pooling from the second time step will not allow bounding when is exactly the second time step of the episode, since is the function function at the first time step of the episode. Due to this "off-by-one" error, the analysis fails.
- Ignoring this off-by-one error and pretending that the analysis goes through, the analysis happens to allow computationally efficient algorithm with double-loop structure trivial.
- 2025v2 fixes the off-by-one error by introducing a new value iteration structure (backward induction instead of "forward induction") such that the subscript is matched to by design, so that no max-pooling is required. However, the analysis on the double-loop structure with computationally efficient algorithm fails, as is discussed in our submission in Section 3.2.
- Our submission introduces the new algorithm structure using the novel deviation-controlled value iteration structure, that allows bounding the difference by the difference in clipping thresholds, which is magically bounded using telescoping sum.
This paper proposed an algorithm for infinite-horizon average-reward reinforcement learning with linear function approximation. The main problem need to be solved is the computation issue arises from minimize the value function in large state space. To address this, the paper proposed a new clipping technique, and proved that this method achieves the same regret order while enjoy computational efficiency under the large state space setting.
给作者的问题
N/A
论据与证据
Yes. The paper clearly provided the computational complexity to support that the proposed algorithm runs in polynomial time.
方法与评估标准
Yes.
理论论述
I checked the proof in the main paper and appendix A, and do not find any issues.
实验设计与分析
N/A
补充材料
I do not fully check the supplementary material. It seems correct to me.
与现有文献的关系
N/A
遗漏的重要参考文献
N/A
其他优缺点
Strength: The explanation of the motivation and idea of why and how to do clipping is clear and easy to understand.
Weakness: The work lack some (at least) toy example to illustrate the effectiveness of the proposed algorithm.
其他意见或建议
N/A
Thank you for your review. Here is our response.
Simulation in toy setting. Thank you for the suggestion for running a simulation in a toy setting. We have confirmed that our algorithm runs in a tabular setting (i.e. linear MDP setting with feature vectors orthogonal) with average regret reasonably low when treating the bonus factor beta as a hyperparameter, as is usually done in UCB-type algorithms.
To our knowledge, no study to date has conducted a simulation under the linear MDP setting. Nevertheless, we agree that providing a simple simulation would be valuable to the community, and we will aim to include a simulation. One reason why simulation under linear MDP is absent in the literature is due to the complication of specifying the vector of measures that ensures is a valid probability measure. However, we believe we can get around the problem by restricting the feature vector to have L1-norm equal to 1, and each of the measures in are probability measures. We will aim to conduct experiments under the restricted linear MDP setting.
This paper proposes a computationally efficient algorithm for infinite-horizon average reward linear MDPs. The paper seeks to improve upon the previously proposed approach -LSVI-UCB by Hong et al'25. The main contribution is that the algorithm proposed in Hong et al.'25 requires to iterate over all the state-space to find the minimum value function for clipping, which might be computationally expensive, especially for linear MDP. Hong et al.'25 assumed some minimum oracle, this paper seeks to improve upon that. The paper proposes a novel idea where one only needs to iterate over the state encountered so far. The paper in particular focuses on bounding the error - . The paper significantly contributes in this direction.
##Post Rebuttal Update##
All my concerns have been resolved, and I am happy to accept this paper.
给作者的问题
- Can the authors summarize briefly the main technical novelties? in particular, how did they bound the deviation?
论据与证据
The proofs of the claims are provided.
方法与评估标准
The paper is theoretical in nature.
理论论述
The theoretical claims seem to be correct.
实验设计与分析
N/A.
补充材料
The reviewer has briefly glossed over the materials which seem to be correct.
与现有文献的关系
The reviewer does not have any major concern.
遗漏的重要参考文献
The reviewer does not have any major concern.
其他优缺点
Strengths:
-
The theoretical contributions are significant.
-
The paper is well-written.
Weakness:
-
The algorithm still needs to iterate over all the states encountered, which can be significant for the large time period.
-
Some simulations would be nice to see the difference in terms of computational time and the regret compared to the state-of-the-art approaches.
其他意见或建议
N/A
Thank you for your time for the review and for the suggestion for improving the paper.
Q1: main technical novelties. We will integrate the following summary in the introduction section and in Section 3.2 when introducing the method.
- The main technical novelty lies in designing a clipped value iteration algorithm that ensures each successively generated value function deviates from its predecessor by no more than the difference between their respective clipping thresholds.
- We show that a naive adaptation of previous work cannot control this deviation in the linear MDP setting.
- A natural workaround would be to clip each newly generated value function with a new threshold so that it deviates from the previously generated value functions by at most the difference in the thresholds. But doing so will make the value functions more and more complex successively, running into an issue when using covering argument for uniform concentration bound.
- To address this, we design a novel way for controlling the deviation without running into covering issues. Specifically, we "pool" latest value functions in such a way that the difference in successive "pooled" value functions is bounded by the the difference in the thresholds. The pooled value function has low complexity so that the function class that captures the pooled value function has low covering number.
I thank the authors for their responses.
Just one clarification question. This is regarding the comment made by the reviewer 3 Auz.
``To my knowledge, there was a previous attempt in the first version of the paper uploaded in arxiv that claimed to achieve the same results but without the double loop structure and therefore with a better time complexity. Unfortunately, the authors of Hong et al. 2025 claimed in their v2 that the previous result contains a mistake.
I think that the mistake in Hong et al 2025 is in the maximum over the past Q functions, this algorithmic technique makes the covering number of the state value functions space to be exponential in . This is indeed avoided by the current submission that replaces the maximum over past value functions with the double loop structure."
You have responded to this comment. Can you please clarify and elaborate what the mistake was? How are you overcoming it in this version? Thanks,
Here is a brief explanation regarding the comment made by reviewer 3Auz:
- Hong et al. 2025v1 uses single-loop structure and uses the scheme of taking maximum of past value functions. It is known that taking maximum of past value functions in linear MDP leads to a covering issue because the function becomes more and more complex and the covering number of the function class that captures such functions become exponential in the number of time steps. As a workaround, they use a scheme of keeping track of the time index that gives the maximum value, i.e., where is some reference time step (exact definition they use is a bit different). However, the analysis has an off-by-one error.
- Hong et al. 2025v2 fixes the problem by introducing a double-loop structure, at the cost of incurring additional factor of in computational complexity. However, the double-loop structure requires computing the minimum of value function over the entire state space when clipping.
- Our submission introduces a novel algorithm structure that allows for computationally efficient clipping.
For more detailed explanation, please see our response to Reviewer 3Auz.
In this paper, the authors have studied the reinforcement learning algorithm for linear MDPs in an infinite-horizon average-reward setting. Previous works approximate the average reward by the discounted one and employ a clipping-based value iteration method. However, it requires the computation of minimum of the value function over the state space. This may be computationally prohibitive. Towards that, in this paper, an efficient clipping technique is introduced for value iteration algorithm. This requires computation of minimum value function over states visited by the algorithm. It has been established that the proposed scheme demonstrates the same regret bound as that of the previous work with a substantial drop in computational complexity which is independent of the size of the state space.
update after rebuttal
The response addresses my major concerns. I have raised my score to 3.
给作者的问题
My comments are as follows:
- What if the underlying Markov chain is irreducible under various policies? In the worst case, the algorithm can still visit all the states and hence, the computational burden may still be high. I expected some simulations too in this direction to verify the efficiency of the proposed scheme.
- Why do we need to assume that ? I guess removing this assumption also does not create any problem.
- It is assume that is known the learner. The authors stated that this assumption can be relaxed to the knowledge of an upper bound on . However, in this case regret will scale with the upper bound. Knowledge of or an upper bound on it can both be difficult to obtain. Also, choice of a loose upper bound can result in poor regret. Is there a way in which such a requirement of knowledge can be avoided completely?
- Please describe what do you mean by covering number.
- Lemma 3.4 holds when . How do one guarantee that?
- In Theorem 3.6, we need choose . However, typically is not in designer’s hand. This makes the algorithm depart from reality.
- It is stated that the computational complexity is independent of the size of the states space. However, is related to the size of the state space. Why is the regret large when is large? As becomes close to the size of the state space, regret should be less. Please elaborate
论据与证据
Yes
方法与评估标准
Yes
理论论述
I have checked the proof , but not in great detail.
实验设计与分析
Experimental results are absent in the paper.
补充材料
Yes
与现有文献的关系
In this paper, the authors have studied the reinforcement learning algorithm for linear MDPs in an infinite-horizon average-reward setting. It has been established that the proposed scheme demonstrates the same regret bound as that of the previous work with a substantial drop in computational complexity which is independent of the size of the state space.
遗漏的重要参考文献
No
其他优缺点
The paper is well-written and contains some interesting ideas. The problem considered by the authors are relevant and interesting. However, the solution is based on some strong assumptions such as knowledge of . Some results need clarification and lacks intuition.
其他意见或建议
Not applicable.
Thank you for your time for the detailed review. Your feedback will help us in improving our paper. Here are our responses to your questions.
Q1: Computational issue & Simulation. You are correct to note that, in the large state space regime, the number of unique states visited over time steps can be very large, potentially with no repeats. However, our algorithm’s computational complexity does not scale with the size of the entire state space, since we do not enumerate all states. Instead, in the worst case, where states never repeat, the complexity scales with . Regarding simulation, due to the theoretical nature of the linear MDP assumption, designing a concrete simulation environment is challenging and, to our knowledge, no simulation has been performed in the linear MDP setting in the literature. we agree that providing a simple simulation would be valuable to the community, and we will aim to include a simulation under a restricted linear MDP setting described in our response to Reviewer w193.
Q2: Why assume reward to be bounded. Your intuition that the assumption can be removed is correct. And in fact, it is this intuition that It is standard in the literature to assume without loss of generality. When the actual reward lines in , say, then we can scale it to , analyze the regret bound under the new reward, and then multiply the final regret bound by to account for the scaling. To avoid carrying the constant around in our derivations, we assumed . We will add a clarifying note on this point in our final version of the paper.
Q3, Q5: Assumption on the knowledge of span. Thank you for bringing this up. Indeed, requiring prior knowledge of is a limitation of our approach. Whether sample-efficient learning is possible without knowing the span in advance has been an open question for a long time, and many existing works rely on this assumption. A recent result by Boone et al. [1] shows for the first time that it can be avoided in the tabular setting. However, extending their technique to the linear MDP setting is non-trivial and will likely require a major breakthrough. We leave this to future work. We will add a remark on this assumption and cite [1] in the final version of the paper.
Q4: Covering number. Thank you for the suggestion. In the paragraph following Lemma 3.3, we will describe the general covering argument that uses epslion-net for covering the function class for getting uniform concentration bound and briefly describe -covering number of a function class and how the log covering number scales the concentration bound.
Q6: Setting discounting factor. The original problem the designer is trying to solve is the average-reward setting. We are proposing to approximate the average-reward setting by a discounted setting with a particular discounting factor. The designer will work with the discounted setting with discounting factor set to the proposed value. So, it is in the hand of the designer to set the discounting factor. We will make it clear in the beginning of Section 3 that the discounting factor is set by the designer to approximate the average-reward setting.
Q7: Size of the state space and the dimension. Thank you for the question. When we say "size of the state space", we mean the number of the states in the state space. In the tabular setting, we do not assume any structure in the MDP that allows generalizing to unseen states, and the computational and statistical complexity both scale with the number of states. However, in the linear MDP setting, we assume a structure in the MDP that allows for the generalization through the feature mapping that maps state-action pairs to a -dimensional feature vector. Due to the feature representation, conceptually, we no longer need to learn about all the states individually. Rather, we only need to learn about "directions", which allows for computational and statistical efficiency. Specifically, the computational complexity and the regret bound scales with , not the size of the state space. With this intuition, we can see that if is large, then there is more "directions" to explore, making both computational complexity and regret large. We will add a summary of this discussion when defining the linear MDP setting in Section 2.3.
[1] Achieving Tractable Minimax Optimal Regret in Average Reward MDPs. NeurIPS 2024.
I thank the authors for their detailed response. However, I am still concerned regarding two aspects:
- Knowledge of : Can this assumption be relaxed?
- The authors stated that when is large, then there is more directions to explore, making both computational complexity and regret large. It is not clear to me. Please explain using two extreme cases, and ? Since there is no approximation involved in the second case, won’t the regret due to approximation be less in the second case?
Another clarification question: Since the original problem considers an infinite horizon average-reward setting, how do you choose to set ? Will any choice of large be fine to represent the infinite horizon setting?
Q1 Relaxing the assumption of the knowledge of span in linear MDPs is an open problem. The authors of "Achieving Tractable Minimax Optimal Regret in Average Reward MDPs, NeurIPS 2024" relaxes the assumption in tabular setting by using a subroutine that estimates the optimal bias function to work without the knowledge of . Their approach is based on the observation that, when an optimal policy is run, the bias difference is roughly the difference in values when starting from the state and . They use the average reward collected in subtrajectories that start at and end at to estimate . It is unclear how to generalize this idea to (linear) function approximation setting where the state space is large, since their idea relies on tracking each pair of to estimate , which can be sample inefficient in function approximation setting.
Designing an algorithm (even one that is computationally inefficient) for linear MDP without the knowledge of the span is an interesting and challenging open problem, and would be a good topic for a future standalone paper.
Q2 In general, in the learning setting where the transition probability is unknown, the algorithm design centers around estimating the unknown or a quantity that depends on the unknown . Typically, in value-iteration based algorithm, which is the algorithm class we use, the quantity of interest is . The regret bound depends on how sample efficiently we can estimate since the regret bound scales with the width of the confidence bound. (in general, regret of UCB type of algorithm scales with sum of confidence bounds). In linear MDP setting, we exploit the fact that for some , and use linear regression to estimate , which gives the concentration bound that scales linearly with (see Lemma 3.3). This suggests that the regret will scale with . To gain more intuition, consider the following two extreme cases.
- . In this case, the transition is the same for all pairs. Hence, we don't have to collect transition data for all pairs separately to be able to estimate for each . We can "pool" data for all pairs of to estimate . Similarly, estimation of can be done by sample average of .
- . In this case the transition can be arbitrary for each pair. Hence, unlike the case, we cannot "pool" data, and requires data for each pair. Similarly, estimating requires data for each pair. Intuitively, this leads to a concentration bound that scales with and hence regret scales with .
We will make it clearer that the concentration bound in Lemma 3.3 scales with , and that the concentration bound eventually enters the regret bound.
Regarding . We first clarify the distinction between the problem setting ("infinite-horizon") and the performance guarantee of an algorithm ("finite-time regret bound"). The notion "infinite-horizon average-reward" describes the problem setting where the criterion for evaluating a policy is through the infinite-horizon average-reward. When analyzing the performance of an algorithm, one typically uses "finite-time regret bound" with a certain time and show a bound on the -step where regret in each step is against the optimal average-reward. Our guarantee allows a designer to choose to achieve -step regret.
As you hinted, a designer may want a single algorithm that guarantees -step regret for any . Algorithms with this guarantee is called anytime algorithms. There is a known reduction called "doubling trick" that uses algorithms with -step regret to design an anytime algorithm. The reduction is as follows.
Suppose we have access to an algorithm that guarantees -step regret for each . In our paper, can be obtained by choosing . Then, we can design an anytime algorithm by running for steps for . That is, run for 1 time step, then run for 2 time steps, then run for 4 time steps, etc. For any , such an algorithm guarantees -step regret bound of
by Cauchy-Schwarz, since .
We realize that discussing anytime algorithm would greatly improve clarity on our guarantee. We will incorporate this discussion into the paper. Also, we think parameterizing the algorithm with instead of may be clearer for this discussion.
The paper studies reinforcement learning in infinite-horizon average-reward settings with linear MDPs. Comparing to the prior work, this work's algorithm relaxes the condition and only requires the algorithm to know the minimum of the value functions over the set of the states visited by the algorithm instead of the entire state space. Thus the work improves prior work in terms of computation complexity.
Reviewers in general are positive about the paper and the authors rebutal also successfully addressed some of the concerns from the reviewers.
I think some suggestions from the reviewers are indeed very good and can make the paper stronger. Please make sure to include the following discussions in the paper: (1) discussion about the comparison to the previous work from Hong et al. 2025v1, and discuss the issues related to the covering number and how the techiniques in this paper handles this issue; (2) discuss how to ensure the boundness assumptions using MVEE and the computation cost from MVEE. I think adding these discussions would make the paper more complete.