Cascading Reinforcement Learning
摘要
评审与讨论
The paper tackles the Cascading Bandits problem over a Markov Decision Process, where time is divided into episode of fixed length , and the attractiveness of items (which determined the expected state transitions and rewards) is dependent on a state at the current step in the episode. At each step in the episode, the system gets to observe the current state and proposed a ranked list of items and the environment provides a click on one of the items according to the cascading click model. For this problem, the paper provides a method for computing the optimal offline strategy when all distributions are known and an introduces the CascadingVI algorithm along with a regret bound of , where is the episode length, is the number of states, the number of episodes and is the number of individual items available to rank. The performance of the proposed algorithm is compared numerically with a baseline algorithm produced by adapting an existing algorithm to the combinatorial space, referred to here as AdaptRM.
优点
- generalizes the cascading bandit problem to MDPs
- provides numerical experiments, even though, in my opinion the baseline is too weak
- provides theoretical guarantees and proofs of correctness
- generally well written, despite the convoluted algorithms I could navigate my way around the paper.
缺点
- The significance of the contribution is not substantial enough to justify acceptance into the venue in my opinion. One of the main contributions is providing an algorithm that does not scale in complexity (both sample or computational) with the space of all possible rankings. I believe we have well-known recipes for how such algorithms can be formulated since the papers on Cascading Bandits from 2015 (Kveton et. al, Combes et al.). Providing a solution scaling with the number of items is expected in my opinion and not a surprising contribution (we know that estimating individual CTRs is enough to construct optimal rankings, we don't need to keep statistics on each individual ranking). This also leads me to believe the baseline used in the numerical experiments is weak.
- The contribution related to the generalization to MDPs is not opening a significant amount of new doors. In my opinion, this generalization does not provide further theoretical insights and has limited additional practical relevance.
- The experimental section use a setting that is too simplistic (very few items and states) and a weak baseline, in my opinion, thus not being representative enough of what we can expect from this algorithm in practice.
问题
On line 8 in Algorithm 2, what is the complexity of computing ?
Regarding BestPerm, to me it feels like the biggest hurdle is the computational complexity of computing the value functions , which BestPerm assumes is an input. How does your proposed algorithm navigate this difficulty?
I would also like to see more intuition regarding the exploration bonus , in particular: how is the optimism of connected to the uncertainty in the estimates of state transition probabilities ? The algorithm is already fairly convoluted and it is hard to see how we can construct optimistic estimates of the value functions in light of estimation noise in the state transition probabilities. I believe it is very valuable to clearly articulate such insights in the main body of the paper.
Thank you very much for your time and effort in reviewing our paper! We have incorporated your comments in the revised version, and highlighted our revision in red font.
1. Contribution of our Generalization to Cascading MDPs
We emphasize that our generalization to cascading MDPs is very different from prior cascading bandit works [Kveton et al. 2015; Combes et al. 2015], since we face a unique challenge on how to efficiently compute the optimal permutation. To tackle this challenge, we make use of special properties of the objective function for cascading MDPs, and develop a novel and efficient oracle BestPerm using a dynamic programming approach for combinatorial optimization.
Specifically, prior cascading bandit works [Kveton et al. 2015; Combes et al. 2015] consider finding the optimal subset of items as follows:
This objective function only depends on the attraction probabilities of items . Thus, prior works only need to select items with the maximum attraction probabilities, which do not involve computational difficulty.
In contrast, our cascading MDP problem considers finding the optimal ordered subset of items as follows (for step ):
Our objective function depends on both the attraction probabilities and the expected future rewards .
This requires us to solve a combinatorial optimization problem for each as follows:
where and
(in the online RL process, we can have the estimates of , and ). This optimization involves two attributes of items and , instead of just depending on , which is far more challenging to efficiently compute than Eq. (i).
To overcome this computational difficulty, we first analyze the special properties of , including the fact that sorting items in descending order of gives the optimal permutation (Section 4.1). Then, based on these properties, we design an efficient oracle BestPerm to solve Eq. (ii), building upon a novel dynamic programming approach for finding the best subset from sorted items (Section 4.2):
Here denotes the optimal objective value that can be achieved by selecting items from items which are sorted in descending order of . Furthermore, we also provide rigorous analyses for the properties of and the correctness of oracle BestPerm (Lemmas 1 and 2).
To sum up, our generalization to cascading MDPs has the following novelties: (1) This generalization introduces a significant computational challenge on how to efficiently find the optimal permutation. (2) To resolve this challenge, we analyze the special properties of the objective function for cascading MDPs, and design an efficient oracle BestPerm based on a novel dynamic programming. Rigorous proofs are provided for the properties of the objective function and the correctness of oracle BestPerm. (3) Building upon oracle BestPerm and the carefully-designed exploration bonuses for estimates, we develop provably computationally-efficient and sample-efficient algorithms for cascading RL.
2. Experiment
Following the reviewer's suggestion, we conduct experiments on a larger-scale real-world dataset MovieLens [Harper & Konstan 2015] (also used in prior works [Zong et al. 2016; Vial et al. 2022]) with more baselines. We present preliminary experimental results here (due to time limit), and will certainly include more results for larger numbers of states and items in our revision.
The MovieLens dataset contains 25 million ratings on a 5-star scale for 62000 movies by 162000 users. We regard each user as a state, and each movie as an item. For each user-movie pair, we scale the rating to [0,1] and regard it as the attraction probability. The reward of each user-movie pair is set to 1. For each user-movie pair which has a rating no lower than 4.5 stars, we set the transition probability to this state (user) itself as 0.9, and that to other states (users) as . For each user-movie pair which has a rating lower than 4.5 stars, we set the transition probability to all states (users) as . We use a subset of data from MovieLens, and set , , , , , (the number of items) and (the number of item lists).
We compare our algorithm CascadingVI with three baselines, i.e., CascadingVI-Oracle, CascadingVI-Bonus and AdaptVI. Specifically, CascadingVI-Oracle is an ablated version of CascadingVI which replaces the efficient oracle BestPerm by a naive exhaustive search. CascadingVI-Bonus is an ablated variant of CascadingVI which replaces the variance-aware exploration bonus by a variance-unaware bonus. AdaptVI adapts the classic RL algorithm [Zanette & Brunskill 2019] to the combinatorial action space, which maintains the estimates for all rather than . The following table shows the cumulative regrets and running times of these algorithms. (CascadingVI-Oracle, which does not use our efficient computational oracle, is slow and still running in some instances with large . We will update its results once it finishes.)
| Regret; Running time | CascadingVI (ours) | CascadingVI-Oracle | CascadingVI-Bonus | AdaptVI |
|---|---|---|---|---|
| 2781.48; 1194.29s | 4097.56; 21809.87s | 8503.45; 1099.32s | 29874.84; 28066.90s | |
| 4630.11; 2291.00s | 6485.36; 71190.59s | 13436.38; 2144.96s | 30694.40; 53214.95s | |
| 6446.51; 2770.94s | 9950.22; 145794.64s | 17472.21; 2731.21s | 34761.91; 81135.42s | |
| 8283.60; 5216.00s | 13170.39; 196125.72s | 21868.54; 5066.27s | 42805.06; 95235.55s | |
| 11412.85; 7334.13s | Still running | 27738.31; 7256.94s | 52505.85; 188589.60s | |
| 16825.65; 18628.83s | Still running | 37606.95; 17829.40s | 56045.01; 319837.83s |
As shown in the above table, our algorithm CascadingVI achieves the lowest regret and a fast running time. CascadingVI-Oracle has a comparative regret performance to CascadingVI, but suffers a much higher running time, which demonstrates the power of our oracle BestPerm in computation. CascadingVI-Bonus attains a similar running time as CascadingVI, but has a worse regret (CascadingVI-Bonus looks slightly faster because computing a variance-unaware exploration bonus is simpler). This corroborates the effectiveness of our variance-aware exploration bonus in enhancing sample efficiency. AdaptVI suffers a very high regret and running time, since it learns the information of all permutations independently and its statistical and computational complexities depend on .
We have also presented the figures of these experimental results in our revised version.
3. Replies to Questions
3.1 Computational Complexity of
The complexity of computing in Algorithm 2 is . (In this response, we use to denote the subscript of due to the formula display issue of OpenReview.)
3.2 Computation of the Value Function
Algorithm CascadingVI performs optimistic value iteration by backward iteration for step . At each step , we already have the estimates and , and then we can compute the optimistic estimate for as (Line 10 of Algorithm 2). The complexity of computing is . After that, for each state , we call oracle BestPerm with to efficiently calculate the optimal value and the optimal permutation, i.e., (Line 11 of Algorithm 2).
Actually, the biggest computational difficulty of BestPerm is how to efficiently solve the combinatorial optimization Eq. (ii) in our Reply 1. BestPerm takes advantage of the special properties of the objective function for cascading MDPs, and adopts a novel dynamic programming approach to solve it (see our Reply 1).
3.3 Intuition of the Exploration Bonus
Our construction of follows that in prior RL work [Zanette & Brunskill 2019], which guarantees the optimistic estimation for . Below we describe the intuition behind (Lemmas 7-8 in Appendix C.2).
According to Bernstern's inequality, we have that with high probability,
Hence, the right-hand side of Eq. (iii) is the Bernstern-type exploration bonus of . However, the algorithm does not know the exact , and thus we want to use the estimate of instead of itself.
Following the analysis in [Zanette & Brunskill 2019], we have that if for any , then with high probability,
Plugging the above inequality into Eq. (iii), we have that with high probability,
Thus, if for any , then is an optimistic estimate of .
In the following, we discuss how to guarantee the optimism of , and the proof for pessimism is similar (Lemmas 14-15 in Appendix C.2).
First, we prove that satisfies the monotonicity property, i.e., if element-wise and the items in are ranked in descending order of , then .
Then, we prove the optimism of by induction. For , it trivially holds that element-wise, and thus is an optimistic estimate of . Using the monotonicity property of , the fact that the items in the optimal permutation are ranked in descending order of weights, and the optimism of , we can prove that the output value of oracle BestPerm is an optimistic estimate of . For , by induction and similarly using the optimism of and , we can prove that is an optimistic estimate of for all .
We have added more explanation in Section 5.1 in our revision.
Thank you for the clarifications. The intuition behind the exploration bonus is more clear to me now.
Thank you for introducing the real world experiments. These baselines are more meaningful in my opinion.
- Contribution of our Generalization to Cascading MDPs
Thank you for the detailed response. Please allow to further articulate my point of view on why I believe this extension's significance is limited relative to what I believe is already known from the vanilla cascading bandits (and please correct me where I am missing something).
Dynamic programming is already an established method for computing optimal policies in an MDP. The proposed novelty here is the reduction of computation time for this algorithm by exploiting the cascading structure, i.e. the fact that "basic" actions are rankings (and hence the number of actual items in our ranking pool is much much smaller than that of actions). But in my opinion, it is already known that we don't need to evaluate every individual ranking, since it is sufficient to know the expected value of each individual item in this state (probability of click and expected future value) to construct the optimal ranking in a given state and avoid complexities in . Therefore I am not convinced this result (expanding the setting to the RL setting and providing an oracle algorithm avoiding computational complexity) represents a sufficiently significant addition to the literature.
I have one other standing curiosity for which I wonder if you can provide any insights: Would a naive algorithm that simply takes the top items as sorted by (the expected return of item weighted by its chance to be clicked when inspected) and further sorts these by (items with the highest expected return at the top, regardless of their attractiveness) produces optimal rankings?
Any insights you can offer are appreciated (I am not expecting a proof, just wondering if you have any counter-example at hand or whether you've seen this fail in your experiments).
Thank you for giving us an opportunity to interact with you!
1. Significance of Our Contribution
Indeed it is already known that one only needs to evaluate each item, instead of each ranking, in classic cascading (combinatorial) bandits [Kveton et al. 2015]. However, the objective function in cascading bandits
only involves one attribute of items, i.e., the attraction probability . This objective function completely depends on , which is obvious to compute. They only need to select items with the maximum attraction probabilities .
However, when generalizing to RL, the objective function at each step is
which involves two attributes of items, i.e., the attraction probability and the expected current and future reward (here given the state and step ). Eq. (ii) itself is a complex combinatorial optimization problem even when one only evaluates the attributes of individual items and ignores the recursion over step , and there is no obvious monotone rule with respect to , or .
To solve Eq. (ii), we design an efficient computational oracle BestPerm, using a novel dynamic programming rule based on the cascading feature of and :
where
denotes the optimal objective value achieved by selecting items from items which are sorted in descending order of . The intuition behind our dynamic programming rule is that: for a sorted item sequence , if we do not select , then the optimal objective value is equal to that achieved by selecting items from , i.e., . Otherwise, if we select , then the optimal objective value is equal to multiplying the optimal value achieved by selecting items from , i.e., , due to the cascading feature of the objective function.
We note that the "dynamic programming" we mentioned in our contribution statements refers to the dynamic programming Eq. (iv) that we design to solve the combinatorial optimization Eq. (ii), which iterates over the indices of items and the number of selected items . We were not referring to the well-known MDP dynamic programming that iterates over step .
In other words, our dynamic programming rule (Eq. (iv)) and oracle BestPerm provide an efficient computational solution to the combinatorial optimization Eq. (ii), which one needs to solve at each step in cascading RL even when only using the attributes of individual items. This is a significant contribution and novel to the combinatorial bandit/RL literature.
2. Counter-example for the Naive Algorithm
A naive algorithm that first sorts items by and further sorts them by cannot find the optimal ranking.
Consider a counter-example as follows: There are 10 items , and we want to find the optimal ranking with the maximum cardinality . The attraction probabilities , weights and their products of items are shown in the following table.
| 0.1 | 0.2 | 0.3 | 0.4 | 0.5 | 0.6 | 0.7 | 0.8 | 0.9 | 1 | |
| 1 | 0.9 | 0.8 | 0.7 | 0.6 | 0.5 | 0.4 | 0.3 | 0.2 | 0.1 | |
| 0.1 | 0.18 | 0.24 | 0.28 | 0.3 | 0.3 | 0.28 | 0.24 | 0.18 | 0.1 |
The naive algorithm the reviewer mentioned will output , while our computational oracle BestPerm will output according to our dynamic programming rule Eq. (iv).
The objective values of and are
This demonstrates that the output of the naive algorithm is not the optimal solution.
Thank you for the clarifications and especially for the insight on the difficulty of computing the best permutation. My concerns about the novelty are now somewhat eased. I recommend that in future revisions of the paper you put more emphasis on the difficulties solved by BestPerm and presenting this contrast between such a naive policy (picking the top items by expected value ) and the actual optimal ranking. The first few reads, I did not pick up on the nuances of this difficulty, my intuition being that it is fairly low hanging fruit.
Thank you very much for your time in interacting with us and insightful suggestions! We will certainly include more discussion on the difficulties solved by BestPerm and the differences between a naive policy and the actual optimal ranking in our future revision.
This paper proposes an extension of the cascading bandits framework to a more general reinforcement learning framework. The authors thereby attempt to model user sessions or historical user behavior and their impact on the click behavior and payoff. While the action space (i.e., combination of items) is combinatorial, the feedback is item-wise so that attraction and transition probabilities can be estimated efficiently. This allows the authors to design algorithms with non-trivial regret and sample complexity guarantees. In particular, the authors rely on monotonicity properties to design an efficient oracle. Finally, the authors support their theoretical findings through experiments which support the efficiency of the proposed algorithms.
优点
- Firstly, I think that the studied problem is interesting and the presentation of the paper clear.
- The extension of the cascading bandit model to an RL formulation is highly non-trivial and the paper contains novel contributions including the RL formulation itself and an interesting algorithm design which relies on properties of the value function.
- The authors do a good job explaining their algorithms and provide intuition for their results.
- The necessity of adopting standard RL algorithms to the proposed cascading model is highlighted in the paper several times and the authors thereby provide proper justification for the proposed algorithms.
缺点
- I am not fully convinced of the practicality of this model. Firstly, historical user behavior can be modeled as part of the context in contextual bandit frameworks. Moreover, "artificially" creating states appears fairly cumbersome compared to the contextual bandit structure and it is also not entirely clear how such states would be defined in practice. However, I could be convinced otherwise.
问题
-
Could you further explain why contextual approaches are insufficient and what the merit of the RL formulation is in contrast to contextual bandits or recommendation approaches based on context trees? In the RL framework it is important what states the user transitions to, as some states may have the potential to yield higher rewards than others. Do you think that this is realistic in practice?
-
It would be great if you could go into a bit more detail in your related work section and more clearly highlight the differences to prior models. For example, in contrast to the classical cascading bandits, the order of items also matters in your case as you usually would like the item with largest reward in state s to be clicked.
Minor things:
- In the 4th contribution, I found the statement about " sufficiently large" slightly confusing. It only becomes clearer later in Theorem 2 when the full bound is stated. Maybe there is a way to state this less conusingly in the introduction.
- Typo in the 3rd paragraph of Section 7: "Regarding the best policy identification objective" instead of "Regarding the regret minimization objective".
Thank you very much for your time in reviewing our paper and your positive comments! We have revised our paper according to your suggestions. Our revision is highlighted in orange font.
1. Merit of the RL Formulation Compared to Contextual Bandits
We agree with the reviewer that contextual bandits can also be used to formulate historical user behaviors as part of contexts. We think that RL is more suitable for long-term reward maximization, since it considers state transition and potential future rewards in decision making.
For example, consider a video recommendation scenario. The recommendation system encounters a user who is interested in funny videos and TV series. Given this context, contextual bandits mainly recommend videos which can maximize the reward in the current context. Hence, contextual bandits may recommend funny videos which usually give higher instantaneous utilities. In contrast, RL recommends videos which maximize the expected cumulative (long-term) reward, which also considers the potential rewards generated by future states. Thus, in this case, RL may recommend TV series videos. Although TV series videos may have lower instantaneous utilities, once the user is attracted to some of them, the user can enter a high-reward successor state, i.e., he/she is obsessed with it and keeps watching subsequent videos of this TV series. Therefore, using the RL formulation can obtain higher rewards in the long term.
2. Related Works
We detail the differences of our cascading RL model from prior works below.
[Kveton et al. 2015; Cheung et al. 2019; Zhong et al. 2021; Vial et al. 2022] study the cascading bandit model. In their model, there is no state (context), and the reward generated by each item (if clicked) is one. Thus, the order of selected items does not matter, and they only need to select items with the maximum attraction probabilities. [Zong et al. 2016; Li & Zhang 2018] consider the contextual cascading bandit problem. In their problem, the agent first observes the context (i.e., the feature vector of each item) at each timestep, and the attraction probability of each item is the inner-product of its feature vector and a global parameter. The order of selected items still does not matter, and they need to select items with the maximum attraction probabilities in the current context. [Li et al. 2016] investigate contextual cascading bandits with position discount factors and a general reward function. In their model, the order of selected items matters, and they assume to have access to an oracle that can output the optimal ordered subset of items for the general reward function.
Different from the above works, our cascading RL formulation further considers state transition, and the attraction probabilities and rewards of items depend on states. Thus, we need to put the items with higher rewards in the current state in the front. In addition, we require to both maximize the attraction probabilities of selected items, and optimize the potential rewards from future states. Furthermore, instead of assuming access to an oracle, we design an efficient oracle to find the optimal ordered subset of items.
3. Replies to Minor Things
Thank you for pointing out the unclear sentence and typo!
We have revised the statement in the fourth contribution to "CascadingBPI is optimal up to a factor of when " to avoid confusion, and fixed the typo.
References:
[1] Branislav Kveton, Csaba Szepesvari, Zheng Wen, and Azin Ashkan. Cascading bandits: Learning
to rank in the cascade model. ICML 2015.
[2] Wang Chi Cheung, Vincent Tan, and Zixin Zhong. A thompson sampling algorithm for cascading
bandits. AISTATS
2019.
[3] Zixin Zhong, Wang Chi Chueng, and Vincent YF Tan. Thompson sampling algorithms for cascading
bandits. JMLR 2021.
[4] Daniel Vial, Sujay Sanghavi, Sanjay Shakkottai, and R Srikant. Minimax regret for cascading bandits.
NeurIPS 2022.
[5] Shi Zong, Hao Ni, Kenny Sung, Nan Rosemary Ke, Zheng Wen, and Branislav Kveton. Cascading
bandits for large-scale recommendation problems. UAI 2016.
[6] Shuai Li and Shengyu Zhang. Online clustering of contextual cascading bandits. AAAI 2018.
[7] Shuai Li, Baoxiang Wang, Shengyu Zhang, and Wei Chen. Contextual combinatorial cascading
bandits. ICML 2016.
Thank you for responding to my comments.
Regarding your first answer. While it does make sense to me that you can model a "customer journey" like this and that this can be useful, defining sensible states seems quite hard and could simply boil down to something like a context tree (which would be simpler than the MDP framework). However, I still appreciate the advantages of "forward" planning and trying to transition to rewarding states, despite the challenges in designing such states in practice. I do not wish to change my original assessment and I am still in favour of accepting the paper.
Thank you very much for your appreciation for our work and raising an insightful future direction! We will further explore how to design sensible states for better practicability in our future work.
The paper presents a new framework for reinforcement learning (RL) called cascading RL, which builds upon the concept of cascading bandits but incorporates state information into the decision-making process. This framework is aimed at applications such as personalized recommendation systems and online advertising, where cascading items are displayed to users one by one.
Key challenges addressed include computational difficulty due to the combinatorial nature of actions and ensuring sample efficiency without relying on the exponential number of potential actions. To overcome these, the authors developed an efficient oracle called BestPerm, which uses dynamic programming to optimize item selection. They also introduced two RL algorithms: CascadingVI for regret minimization and CascadingBPI for sample complexity, both of which utilize BestPerm to achieve polynomial regret and sample complexity.
CascadingVI achieves near-optimal regret matching a known lower bound. CascadingBPI offers efficient computation and sample complexity, reaching near-optimal performance.
In summary, the paper contributes a new cascading RL framework that efficiently handles the computational and sample complexity challenges of stateful item selection, supported by theoretical guarantees and demonstrated effectiveness through experiments.
优点
- As far as I know, this would be the first theory RL work with cascading actions. The formulation is sound.
- The paper proposes an efficient optimization oracle for cascading RL, providing its correctness (it would be better to show it in the main text rather than the appendix).
- Although it is not groundbreaking, the paper provides both regret analysis and sample complexity analysis.
缺点
- It seems that the techniques used for regret analysis and sample complexity are largely borrowed from the previous literature. I wonder if there are sufficient technical challenges once you know how to solve Problem (2).
- The readability of the algorithms can be improved. e.g., Update rule of and in Line 6 of Algorithm 2, Line break in Line 8 of Algorithm 2, etc.
问题
Where does the gap appear? Can you elaborate on this? Any possible direction to carve off this given that the tabular RL methods with a single action can achieve minimax optimality?
Thank you very much for your time and effort in reviewing our paper! We have revised our paper according to your comments. Our revision is highlighted in orange font.
1. Presentation on the Correctness of Oracle BestPerm
Thank you for your helpful suggestion! We have moved the correctness guarantee of oracle BestPerm to the main text (Lemma 2) in our revision.
2. Technical Challenges
Below we elaborate the technical challenges and novelties in addition to the oracle for solving Problem Eq. (2).
-
Even if we know how to solve Problem Eq. (2), in the online RL process, we only have the estimates of attraction probabilities, transition distributions and value functions. Then, how to guarantee the optimism of the output value by oracle BestPerm with optimistic estimated inputs is a crucial challenge. To handle this challenge, we prove the monotonicity property of the objective function for cascading RL by induction, leveraging the fact that the items in the optimal permutation are ranked in descending order of (Lemma 14). This monotonicity property guarantees the optimism of our estimation under the use of oracle BestPerm with optimistic estimates (Lemma 15).
-
Since the objective function for cascading RL involves two attributes of items, i.e., attraction probabilities and the expected future rewards , how to construct exploration bonuses for them is an important step in algorithm design. We build the exploration bonuses for and separately for , rather than adding an overall exploration bonus for each . This exploration bonus design enables us to achieve regret and sample complexity bounds that depend only on the number of items , instead of the number of item lists .
-
Another critical challenge is how to achieve the optimal regret bound when our problem degenerates to cascading bandits [Vial et al. 2022] given that cascading RL has a more complicated state-transition structure. To tackle this challenge, we use a variance-aware exploration bonus , which can adaptively shrink when the attraction probability is small. In regret analysis, when performing summation over , this variance-awareness saves a factor of , and enables our regret bound to match the optimal result in cascading bandits [Vial et al. 2022].
3. Readability
Thank you for your valuable suggestions! We have revised Algorithm 2 according to your comments to improve readability in our revision.
4. The Gap
The gap appears because we construct the exploration bonuses for and separately, which leads to an additional factor when summing up and in regret analysis.
Below we elaborate this in detail. In regret analysis, using value difference lemma, the regret decomposition contains the following term
Since we construct exploration bonuses for and separately, in Eq. (i) above, the regret contains a product , while the regret for classic RL only contains the term . To handle this product, we bound it by . Here the term directly leads to a regret.
A straightforward idea to improve the gap is to combine the attraction probability and transition distribution as an integrated transition distribution for each item list , and construct an overall exploration bonus for . However, this strategy forces us to maintain the exploration bonus for each , which will incur an additional dependency on in the regret bound and is computationally inefficient.
We think that a more fine-grained analysis for bounding or a more advancing bonus construction approach may be helpful to close the gap. We plan to further investigate it in our future work.
Reference:
Daniel Vial, Sujay Sanghavi, Sanjay Shakkottai, and R Srikant. Minimax regret for cascading bandits.
NeurIPS 2022.
Thanks for the responses. I will keep my positive score.
Thank you very much again for your positive comments and time in reviewing our paper!
This paper studies cascading reinforcement learning, which is a natural extension of cascading bandits to the episodic reinforcement learning setting. In particular, this paper has
-
proposed the cascading reinforcement learning framework;
-
developed a computationally efficient algorithm to solve the "cascading MDPs" (i.e. the cascading reinforcement learning problems with known models). The key ideas are summarized in the BestPerm algorithm (Algorithm 1).
-
developed a learning algorithm for the cumulative regret minimization setting, which is referred to as CascadingVI (Algorithm 2). A regret bound is also established (Theorem 1). This paper has also discussed the tightness of this regret bound.
-
also developed a learning algorithm for the best policy identification setting, which is referred to as CascadingBPI. A sample complexity bound is established (Theorem 2). This paper has also discussed the tightness of this regret bound.
-
demonstrated preliminary experiment results (Section 7).
优点
In general, I think this paper is a strong theoretical paper, for the following reasons:
-
This paper studies a natural extension of a well-studied problem. Moreover, as summarized above, the contributions of this paper are clear.
-
The main results of this paper, summarized in Section 4 (efficient oracle), Section 5 (regret minimization), and Section 6 (best policy identification), are interesting and non-trivial. In particular, the regret bound in Theorem 1 and the sample complexity bound in Theorem 2 are non-trivial. Moreover, this paper has also discussed the tightness of these two bounds by comparing with existing lower bounds. Both bounds are near-optimal.
-
The paper is well-written in general, and is easy to read.
缺点
-
I am wondering if the developed algorithms are useful for practical recommendation problems. The reason is that, this paper considers a tabular setting, thus the developed regret bound and the sample complexity bound depend on the number of states . However, my understanding is that for most practical recommendation problems, will be exponentially large. Might the authors identify a setting where is not too large and hence the proposed algorithms are practical?
-
Currently the experiment results are very limited. In particular, this paper has only demonstrated experiment results in small-scale synthetic problems. I think experiment results on large-scale practical problems will further strengthen this paper.
问题
Please address the weaknesses above.
Thank you very much for reviewing our paper and your positive comments! We have incorporated your suggestions in the revised version, and highlighted our revision in teal font.
1. The Scenario Where Is Not Too Large
Consider a video recommendation scenario where the videos are categorized into multiple types, e.g., news, education, entertainment and movies. Here each item is a video, and each action (item list) is a list of videos. We can define each state as the types of the last one or two videos that the user just watched, which represents the recent preference and focus of the user. Then, the recommendation system suggests a list of videos according to the recent viewing record of the user. After the user chooses a new video to watch, the environment transitions to a next state which represents the user's latest interest. In such scenarios, the number of states is polynomial in the number of types of videos, which is not too large. Therefore, our cascading RL formulation and algorithms are useful and practical for these applications.
2. Experiment
Following the reviewer's suggestion, we conduct experiments on a larger-scale real-world dataset MovieLens [Harper & Konstan 2015] (also used in prior works [Zong et al. 2016; Vial et al. 2022]) with more baselines. We present preliminary experimental results here (due to time limit), and will certainly include more results for larger numbers of states and items in our revision.
The MovieLens dataset contains 25 million ratings on a 5-star scale for 62000 movies by 162000 users. We regard each user as a state, and each movie as an item. For each user-movie pair, we scale the rating to [0,1] and regard it as the attraction probability. The reward of each user-movie pair is set to 1. For each user-movie pair which has a rating no lower than 4.5 stars, we set the transition probability to this state (user) itself as 0.9, and that to other states (users) as . For each user-movie pair which has a rating lower than 4.5 stars, we set the transition probability to all states (users) as . We use a subset of data from MovieLens, and set , , , , , (the number of items) and (the number of item lists).
We compare our algorithm CascadingVI with three baselines, i.e., CascadingVI-Oracle, CascadingVI-Bonus and AdaptVI. Specifically, CascadingVI-Oracle is an ablated version of CascadingVI which replaces the efficient oracle BestPerm by a naive exhaustive search. CascadingVI-Bonus is an ablated variant of CascadingVI which replaces the variance-aware exploration bonus by a variance-unaware bonus. AdaptVI adapts the classic RL algorithm [Zanette & Brunskill 2019] to the combinatorial action space, which maintains the estimates for all rather than . The following table shows the cumulative regrets and running times of these algorithms. (CascadingVI-Oracle, which does not use our efficient computational oracle, is slow and still running in some instances with large . We will update its results once it finishes.)
| Regret; Running time | CascadingVI (ours) | CascadingVI-Oracle | CascadingVI-Bonus | AdaptVI |
|---|---|---|---|---|
| 2781.48; 1194.29s | 4097.56; 21809.87s | 8503.45; 1099.32s | 29874.84; 28066.90s | |
| 4630.11; 2291.00s | 6485.36; 71190.59s | 13436.38; 2144.96s | 30694.40; 53214.95s | |
| 6446.51; 2770.94s | 9950.22; 145794.64s | 17472.21; 2731.21s | 34761.91; 81135.42s | |
| 8283.60; 5216.00s | 13170.39; 196125.72s | 21868.54; 5066.27s | 42805.06; 95235.55s | |
| 11412.85; 7334.13s | Still running | 27738.31; 7256.94s | 52505.85; 188589.60s | |
| 16825.65; 18628.83s | Still running | 37606.95; 17829.40s | 56045.01; 319837.83s |
As shown in the above table, our algorithm CascadingVI achieves the lowest regret and a fast running time. CascadingVI-Oracle has a comparative regret performance to CascadingVI, but suffers a much higher running time, which demonstrates the power of our oracle BestPerm in computation. CascadingVI-Bonus attains a similar running time as CascadingVI, but has a worse regret (CascadingVI-Bonus looks slightly faster because computing a variance-unaware exploration bonus is simpler). This corroborates the effectiveness of our variance-aware exploration bonus in enhancing sample efficiency. AdaptVI suffers a very high regret and running time, since it learns the information of all permutations independently and its statistical and computational complexities depend on .
We have also presented the figures of these experimental results in our revised version.
Thanks for the experiment results! They will definitely further strengthen the paper!
Thank you very much again for your time in reviewing our paper and reading our rebuttal!
Thanks for the response! It has addressed one of my concerns.
This paper generalizes online learning to rank in the cascade model to reinforcement learning (RL). The main difference from prior works in the bandit setting is that the agent also optimizes for future states. The paper was a clear accept after the author-reviewer discussion. Some of the initial concerns, that the model does not apply to more realistic problems and that the technical contribution is limited, were addressed in the rebuttal.
为何不给更高分
Although this work can have a reasonably large follow-up, the portion of the RL community interested in ranking, or vice versa, is small.
为何不给更低分
The cascade model is one of the two most popular models for learning to rank. Therefore, this work can have a reasonably large follow-up, where more modern and realistic models for learning to rank are extended to RL. This could have a major impact on sequential search and recommendations.
Accept (spotlight)