Beyond task diversity: provable representation transfer for sequential multitask linear bandits
We study lifelong learning in linear bandits, where a learner interacts with a sequence of linear bandit tasks that share a low-rank representation without requiring the Task Diversity assumption.
摘要
评审与讨论
This paper extends the existing work on multi-task linear bandits where the task parameters lie in a low rank subspace. Specifically, this paper assumes -dimensional linear bandits with parameters lie in an -dimensional subspace. For such a setting, the classical approaches yield a regret linear in and . This paper then proposed a new algorithm that improves the regret with mild assumptions of well-conditioned ellipsoids of the tasks and bounded task parameters. The algorithm is based on a reduction to a bandit online subspace selection problem. The key idea of the proposed approach is bi-level. At the lower level, the algorithm performs either an meta-exploration algorithm or a meta-exploitation algorithm for each of the tasks. At the upper level, the learner either chooses exploration/exploitation or chooses a subspace to use for exploitation. The effectiveness of the proposed algorithm is verified empirically on synthetic adversarial settings.
优点
Significance: This paper proposed a new algorithm for multi-task linear bandit with provable low regret without making strong assumptions on task parameters. The sequential setting studied is more challenging than parallel settings due to the slow revelation of the underlining subspace. Compared with a few related existing works, this paper does not need assumptions such as task diversity.
Quality: this paper is in general good quality. The algorithm is well described. The theoretical part of the work is clearly formulated. Thought the experiments are synthetic, the result still verifies the effectiveness of the proposed algorithm.
Clearity: The design of the algorithm and intuition behind the design is clearly explained.
缺点
There is no major weakness in the paper. I do have some questions regarding the following:
1. This paper assumes knowledge of the parameters corresponding to the task number, the dimension, and the number of rounds, which is not ideal.
2. [Originality] The algorithmic design and the guarantee is closely related to a a few works, including the PEGE by [Rusmevichientong and Tsitsiklis 2010] (related to lemma 3 in this paper), the analysis in Yang et al [2020] (related to lemma 4), EWA algorithm (related to lemma 5). Therefore, the technical novelty of the proposed algorithm and analysis is not very obvious to me.
问题
Please see weaknesses.
Also, I am curious about the following:
-
This paper assumes knowledge to a few hyper-parameters. Among these, I wonder whether the assumption on the prior knowledge of the parameter can be relaxed with certain adaptive design. Could the author elaborate on why knowledge is necessary for the current algorithm to work?
-
Is there a corollary or simple extension that can readily extend the current result to the parallel setting?
局限性
Yes.
- Originality: The algorithmic design and the guarantee is closely related to a a few works, including the PEGE by [Rusmevichientong and Tsitsiklis 2010] (related to lemma 3 in this paper), the analysis in Yang et al [2020] (related to lemma 4), EWA algorithm (related to lemma 5). Therefore, the technical novelty of the proposed algorithm and analysis is not very obvious to me.
As discussed in the Introduction and Table 1, Table 2, our work fills the gap of the sequential representation transfer for linear bandit, without the Task Diversity assumption. Before our work, no algorithms could provably beat the naive individual single-task baseline. The absence of task diversity assumption makes meta-learning the task representation matrix nontrivial.
Our proposal designs a bi-level approach for this problem and shows that, under mild assumptions, we can efficiently learn the underlying subspace in an online sense and maintain the correct uncertainty over possible experts when the environment acts adversarially (i.e. not revealing the full underlying subspace till the end). We believe our reduction to online subspace selection is new to sequential representation transfer and may benefit other meta-learning applications such as in RL or supervised learning.
I thank you authors for their efforts in the rebuttal. I will take all the rebuttals into consideration during the discussion period between reviewers.
The paper studies lifelong learning in linear bandits and designs a two-level algorithm with provable low regret without task diversity assumptions. The paper assumes the tasks share a low-rank representation and provides a regret upper bound.
优点
The paper uses a two-level approach to solve this problem, which is novel. A theoretical guarantee is given with numerical experiments performed.
缺点
There are some statements in the theorems that seem wrong and the upper bound provided does not seems tight. See Questions for details.
问题
- In line 109, for a matrix , isn't not unique? This should be defined as a set or stated as one of the matrices.
- The Assumption 2 is stated as linear bandits. I didn't get the reason for this title.
- In line 3 in Algorithm 1, it should be
- The statement in Lemma 4 seems problematic. As stated in 1), the matrix is not unique, hence the RHS is a matrix-dependent bound. Thus, the condition is also dependent on the choice of the matrix. This discussion is missing in the paper. This will also affect the definitions in (2) and (3), and the following results.
- In line 173, the definition of the expert set is in the Appendix, which makes the paper not self-contained.
- What does the EWA algorithm do in Algorithm 3? Which part of the results is directly from the EWA algorithm?
- Overall, the paper provides the regret bounds of order in equation (5), times of interactions. Compared to the baseline, which is , I didn't see a clear improvement in the regret/ Firstly, the regret scales linear with when fix , which is known to be suboptimal in linear bandit. At least some truncation can be done in the first tasks to get rid of the linear term. This term cannot be hidden from the abstract from my point of view. Secondly, to achieve similar upper bounds, one needs and , which means the algorithm can only learn for a finite time and the tightness discussion is missing the paragraph Comparison with lower bounds with only a conjecture.
- The author answer Yes to the Question about Open access to data and code in the Checklist, but the code is not released during the submission. And I am curious how Figure 1B is generated.
- The labels on the y-axis in the figures need to be fixed.
局限性
The regret upper bounds has a hugh gap to the lower bound.
- ``In line 109, for a matrix , isn’t not unique? This should be defined as a set or stated as one of the matrices.''
We thank the reviewer for pointing out this impreciseness. Indeed, the choice of is not unique according to our writing in the submitted version.
To make the definition well-defined, it suffices to fix any orthonormal basis ; one may break ties in dictionary order. We will clarify this in the final version of the paper.
Indeed, the specific choice of does not affect the correctness of our proof -- e.g., are always the same regardless
of the specific choice of , as long as its columns form a orthonormal basis of :
For any two choices of
(denoted by ,
, respectively)
there exists orthonormal such that
.
Therefore,
To show that all valid choices of are equivalent up to a orthogonal transformation, we have:
Let be a -dimensional subspace of . Let be matrices whose columns form an orthonormal basis of . Then, there exists an orthogonal matrix such that .
Proof: Since is a basis of , there exists some such that . Since , is an orthogonal matrix.
- ``The statement in Lemma 4 seems problematic. As stated in 1), the matrix is not unique, hence the RHS is a matrix-dependent bound.
Thus, the condition is also dependent on the choice of the matrix. This discussion is missing in the paper. This will also affect the definitions in (2) and (3), and the following results.''
As we mention in our previous item, the value of is invariant to the specific choice of matrix
. Still, we agree that this can be made clearer in the main paper, so we will make the necessary edits in the final version.
- ``What does the EWA algorithm do in Algorithm 3? Which part of the results is directly from the EWA algorithm?''
The algorithm uses Exponentially Weighted Algorithm (EWA) in line 8 of Algorithm 3. The intuition of using EWA is to choose subspaces online such that their linear spans capture 's over tasks . The idea is to use the feedback from the meta-exploration tasks to update the weights of all possible experts from the expert set. Since this expert set -cover the true , the EWA guarantee would ensure that BOSS would efficiently learn the closest expert to the true while maintaining the correct uncertainty over all possible experts when the subspace dimensions are not fully revealed, thus, neatly dealing with the non-Task Diversity setting.
Specifically speaking, the EWA's guarantee is used in the proof of Lemma 5, line 455. We will provide a full description of the algorithm with weight updates in the Appendix in the final version.
- ``Overall, the paper provides the regret bounds of order in equation (5), times of interactions. Compared to the baseline, which is , I didn’t see a clear improvement in the regret. ... the regret scales linear with when fix N, which is known to be suboptimal in linear bandit. At least some truncation can be done in the first tasks to get rid of the linear term. This term cannot be hidden from the abstract from my point of view''
We agree that the term cannot be ignored and will add it to the abstract. Note that we are studying the multi-task problem; thus, the interesting regime is when the number of tasks is large enough to allow useful transfer learning across tasks. The part of the regret can be understood as the learner ``sacrifices'' a small number of tasks to learn transferable knowledge. This last term is of lower order than when .
- ``The definition of the expert set is in the Appendix, which makes the paper not self-contained''
We agree. At first, we moved these definitions to the appendix to not distract the reader. We realized that they may be required for the main paper to be self-contained, so we will make the necessary edits.
- ``... the code is not released during the submission. And I am curious how Figure 1B is generated.''
Correct. We want to add some additional experiment results. The code is in https://anonymous.4open.science/r/Serena-C5F1.
For Fig 1b, since we use a simulated environment, we have access to the , the subspace spanned by thus far and the estimated subspace of the learners for each task. Thus, we plot this figure on this information.
- ``The Assumption 2 is stated as linear bandits. I didn’t get the reason for this title.''
We will change it to ``linear bandits with ellipsoid action sets'' in the final version.
Thanks for clarifying some of the points I over-interpreted. I would like to raise the score to 5 but reduce the confidence to 3. The main weakness in my opinion is still the linear term in . Consider any real-world tasks (such as search systems or recommender systems like movie recommendations), there will be a large volume of queries and will be arbitrarily larger than poly.
The paper proposes a strategy for the multi task linear bandit problem, where the tasks are assumed to share a low rank representation, which is ought to be learnt via a new meta learning strategy, where meta exploration and exploitation needs to be balanced. The low rank representation is learnt via optimizing a newly constructed loss function over an epsilon-net. The tasks appear sequentially to the learner and the authors explicitly do not make any assumption on the task diversity of their setting. The authors provide a meta regret bound for their algorithms and synthetic data results.
优点
The paper is well written and easy to follow
Original idea for solving the multi-task/meta learning setting with low rank representation
The assumption on the action set is weaker than in comparable works (constant action set vs iid action set)
缺点
The EWA algorithm for the meta learning procedure should not just be referenced but actually appear in the paper (or at least in the supplementary)
The presentation of the actual meta learning could be made clearer. For that matter, definition 7 and 8 should be moved to the main paper so it becomes clearer what the uniform distributions D_n are supposed to be.
The experiments would benefit if there were another baseline to compare to (Cella et al or Bilaj et al as they appear in the related work section)
The plot labels of figure 1 b) and c) should have latex style writing
问题
In Figure 1a) it looks like the Boss algorithm performs worse after a new dimension for the subspace is incremented at n= 2501, does that mean that the dimension of the subspace was falsely estimated?
at n=1 the the low rank should be estimated as m=1 as well, which makes it hard to minimize , (Since the n+1th task parameter might be outside the subspace selected for the first n tasks), Intuitively, the algorithm should work best if at least m task parameters are already bias-less estimated for a m-dimensional low rank structure. Could you explain how you mitigate this issue?
Shouldn't a bound on the term or a term that showcases the estimation error on the representation, appear in the final regret?
Does p need to be constant? after enough tasks explored, p could gradually vanish. I see in the final bound that , which is still constant.
局限性
Aside from the assumptions on their setting made, I did not see any limitations explicitly addressed.
- ``Shouldn’t a bound on the term or a term that showcases the estimation error on the representation, appear in the final regret?''
Yes. Indeed, our regret bound has an implicit dependence on - see Eq. (2).
This is a key quantity like the you mentioned.
- ``At n=1 the the low rank should be estimated as m=1 as well, which makes it hard to minimize , (Since the n+1th task parameter might be outside the subspace selected for the first n tasks), Intuitively, the algorithm should work best if at least m task parameters are already bias-less estimated for a m-dimensional low rank structure. Could you explain how you mitigate this issue?''
If we understand correctly, the reviewer was referring to that the algorithm needs to overcome additional challenges without the task diversity assumption. Without task diversity, we cannot hope to obtain estimate 's that converge to the underlying . Instead, we use online learning with randomized meta-exploration (using the learned 's to estimate the underlying representation ) to ensure that our learned 's can capture 's for most task , in an average sense.
- ``In Figure 1a) it looks like the Boss algorithm performs worse after a new dimension for the subspace is incremented at n= 2501, does that mean that the dimension of the subspace was falsely estimated?''
Our estimator is always in , so the subspace dimension estimate is always . If the reviewer meant ``the newly revealed direction of the subspace was falsely estimated'', we agree.
BOSS uses the estimated to efficiently estimate . When a new dimension is revealed at n=2501, it takes a while for BOSS to have an accurate estimation with respect to , the subspace spanned by all shown so far. Fig 1b and 1c show that if the expert set covers the underlying (green plot), BOSS can adapt and learn very fast.
- ``The presentation of the actual meta learning could be made clearer. The EWA algorithm for the meta learning procedure should not just be referenced but actually appear in the paper''
Thank you for your suggestion. We will include the full EWA algorithm in the appendix in the final version.
Thank you for your detailed answers to my questions and concerns. I will take them into consideration for the reviewers discussion period.
This paper addresses the problem of representation transfer in a sequential multi-task linear bandit problem. The main objective in the paper is to remove the task diversity assumption made in Qin et al (2022), which places a constraint on any subsequence of tasks observed. The paper proposes an algorithm that for each task performs exploration with a constant probability, and uses the exponential weighted average algorithm to choose the exploration probability over a large set of arm vectors. The paper proves an upper bound on the cumulative regret without making the task diversity assumption and demonstrate the performance in an experiment.
优点
- The paper presents the results in an engaging and clear manner and the algorithms proposed are intuitive.
- Removal of the task diversity assumption appears to be significant advance.
缺点
- The problem setting emphasizes the sequential nature of the tasks. Specifically, the algorithm needs to collect all samples from task before moving on to , i.e., the tasks cannot be interleaved. It would be helpful to outline some situations where this aspect of the problem setup is important.
- The algorithm needs to know the value of to choose the exploration probability. The proposed algorithm performs better than the "individual task baseline" only under a specific parameter regimes (see line 242).
- The theoretically valid size of the set of experts is extremely large for reasonable parameter values (lines 408 and 284). Is that one of the reasons why BOSS only performs better when we have very large number of tasks?
- There is a gap to the known lower bound, which is suggested to be due to the task diversity assumption not being met.
问题
- If possible, it would be nice to demonstrate the impact of the task diversity assumption. Specifically, could we construct a simulated environment where the assumption is not met by a significantly large fraction of subsequences, and the performance of BOSS is better than SeqRepL Qin et al (2022) ? Maybe your experiment already demonstrates that but I missed it. Could you clarify?
局限性
NA
- ``There is a gap to the known lower bound, which is suggested to be due to the task diversity assumption not being met.''
Before our work, no regret bounds that are were known for Sequential representation transfer multi-task linear bandits without Task Diversity assumption.
In addition, we don't know if our upper bound can be improved. Even with task diversity, the upper and lower bounds of Qin et al. do not exactly match (see Table 1)
- ``The theoretically valid size of the set of experts is extremely large for reasonable parameter values, is that one of the reasons why BOSS only performs better when we have very large number of tasks?''
Yes. directly affects the term in the regret bound, and will subsequently affect the term in the regret bound. If we can make this term smaller, we can broaden the parameter regimes when our regret bound is better than the baseline. We are also working on a method that is more computationally efficient in the future.
The reason why we need large is explained in the global answer above.
- ``It would be nice to demonstrate the impact of the task diversity assumption''
Our experiments in the submission are designed without the Task Diversity assumption, as shown in the caption of Fig 1 (a new dimension of the underlying subspace is revealed at tasks 1, 2500, and 3500). We also add some extra experiment results to highlight how BOSS outperforms SeqRepL when the parameters length is larger () in the attached PDF file of the global answer. Furthermore, we also add another set of experiments when the Task Diversity assumption is satisfied, where SeqRepL outperforms BOSS as expected.
The first figure is the experiment results without the Task Diversity assumption and the second is with Task Diversity. Different from the main paper, we ensure that highlights how badly SeqRepL performs when the Task Diversity assumption is violated. We also made some small changes to the hyper-parameters, as shown in the figures' captions.
Thank you for the response, I have increased my score.
As the author-reviewer discussion period is coming to a close, we kindly ask for your review of our response. This ensures that any additional questions or feedback you may have can be addressed before the discussion period ends. Thank you for your time!
We thank the reviewers for their insightful feedback.
In this work, we introduce the first provable representation transfer algorithm for sequential linear bandits without relying on the task diversity assumption. Unlike in parallel settings or sequential settings that assume task diversity, the learner now has to address the in meta-exploration, which involves carefully deciding when to acquire more information on the low-dimensional representation.
We are encouraged that the reviewers found our work clear (R1, R2, R4), intuitive (R1, R4), and novel (R2, R3). We are also glad that the reviewers recognized the significance of removing the task diversity assumption (R1, R2, R4) in exchange for some mild assumptions (R2) within the more challenging sequential setting (R4).
In addition, our theoretical guarantee and experiment are also appreciated (R1, R3, R4). We address some common questions below and the specific questions in each reviewer's section.
- ``The proposed algorithm performs better than the ”individual task baseline” only under a specific parameter regimes (R1, R4), and , which means the algorithm can only learn for a finite time (R3). [Does] BOSS only performs better [than the individual single-task baseline] when we have very large number of tasks? (R1)''
Before our work, no regret bounds better than individual single-task baseline (i.e. ) were known for this setting.
Our main motivation for this paper was to show that sequential representation transfer is indeed possible without the task diversity assumption. Proving this opens the floodgates for developing more sample-efficient algorithms.
We can rephrase the parameter regime where our regret bound is better than the individual single-task baseline as: and . So, indeed, BOSS performs better when is large. As mentioned in the paper, we think that broadening the parameter regime for improvement is an important problem.
If we interpret R3's concern correctly, being capped at is a limitation of our paper. We agree -- this is due to the existence of the burn-in term (when , this is greater than ). Removing this burn-in term is an interesting open problem.
- ``The algorithm needs to know the value of . Can this be avoided?'' (R1, R4)
. The requirement for knowledge of can be relaxed to knowing an upper bound of . Removing this knowledge requires a change of approach, such as low-rank matrix optimization, as in Cella et al. 2023 or additional assumption, as in Bilaj et al. 2024. Cella et al. 2023 is in the parallel setting, which is not applicable here, and Bilaj et al. 2024's guarantee can be as large as as discussed in line 91.
The requirement for and is not too bad since the action space is given to us ahead of time, so is known, and can be known after finishing the first task.
Using an adaptive design to not be dependent on is likely to be possible with the doubling-trick. The details are in the next question.
- ``Can we use an adaptive design (R4)? Can gradually vanishes (R2)?''
. We think achieving adaptivity in can be done using a doubling trick.
Specifically, in phase , run our algorithm with the assumption that there are total tasks in this phase. The modified algorithm has a meta-regret guarantee that is within a constant of the algorithm that knows . This implicitly gives an adaptive setting of meta-exploration probability that is decaying over time.
- ``Can our results translated to the interleaved (R1) or Parallel (R4) setting?''
Our algorithm is designed to tackle the unique challenges in sequential transfer without the task diversity assumption, and so applying it to the parallel setting would result in losing the intended benefits.
We don't think that our result can be easily adapted to the parallel setting and achieve a better guarantee than Hu et al. 2021.
The interleaved setting can be seen as a unified model for both the Sequential and Parallel settings. Since these two are fundamentally different, investigating the interleaved setting is left for future work.
- ``The labels on the y-axis in the figures need to be fixed (R2, R3)''
We will correct it as you suggest.
This paper addresses the problem of representation transfer in a sequential multi-task linear bandit problem. The main objective in the paper is to remove the task diversity assumption made in Qin et al (2022), which places a constraint on any subsequence of tasks observed. The paper proposes an algorithm that for each task performs exploration with a constant probability, and uses the exponential weighted average algorithm to choose the exploration probability over a large set of arm vectors. The paper proves an upper bound on the cumulative regret without making the task diversity assumption and demonstrate the performance in an experiment.
All reviewers believe this paper is above the bar of acceptance. The AC agrees and recommends acceptance.