PaperHub
6.2
/10
Poster5 位审稿人
最低6最高7标准差0.4
6
6
7
6
6
2.8
置信度
正确性3.0
贡献度3.2
表达2.4
NeurIPS 2024

On the Minimax Regret for Contextual Linear Bandits and Multi-Armed Bandits with Expert Advice

OpenReviewPDF
提交: 2024-05-15更新: 2025-01-16

摘要

This paper examines two extensions of multi-armed bandit problems: multi-armed bandits with expert advice and contextual linear bandits. For the former problem, multi-armed bandits with expert advice, the previously known best upper and lower bounds have been $O(\sqrt{KT \log \frac{N}{K} })$ and $\Omega( \sqrt{KT \frac{ \log N }{\log K }} )$, respectively. Here, $K$, $N$, and $T$ represent the numbers of arms, experts, and rounds, respectively. We provide a lower bound of $\Omega( \sqrt{KT \log \frac{N}{K}} )$ for the setup in which the player chooses an expert before observing the advices in each round. For the latter problem, contextual linear bandits, we provide an algorithm that achieves $O ( \sqrt{d T \log ( K \min\{ 1, \frac{S}{d} \} )} )$ together with a matching lower bound, where $d$ and $S$ represent the dimensionality of feature vectors and the size of the context space, respectively.
关键词
contextual linear banditmulti-armed bandit with expert advice

评审与讨论

审稿意见
6

This paper studies the problem of MAB with expert advice and the contextual linear MAB. Specificially, the authors proposed a lower bound of Ω(KTlogNK)\Omega(\sqrt{KT\log\frac{N}{K}}), improving upon previously known lower bound Ω(KTlogNlogK)\Omega(\sqrt{KT\frac{\log N}{\log K}}) and showing that the previous upper bound O(KTlogNK)O(\sqrt{KT\log\frac{N}{K}}) is tight. For the second problem, the authors design an algorithm based on FTRL with Tsallis entropy regularizer and achieve O(dTlogK(min(1,S/d)))O(\sqrt{dT \log K(\min(1,S/d))}) regret bound. A matching lower bound is also proven (under certain condition of KK) based on an extension of the previous lower bound constructed for linear bandits.

优点

  • The paper bridges the gap between the lower and the upper bound for MAB with expert advice, solving an open problem in this field.
  • While it is not technically hard to show that a regret minimization algorithm can be transformed to a best expert identification algorithm, the reduction with a specific construction of the expert problem instance leads to a matching lower bound, which is interesting to me.
  • While both the algorithm and the lower bound construction for contextual linear case are kind of standard from a technical perspective, the matching regret bound is good to know.

缺点

While I do not have much concern for the MAB with expert advice problem, for the contextual linear bandit problem, one of my concern is that the algorithm only works for the finite context case, which is very restrictive in real applications. In addition, the implementation of the algorithm requires the knowledge of the context distribution, which is also a limitation of the algorithm. As for the regret bound, while the leading term matches the lower bound, the optimality of the additonal term is unclear. Specifically, in the second bound, LL may also be related to TT (e.g. TcT^{-c}), making the additional term dominant.

问题

  • Whether Algorithm 1 can be extended to infinite / unknown context case?
  • Whether the additional term in the upper bound for contextual linear bandits is unavoidable?

局限性

See "weakness" and "questions".

作者回复

one of my concern is that the algorithm only works for the finite context case, which is very restrictive in real applications.

The proposed algorithm (Algorithm 1) can also work for the infinite context case, in which it enjoys the second regret upper bound in Corollary 1 (line 222): RT=O(dTlogK+λ0T/(dlogK))R_T = O \left( \sqrt{dT \log K} + \lambda_0 \sqrt{T / (d \log K)} \right). In fact, this regret upper bound does not include S=XS = |\mathcal{X}| or LL, and the value of g(x)g(x) is not required to define η(x)\eta(x) in showing this second upper bound. Hence, we do not require Assumption 1 (line 192) to show this bound.

In the infinite context case, however, further challenges regarding the computational complexity and practicality of the algorithm should be noted. For example, we need to compute V(pt)V(p_t) (defined in line 195) in the algorithm as it appears in the definition of θ^t\hat{\theta}_t in (3), which tend to be computationally expensive, depending on the computational model and the setup of distributions. We also require Assumption 2 (line 196) as well.

In addition, the implementation of the algorithm requires the knowledge of the context distribution, which is also a limitation of the algorithm.

We agree with this comment. Relaxing this assumption is an important future research direction. We believe that the technique of Matrix Geometric Resampling [NO20] can be used to relax the assumptions that the context distribution is known, and to consider settings where any number of samples from the distribution can be accessed. However, as long as this is employed, an extra logT\sqrt{\log T} factor seems inevitable. To further relax the assumption, [LWZ23] have proposed algorithms that do not even require access to a simulator from which the learner can draw samples from the context distribution. For such a setup, there is no known algorithm that achieve a (logK)(\log K)-dependent bound or that avoid extra logT\sqrt{\log T} factors, to our knowledge.

As for the regret bound, while the leading term matches the lower bound, the optimality of the additonal term is unclear. Specifically, in the second bound, LL may also be related to TT (e.g. TcT^{-c}), making the additional term dominant.

The parameter LL is defined on the basis of the context distribution, independent of TT, in Assumption 1 and does not grow with TT after fixing the problem instance. Thus, after fixing the context distribution arbitrarily, the additional term will not be dominant in a regime in which TT is sufficiently large. However, conversely, if we consider the regime of considering the worst-case context distribution after fixing TT, then the additional term can be dominant, as the reviewer suggested.

We here note that the additional term in the second bound of Corollary 1 is the minimum of λ0TdlogK\lambda_0 \sqrt{\frac{T}{d \log K}} and λ0Lβ/αT1β\lambda_0 L^{\beta/\alpha} \sqrt{T^{1-\beta}}, and hence large LL is not necessarily problematic. However, the parameter λ0\lambda_0, which is defined in Assumption 2 on the basis of the context distribution and exploration policy p0p_0, can be arbitrarily large depending on the context distribution, which might make the additional term dominant.

Whether Algorithm 1 can be extended to infinite / unknown context case?

Please refer to the response above.

Whether the additional term in the upper bound for contextual linear bandits is unavoidable?

This is an open question at this time to our knowledge. However, if the main term can be O(d2TlogT)O(\sqrt{d^2 T \log T}) instead of O(dTlog(Kmin1,S/d))O(\sqrt{d T \log (K \min \\{ 1, S/d \\}})), then we can remove the terms regarding λ\lambda and LL, as shown in Theorem 4 of [LWZ23].

Reference

  • [LWZ23]H. Liu, C.-Y. Wei, and J. Zimmert. Bypassing the simulator: Near-optimal adversarial linear contextual bandits. NeurIPS2023.
  • [NO20] G. Neu and J. Olkhovskaya. Efficient and robust algorithms for adversarial linear contextual bandits. COLT2020.
审稿意见
6

This paper studies the problem of bandit learning with expert advice. The main contributions are two refined regret bounds: (1) A matching lower regret bound of Ω(KTlog(N/K))\Omega(\sqrt{KT\log(N/K)}) for multi-armed bandit problem; (2) Θ(dTlog(K,min{1,S/d}))\Theta(\sqrt{dT\log(K,\min\{1,S/d\})}) regret bound for contextual linear bandits.

优点

This authors prove matching regret bounds for bandit problem with expert advice, which is significant enough for an acceptance.

缺点

My concern is about the writing. After reading the paper, I think the high-level intuitions of the construction for lower bounds could be clarified in one or two pages.

I also have a question about the proof of Lemma 3. In equation (15), a lemma in the bandit algorithm book is introduced to derive regret bounds. As far as I can see, the lemma only works for fixed X0X_0. However, in the learning process ptp_t might depends on the historical contexts X0,X1,X2,...,Xt1X_0,X_1,X_2,...,X_{t-1}. I am confused about the usage of the lemma and hope for some explanations.

问题

Please find my question in the comments above.

局限性

Yes.

作者回复

I also have a question about the proof of Lemma 3. In equation (15), a lemma in the bandit algorithm book is introduced to derive regret bounds. As far as I can see, the lemma only works for fixed X0X_0. However, in the learning process ptp_t might depends on the historical contexts X0,X1,X2,,Xt1X_0, X_1, X_2, \ldots, X_{t-1}. I am confused about the usage of the lemma and hope for some explanations.

As mentioned in lines 227--229, the variable X0X_0 is a kind of dummy random variable that does not appear in the decision-making process or algorithms, but appear only in the analysis, and is defined to be independent of X1,X2,,XTX_1, X_2, \ldots, X_{T} (and therefore is independent of any other variables including ptp_t). The approach of such an analysis is proposed by [NO20], in which this is referred to as the ghost sample. The idea and usage of this technique are explained around (1) of Section 2 of [NO20]. Since X0X_0 is independent of all other random variables, it justifies the approach of marginalizing w.r.t. X0X_0 after evaluating for a fixed X0X_0. The revised version explains this more explicitly.

Reference

  • [NO20] G. Neu and J. Olkhovskaya. Efficient and robust algorithms for adversarial linear contextual bandits. COLT202.
评论

I am still confused about the proof so I hope for more detailed analysis. The linear contextual bandit problem seems too dirty so we can consider contextual bandit problem instead. For example, we assume there are KK arms in XX. In each round. Let the context distribution DD be the uniform distribution over all subsets of XX with size K/2K/2. Consequently, with high probability XiXjX_i\neq X_j for any iji\neq j. We then choose the reward function rt(x)r_t(x) as a Bernoulli variable with mean 1/21/2. Then the best response π\pi^* is the policy such that π(Xt):=argmaxxXtrt(x)\pi^*(X_t): = \arg\max_{x\in X_{t}}r_{t}(x). With hight probability, we have t=1TmaxxXtrt(x)=T\sum_{t=1}^T \max_{x\in X_{t}}r_t(x)=T. However, to reach a sublinear regret, one needs to identify the arms with reward 11 in all but o(T)o(T) rounds without any prior information. This seems impossible.

Could you please explain more about this?

评论

Thank you for your time and the opportunity to engage in further discussion.

The construction you proposed would not suffice as a proof of the lower bound on regret. As can be seen in Line 189, in the definition of regret, the operator of supπ\sup_{\pi^*} is placed outside of the expectation operator E[]\mathbb{E}[\cdot]. This means that we take the maximum with respect to π\pi^* after taking the expectation with respect to the context. Consequently, the "optimal policy" π:X[K]\pi^*: \mathcal{ X } \rightarrow [K] cannot depend on the realization of contexts but is the policy that maximize the expected total rewards.

In the case of the problem instance you presented, the policy π\pi^* is chosen to ensure that π(Xt)argmaxxX_tr_t(x)\pi^*(X_t) \in \arg\max_{x \in X\_t} r\_t(x) for (almost) all t[T]t \in [T]. Such a policy π:X[K]\pi^*: \mathcal{ X } \rightarrow [K] necessarily depends on the sequence X_tt[T]\\{ X\_t \\}_{t \in [T]}. Indeed, for most randomly generated sequences r_t_t[T]\\{ r\_t \\}\_{t \in [T]}, there exists no policy π\pi^* such that π(Xt)argmax_xX_tr_t(x)\pi^*(X_t) \in \arg\max\_{x \in X\_t} r\_t(x) (t[T])(t \in [T]) for (almost) every possible realization of contexts X_t_t[T]\\{ X\_t \\}\_{t \in [T]}.

In the literature on adversarial linear contextual bandits, it appears standard to define regret by taking the maximum with respect to the comparator policy after taking the expectation, as we have done in our paper. For reference, please see Section 2 of [LWZ23] and Section 2 of [NO20]. This can be interpreted as a comparison with the optimal policy chosen without knowledge of the realization of randomly generated contexts.

Reference

  • [LWZ23] H. Liu, C.-Y. Wei, and J. Zimmert. Bypassing the simulator: Near-optimal adversarial linear contextual bandits. NeurIPS2023.
  • [NO20] G. Neu and J. Olkhovskaya. Efficient and robust algorithms for adversarial linear contextual bandits. COLT2020.
评论

Thanks for the response. I think there is some misunderstandings. π\pi^* is a mapping from the context to the arm, i.e., a mapping from 2[K]2^{[K]} to [K][K], so why can't π\pi^* depend on the context?

评论

Also another more concrete question: what is the optimal policy for the contextual bandit problem above?

评论

Thanks for the reply, and sorry for the confusion. We understand that π\pi^* is a mapping from 2[K]2^{[K]} to [K][K], so of course, the value of π(X_t)\pi^*(X\_t) depends on X_tX\_t. My point is whether the mapping π\pi^* itself can change according to the realization of X_tt[T]\\{ X\_t \\}_{t \in [T]}.

As we mentioned, in the definition of regret, maxπ\max_{\pi^*} is placed outside of E[]\mathbb{E}[ \cdot ], where the expectation is taken over all randomness. Thus, when considering bounds on regret, we focus on the value of

maxπ:2[K][K]E_X_t[t=1Tr_t(π(X_t))],\max_{\pi^*:2^{[K]} \rightarrow [K]} \mathbb{E}\_{\\{ X\_t \\}}\left[\sum_{t=1}^T r\_t( \pi^*(X\_t))\right],

where the expectation is taken with respect to the contexts X_t_t[T]DT\\{ X\_t \\}\_{t \in [T]} \sim {D}^T. The "optimal policy" here is the policy that attain the maximum above. In contrast, we guess that you are focusing on the following value:

E_X_t[maxπ:2[K][K]t=1Tr_t(π(X_t))].\mathbb{E}\_{\\{ X\_t \\}}\left[ \max_{\pi^*:2^{[K]} \rightarrow [K]} \sum_{t=1}^T r\_t( \pi^*(X\_t))\right].

In the latter, one is allowed to choose a different mapping π\pi^* depending on X_t_t[T]\\{ X\_t \\}\_{t \in [T]}, but not in the former. We consider this is an important difference. In the literature of adversarial linear contextual bandits, regret is defined with the former concept.

Also another more concrete question: what is the optimal policy for the contextual bandit problem above?

The optimal policy is one that maximizes E_X_t[t=1Tr_t(π(X_t))]\mathbb{E}\_{\\{ X\_t \\}}\left[\sum_{t=1}^T r\_t( \pi^*(X\_t))\right]. As a simple example, let us consider the case of T=2T = 2, K=4K = 4, X=1,2,3,4X = \\{ 1, 2, 3, 4 \\}, r1=[1,1,1,1]r_1 = [1,1,-1,-1] and r2=[1,1,1,1]r_2 = [-1,-1,1,1]. Suppose that XtX_t is chosen from a uniform distribution over all subset of XX with size 22. Then we can see that

E_X_t[r_1(π(X_1))+r_2(π(X_2))]=0\mathbb{E}\_{\\{ X\_t \\}}\left[r\_1( \pi^*(X\_1)) + r\_2( \pi^*( X\_2 ))\right] = 0

for any π:2[K][K]\pi^*:2^[K] \rightarrow [K], which means that every π\pi^* is an optimal policy. In fact, as X_1X\_1 and X_2X\_2 follow an identical distribution, we have

E_X_t[r_1(π(X_1))+r_2(π(X_2))]=E_X_t[r_1(π(X_1))+r_2(π(X_1))].\mathbb{E}\_{\\{ X\_t \\}}\left[r\_1( \pi^*(X\_1)) + r\_2( \pi^*( X\_2 ))\right] = \mathbb{E}\_{\\{ X\_t \\}}\left[r\_1( \pi^*(X\_1)) + r\_2( \pi^*( X\_1 ))\right].

If π(X_1)1,2\pi^*(X\_1) \in \\{ 1 , 2 \\}, we have r_1(π(X_1))+r_2(π(X_1))=1+(1)=0r\_1( \pi^*(X\_1)) + r\_2(\pi^*( X\_1 )) = 1 + (- 1) = 0. Otherwise, i.e., if π(X_1)3,4\pi^*(X\_1) \in \\{ 3 , 4 \\}, we have r_1(π(X_1))+r_2(π(X_1))=(1)+1=0r\_1( \pi^*(X\_1)) + r\_2(\pi^*( X\_1 )) = (-1) + 1 = 0. Hence, the expected cumulative reward is 00 for any π\pi^*.

评论

Thanks for the explanation. It is much more clear.

审稿意见
7

In this work, the authors tackle the existing gap between upper and lower bounds in bandits with expert advice. An existing lower bound scaled as Ω(KTlogNlogK\Omega(\sqrt{KT \frac{\log N}{\log K}}, while the state of the art of only provided a O(KTlog(N/K)O(\sqrt{KT \log (N/K)} bound (Kale 2014) The authors close this gap by proposing a novel lower bound that matches the result of Kale 2014 and improves upon the lower bound of Seldin and Lugosi (2016).

Then, the authors also consider the problems of linear bandits and contextual linear bandits. For linear bandits, the authors improve upon the state of the art and provide tighter upper bounds, which match existing lower bounds for various relations of the dimension dd and the number of arms KK. The study of contextual linear bandits in this setting is fairly recent and there, the authors propose novel algorithm that improves upon the state of the art and a novel lower bound that matches their upper bound for certain combinations of dimensions dd and number of arms KK.

优点

This paper provides several significant improvements for variations around the multi-armed bandits. Notably, several of their contributions are lower bounds, one of which closed an open problem that had been open for 8 years. The strategy used to solve this problem is a reduction from the best expert identification problem to the bandits with expert advice setting, and this approach will likely lead to other improvements in different fields.

In the field of linear bandits, the approach using FTRL with the correct tuning of Tsallis entropy has proved crucial to solving many bandit problems, not just in the adversarial setting but in the more challenging best-of-both-worlds regime.

In all of these sections, the authors provide detailed proofs that appear correct.

缺点

If anything, this paper could benefit from some discussions about the use of FTRL with Tsallis-Inf with tuning of α(1/2,1)\alpha \in (1/2, 1) which is used to solve many variations of the multi-armed bandits problem (for example but not limited to: bandits with feedback graphs (Zimmert and Lattimore (2019), Ito et al. (2022), Dann, Wei and Zimmert (2023) or Decoupling explorations and exploitation in MAB (Rouyer and Seldin (2020), Jin, Liu and Luo (2023)).

Using this framework is very beneficial in best-of-both-worlds settings rather than in the purely adversarial or purely stochastic regimes. Have you considered generalizing your results to that setting?

问题

See weaknesses.

局限性

Purely theoretical work. The scope of the results is clearly indicated in the theorems.

作者回复

Using this framework is very beneficial in best-of-both-worlds settings rather than in the purely adversarial or purely stochastic regimes. Have you considered generalizing your results to that setting?

Thanks for your suggestion. We believe that our results can be extended to the best-of-both-worlds (BOBW) setting if some difficulties are overcome. Challenging factors are inferred from the fact that there are only limited number of results achieving BOBW with FTRL using Tsallis entropy with α1/2\alpha \neq 1/2. The few such examples are [ITH24], [JLL23] and [RS20], which seem to require careful adjustment of learning rates that may depend on the feedback and output distributions so far, as is given, e.g., in Algorithm 1 of [JLL23]. They also seem to require some sort of stability condition, as given in equation (25) of [ITH24] and in Section C.3 of [JLL23]. If we can overcome these challenges, we believe our results can be extended to the BOBW setting.

Reference:

  • [ITH24] Shinji Ito, Taira Tsuchiya, and Junya Honda. Adaptive learning rate for follow-the-regularized-leader: Competitive analysis and best-of-both-worlds. COLT2024.
  • [JLL23] Tiancheng Jin, Junyan Liu, and Haipeng Luo. Improved best-of-both-worlds guarantees for multiarmed bandits: FTRL with general regularizers and multiple optimal arms. NeurIPS2023.
  • [RS20] Chloé Rouyer and Yevgeny Seldin. Tsallis-INF for decoupled exploration and exploitation in multiarmed bandits. COLT2020.
审稿意见
6

The paper investigates two significant extensions of multi-armed bandit problems: multi-armed bandits with expert advice (MwE) and contextual linear bandits (CLB). For MwE, the authors close the gap between previously known upper and lower bounds, establishing a matching lower bound of Omegaleft(sqrtKTlogfracNKright)\\Omega\\left(\\sqrt{KT \\log \\frac{N}{K}}\\right), where KK, N>KN>K, and TT denote the number of arms, experts, and rounds, respectively. This claim seemingly resolves an open question posed by Seldin and Lugosi (2016), where a Omegaleft(sqrtKTfraclogNlogKright)\\Omega\\left(\\sqrt{KT \\frac{\\log N}{\\log K}}\\right) regret lower bound was shown. For CLB instead, the authors introduce an algorithm that achieves an improved upper bound of O\\left( \\sqrt{dT \\log\\left(K \\min\\{1, \\frac{S}{d}\\}\\right)\} \right), where dd is the dimensionality of feature vectors, and SS is the size of the context space. The authors also provide a matching lower bound, confirming the minimax regret is Thetaleft(sqrtdTlogleft(Kmin1,fracSdright)right)\\Theta\\left( \\sqrt{dT \\log\\left(K \\min\\{1, \\frac{S}{d}\\}\\right)} \\right). The results are achieved using the follow-the-regularized-leader (FTRL) approach using the negative Tsallis entropy regularizer for an appropriate tuning of its parameter, and carefully designed context-dependent learning rates.

优点

This work provides relevant contributions to the theoretical understanding of multi-armed bandit problems in terms of the minimax regret. Even if the use of the follow-the-regularized-leader (FTRL) approach with negative Tsallis entropy regularization follow from a previous line of work of improved regret bounds in other bandit problems, as made explicit by the authors in the introduction, the regret analysis of the proposed algorithm for CLB requires novel ideas in tuning the learning rate and more care in leveraging the possibility to tune the parameter of the Tsallis entropy.

More interestingly, the authors manage to provide an improved lower bound for a harder version of bandits with experts advice that matches the regret of the algorithm proposed by Kale (2014). They do so via a proof argument that adapts ideas from the previous work on bandits with feedback graphs by Chen et al. (2023) to MwE, requiring a careful construction of hard problem instances for the latter problem. This construction is nontrivial, interesting, and could potentially help in the design of hard problem instances for proving lower bounds of other related problems. Furthermore, introducing logarithmic factors in regret lower bounds is typically challenging and doing so enables a more complete understanding of the problems. Similarly, the lower bound for CLB requires a clever adaptation of previous results.

缺点

While the lower bound for the MwE problem improves on prior results, it requires to consider a harder setting than the standard one by restricting the learner. Specifically, they restrict the learner to selecting only experts rather than possibly choosing actions directly, and the learner can observe the expert advice only after making a decision. The presentation is also somewhat misleading, as the authors claim to resolve the open problem left by Seldin and Lugosi (2016) for the “classical” problem of multi-armed bandits with expert advice while it is not exactly the case. To fully address that open question, one needs to consider learners that may select actions irrespectively of the expert advice, and that observe the expert advice before choosing an action. Nonetheless, the authors themselves make a clear point (rf. Remark 1) that this is indeed a more challenging setting, and my concern mainly revolves about the claims in the abstract and introduction. It should also be clarified that proving the same lower bound for the standard setting remains an open problem that could be interesting to investigate.

Some proofs in the paper feel more like sketches rather than complete proofs. While they provide a good overview of the approach and main ideas, they lack detailed steps and thorough explanations. This makes it challenging for readers to fully verify the results and understand all nuances of the arguments. For instance, the proof of Corollary 1 would benefit from a more detailed explanation of the steps; additionally, an intuitive explanation of the choice of the context-dependent learning rate would help the reader while possibly helping in adapting a similar idea in other related problem settings. Another example is provided by the proof of Theorem 4, which is missing a more formal and explicit computation of the lower bound.

Finally, the results in this current work still do not achieve minimax optimality for the non-contextual linear bandits problem. Indeed, the only known lower bounds that match the upper bound provided in this work hold for the cases K=dK=d and K=2dK=2^d, respectively. The problem of proving a lower bound for arbitrary values of KK is still open and stating this explicitly would make the exposition of the contribution of this paper more transparent and clearer for the reader. A similar question could be made for the contextual version of the problem since the lower bound proved in this paper requires 2d/SleKle2d2^{d/S} \\le K \\le 2^d.

问题

  • Could you clarify the doubts that were raised in the weaknesses above?
  • Could you expand a bit on how the argument at lines 278-279 allows us to keep the generality of the values of SS, KK, and dd as claimed?
  • Do you think the ideas from the different lower bounds for linear bandits could be adapted and combined to prove a regret lower bound for arbitrary values of KK? Or is there a much larger technical hurdle that needs to be overcome?

Minor comments/typos:

  • Line 99: The argmin is missing mu_j\\mu\_j
  • Line 104: “denote” instead of “denotes”
  • Line 130: “and the set” instead of “the set”
  • Line 143: “provability” instead of “probability”
  • Line 180: “Lemma” instead of “to Lemma”
  • Line 187: “drawn” instead of ”drown”
  • Line 188: “gets” instead of “get”
  • Line 192: “contexts” instead of “context”
  • In assumption 1, it could be clearer to specify that LgeSL \\ge S instead of just L>0L>0 for it to make sense, otherwise it might not be satisfiable
  • Math display after line 195: Isimp(X)I \\sim p(X) instead of Isimp(x)I \\sim p(x)
  • Line 196: “that there exists an” instead of “that an”
  • Line 252: the choice of naming the top-dd subset of pairs (x,i)(x,i) as SS might cause confusion in the reader, as SS is already defined to be the number S=mathcalXS = |\\mathcal{X}| of contexts

局限性

N/A

作者回复

While the lower bound for the MwE problem improves on prior results, it requires to consider a harder setting than the standard one by restricting the learner.

We deeply appreciate your suggestion. We agree that this is an important issue, and we will add this point to the abstract and introduction. In addition, we will mention in the revised version that showing a similar results in the "classical" problem setting remains as an open question.

For instance, the proof of Corollary 1 would benefit from a more detailed explanation of the steps; additionally, an intuitive explanation of the choice of the context-dependent learning rate would help the reader while possibly helping in adapting a similar idea in other related problem settings.

The revised version will includes a step-by-step explanation of how to calculate to show the bounds in Corollary 1. The context-dependent learning rate is designed so that the regret upper bound of Theorem 3 is bounded well. In fact, if we consider minimizing xg(x)η(x)\sum_{x} \frac{g(x)}{\eta(x)} subject to the constraint of supxη(x)g(x)β\sup_{x} \frac{\eta(x)}{ g(x)^{\beta} } \le const, the optimal solution is given as η(x)(g(x))β\eta(x) \propto (g(x))^{\beta}.

Finally, the results in this current work still do not achieve minimax optimality...

As you pointed out, for linear bandits and contextual linear bandits, minimax regret bounds have only been identified when KK is within a specific setting or range. The revised version will place more emphasis on this fact and make it clear that relaxing these assumptions is an open question to be resolved in future studies.

Could you expand a bit on how the argument at lines 278-279 allows us to keep the generality of the values of SS, KK, and dd as claimed?

We first note Theorem 4 implies that, if some dd' satisfies K2dK \ge 2^{d'} and ddSd \ge d'S, we can obtain a regret lower bound of RT=Ω(dST)R_T = \Omega(d'\sqrt{ST}). Let (d,K,S)(d, K, S) be an arbitrary given parameter set satisfying K2dKSK \le 2^d \le K^S. We then have log2KdSlog2K\log_2 K \le d \le S \log_2 K. Define d:=log2Kd' := \lfloor \log_2 K \rfloor and S:=d/log2KSS' := \lfloor d / \log_2 K \rfloor \le S. Then, as we have K2dK \ge 2^{d'} and dSlog2KSdd \ge S' \log_2 K \ge S'd', from Theorem 4, we obtain a regret lower bound of RT=Ω(dST)=Ω(dSTd)R_T = \Omega(d' \sqrt{S'T}) = \Omega( \sqrt{d' S' T d'} ) for a problem instance with parameters (d,K,S)(d, K, S'). By combining this with dS=Ω(d)d' S' = \Omega( d ) and d=Ω(logK)=Ω(log+(Kmin1,Sd))d' = \Omega( \log K ) = \Omega( \log_+ (K \min \\{ 1, \frac{S}{d} \\})), we obtain RT=Ω(dTlog+(Kmin1,Sd))R_T = \Omega( \sqrt{ d T \log_+ ( K \min \\{1, \frac{S}{d} \\}) }). As we have SSS' \le S, the same lower bound applies to the problem with parameters (d,K,S)(d, K, S) as well. The revised version will include a more detailed explanation like the above.

Do you think the ideas from the different lower bounds for linear bandits could be adapted and combined to prove a regret lower bound for arbitrary values of KK?

We believe there is hope, and we have tried that approach, but have not yet succeeded. What we have considered so far is a generalization based on the combination of standard multi-armed bandits (K=dK = d) and hypercube bandits (K=2dK = 2^d). For example, we have considered a problem in which we choose one of the the kk hypercubes and then choose a point in that hypercube, i.e., we set d=dkd = d' k and define an action set as A=({1,1}d×{0}dd)({0}d×{1,1}d×{0}d2d)({0}dd×{1,1}d)\mathcal{A} = (\{ -1, 1 \}^{d'} \times \{ 0 \}^{d-d'}) \cup ( \{ 0 \}^{d'} \times \{ -1, 1 \}^{d'} \times \{ 0 \}^{d - 2d'}) \cup \cdots \cup (\{ 0 \}^{d-d'} \times \{ -1, 1 \}^{d'}). We then have K=A=2dk=d2d/dK = |\mathcal{A}| = 2^{d'} k = d 2^{d'} / d'. This construction almost covers the range of K[d,2d]K \in [d , 2^d] by adjusting the parameter d[1,d]d' \in [1, d]. We conjecture that a lower bound of Ω(dkT)\Omega(d'\sqrt{kT}) holds in this setting. We have been attempting to prove a lower bound using hard instances of hypercube bandits [DKH07], but the existing methods of analysis (e.g. [CHZ24]) do not seem directly applicable and have not yet been able to show the lower bound.

Minor comments/typos:

Thank you so much for your many helpful comments. Each will be reflected in the revised version.

Reference:

  • [DKH07] V. Dani, S. M. Kakade, and T. Hayes. The price of bandit information for online optimization. NeurIPS2007.
  • [CHZ24] H. Chen, Y. He, and C. Zhang. On interpolating experts and multi-armed bandits. ICML2024.
评论

I thank the authors for the detailed response. Since my main concerns have been addressed, I am raising my score as I believe this paper provides valuable insights towards studying the minimax optimal rate of multiarmed bandit problems.

审稿意见
6

This paper presents new bounds for regret minimization problem in multi-armed bandits with expert advice (MwE) and contextual linear bandits (CLB). For MwE, This paper bridges the gap between existing upper and lower bounds by establishing a new matching minimax optimal lower bound. In the case of CLBs, the authors propose a lower bound and devise an algorithm based on the follow-the-regularized-leader approach, which achieves a matching upper bound.

优点

• Novel contribution: The paper introduces new minimax optimal lower bounds for MwE and CLB. Additionally, for CLB, it devises an algorithm that achieves a matching upper bound. • Theoretical results: The paper provides comprehensive proofs for each of the newly introduced bounds, offering rigorous theoretical validation of the results.

缺点

• Motivation: The paper lacks an exploration of practical applications of the proposed work, thus indicating a deficiency in motivation. Incorporating a discussion on potential practical applications would significantly add value of the results presented in the paper. • Future Work: The paper does not discuss potential directions for future research. • No empirical results verifying the bounds.

I found some typos that the author(s) might want to correct.

• In line 84, ”the player chooses an expert Ji[K]J_i \in [K] should be replaced by ”the player chooses an expert Jt[N]J_t \in [N], correct?

• In line 99,”jargminj[N]j^*\in\argmin_{j\in [N]} is missing μj\mu_j

• In line 213: "This section provide" to "This section provides"

问题

The proposed algorithm for contextual linear bandits (CLB) requires an initial exploration policy p0 as one of its input parameters. How is this policy determined at the start of the algorithm? Are there any specific assumptions or conditions on p0 necessary for it to adhere to the presented upper bound?

局限性

The authors discuss the assumptions needed for their results to hold.

作者回复

Motivation: The paper lacks an exploration of practical applications of the proposed work, thus indicating a deficiency in motivation. Incorporating a discussion on potential practical applications would significantly add value of the results presented in the paper.

We appreciate your suggestion. Since our results on linear bandits and contextual linear bandits imply improvements in the regret upper bound compared to existing algorithms, we expect improved performance in real-world applications and believe they will be useful in practice. However, this study focuses primarily on the goal of revealing the limits of achievable regret upper bounds for the fundamental online decision-making problems, rather than on specific applications.

Future Work: The paper does not discuss potential directions for future research.

The revised version describes open questions and future research direction more explicitly, as we mentioned in "Author Rebuttal" above.

No empirical results verifying the bounds.

This paper does not include empirical results because, in general, results of some specific numerical experiments are not very informative for the purpose of validating the analysis of minimax regret. Indeed, the minimax regret corresponds to the worst-case scenario, i.e., the problem instances that are most unfavorable loss-sequences for the algorithm. Numerical experiments on specific data are rarely the worst-case input, and it is difficult to know how close to the worst-case it is. The same reason may explain why many existing studies focusing on minimax regret do not include numerical experiments.

I found some typos that the author(s) might want to correct.

We deeply appreciate your pointing out the typo. We consider that all of them need to be corrected as you pointed out, and we will reflect them in the revised manuscript.

The proposed algorithm for contextual linear bandits (CLB) requires an initial exploration policy p0 as one of its input parameters. How is this policy determined at the start of the algorithm? Are there any specific assumptions or conditions on p0 necessary for it to adhere to the presented upper bound?

The necessary assumption on p0p_0 is described in Assumption 2 (line 196), where λ(p0)\lambda( p_0 ) is defined just above it. As long as this assumption is satisfied, p0p_0 can be determined in any way. In particular, if p0p_0 is determined using g-optimal design, then λ(p0)Ld\lambda(p_0) \le Ld holds, and hence this assumption is satisfied, as explained just after Assumption 2. The parameter λ0=λ(p0)\lambda_0 = \lambda(p_0) depending on p0p_0 affects the asymptotically non-dominant terms (in the regime of TT \rightarrow \infty) of the regret upper bounds provided in Theorem 3 and Corollary 1.

评论

Thank you for addressing my questions.

The phrasing of Assumption 2 is a bit confusing to me. You write "We assume that an exploration policy p0:XP(K)p_0 : X \to \mathcal{P}(K) such that λ(p0)<\lambda(p_0) < \infty." It feels like a word is missing. You are saying that a policy p0p_0 exists**exists** with λ(p0)<\lambda(p_0) < \infty, correct?

I see now why such a policy should exist. Incidentally you wrote "g-optimal" and usually I see the "G" capitalized.

评论

Thank you for your review and feedback on the manuscript and rebuttal.

You are right about both of the points you raised. In the revised version, we will add the phrase "there exists" in the description of Assumption 2 and correct the "g" in "g-optimal" to the "G" capitalized.

We are deeply grateful to you for bringing these errors to our attention.

作者回复

Thank you and future revisions

First of all, we would like to express our deepest gratitude to the reviewers who spent a great deal of time reviewing this paper. Thanks to the valuable peer review comments we received, we are confident that the quality of our manuscript will be greatly improved. The following revisions are planned as major changes:

Note on the setup of multi-armed bandits with expert advice

In the revised version, the notes on problem setup explained in Remark 1 of the current manuscript will also be mentioned in the abstract and introduction. As noted in Remark 1 and in the comment by Reviewer FSFk, the problem setting of BwE considered in this paper is more challenging than the "classical" setting where the player can observe all expert advice before selecting an arm, while almost all known existing algorithms, including those in Table 1, work for this more challenging setting. Although these facts were written in Remark 1, we believe that the current abstract and introduction were misleading, as Reviewer FSFk pointed out. To avoid such confusion, we revise the manuscript so that readers will notice this fact just by reading the intro and abstract.

Open question and future work

In the revised version, we will add descriptions of remaining open questions and potential future research directions. For examples:

  • Tight regret bounds for "classical" setting of multi-armed bandits with expert advice
  • Tight bounds for linear bandits and contextual linear bandits with more general parameter setting of KK, SS, and dd (please see the comment by Reviewer FSFk and our response to it)
  • Relaxing the assumption on the context distribution (please see the comment by Reviewer nf4R and our response to it)

Correction of typos and more more detailed explanations

We will correct the typos the reviewers pointed and add intuitive explanations for the unclear points you commented on, as well as more detailed explanations of the steps in the analysis.

In addition to the above, we will revise the manuscript in response to each of the comments received. Once again, we are deeply grateful for the tremendous amount of helpful feedback.

最终决定

This paper studies two related problems. First the authors show a novel lower bound for the setting of the CMAB problem which tightly matches the known upper upper. Second the authors propose a new algorithm for the Contextual Linear Bandit problem together with a matching lower bound.

Reviewers are mostly positive on this work, however, there are multiple important problems that need to be addressed before the final version. First the authors have to discuss the fact that the lower bound for MwE problem is in a harder setting than the standard. Next, the authors have to discuss the range of K for which their lower bounds hold for the linear bandit problem. Finally, the authors have to discuss the assumption that the proposed algorithm for the linear bandit setting requires knowledge of the contextual distribution and the computational problems in the case of an infinite number of contexts. Further, the presentation of the paper needs to be improved as multiple reviewers have pointed out. Nevertheless the works contributions are non-trivial and I am leaning towards recommending it for acceptance to the program.