Thompson Sampling for Multi-Objective Linear Contextual Bandit
We propose the first Thompson Sampling algorithm with Pareto regret guarantees in multi-objective linear contextual bandit.
摘要
评审与讨论
In this paper, the authors study stochastic linear contextual bandits under the multi-objective setting, in which each arm corresponds to a vector of objective rewards, whose expected reward is the inner product between the hidden objective value and the arm’s context. As the Pareto front is non-convex, they define the effective Pareto optimality, which is essentially the convex hull of all arms. They measure performance using the cumulative minimal effective Pareto sub-optimality, defined as the distance between the performance and the convex hull.
They present a Thompson Sampling-based algorithm to solve this problem efficiently. Specifically, for each objective, they sample a logarithmic number of hidden objective parameters and estimate the empirical objective reward of each arm using the maximum inner product between the sampled parameters and the context. The algorithm then randomly plays an action sampled uniformly from the empirical Pareto front. They establish a regret bound of order for their algorithm. Experiments are provided to demonstrate the benefits of the algorithm.
优缺点分析
Strength:
-
The multi-objective linear contextual bandit setting is definitely significant and interesting. Applying Thompson Sampling to this setting is a natural idea. The authors propose a clean algorithm for this setting, with a regret upper bound that matches the performance of Thompson Sampling in the linear contextual bandit setting.
-
I think optimistic sampling is a novel idea, as it allows one to perform optimistic estimation in a computationally efficient way compared to the UCB bounds ( for optimistic sampling per action vs. in UCB). The corresponding intuition and proof sketch look correct to me. This analysis may shed light on future work in the bandit literature.
Weakness:
- I feel that the main reason one would prefer a TS-based approach over a UCB-based approach is its computational efficiency, as suggested by Agrawal et al. (2012), especially when there are many arms. Unfortunately, I don't see the authors highlighting this point. I would suggest that the authors add a section discussing the computational complexity of the proposed approach.
Agrawal, Shipra and Navin Goyal. “Thompson Sampling for Contextual Bandits with Linear Payoffs.” International Conference on Machine Learning (2012).
问题
- How does affect the constant in the regret bound of Theorem 2? For example, what happens if doubles?
局限性
Yes
最终评判理由
I would maintain my positive evaluation of this work and recommend it for acceptance.
格式问题
No
We thank the reviewer for your time and thoughtful review of our paper. We greatly appreciate the insightful feedback, especially the positive recognition of our use of optimistic sampling.
Computational Cost : We will gladly discuss the computational complexity of our proposed approach. Unlike previous works that require computing the entire Pareto front, which involves comparing all pairs of arms and has a computational complexity of (where is the number of arms), our method significantly reduces the computational cost. Instead of explicitly identifying all effective Pareto optimal arms, we play an arm based on a random weight vector, thus selecting an arm that is optimal with respect to the random weight, which is effective Pareto optimal arm. This approach takes the computational complexity to . Our method saves computational time since .
Discussing the computational complexity involves several factors, including both the selection of arms and the method for computing the Pareto front. Beside comparing computational complexity between UCB and TS which involves considering multiple factors of each algorithm, our experiments show that the proposed Thompson Sampling algorithm is computationally faster. We believe that the computational efficiency of TS, as demonstrated empirically, arises from the way it avoids the explicit computation of the entire Pareto front and instead samples arms based on a random weight vector, leading to lower computational overhead. The following table shows the average time complexity with , and .
| Time Complexity (s) | (ours) | |
|---|---|---|
| 16.73 | 5.09 | |
| 17.39 | 6.92 | |
| 34.64 | 5.97 | |
| 37.25 | 7.29 | |
| 40.87 | 11.43 | |
| 79.95 | 7.81 | |
| 88.08 | 11.69 |
It is also important to mention that anthor main reason of using TS over UCB is the superior empirical regret performances which are presented in the experiments in the paper (Figure 2~13).
The constant dependency with respect to : The regret bound in Theorem 2 depends on through the term , which arises from the fact that multiple sampled parameters need to satisfy the concentration property [1]. Doubling results in a logarithmic increase only in the regret bound.
[1]Abeille and Lazaric. "Linear thompson sampling revisited" (2017).
Thanks to the authors for their answer. I have no further questions.
The paper proposes a Thompson sampling (TS) based approach for multi-objective linear bandits, with the goal of optimizing the effective Pareto regret. The authors define this new notion to overcome the issues associated with measuring cumulative regret using the Pareto regret idea. The algorithm (MOL-TS) essentially takes multiple samples from the posteriors to construct a randomized optimistic confidence bound for each arm, based on which the effective Pareto front is determined at that round, and action is taken randomly from this front. The paper establishes a worst-case regret bound, which matches the lower bound for single-objective linear bandit TS.
优缺点分析
The paper is well-written and the results are promising. However, due to the missing appendix, the proofs cannot be verified and experimental details are missing at present.
问题
- The optimistic TS using samples basically behaves like a UCB-type algorithm, by taking large number of samples and taking the maximum, this maximum (for large ) behaves like the mean added with an upper confidence term, depending on the Gaussian tail and , however unlike UCB it is slightly randomized (amount of randomization controlled by ) - can the authors provide any comment on this connection?
- Can you explain why for each individual objective, greedy outperforms UCB, but somehow the Pareto regret is way better for UCB (in Figure 2)? This figure is misleading and leads to question the effectiveness of the Pareto regret (and effective PR) notion. Appendix is missing - proofs and other experimental details.
- Can you be more explicit in the regret bounds in the orders of dependence on and ? Is the bound tight in these? Also, it might be good to show the dependence on (or ) in theorem 2 and corollary 1.
- [Contextual Bandits with Linear Payoff Functions by Chu et al. 2011] shows lower bound for linear contextual bandits (single-objective) of the order and a matching upper bound (up to log terms). For TS, the recent [Geometry-Aware Approaches for Balancing Performance and Theoretical Guarantees in Linear Bandits by Luo et al] used a slight modification of TS to achieve this regret bound for randomized algorithms. Even in the high-dimensional case, the dependence on dimension and sparsity for TS is (e.g. [Thompson Sampling for High-Dimensional Sparse Linear Contextual Bandits by Chakraborty et al.]), demonstrating better dependence on dimension (stochastic setting). - in the current paper, the dependence on seems to be slightly higher at - can the authors comment on this? At least the authors should include a brief discussion on single-objective regret bounds in the literature to claim “best known order for randomized linear bandit algorithms for single objective”.
- What is the parameter in the algorithm 1?
- From a practician’s perspective, how do you tune all the hyper parameters in algorithm 1. Theoretical results give some indication on , but what about the others, in particular and ? Did you tune these parameters for the numerical experiments? If so, did you tune the associated parameters for greedy and UCB based methods for a fair comparison?
局限性
See the questions.
最终评判理由
I increased my score based on the authors responses.
格式问题
Clarify the setting, whether the contexts are adversarial or stochastic. It seems the former given there are no stochastic assumptions, still it would be good to clarify your setup.
We thank the reviewer for your time reviewing our paper. We sincerely hope that our answers can help clarify your questions.
The appendix has been uploaded. We would like to respectfully clarify that the full appendix, containing all proofs, experimental details, and additional experimental results, has been submitted as part of the supplementary material. Every theoretical claim made in the main paper is rigorously proved in the appendix, and our empirical evaluations are documented in detail.
We sincerely hope that, in light of this information, the reviewer might reconsider any concerns based on the (unintended) mistaken assumption that proofs or technical details are missing. We are concerned that an evaluation influenced by this misunderstanding could inadvertently affect the overall assessment of the well-prepared paper. We greatly appreciate the reviewer’s time and thoughtful engagement, and we are happy to point to any specific section of the appendix upon request.
We provide answers to your questions below:
Does the optimistic TS using samples behaves like UCB-type algorithm? : No, they are fundamentally different. UCB algorithm is deterministic, as they use upper confidence bound to evaluate arms, and the optimism comes from deterministically evaluating arms within the estimated confidence bounds. This algorithm is a fixed strategy for exploration.
Thompson Sampling is a randomized algorithm. The optimism in TS is handled through multiple sampling from the posterior distribution where the number controls the randomness of optimism and controls the boundary that the sampled parameters following concentration property [1] (more specifically, for every , holds with probability at least ) . Even with large , the maximum of the sampled parameters does not guarantee an optimistic value in the same way as UCB does.
[1] Abeille and Lazaric. "Linear thompson sampling revisited" (2017).
Comparison between greedy and UCB algorithm : We are more than happy to clarify this question. The single-metric regret measure may not fully explain the observed behavior in Figure 2 (also Figure 3~13 in Appendix F) because it is difficult to calculate a single value of regret that captures the full complexity of the reward vector across multiple objectives. It becomes more difficult as the number of objectives grows, which may lead to misleading interpretations in certain cases. To the best of our knowledge, for the first time, our paper demonstrates the experiments of objective-wise total rewards.
Pareto regret focuses on the trade-offs between objectives, but it does not account for the total rewards across all objectives. Our newly defined effective Pareto regret covers this problem, which offers a more comprehensive measure of performance. This new definition represents an important step forward in overcoming the difficulties associated with measuring the regret in multi-objective settings.
Explicit regret equation : In the Appendix B.2, Theorem 2 shows that algorithm has the worst-case regret upper bound by , where The regret depends on the number and up to logarithmic factor, and it does not depend on the number of arms .
Comparisons with LinTS-type algorithms for single objective : We are happy to elaborate on this. First of all, the well-known minimax lower bound for the linear contextual bandit is [2], so there is a gap between ours and the lower bound. (Please note that bounds such as Chu et al. 2011 possess factors for finite arms and regret algorithms including ours do not depend on the number of arms. We hope that we are discussing the results based on the relevant context.) The regret bound in our paper is , which matches the best-known regret bound of the linear Thompson Sampling (LinTS) type algorithms (e.g., [1]) in single objective setting. It is known that this bound cannot be improved for LinTS and its variants due to the inevitable requirement of posterior variance inflation by a factor of , a limitation that arises in the analysis of worst-case regret for Thompson Sampling compared to fixed optimism-based algorithms [3]. Our work extends this well-known result to the multi-objective setting, and the core regret bound still aligns with the fundamental limitations of LinTS in single objective setting.
Our work extends this well-known result from single objective to multi-objective setting, where we also face the same fundamental limitations in terms of the scaling with . Therefore, while the regret bound in the multi-objective setting grows more complex due to the additional factor of (the number of objectives), the worst-case regret bound is .
[2] Dani, Hayes, and Kakade. "Stochastic Linear Optimization under Bandit Feedback" COLT (2008).
[3] Hamidi and Bayati. "On Frequentist Regret of Linear Thompson Sampling" (2020).
What is ? : In Algorithm 1, the input represents a confidence bound input that plays two key roles, RLS estimations and concentration property. As described in Appendix C, measures the probability bound on the distance between the true parameter and the RLS estimates [4]. This ensures that the estimated parameters are close enough to true values with probability at least . Also, it controls the probability that the sampled parameters follow the concentration property [1]. This property ensures that sampled parameters are close enough to RLS estimate with high probability. It is crucial property for random exploration in the Thompson Sampling algorithm, as it helps balance the trade-off between exploration and exploitation, while maintaining reliable confidence bounds.
[4] Abbasi-Yadkori et al. "Improved algorithms for linear stochastic bandits" (2011).
Extra experiments of , , : The main focus of our experiments is to compare the performance of the algorithms with different values of and . The other hyperparameters, including , and were kept fixed in our experiments, with the value set to , and . We did explore additional results for different initial values of these hyperparameters. But these values were not central to the core ideas of our paper, and varying these parameters does not change the results. Regarding the comparison with greedy and UCB algorithm, we used the same fixed hyperparameters for fairness, ensuring that the comparison is made under consistent restriction. The following table measures the average of total regret over 10 different instances in with different , and .
| -Greedy | (ours) | ||
|---|---|---|---|
| Pareto Regret | 240.71 | 178.29 | 73.93 |
| Effective Pareto Regret | 285.38 | 268.92 | 95.50 |
| -Greedy | (ours) | ||
|---|---|---|---|
| Pareto Regret | 235.12 | 196.08 | 105.86 |
| Effective Pareto Regret | 279.80 | 289.89 | 133.27 |
| -Greedy | (ours) | ||
|---|---|---|---|
| Pareto Regret | 239.18 | 211.66 | 77.42 |
| Effective Pareto Regret | 285.00 | 310.29 | 99.16 |
Adversarial setting : We are happy to modify Section 3 of the paper, that the contexts are adversarially given.
Thank you for the response. Most of the questions were answered by the authors. I went through the appendix and my concerns regarding this are satisfied. The paper is pretty good in terms of the theoretical contributions and I raised my score to reflect that. For the final version, it might be good to state delta used in Algorithm 1 in the main text, at least a brief note on what this parameter indicates.
While I understand that TS and UCB are fundamentally different, I meant that in this case they can be seen connected in the following way. For linUCB type methods, . In your case, you take samples from the posterior and take the maximum - if you look at the distribution of this, this is centered around a term roughly similar to the UCB-type, where controls the exploration-exploitation trade-off (). If you take one sample, then you cannot control this - taking a correct choice of properly balances this trade-off, as you show.
I am still a bit confused about Figure 2 - I understand that it is harder to metrize performance across all objectives by a single measure, but it seems greedy outperforms the MOL-UCB for ALL objectives, however, the pareto regret shows a different story. So with this new metric, it seems it is possible to do better overall (in the new metric sense) when individually you perform worse for all underlying objectives. Can the authors provide an explanation for this?
Thank you very much for engaging with our rebuttal and for taking the time to re-evaluate your score. We truly appreciate the increased score and your acknowledgment after reviewing the appendix.
We also value your perspective on the connection between UCB methods and optimistic sampling in Thompson Sampling, particularly regarding the role of (or in our setting) in controlling the exploration–exploitation trade-off.
To address your question on Figure 2, we are happy to further explore scenarios that could lead to the results shown in Figure 2.
Consider a setting with four arms, each associated with a deterministic two-dimensional reward:
Arms , , and are Pareto optimal (and also effective Pareto optimal), whereas arm is strictly dominated, resulting in a Pareto regret (and effective Pareto regret) of per round when is selected.
Now compare two algorithms:
-
The first algorithm selects arm in half of the total rounds and arm in the other half (order does not matter). This results in a per-round reward of and zero Pareto regret.
-
The second algorithm selects arm in all rounds, achieving a reward of but incurring a constant regret of per round.
Despite the first algorithm achieving no Pareto regret and optimal total rewards in each individual run, its average reward vector is lower than that of the second algorithm. This highlights the unintuitive possibility that an algorithm with lower regret can yield lower average rewards than another.
While this phenomenon may occur under both Pareto regret and effective Pareto regret, it is less pronounced with effective Pareto regret (e.g., Figure 2, the difference between Pareto regret and effective Pareto regret). This is because the effective Pareto front may exclude Pareto optimal arms that lie inside the convex hull. For example, as illustrated in the left subfigure of Figure 1, some arms that are Pareto optimal may still fall outside the effective Pareto front.
To see this more concretely, consider a similar setting with reward vectors:
In this case, choosing incurs zero Pareto regret (with losses in rewards in both dimension) but non-zero effective Pareto regret. Effective Pareto regret better copes with these cases than the plain Pareto regret does.
We greatly appreciate your insightful question and the opportunity to elaborate on the distinctions between Pareto regret and effective Pareto regret, and their relationship to singleton reward performance.
If you have any further questions, we would be happy to provide additional clarification. Thank you again for your valuable time and thoughtful feedback.
Thank you for the clarification. It might be good to elaborate on this example regarding the seemingly strange behavior in Figure 2 in the revised version.
This paper studies the multi-objective linear bandit problem, which extends traditional linear contextual bandit problems by requiring optimization across multiple, often conflicting objectives simultaneously. In this paper, the authors proposes MOL-TS as the first randomized algorithm for multi-objective contextual bandits with Pareto regret guarantees, achieving regret. Also, this paper firstly introduces "effective Pareto optimal arm" that not only satisfies standard Pareto optimality but also ensures total rewards across all objectives remain Pareto optimal with repeated selections. Numerical experiments demonstrate improved performance in both regret minimization and objective-wise total reward maximization
优缺点分析
Strengths
-
This research addresses multi-objective bandit literature by introducing the first Thompson Sampling approach with theoretical guarantees.
-
The algorithm sample multiple models from the posterior distribution and compute an optimistic reward estimation to adapt to the multi-objective setting, which is novel.
-
The effective Pareto optimality concept provides a more robust framework for multi-objective optimization, ensuring better long-term performance across all objectives rather than just instantaneous optimality.
Weaknesses
-
The regret bound achieved by this paper is , which is worse than UCB-based approaches for multi-objective linear bandits.
-
It would be better if the authors also discuss how to modify the previous approaches so that they could work for the new definition of Pareto optimality.
问题
In feel-good TS[1], it is shown that for standard linear bandits, the regret upper bound is , which matches with the UCB-based approaches. Can similar techniques be applied in multi-objective linear bandit settings to achieve a tighter bound?
[1] Tong Zhang. Feel-good thompson sampling for contextual bandits and reinforcement learning.
局限性
Yes.
最终评判理由
The authors' rebuttal helps me assess this work better. I will keep my original score.
格式问题
NA
We sincerely thank the reviewer for taking the time to carefully review our paper. We especially appreciate the positive comments on the strength of our work, especially acknowledging optimistic sampling.
Regret bound of and possible adaptation of Feel-Good TS
We are happy to answer this. First, the regret bound of for our algorithm—viewed as a member of the LinTS family—is not improvable in terms of the dependence. Hamidi & Bayati (2023) [1] show that Linear Thompson Sampling incurs an inherent posterior-variance inflation by a factor of , leading to a frequentist regret lower bound of is the best possible for LinTS-type algorithms. If not for such an inflation, LinTS-type algorithms would incur a linear regret in [1]
Our work extends this single-objective result to the multi-objective setting, and the bound we obtain remains consistent with these fundamental limitations of LinTS.
[1] Hamidi & Bayati. “On Frequentist Regret of Linear Thompson Sampling” (2023).
Adapting Feel-Good TS is indeed intriguing. However, even in the single-objective case, adding the Feel-Good exploration term typically requires approximate MCMC to sample from a non-conjugate posterior—unless strong distributional assumptions are imposed. Because we assume only classical sub-Gaussian noise, such sampling would introduce significant computational overhead.
For these reasons we base our algorithm on LinTS techniques. Extending Feel-Good TS to the multi-objective setting is an interesting avenue for future research, but lies beyond the scope of the present work.
Previous methods and our new notion of Pareto optimality: We are happy to revise the discussion to clarify how our newly defined concept—effective Pareto optimality—can benefit existing algorithms as well. Existing multi-objective bandit methods typically sample directly from the standard Pareto front, which can lead to inferior cumulative rewards across objectives. Replacing that step with our notion of effective Pareto optimality could improve the performance of those algorithms as well as our own, further underscoring the contribution of this work.
Thanks for the answers. It helps me assess this work better. I will keep my original score.
The paper studies the multi-objective linear contextual bandit problem, where a decision-maker (or learner) must simultaneously optimize several potentially conflicting objectives. The goal is to learn a policy for selecting an effective Pareto optimal arm (Definition 5) for each given context, whose mean reward vector is Pareto optimal and maximizes the cumulative rewards across all objectives. The performance of the learned policy is measured in terms of minimizing effective Pareto regret (Definition 7), which is the sum of effective Pareto sub-optimality gap (defined in Eq. (1)).
The authors propose a Thompson sampling-based algorithm, MOL-TS (Thompson Sampling for Multi-Objective Linear Bandits), for this setting with sub-linear effective Pareto regret guarantees. This paper introduces the notion of effective Pareto front, which consists of arms whose selection maximizes the total reward. The authors have also validated the performance of MOL-TS using synthetic problem instances and show that MOL-TS outperforms baseline algorithms in both regret minimization and cumulative objective-wise rewards.
优缺点分析
The following are the strengths of the paper:
-
This paper studies the multi-objective multi-armed bandit problem, which has applications in many areas, especially having several potentially conflicting objectives.
-
This paper proposes a Thompson sampling-based algorithm, MOL-TS, to learn a policy to select the best effective Parato optimal arms for each context. They have theoretically shown that MOL-TS enjoys a sub-linear regret guarantee while choosing a better arm that maximizes the total reward.
-
The authors have also empirically validated the performance (lower regret and higher cumulative rewards) of MOL-TS using synthetic problem instances.
The following are the weaknesses of the paper:
-
Assumption: Assuming a linear reward function may not be practical in real-life applications. Since there has been a lot of work considering the non-linear bandit setting, it would be much better to have a more practical algorithm by considering a general reward function.
-
Selecting the best arm: Selecting an effective Pareto optimal arm can lead to a better performance than simply selecting a Pareto optimal arm. Since there can be multiple effective Pareto-optimal arms, it is unclear how to choose a single best arm among them. Doing this may need additional constraints, but it may make it more practical, as one can explain why a specific effective Pareto-optimal arm is selected. Otherwise, different effective Pareto-optimal arms may lead to different values of the objectives.
-
Analysis novelty: The multi-objective linear contextual bandits and Thompson sampling-based bandit variants are now separately well-studied problems. From the paper, it is unclear what the key challenges are faced in the regret analysis, especially after getting an upper bound on the effective Pareto sub-optimality gap in Section 5.2.
问题
Please address the weaknesses raised in Strengths And Weaknesses*.
局限性
I have raised a few limitations of the paper in my response to the Strengths And Weaknesses*. Since the paper is a theoretical contribution to bandit literature, I do not find any potential negative societal impact of this work.
最终评判理由
The authors’ rebuttal has addressed my concerns. Overall, this paper proposes, for the first time, a Thompson sampling–based algorithm for the multi-objective contextual bandits setting. This work can serve as a fundamental building block for designing bandit algorithms in more complex real-world applications, including modeling nonlinear objectives or incorporating refined criteria for the optimal arm from Pareto optimal arms.
格式问题
I found no major formatting issues in this paper.
We would like to sincerely thank the reviewer for taking the time to review our paper. We address the feedback you provided point by point below.
Why linear assumption? : We appreciate the reviewer’s feedback regarding the assumption of a linear reward function, but we believe this is not a fundamental weakness of the paper. Analyzing Thompson Sampling in multi-objective bandits (particularly in contextual bandits including linear and generalized linear settings) has remained as an open problem, and to the best of our knowledge, for the first time, our work presents a fundamental milestone in the development of randomized algorithms with the (worst-case) Pareto regret guarantees in multi-objective linear contextual bandits.
The focus of our paper is providing the first randomized algorithm, with Pareto regret analysis and empirical validation in linear reward function. Extending our approach to the (non-linear) GLM is straightforward. Hence, our results aren't limited to linear settings. Further extension to handle more general function class is certainly an interesting direction. However, it is important to note that the optimistic sampling that we propose for Pareto regret analysis can be naturally generalized to other non-linear reward functions. Our work is a noteworthy contribution that sets the foundation for future work with more complex settings. Overall, the fact that we have solved this long-standing open problem in the multi-objective linear contextual bandit should be considered a notable contribution, rather than considered as a weakness.
What is a best single arm? : In multi-objective bandits—and, more broadly, in multi-objective optimization—a single best optimal solution is generally not guaranteed to exist. As noted in our Introduction, multiple objectives often conflict, so improvement in one objective can degrade another. The appropriate notion of optimality is therefore Pareto optimality: the set of solutions that cannot be improved in any objective without worsening at least one other.
If a unique “best” arm did exist under some scalarized criterion, the problem would degenerate to a single-objective bandit setting—contradicting the essence of multi-objective learning. Prior work thus focuses on identifying the Pareto-optimal set (the Pareto front), where each arm is non-dominated by the rest. However, simply enumerating Pareto arms does not address how to maximize cumulative reward across all objectives. Our contribution is to introduce effective Pareto-optimal arms, together with an algorithm that attains sub-linear regret for this richer goal. This adds a fresh perspective to the multi-objective bandit literature.
Should one wish to impose problem(or user)-specified weights and reduce the problem to a single scalar objective, our algorithm naturally specializes to that case and still guarantees sub-linear regret. Nevertheless, that scenario is not the focus of this paper, and therefore it should not be viewed as a weakness of our work.
Technical challenges in analysis : We are happy to elaborate on this. First of all, a simple, straightforward combination of Thompson Sampling (TS) techniques with existing analyses for multi-objective linear contextual bandits is not sufficient for multi-objective linear contextual bandits This gap explains why no worst-case regret analysis of TS currently exists for this setting.
As explained in Section 5.3, one of the key technical challenges in deriving the regret bound of Thompson Sampling algorithm for multi-objective linear contextual bandits lies in ensuring the probability of optimism. Previously, there are many papers of multi-objective UCB-type algorithm with Pareto regret analysis (Drugan & Nowe (2013), Tekin & Turgay (2018), Turgay et al. (2018), Lu et al. (2019), Xu & Klabjan (2023)). The analysis of UCB algorithm is almost similar to that of single objective setting, attaining Pareto regret bound where the number of objectives depend up to logarithmic factor. By contrast, deriving comparable guarantees for TS in the multi-objective linear contextual setting remains an open problem, and our work is the first to tackle these technical obstacles directly.
Suppose we follow standard Thompson Sampling algorithm, by setting . Widely known in single objective setting , the probability that is sufficiently optimistic is , where is a constant probability. In multi-objective setting , as the arm is randomly selected, this optimism must hold for all objectives. For any randomly selected arm, we need to ensure this optimism for all parameters , i.e., . In the worst-case, this results in the regret bound , which become exponentially large as the number of objectives increases.
The optimistic sampling resolves this problem with multiple sampled parameters. Suppose . For each objective , there are sampled parameters . The arm is evaluated with the sampled parameter that maximizes the evaluation, i.e., . Then, the probability bounds for optimism become . Following above steps, we have the probability of optimism at least . We choose large enough so that this probability is bounded below by .
We modify Thompson sampling algorithm with optimistic sampling to ensure the optimism and attaining the regret bound . Addressing this challenging problem requires the development of a novel approach and perpectives. We would be more than happy to include more explanations in Section 5.3 in the final version to further strengthen our contributions.
Overall remarks : We appreciate the reviewer’s feedback, but we do not believe the points raised constitute fundamental weaknesses of our paper. We are confident that our work makes substantial contributions to the multi-objective contextual bandit literature and respectfully ask the reviewer to re-evaluate our work in light of our clarifications.
If any questions remain, please let us know, and we would be more than happy to provide further explanations.
Dear Reviewer fjDD,
Thank you again for your service and the time you invested in reviewing our submission.
We have addressed each of the points you raised in our rebuttal and provided detailed clarifications. With fewer than 36 hours remaining in the discussion period, we would be grateful if you could consider our responses and let us know whether any questions remain.
We would sincerely appreciate your re-evaluation of the paper in light of the clarifications provided, and we are ready to answer any further questions you may have before the discussion window closes.
Sincerely,
Authors
Thank you for your detailed rebuttal. Since all my concerns have been addressed, I am increasing my rating. Please consider incorporating these points into the revised paper, especially highlighting your main technical contributions.
The paper studies the multi-objective linear contextual bandit problem. The paper introduces a new concept effective Pareto regret and proposes a Thompson sampling based algorithm relying on the principle of optimism, The paper proves the effective Pareto regret (hence the Pareto regret) of the algorithm. The paper concludes with experiments.
优缺点分析
Strength:
- Significance: The paper establishes the first TS based algorithm with worst case Pareto regret guarantee, and the regret bound matches the best known order that in the single objective setting. The multi-objective is very common in real word scenarios but less studied.
- Clarity: The presentation leading up to the introduction of the algorithm is very clear for someone like me who do not have the extensive background in multi-objective settings.
- Originality: I really like the optimistic sampling intuition
Weakness:
- I think this is minor but I think it would be interesting to see some stretch analysis on the number in the appendix (or in the main text). Is there any reason that you omitted the comparison between and when in the appendix? What if keeps getting larger.
问题
- This is still on the problem of . I am curious why in the appendix if we keep the same, the difference between and is larger when is larger?
- I am also a bit confused about why does not seem to be that bad given the algorithm relies on optimistic sampling.
局限性
Yes
最终评判理由
All my concerns are resolved and I vote for acceptance.
格式问题
None
We thank the reviewer for taking the time to review our paper. We sincerely appreciate the insightful comments, particularly the recognition of the optimistic sampling. We view this as an excellent opportunity to further clarify and strengthen our contributions, ensuring a more thorough understanding of our approach and its implications.
Streching Analysis on the number : We thank the reviewer for highlighting this issue. We will gladly revise the technical lemmas in Appendix C to provide a more detailed explanation of how the number of objectives affects the regret analysis of our algorithm.
The Comparison between and , for : We are also grateful for this feedback and happy to show our results. Due to the limited space (by large number of objectives), we omitted the comparison between and . We will add the results of additional experiments comparing these two settings. However, when , the results demonstrate that Thompson Sampling with optimistic sampling clearly exhibits lower effective Pareto regret. The following table is the experimental results of total regret when and .
| Pareto Regret | ||
|---|---|---|
| 3.72 | 3.29 | |
| 10.53 | 8.81 | |
| 24.39 | 17.84 |
| Effective Pareto Regret | ||
|---|---|---|
| 4.53 | 3.89 | |
| 13.81 | 11.39 | |
| 32.16 | 23.96 |
| Pareto Regret | ||
|---|---|---|
| 2.42 | 1.57 | |
| 6.59 | 4.15 | |
| 11.87 | 8.73 |
| Effective Pareto Regret | ||
|---|---|---|
| 2.96 | 1.91 | |
| 8.86 | 5.43 | |
| 16.45 | 11.97 |
The difference with respect to : As the dimension increases, the theoretical worst-case regret increases as well, due to the higher uncertainty in the parameter estimation. Empirically, this behavior is consistent with our experiments, as the regret gap between and grows more significant.
Why does not seem to be bad? : The worst-case theoretical regret bound does not imply that the algorithm must fail to the worst-case empirical results. It is essential to note that without optimistic sampling, the theoretical worst-case regret could grow exponentially. We resolve this issue by adopting optimistic sampling in , which mitigate the potential for exponential growth in regret. However, the empirical results do not necessarily be the worst-case scenario. While optimistic sampling still achieving better outcomes empirically, optimistic sampling is employed in to resolve the theoretical issue and ensure better performance in the worst-case scenario.
Thanks for the response. I don't have any remaining issues.
This paper studies the problem of cumulative reward maximization in multi-objective stochastic linear contextual bandits. The paper observes that the conventional definition of Pareto regret does not take into account the optimality of cumulative rewards. To address this issue, the authors introduce the notion of Effective Pareto optimality of cumulative reward vectors. Building on this, they propose an algorithm called MOL-TS, based on Linear Thompson Sampling, which aims to minimize the Effective Pareto regret. In MOL-TS, regularized least squares are performed for each objective, and for each objective samples are drawn from the posterior distribution. By taking the maximum, each parameter is evaluated optimistically. The authors also prove a worst-case regret bound. The key algorithmic and analytical novelty lies in incorporating this form of optimism, which prevents the regret coefficient from growing exponentially with respect to . However, this appears to be essentially the only substantial contribution of the paper.
After the review and discussion phases, all reviewers expressed a positive evaluation of this paper. While there were questions about the technical novelty, comparisons with UCB-type algorithms, and whether the dependency on is optimal, the authors’ responses were found satisfactory, and several reviewers raised their scores accordingly.
As also pointed out by Reviewer EmBN, the current theorem statement (Theorem 2) in the main part of the paper does not make the explicit dependence on important parameters such as entirely clear. If possible, presenting such dependencies directly in the main part would help make the paper’s main contributions more transparent.