PaperHub
5.5
/10
Poster4 位审稿人
最低5最高6标准差0.5
5
5
6
6
2.8
置信度
正确性3.0
贡献度2.8
表达2.8
NeurIPS 2024

Contextual Decision-Making with Knapsacks Beyond the Worst Case

OpenReviewPDF
提交: 2024-05-14更新: 2024-12-19

摘要

We study the framework of a dynamic decision-making scenario with resource constraints. In this framework, an agent, whose target is to maximize the total reward under the initial inventory, selects an action in each round upon observing a random request, leading to a reward and resource consumptions that are further associated with an unknown random external factor. While previous research has already established an $\widetilde{O}(\sqrt{T})$ worst-case regret for this problem, this work offers two results that go beyond the worst-case perspective: one for the worst-case gap between benchmarks and another for logarithmic regret rates. We first show that an $\Omega(\sqrt{T})$ distance between the commonly used fluid benchmark and the online optimum is unavoidable when the former has a degenerate optimal solution. On the algorithmic side, we merge the re-solving heuristic with distribution estimation skills and propose an algorithm that achieves an $\widetilde{O}(1)$ regret as long as the fluid LP has a unique and non-degenerate solution. Furthermore, we prove that our algorithm maintains a near-optimal $\widetilde{O}(\sqrt{T})$ regret even in the worst cases and extend these results to the setting where the request and external factor are continuous. Regarding information structure, our regret results are obtained under two feedback models, respectively, where the algorithm accesses the external factor at the end of each round and at the end of a round only when a non-null action is executed.
关键词
Contextual Decision-Making with KnapsacksRe-Solving

评审与讨论

审稿意见
5

This paper considers the problem of dynamic decision making with resource constraints. They show that under certain conditions they can achieve O(1) regret with respect to the time horizon. However, I am not sure whether this is true regret (i.e. with respect to the best policy in hindsight) or regret with respect to a number VFLV^{FL}. My final score is highly dependent on the author’s response to this question.

优点

I am wondering if in Theorem 3.1 the display equation should say VONV^{ON} instead of VFLV^{FL}. If it is indeed VFLV^{FL} then your goal is only comparing to VFLV^{FL} so what is the point of the “worst case” discussion comparing VFLV^{FL} to VONV^{ON}? If is is meant to be VONV^{ON} in Theorem 3.1 then I think that this is an interesting result. However, if it is indeed VFLV^{FL} then the result is much weaker as you have not shown any improvement over O(T)O(\sqrt{T}) with respect to the optimal policy. My final score will be highly dependent on your answer to this question.

缺点

If Theorem 3.1 is written correctly then please see the “strengths” section for a major weakness.

In Theorem 3.1 I believe that the O(1) must be hiding problem-dependent constants. I don’t just mean constants like the sizes of the sets but constants that are based on the linear program itself. I.e. are there some non-degenerate linear programs in which the hidden constant factor blows up arbitrarily high - e.g. when we limit to degeneracy. This is the same concept as in stochastic bandits: the O(Kln(T))O(Kln(T)) factor hides the gap between the mean rewards of the optimal and second-optimal arms - which limits to infinity as the gap limits to zero (although the regret actually never goes above O(KT)O(\sqrt{KT})). You should certainly include any problem-dependent constants in your bound (although I do understand they may be complex to bound - so at least point out that they exist (if they do)).

Line 226 (and line 78): You say that any LP can easily escape from degeneracy. Doe this mean that you can improve on the sqrt{T} bound when J(ρ1)J(\rho_1) is degenerate? Otherwise this statement (which appears twice) is highly misleading.

Line 171: Theorem 2.1 does not show an \Omega(\sqrt{T}) regret lower bound when you define the regret relative to VFLV^{FL}.

问题

Can you give examples to show that the non-degeneracy of J(ρ1)J(\rho_1) is common.

局限性

N/A

作者回复

Thanks for your review and comments. We now respond to your concerns and questions.

Theorem 3.1.
It is completely correct for you to say that in the equation of Theorem 3.1, VFLV^{\mathrm{FL}} could be replaced by VONV^{\mathrm{ON}}. In other words, we are talking about the true regret, and we indeed obtain an O(1)O(1) true regret under Assumption 3.1. This is also true for all our regret results, including Theorems 4.1, 5.1, 5.2, B.1, and B.2. We hope that our claim here could eliminate your concerns.

Nevertheless, we should notice that from Proposition 2.1, we know that VFLVONV^{\mathrm{FL}} \geq V^{\mathrm{ON}}, and of course, we have VONRewV^{\mathrm{ON}} \geq Rew. Thus, the formulae we write in this paper are, in fact, stronger in the sense that if VFLRew=O(f(T))V^{\mathrm{FL}} - Rew = O(f(T)), then VONRew=O(f(T))V^{\mathrm{ON}} - Rew = O(f(T)) also holds. For Theorem 2.1, we only want to know how far VFLV^{\mathrm{FL}} could be above VONV^{\mathrm{ON}} at worst and the corresponding conditions so as to make our work more complete.

Problem-dependent constants in Theorem 3.1.
It is true that our regret bound consists of problem-dependent variables, as we discussed in Lines 233--258. More specifically, as you mentioned, we have a constant DD in the bound based on the linear program J(ρ1)J(\rho_1). This constant represents the minimum LL_\infty distance between the LP J(ρ1)J(\rho_1) and any LP with multiple optimal solutions or a degenerate optimal solution (Lines 237--238). When our CDMK problem degenerates to stochastic decision-making by removing the context and resource constraints, DD is precisely half the gap between the mean rewards of the best and second-best arm. Thus, DD resembles the reward-gap-like parameter in the multi-armed bandit literature (Lines 252--253). And our bound goes quadratically with 1/D1/D (Line 258, with a typo, we will correct that.). We will work on to make our arguments clearer.

Escaping from degeneracy.
In fact, we are not saying that we can improve the O(T)O(\sqrt{T}) regret under degeneracy. We mean that in practice, we can prevent the real-world problem from J(ρ1)J(\rho_1) being degenerate by minorly changing the resource constraints, e.g., neglecting some of the resources. We will show in the later response that non-degeneracy for J(ρ1)J(\rho_1) is common. We will clarify this part, of course, to eliminate any misleading information. Thanks a lot for pointing that out.

Regret lower-bound.
You are correct. We will surely clarify these arguments by clearly differentiating the "true regret" and the regret we define in this work's main body.

Non-degeneracy for J(ρ1)J(\rho_1) is common.
We simplify the example that we use in our numeric validations (Appendix G) by disregarding the third context, renormalizing, and letting the first resource constraint be 0.750.75 and the second be yy. Then the LP J(ρ1)J(\rho_1) is now maximizing 0.3x1+0.36x20.3x_1 + 0.36x_2, under the constraint that 0.3x1+0.6x20.750.3x_1 + 0.6x_2 \leq 0.75, 0.6x1+0.3x2y0.6x_1 + 0.3x_2\leq y, and 0x1,x210\leq x_1, x_2\leq 1. This LP always has a unique solution, and the only value y>0.6y > 0.6 such that J(ρ1)J(\rho_1) is degenerate is y=0.825y = 0.825, with the degenerate solution being (x1,x2)=(1,0.75)(x_1, x_2) = (1, 0.75). That is to say, if we suppose that yy is uniformly located between 0.6 and 1, for example, then J(ρ1)J(\rho_1) is almost surely non-degenerate. I hope this can address your question, and this is the reason why we say that any LP can easily "escape" from being degenerate. Nevertheless, we will correct the misleading sentences.

评论

Hi, we are looking forward to seeing your opinion on our rebuttal!

评论

Yes - sorry - of course you can replace VFLV^{\mathrm{FL}} with VONV^{\mathrm{ON}} - I was thinking "losses" rather than "rewards". I am therefore increasing my score - but I feel strongly that, if this paper is accepted, you should include all problem-dependent constants in Theorem 3.1

评论

Thanks a lot! We will surely improve our manuscript on noting more about problem-dependent constants.

审稿意见
5

This paper studies contextual decision making with Knapsack constraints assuming that the requests and the contexts follows some distributions. It studies the full information setting when each context is revealed after the decision is made and the partial information setting when the context is revealed only when a non-null decision is taken. The paper provides regret bounds under various settings.

优点

Under the unique non degenerate LP assumption, the paper has provided regret bounds that are better than the worst cases in both full information and partial information setting, which is novel. It further shows that without this assumption, the regret bound is indeed O(T)O(\sqrt{T}) that matches the lower bounds. The paper further provides regret bounds when the request and context are continuously distributed. It greatly enhances the understanding of the problem.

缺点

  1. I found the presentation can be further improved, particularly about why the unique non-degenerate solution can greatly improve the regret bound and provide a proof sketch. This is the main contribution of the work yet the intuition is not fully clear.
  2. See questions below.

问题

  1. In the motivating example, it is unclear why for the dynamic bidding problem, the highest competing bid would follow a distribution. It seems to be more adversarially chosen.
  2. In the partial information setting, since choosing a null action will not be able to observe the context, does the v^t(γ)\hat v_t(\gamma) still provide an unbiased empirical estimator of the distribution of γ\gamma? Since the decision depends on previous observed contexts, observing the next context or not depends on the previous observations. Thus v^t(γ)\hat v_t(\gamma) does not seem to be an unbiased empirical estimator anymore. What are the techniques used to address such bias?
  3. Since the context γ\gamma is revealed after the decision is made, is it worthy to take more non-null decisions in the early stages and try to learn the distribution faster?
  4. The relationship of the proposed problem comparing to the NRM problem is not explicitly discussed. Please compare the similarities and differences, particularly regarding the measurements.
  5. The linear programming is only solvable when θ\theta and aa admit finite support. What if they are continuous variables? Continuous variables are quite common in both inventory and bidding situation.

I am happy to raise the score if these issues are well addressed.

局限性

Comparison to continuous optimization methods in NRM problems are not discussed.

作者回复

Thanks for your comments and questions. We will now answer them.

Why uniqueness and non-degeneracy is important.
The corresponding discussion is located in Lines 233--243 in our paper. In short words, under this assumption (Assumption 3.1), we have the stability property that if all estimated parameters in J^(ρt,Ht)\hat{J}(\rho_t, \mathcal{H}_t) are not far away from the true parameters, then J^(ρt,Ht)\hat{J}(\rho_t, \mathcal{H}_t) and J(ρ1)J(\rho_1) have the same set of basic variables and binding constraints, respectively. This property is given in Lemma D.2 in the appendix and plays a crucial role in our analysis. We will work on to make that clearer.

Stochastic highest competing bids in the dynamic bidding problem.
This is a good question. Suppose we stand in the view of a single bidder and consider a large auction market. In this scenario, even if each of the other bidders bid strategically, their highest bid can still be regarded as being stochastic due to the large market. This is sometimes referred to as the "mean-field approximation" in economics. In the related literature, such a stochastic assumption on the highest competing bids is also widely taken, for example, Balseiro and Gur (2019), Feng, Padmanabhan, and Wang (2023), and Chen et al. (2024).

Unbiased estimation of v^t\hat{v}_t.
This is an important question. In fact, V^t\hat{\mathcal{V}}_t is still an unbiased estimation of V\mathcal{V} even under partial feedback. The reasoning here is that γt\gamma_t is drawn independently with the context θt\theta_t, and the action ata_t is also chosen before γt\gamma_t is revealed. Thus, we have that Pr[γtθt,at]=Pr[γt]\Pr[\gamma_t | \theta_t, a_t] = \Pr[\gamma_t], or that γt\gamma_t is independent with (θt,at)(\theta_t, a_t). This guarantees that V^t\hat{\mathcal{V}}_t is unbiased.

More non-null decisions in the beginning.
This is a very good thought experiment. However, although it seems delightful to make more non-null decisions initially, we have to ensure that we are not "exploring" too much or that we could be too far from the optimal and may not "rescue back" in the later time slots. In fact, our key ingredient for the proof is Lemma 4.1, which indicates that under the re-solving heuristic, we explore in every O(logT)O(\log T) rounds and have an O(1)O(1) overall exploring frequency in the beginning. (Please also refer to Figure 1.) This is already asymptotically much efficient for getting samples. You raised a very important question, and we will see if we can do better by slightly modifying the algorithm and obtaining a better regret bound.

Relationship between CDMK and NRM.
Due to the space limit, we defer the discussion to Appendix A, Lines 499--512. Put simply, NRM is a sub-problem of our CDMK problem with no external factors and only a binary action space A={0,1}A = \{0, 1\}, where 00 is the null action and 11 is the "accept" action. Therefore, our CDMK is generally "harder" than the NRM problem. As for the measurements, our fluid benchmark is identical to the deterministic LP (DLP) benchmark introduced in classical works on NRM, e.g., Jasin and Kumar (2012) and Balseiro, Besbes, and Pizarro (2023). Thus, we believe that our work is built on a strong literature basis.

Continuous variables.
This is also an important question. In Appendix B, we suppose that there is an outside oracle that can help us solve the programming in each round for continuous variables. Yet, this could be hard in practice if we are not facing a convex continuous programming. Related works in the literature would mostly suppose a finite context and external factor set or similarly suppose an oracle for continuous variables, e.g., Balseiro, Besbes, and Pizarro (2023). Also, in real-life scenarios, e.g., inventory or bidding, even though the action space is large, it is still finite since the allocation/bid is usually required to be a multiple of a minimum unit.

Reference:
Balseiro, S. R., & Gur, Y. (2019). Learning in repeated auctions with budgets: Regret minimization and equilibrium. Management Science, 65(9), 3952-3968.
Feng, Z., Padmanabhan, S., & Wang, D. (2023, April). Online Bidding Algorithms for Return-on-Spend Constrained Advertisers. In Proceedings of the ACM Web Conference 2023 (pp. 3550-3560).
Chen, Z., Wang, C., Wang, Q., Pan, Y., Shi, Z., Cai, Z., ... & Deng, X. (2024, March). Dynamic budget throttling in repeated second-price auctions. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 38, No. 9, pp. 9598-9606).
Jasin, S., & Kumar, S. (2012). A re-solving heuristic with bounded revenue loss for network revenue management with customer choice. Mathematics of Operations Research, 37(2), 313-345.
Balseiro, S. R., Besbes, O., & Pizarro, D. (2023). Survey of dynamic resource-constrained reward collection problems: Unified model and analysis. Operations Research.

评论

Thanks for your responses! It addresses my questions well. I will raise the score to 5.

评论

Thanks a lot!

审稿意见
6

This paper studies an online contextual optimization problem with resource constraints. The paper provides a sufficient condition (worst case condition) under which the fundamental limit on the regret bound is reached. The paper further provides an algorithm that can achieve \tilde O(1) regret when the worst case condition does not hold. Numerical results are also provided to validate the theory.

优点

The paper is well written. The relation between the worst case condition and the degeneracy of linear constraints is novel and inspiring. The intuition behind the proposed algorithm is clearly explained.

缺点

  1. Is TT known beforehand? If TT is unknown, how to compute the leftover budget ρt\rho_t? If TT is known, how to design an algorithm with unknown TT?

  2. What's the difference between Rew and V^{on}?

  3. if there is ρi\rho^i amount of resources per stage, why does the problem formulation only respect the resource constraint at stage TT, instead of considering resource constraint at every stage tt, i.e. total resources available to be used in stages tt <= ρit\rho^i t for all tTt\geq T?

  4. In the simulation, how does the proposed algorithm compare with the existing algorithms proposed for this problem?

问题

  1. It is slightly confusing to use ρi\rho^i to denote the iith entry since it also means exponent.

局限性

NA

作者回复

Thanks for appreciating our work. We will now answer your questions.

Known TT.
TT is known beforehand in this work, which is a common assumption in the literature, as supposed by the survey of Balseiro, Besbes, and Pizarro (2023). We have not yet considered the problem of unknown TT, which is certainly an important future direction.

RewRew and VONV^{\mathrm{ON}}.
RewRew is the expected total reward of our algorithm, while VONV^{\mathrm{ON}} is the expected total reward of the optimal algorithm.

Resource constraint for each stage.
This is a very interesting question. In practice, only the global constraint is taken since we usually have an initial inventory of resources and only require that all the allocations should be done concerning this initial inventory. Thus, a resource constraint for the later stages is unnecessary. Considering resource constraints for every stage will make the problem much more difficult, as an algorithm cannot sufficiently explore all actions in the beginning stages. This is certainly a future extension of our model.

Simulation.
In this work, we do not compare our algorithm with other existing algorithms mainly because CDMK with partial feedback is a new model. Further, the regret results highly rely on instance-dependent factors for both existing algorithms and ours. Altogether, it is a bit unfair to conduct such a comparison under only some instances since they could be biased. Our simulation results mainly aim to justify our theoretical regret bounds, and we do see a match there.

Usage of ρi\rho^i.
We are sorry for causing such confusion. We will work on improving the notation.

Reference:
Balseiro, S. R., Besbes, O., & Pizarro, D. (2023). Survey of dynamic resource-constrained reward collection problems: Unified model and analysis. Operations Research.

评论

Thanks for your responses! I will keep my score.

评论

Thanks a lot!

审稿意见
6

This paper considers a new contextual decision-making model with knapsack constraints, which is highly related to the CBwK setting but features a different information feedback structure. Under this model, the authors nearly characterize the conditions under which O~(1)\tilde{O}(1) regret can be achieved based on the degeneracy of the optimal solution of the fluid LP. Specifically, the re-solving heuristic is proposed to achieve O~(1)\tilde{O}(1) regret under these conditions. Additionally, the O~(T)\tilde{O}(\sqrt{T}) worst-case regret of the algorithm is provided.

优点

  1. The model considered in the paper is quite general to cover several interesting problems.

  2. The results provided under the proposed model is relative sharp and complete.

缺点

  1. The assumption on the randomness is stronger than those made in the contextual decision-making literature.

  2. While I understand that this paper considers a new setting and that the works in CBwK are the most related, it is acceptable that the authors mainly compare their results with those in CBwK. However, since the information feedback assumption in this paper is strictly stronger than in CBwK, I suggest the authors provide more comments regarding this difference to make the comparison fairer.

问题

Both the feedback structure and the degeneracy-based condition for breaking O(T)O(\sqrt{T}) regret make me think it is similar to the corresponding results in online linear programming [1,2,3]. While several structures are different (e.g., the context and the action set), are there any high-level heuristics that can explain such similarity? Moreover, [3] provides a condition for breaking O(T)O(\sqrt{T}) regret beyond degeneracy. Can the idea in [3] be utilized in this setting to further characterize the conditions for breaking O(T)O(\sqrt{T}) regret in this paper?

Ref:

[1] Arlotto A, Gurvich I. Uniformly bounded regret in the multisecretary problem[J]. Stochastic Systems, 2019, 9(3): 231-260.

[2] Bray R L. Logarithmic Regret in Multisecretary and Online Linear Programs with Continuous Valuations[J]. arXiv preprint arXiv:1912.08917, 2019.

[3] Jiang J, Ma W, Zhang J. Degeneracy is ok: Logarithmic regret for network revenue management with indiscrete distributions[J]. arXiv preprint arXiv:2210.07996, 2022.

局限性

N.A.

作者回复

Thanks for your kind comments and suggestions. We now respond to your concerns and questions.

Stronger assumptions on the randomness.
Indeed, our model requires an explicit form of randomness γ\gamma for the external factor. This explicit model is crucial for us in this work to learn its distribution. We will work on to see how to relax this assumption.

Comparison with CBwK.
Due to the space limit, we have to delay the comparison between our model and CBwK to Appendix A, Lines 469--498. There, we emphasize the difference between our results and results in CBwK, as well as the difference in feedback models. In short, although our full/partial feedback models are stronger than the bandit feedback model, techniques in the literature for CBwK would require assumptions on the problem structure, e.g., linear dependence between expected rewards/cost vectors and the action (linear CBwK). With these model assumptions, existing regression modules can help to address the problem of learning unknown parameters. In our work, we do not take any of these simplifying assumptions and no learning oracles (Line 492); therefore, we would require a stronger feedback model to ensure the efficient learning of randomness.

In fact, we give a novel, important, and non-trivial analysis for the learning guarantee under partial feedback (Lemma 4.1), which is crucial for our regret results in this model. Lemma 4.1 itself could also be of independent interest.

Degeneracy-based conditions, and other heuristics.
This is a very interesting topic. In general, we believe that non-degeneracy is a critical factor to an O(T)O(\sqrt{T}) regret bound in the CDMK/CDwK problems, at least for the re-solving technique we use in this work (e.g., Jasin and Kumar (2012)), as pointed out by Bumpensanti and Wang (2020).

To break the O(T)O(\sqrt{T}) regret without the degeneracy assumption, we should explore other heuristics that are already discovered in the online linear programming problem, as you mentioned, and see whether other conditions could be characterized for breaking the O(T)O(\sqrt{T}) regret. Despite this, our work is an important first trial of extending the re-solving technique to the CDMK problem.

Reference:
Jasin, S., & Kumar, S. (2012). A re-solving heuristic with bounded revenue loss for network revenue management with customer choice. Mathematics of Operations Research, 37(2), 313-345.
Bumpensanti, P., & Wang, H. (2020). A re-solving heuristic with uniformly bounded loss for network revenue management. Management Science, 66(7), 2993-3009.

评论

Thank you for your response. I will keep my score positive.

评论

Thanks a lot!

评论

Hi, we are looking forward to seeing your opinion on our rebuttal!

最终决定

This paper studies the problem of maximizing the total reward under resource constraints in the online setting. The paper establishes an Ω(T)\Omega(\sqrt{T}) lower-bound on the worst-case regret. Under a non-degeneracy assumption (that helps avoid the worst-case scenario), the paper proposes an algorithm that achieves O(1)O(1) regret in the full information setting, and O(log(T))O(\log(T))) regret in the partial information setting.

The reviewers agree that the paper is well-written, and its contributions merit acceptance. After carefully reading the paper and the corresponding discussion, I tend to agree. Please incorporate the reviewers' feedback. In particular, addressing the following concerns will help strengthen the paper:

  • Move the comparison to the CBwK from the Appendix to the main paper (response to Rev. c8Hc)
  • Add a more detailed comparison to the papers (pointed by Rev. c8Hc) that break the O(T)O(\sqrt{T}) lower-bound without the non-degeneracy assumption.
  • More intuition/explanation on why uniqueness and non-degeneracy help break the O(T)O(\sqrt{T}) lower-bound (response to Rev. TgME).
  • Move the comparison to NRM from Appendix A to the main paper (response to Rev. TgME).
  • Include problem-dependent constants in the theorems, and explain their dependence (response to Rev. ZSDZ)