PaperHub
6.8
/10
Poster4 位审稿人
最低4最高5标准差0.4
4
4
5
4
3.0
置信度
创新性2.8
质量2.8
清晰度3.0
重要性2.5
NeurIPS 2025

Variance-Reduced Long-Term Rehearsal Learning with Quadratic Programming Reformulation

OpenReviewPDF
提交: 2025-05-03更新: 2025-10-29
TL;DR

We present the first long-term rehearsal learning approach, which demonstrates favorable properties such as variance reduction and optimality.

摘要

In machine learning, a critical class of decision-making problems involves *Avoiding Undesired Future* (AUF): given a predicted undesired outcome, how can one make decision about actions to prevent it? Recently, the *rehearsal learning* framework has been proposed to address AUF problem. While existing methods offer reliable decisions for single-round success, this paper considers long-term settings that involve coordinating multiple future outcomes, which is often required in real-world tasks. Specifically, we generalize the AUF objective to characterize a long-term decision target that incorporates cross-temporal relations among variables. As directly optimizing the *AUF probability* $\mathbb{P}_{\operatorname{AUF}}$ over this objective remains challenging, we derive an explicit expression for the objective and further propose a quadratic programming (QP) reformulation that transforms the intractable probabilistic AUF optimization into a tractable one. Under mild assumptions, we show that solutions to the QP reformulation are equivalent to those of the original AUF optimization, based on which we develop two novel rehearsal learning methods for long-term decision-making: (i) a *greedy* method that maximizes the single-round $\mathbb{P}_{\operatorname{AUF}}$ at each step, and (ii) a *far-sighted* method that accounts for future consequences in each decision, yielding a higher overall $\mathbb{P}_{\operatorname{AUF}}$ through an $L/(L+1)$ variance reduction in the AUF objective. We further establish an $\mathcal{O}(1/\sqrt{N})$ excess risk bound for decisions based on estimated parameters, ensuring reliable practical applicability with finite data. Experiments validate the effectiveness of our approach.
关键词
decision-makingprobabilistic optimizationstructural model

评审与讨论

审稿意见
4

The paper introduces the AUF (avoiding undesirable futures) problem for maximizing the probability of a sequence of future outcomes being in a certain set S. They work on a simple case of the problem - a linear markovian transition structure that is essentially a linear dynamical system. This allows them to formulate the problem as a QP, which they solve with both a greedy and a far-sighted method.

优缺点分析

Strengths:

  1. The paper extends rehearsal learning from single-round AUF to multi-round, long-term settings, which better reflects many practical problems (e.g., portfolio management, sequential decision-making).
  2. It establishes optimality of the QP solution under reasonable (but strong) assumptions and derives variance reduction rates (Theorem 3.5) that are directly relevant for sequential AUF objectives.
  3. The quadratic programming reformulation provides a computationally efficient way to handle the long-term AUF problem, avoiding sampling-based or MILP approaches that scale poorly with the dimensionality of variables.
  4. The experiments, while limited, clearly show that the proposed far-sighted method achieves substantially higher AUF probabilities than both baselines and existing rehearsal learning methods across multiple datasets.

Weaknesses:

  1. I think the linear assumption that crucially allows the paper to design a QP formulation of the AIUF problem is quite strong, and given the lack of experiments, it is unclear how well it holds up.
  2. The linear assumption also turns this problem into a linear dynamical system, unless I am mistaken. There is extensive literature in control theory, to my understanding, which addresses the problem of controlling a linear dynamical system to let it end up in a desirable set. I am not sure if the exact problem in this work is addressed in control theory literature, but the work is worth diving into and reference.
  3. Many arguments about the ineffectiveness of RL are about its stated “inability to use existing knowledge about systems.” I think this is incorrect for model-based RL. In fact, an experimental comparison with model-based RL methods under the same linear model is missing. The argument for reward drift seems incorrect to me - you can just have the reward be the indicator function of the desired set S, making it a deterministic reward with no distribution shift. The actual value of the reward now just depends on the observed Y_t.

问题

  1. Could you address the implicit questions within my discussion of the paper’s weaknesses?
  2. The noise in the trajectory is non-isotropic, so I find it a bit odd that the way to maximize the probability is to optimize an objective with L2 loss, instead of an L2 loss transformed by a matrix M.

局限性

The paper does not address the strength of the linearity assumption, and I think it also needs to spend some time verifying to what extent the assumption holds in its dataset (to what extent can the best fit linear transition model explain the observations in the dataset).

最终评判理由

The empirical justification of their linearity assumption and the discussion of their relation to MPC have largely satisfied me, but I don’t think they have faithfully addressed their relation to causal RL, as I discuss in my response. So my score had gone from a 3 to a 4 but not to a 5.

格式问题

None

作者回复

Thanks for your detailed feedback and for recognizing the strengths of our work! We hope our responses below can address your concerns.


W1&Limitation. Concerns about the linear assumption.

A1. Thanks for your constructive suggestion. First, we would like to clarify that our linear assumption generalizes the noise distribution to the log-concave family, which is a broader class than that considered in prior works [1, 2]. While this assumption is essential to enabling our QP reformulation, we also further discuss its generality and empirical validity as follows:

  • Potential extensions to non-linear settings.
    The core ideas behind our algorithm are not limited to linear models. Specifically, our decision-making objective only requires that the relationship between the aggregated outcome iYi/T\sum_i Y_i/T and the sequence of alterations zi\\{z_i\\} can be parameterized with additive aggregated noise. That is, if there exists a functional family ff such that iYi/T=f(θ;zi)+ε~,\sum_i Y_i/T = f(\theta; \\{z_i\\}) + \tilde{\varepsilon}, for some parameters θ\theta and aggregated noise ε~\tilde{\varepsilon}, then the optimization at each decision round kk can proceed by minimizing the surrogate loss f(θ;zi)sk22,\\|f(\theta; \\{z_i\\}) - s_k\\|_2^2, where sks_k denotes the adjusted center of the desired target region. This observation suggests that our formulation is not inherently tied to linear dynamics and may naturally extend to some general non-linear settings.

  • Empirical validation of the linearity assumption.
    To assess the adequacy of the linear assumption, we fitted the model using least squares and evaluated its fit using standard metrics: (i) per-coordinate R2R^2 scores, (ii) normalized mean squared error (NMSE), and (iii) normalized mean absolute error (NMAE).

    The linear model achieves average R2R^2 scores of 0.869/0.686, NMSE of 0.212/0.401, and NMAE of 0.305/0.468 on the synthetic / real-world (Bermuda) datasets, respectively. For other datasets, this validation step can also be performed to verify the suitability of the linear assumption.

Taken together, these points support both the current validity and future extensibility of our formulation. We thank the reviewer again for raising this important issue.


W2. Linear dynamical system and control theory.

A2. Thanks for your insightful question. Our approach differs from classical control methods, such as online linear quadratic control (LQR) [3, 4], in several key ways:

  • Different Motivations. Traditional LQR assumes a linear dynamical system of the form xt+1=Axt+But+wtx_{t+1} = Ax_t+Bu_t+w_t and aims to control the state trajectory {xt}\{x_t\} over time. In contrast, our goal is to maximize the probability that the aggregated outcome iYi/T\sum_iY_i/T lies within a desired region, without being concerned about the specific trajectory of XX, YY, or ZZ at individual time points.

  • Distinct Formulations. Our system cannot be reduced to the standard LQR formulation, as the alteration can also affect variables within the same time point. Moreover, our problem is fundamentally defined by Eq. (3), and Eq. (5) is a heuristic reformulation under certain assumptions. Even so, the resulting QP in Eq. (5) differs from that in LQR: while online LQR decomposes into per-step QPs (w.r.t. control variables like t(xtQxt+utRut)\sum_t(x_t^\top Qx_t+u_t^\top Ru_t)) due to additive structure, our objective involves aggregation of altered variables ztz_t through matrix FF, making the problem cannot be decomposed thus more challenging.

  • Connection to MPC. While our method shares a high-level similarity with Model Predictive Control (MPC) [5], in that we optimize a future sequence of actions but only execute the first, our predictive model is different from classic ones. We rely on structured dependencies (see A3) that go beyond the scope of standard MPC formulations.

We will incorporate this discussion into the revised version. Thanks again!


W3. Comparison with RL methods.

A3. Thanks for your thoughtful comment. It appears there may be a misunderstanding of our problem setting, and we appreciate the opportunity to clarify the differences between our structural approach and conventional RL, including model-based RL.

Your question highlights a key advantage of the rehearsal learning framework. Simply formulating a reward function with an indicator I(YS)\mathbb{I}(Y \in S) is insufficient for addressing decision problems with structural dependencies (like causal relations), especially those involving confounding. Even when the indicator function is deterministic, it is crucial to distinguish between samples collected from different data distributions (observation or interaction/intervention), as this distinction fundamentally impacts the decision-making process.

  • Observation vs. Intervention.
    Our setting relies solely on observational data to make immediate decisions, where structural (e.g., causal) relations exist. In contrast, standard RL methods assume interventional data from agent-environment interactions. Treating observational samples x,z,I(YS)\langle x, z, \mathbb{I}(Y \in S) \rangle as if they were standard RL tuples s,a,r\langle s, a, r \rangle implicitly assumes that the observational and interventional distributions are identical, an invalid assumption in the presence of confounding.

    E.g., suppose the structure includes: XYX\rightarrow Y, XZaYX\rightarrow Z_a\rightarrow Y, ZaZuZ_a^\prime\rightarrow Z_u and ZaZuYZ_a\leftarrow Z_u\rightarrow Y, where XX can be viewed as the state in RL, Z=(Za,Za)Z=(Z_a,Z_a^\prime) are two actionable variables, and ZuZ_u is an observed but unactionable confounder. Observational data might show a strong correlation between Z=zZ = z and YSY\in S given X=xX=x, leading an RL agent to learn zz as the optimal action incorrectly. In contrast, our method can identify the optimal action zzz^\star\ne z without requiring costly exploration by leveraging structural information and applying backdoor criteria [10].

  • Structural vs. Statistical Knowledge.
    While model-based RL can incorporate prior knowledge, it typically captures statistical transition dynamics (e.g., p(st+1st,at)p(s_{t+1} \mid s_t, a_t)) rather than fine-grained structural dependencies. Our method leverages the latter, which is essential to distinguish causation/influence relations from purely correlation.

  • New Experiments. To empirically illustrate this distinction, we compared our method with both model-free (DDPG [6], SAC [7]) and model-based (MBPO [8]) RL agents. Each was allowed 100 environment interactions, the same sample budget available to our method from observational data.

    MethodOurs[6][7][8]
    AUF Probability0.827±0.0440.827\pm 0.0440.207±0.0550.207\pm 0.0550.182±0.0390.182\pm 0.0390.228±0.0270.228\pm 0.027

All experiments were run on the same dataset, with the same region SS and horizon 1, across 100 rounds and 10 random seeds. RL methods underperform due to their inability to exploit structural information, resulting in slow convergence. Notably, SAC can eventually reach comparable performance, but only after ~6,000 interactions, underscoring the sample inefficiency of RL in our context.

In summary, our work is motivated by the need to incorporate structural knowledge, which is essential for making decisions in the presence of confounding. We hope this clarifies the distinction and illustrates why our approach is better suited to this class of problems than standard RL pipelines.


Q2. Non-isotropic noise.

A4. Thank you for this insightful question. Your observation is very sharp and touches upon a key aspect of our theoretical contribution.

Indeed, when the target variable YY is multi-dimensional, the noise in the aggregated outcome iYi/T\sum_i Y_i/T is possible to be non-isotropic. However, this does not affect our decision-making strategy. Let us revisit the core objective of our AUF problem: given an observation X=xX=x, we aim to choose an alteration sequence zi\\{z_i\\} to maximize the probability of iYi/TS\sum_i Y_i/T \in S. In this context, our Prop. 3.1 guarantees that the entire stochasticity of the aggregated outcome is captured by the aggregated zero-mean noise term, Fe~F\tilde{e}, which is unaffected by the choice of the choosed action variables zi\\{z_i\\}.

Consequently, our optimization in Eq.(5) is performed over the sequence of action variables zi\\{z_i\\}. The choice of norm for this optimization does not alter the stochastic term Fe~F\tilde{e} that governs the randomness of the outcome iYi/T\sum_i Y_i/T. The reason our method successfully maximizes the probability of the outcome falling into the desired region SS is guaranteed by established results in probability theory, involving key lemmas such as Anderson's Theorem [9]. Should you be interested, a detailed proof is provided in Appx. D.3.


We hope our detailed responses have addressed your concerns. If any points remain unclear, we would be happy to elaborate further. In light of these clarifications, we would be grateful if you would reconsider your evaluation. Thank you again for your thoughtful feedback!


Reference

[1] Rehearsal Learning for Avoiding Undesired Future. NeurIPS 2023.

[2] Avoiding Undesired Future with Minimal Cost in Non-Stationary Environments. NeurIPS 2024.

[3] Online Linear Quadratic Control. ICML 2018.

[4] Naive Exploration is Optimal for Online LQR. ICML 2020.

[5] Model predictive control: Theory and practice—A survey. Automatica 1989.

[6] Continuous control with deep reinforcement learning. ICLR 2016.

[7] Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. ICML 2018.

[8] When to Trust Your Model: Model-Based Policy Optimization. NeurIPS 2019.

[9] The integral of a symmetric unimodal function over a symmetric convex set and some probability inequalities. American Mathematical Society 1955.

[10] Causality: Models, Reasoning and Inference.

评论

Dear Reviewer 2xvd,

We sincerely thank you for your time and feedback on reviewing our paper. As the discussion period is nearing its end, we would like to politely remind you and would greatly appreciate it if you could take a moment to consider our responses.

If there are any remaining concerns about our work, we would be happy to engage in further discussion.

Yours sincerely,
The Authors

评论

I think the reviewers have satisfied my questions about the linearity assumptions with their experiment, convincing me to raise my score from 3 to 4. I must apologize for the delay in my response.

However, I disagree that this problem cannot be viewed as RL. There is rich literature on offline RL under a causal structure (see, for example, Offline Policy Evaluation and Optimization under Confounding by Kausik et al). I think your paper’s qualms about work on causal RL is that it works on reward data, while your qualm with vanilla RL is that when applied to the indicator function, it cannot utilize the causal structure appropriately. Surely, applying causal RL to the indicator function solves both problems. I urge you to faithfully discuss this.

I am largely convinced by your claims about the distinction from MPC. I still think that there is a plausible way to phrase this as a control problem by combining the controls at many time points into one vector, but I suspect that this is high dimensional and not structured enough to be efficient or useful for real life usage.

I am keeping the confidence of my borderline accept score low due to the delay in my response, for which I apologize again. I am still quite confident that you are not faithfully addressing your relation to causal RL, and I strongly encourage you to change this.

评论

Dear Reviewer 2xvd,

We are very glad that our rebuttal was able to address some of your key concerns, and we truly appreciate your constructive suggestions. Regarding the discussion on MPC and control, we are pleased to reach a consensus with you that our method, by leveraging structural information, can address efficiency and practicality issues present in traditional control approaches.

On the point you raised about causal RL (e.g., Offline Policy Evaluation and Optimization under Confounding by Kausik et al.), we will certainly include a careful discussion in the revised version to elaborate on its feasibility as well as its distinctions from our approach.

Thank you again for your time and valuable feedback.

Sincerely yours,
The Authors

审稿意见
4

This paper introduces a novel approach to the Avoiding Undesired Future (AUF) problem by extending the rehearsal learning framework to long-term decision-making scenarios. It reformulates the intractable probabilistic AUF objective, maximizing the likelihood that the average of future outcomes falls within a desired region, into a tractable quadratic programming (QP) problem under mild assumptions such as linearity and log-concave noise. The authors propose two algorithms: a greedy method that optimizes each decision step independently, and a far-sighted method that accounts for future consequences to achieve better variance reduction and overall performance. Theoretical guarantees are provided for the optimality of the QP solution, variance reduction, and robustness when using estimated parameters. Empirical evaluations on synthetic and real-world datasets confirm the proposed methods significantly outperform existing approaches in both accuracy and efficiency.

优缺点分析

Strengths: (1) The paper is well-grounded in theory. The authors provide a principled extension of the AUF framework to long-term settings, supported by a rigorous QP reformulation. (2) The paper is generally well-organized and clearly written. Key concepts are introduced with sufficient motivation, and the technical derivations are presented with clarity. (3) The long-term AUF setting is practically important and aligns with real-world needs, where decisions often have delayed and cumulative consequences.

Weaknesses: (1) While the QP reformulation is elegant, its optimality depends on a set of assumptions (e.g., linearity, unique targets, symmetric desired region). These are practical in many settings, but the paper could benefit from more detailed discussion and empirical evaluation of how violations of these assumptions affect performance. While a counterexample is provided, its role is limited to theoretical discussion. (2) The far-sighted method achieves superior performance but with increased computational cost. While this is acknowledged, a deeper analysis of scalability with respect to horizon length, dimensionality, and system size would improve practical understanding. (3) The comparison excludes reinforcement learning methods based on an assumption of interaction constraints. While justified, the paper would be stronger with at least some controlled comparisons or sensitivity analyses to confirm this assumption empirically, especially since many AUF problems in practice involve some feedback or batch interaction.

问题

  1. What is the practical computational limit (in terms of time horizon 𝑇𝑇, dimensionality Z|Z|, and data size) for applying the far-sighted method in real-time or near real-time applications?
  2. Can the method accommodate settings where the decision horizon TT is not known in advance or varies over time?
  3. The paper proposes algorithms that compute optimal alterations in closed form (via QP), but it remains unclear how interpretable these solutions are. Given the practical relevance (e.g., in policy or finance), can the authors discuss whether the alteration vectors can be explained to domain experts, or whether some sparsity constraints or diagnostics could be incorporated?
  4. While the paper benchmarks against previous rehearsal-based methods (QWZ23, MICNS), the experimental comparison omits broader classes of structured decision-making or robust control algorithms. Can the authors clarify why alternatives were not considered? Could a simplified comparison or qualitative discussion be added to contextualize where rehearsal-based approaches stand in the broader decision-making landscape?

局限性

The authors have adequately addressed the limitations of their work in the Appendix B (Discussion) section. They explicitly acknowledge that their proposed QP reformulation may be sub-optimal under certain conditions. This reflects a clear effort to communicate the boundaries within which their methods can be expected to perform reliably.

One suggestion for improvement: It would be helpful to explicitly discuss potential negative consequences of automation in sensitive domains (e.g., economic systems, healthcare, or policy decision-making) where biased or poorly specified “undesired regions” (AUF targets) might lead to unintended harms if not designed responsibly.

最终评判理由

I have reviewed their responses to other reviewers regarding non-linear settings, which have addressed some of the concerns I initially raised. Based on these clarifications and the improvements made, I am increasing my score to 4.

格式问题

In line 23 and 45, The references include author name 'Zhou', but this is not the case for other references. Please use one reference type consistently.

作者回复

We sincerely thank you for your detailed feedback and for recognizing the strengths of our work, including its strong theoretical grounding, clear presentation, and the practical importance of the long-term AUF setting. We hope our responses below fully address your concerns.


W1. Empirical evaluation of the assumptions.

A1. Thanks for this constructive suggestion. To empirically assess the impact of violating our assumptions, we conducted new experiments where the (i) symmetric desired region and/or (ii) unique target assumptions are violated. The table below presents results on a synthetic dataset (with dimensions of XX, ZZ, and YY set to 3, 2, and 2, respectively, and a decision horizon of 9). To ensure a fair comparison, when violating assumption (i), we chose a desired region with a total area similar to that in the no-violation setting.

ViolationQWZ23 [1]MICNS [2]Ours (Far-Sighted)
No0.342±0.0170.342 \pm 0.0170.332±0.0250.332 \pm 0.0250.765±0.1280.765 \pm 0.128
(i)0.225±0.0110.225 \pm 0.0110.250±0.0370.250 \pm 0.0370.603±0.0150.603 \pm 0.015
(ii)0.094±0.0270.094 \pm 0.0270.052±0.0200.052 \pm 0.0200.278±0.1660.278 \pm 0.166
(i)&(ii)0.032±0.0280.032 \pm 0.0280.062±0.0280.062 \pm 0.0280.192±0.1230.192 \pm 0.123

These results show that although performance degrades when assumptions are violated, our method remains substantially more robust than existing approaches [1, 2], particularly when the unique target assumption fails. This offers strong empirical support for the practical utility of our approach, even beyond the scope of our theoretical guarantees.


W2&Q1. A deeper analysis of scalability.

A2. Thanks for raising this important point. We agree that understanding the computational scalability of our method is crucial, especially for large dimensions and long horizons. As shown in Fig. 4 (middle) of our paper, our approach is already significantly faster than prior rehearsal-based methods [1, 2] for long horizons with fixed dimensionality.

To further address scalability, we propose two practical strategies: (i) In high-dimensional settings, heuristic methods can be developed to select a key subset of variables ZsZZ_s \subset Z and approximately solve Eq. (5) over ZsZ_s; (ii) For long horizons, the matrices (MM, NN, HH in Eq. (5)) that depend on the horizon length can be precomputed and cached offline to accelerate real-time inference.

We will incorporate this discussion into the revised version. Thank you!


W3&Q4. Empirical comparison with other methods.

A3. Thanks for this insightful question. We agree that positioning our method within the broader decision-making landscape is important. While we primarily compare with rehearsal-based methods due to (i) the specific problem setting (detailed in Appx. A), and (ii) prior empirical evidence showing that alternative methods (e.g., RL) achieve limited effectiveness in this setting [1, 2], we conducted additional experiments in response to your suggestion.

We include comparisons with representative RL methods (DDPG [3], SAC [4]) and a recent causal bandit algorithm [5]. For fairness, we allowed the RL agents to interact with the environment for 100 episodes—the same sample budget available to our method from observational data.

Ours[3][4][5]
0.827±0.0440.827 \pm 0.0440.207±0.0550.207 \pm 0.0550.182±0.0390.182 \pm 0.0390.166±0.0220.166 \pm 0.022

All experiments were conducted on the same dataset, with the same desired region SS and a fixed horizon 1, over 100 rounds and 10 random seeds. The table reports the AUF probability mean and variance. Our method clearly outperforms these baselines, as the rehearsal-based approach is more effective at leveraging fine-grained structural dependencies among variables and observed context XX. We also observed that SAC can eventually reach comparable performance after ~6,000 training steps. However, such extensive interaction is often not feasible in our problem setting.


Q2. Applicability to varying decision horizons.

A4. Thanks for the thoughtful question.

We would like to respectfully clarify that the decision horizon TT is a user-specified parameter (thus should not be unknown). For instance, if a trader aims to maximize profit over the next 10 months, they would simply set T=10T=10.

For a varying decision horizon setting, our main theoretical result (Thm. 3.4) can be straightforwardly applied to such scenarios. Thm. 3.4 provides a strong guarantee: for any finite future horizon TT and any decision step tt, the solution to Eq.(5) is optimal in maximizing the AUF probability of the aggregated target iYi/T\sum_i Y_i/T, assuming Ass. 3.3 holds. If decision-makers choose to specify different horizons TT at different steps tt, our method remains applicable and optimal in this context. One only needs to recompute the corresponding matrices (MM, NN, HH) in Eq.(5) for the specified TT.


Q3. Interpretability and sparsity constraints.

A5. Thank you for highlighting this important issue. When the horizon is 1, the solution is easy to interpret: it directly maximizes the probability that the immediate outcome YY lies within the desired region SS. For longer horizons, interpretation becomes more subtle: the optimal alteration at each step maximizes the probability that the average outcome across the horizon falls in SS. Using a Texas Hold'em analogy: if the goal is to win the current hand, the strategy should reflect immediate hand strength. But if the goal is to maximize winnings over 100 hands, one may initially build a specific table image (e.g., tight or loose) that improves long-term return, even if it sacrifices short-term gains.

Regarding sparsity, since Eq. (5) is a convex QP, we can include sparsity-inducing regularizers (e.g., L1 penalty, as in LASSO) without losing convexity. This helps generate more interpretable alteration vectors for domain experts.

We will add this discussion to the revised version. Thanks!


Other questions. Suggestions and formatting.

A6. Thanks for your detail and helpful suggestions. We will include a discussion of potential negative consequences in the Appendix and Checklist of the revised version.

Regarding reference formatting: we would like to respectfully clarify that our use of \citet and \citep is intentional, following standard citation practices to distinguish between in-text (e.g., Author [X]) and parenthetical ([X]) citations. Meanwhile, we have used the \bibliographystyle{unsrtnat} style consistently throughout the paper. Thanks!


We hope that our responses and additional experiments address your concerns. We would greatly appreciate your reconsideration of the paper in light of these clarifications. Thank you again for your valuable and constructive feedback.


Reference

[1] Rehearsal Learning for Avoiding Undesired Future. NeurIPS 2023.

[2] Avoiding Undesired Future with Minimal Cost in Non-Stationary Environments. NeurIPS 2024.

[3] Continuous control with deep reinforcement learning. ICLR, 2016.

[4] Soft actorcritic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. ICML, 2018.

[5] Additive Causal Bandits with Unknown Graph. ICML, 2023.

评论

Dear Reviewer qTb1,

We sincerely thank you for your time and feedback on reviewing our paper. As the discussion period is nearing its end, we would like to politely remind you and would greatly appreciate it if you could take a moment to consider our responses.

If there are any remaining concerns about our work, we would be happy to engage in further discussion.

Yours sincerely,
The Authors

评论

The authors addressed some of key weaknesses identified in the initial review. For Weakness (1), they added new experiments examining the effect of violating the symmetric desired region and unique target assumptions, as well as their combination. This is exactly the type of empirical evidence I requested. For Weakness (3), they extended the comparison to include RL methods (DDPG, SAC) and a causal bandit baseline under matched interaction budgets. This situates their method more clearly within the broader decision-making landscape; however, the authors should acknowledge the limitations of RL in low-interaction settings while noting RL’s competitiveness with more interactions. The paper still does not empirically test the effect of violating the linearity assumption, nor does it address scalability analysis.

Overall, the rebuttal strengthens the empirical case for the method’s robustness and relevance, but I am maintaining my original score to reflect the remaining limitations in breadth and practical analysis.

评论

Dear Reviewer qTb1,

We are pleased that our rebuttal was able to address your key concerns. Regarding the new points you raised, we would like to clarify the following:

  • On the effectiveness of RL in high-interaction settings, as noted in A3 of our rebuttal, We also observed that SAC can eventually reach comparable performance after ~6,000 training steps. However, such extensive interaction is often not feasible in our problem setting.
  • On the applicability of our method in non-linear settings, we have discussed this in detail in A1 of our response to Reviewer JXfA and A1 of our response to Reviewer 2xvd. We would politely encourage you to review those responses for a fuller explanation. Due to time constraints, we were unfortunately unable to conduct additional experiments to demonstrate this further, and we hope for your understanding.

We sincerely thank you again for your time and feedback.

Sincerely yours, The Authors

审稿意见
5

This paper introduces a new approach for "Avoiding Undesired Future" (AUF) problems in machine learning, extending existing methods to handle long-term decision-making scenarios. The authors propose a variance-reduced rehearsal learning framework that uses quadratic programming (QP) reformulation to transform intractable probabilistic optimization into a tractable one when making decisions to prevent predicted undesired outcomes over extended time horizons.

优缺点分析

Strengths:

  1. The paper provides rigorous theoretical guarantees including optimality of the QP reformulation under mild assumptions, variance reduction analysis, and O(1/√N) excess risk bounds

  2. The quadratic programming reformulation transforms an intractable probabilistic optimization into an efficiently solvable problem

  3. The far-sighted method achieves substantial performance gains (e.g., 95% vs 32% AUF probability on synthetic data)

Weaknesses:

  1. The optimality guarantees require three specific assumptions (linear systems, unique targets, symmetric desired regions) that may not hold in many real-world scenarios

  2. Only two datasets (one synthetic, one environmental) are used for evaluation

  3. While theoretically efficient, the paper doesn't adequately address how the methods scale to high-dimensional problems or very long time horizons

问题

  1. How should practitioners handle model uncertainty and potential distribution shift over time?

  2. What monitoring and adaptation strategies are recommended when the system is deployed in dynamic environments?

  3. How can the approach be integrated with existing decision-making systems that may have additional constraints not captured in the current formulation?

局限性

yes

格式问题

The format of the paper seems in accordance with the format stated in the website.

作者回复

Thanks for your valuable feedback and appreciation of our work! We hope that our responses can address your concerns.


W1. Concerns about the assumptions.

A1. Thanks for the insightful question. While our theoretical guarantees (Thm. 3.3) rely on certain assumptions, our method remains effective in practice even when these assumptions are partially violated, as evidenced by the new experimental results in A2.

The latter two assumptions, i.e., unique targets and symmetric desired regions, are often satisfied in real-world applications, as they are typically determined by the decision-maker. As for the linear system assumption, we would like to clarify that we have already generalized the noise distribution beyond the Gaussian assumption used in prior works [1, 2], allowing for symmetric log-concave noise. Moreover, potential non-linearities could be addressed using standard techniques such as kernel methods [3, 4]. Due to space constraints, we kindly refer you to our response to Reviewer JXfA (A1) for a more detailed discussion of the linearity assumption.

We will incorporate this discussion into the revised paper. Thanks!


W2. Concerns about the experiments.

A2. Thanks for your question. To address your concern, we conducted additional experiments on a new dataset to evaluate our method's performance when the assumptions are violated. The table below shows the performance on this new dataset (where dimensions of XX, ZZ, and YY are 3, 2, and 2, respectively and the horizon is set to 9) when assumption (i) symmetric desired region and/or (ii) unique target assumptions are violated. To ensure a fair comparison, when violating assumption (i) (symmetric desired region), we used a desired region with a total area similar to that of the no violation setting.

ViolationQWZ23 [2]MICNS [3]Ours(Far-Sighted)
No0.342±0.0170.342\pm 0.0170.332±0.0250.332\pm 0.0250.765±0.1280.765\pm 0.128
(i)0.225±0.0110.225\pm 0.0110.250±0.0370.250\pm 0.0370.603±0.0150.603\pm 0.015
(ii)0.094±0.0270.094\pm 0.0270.052±0.0200.052\pm 0.0200.278±0.1660.278\pm 0.166
(i)&(ii)0.032±0.0280.032\pm 0.0280.062±0.0280.062\pm 0.0280.192±0.1230.192\pm 0.123

As the results show, while not explicitly discussed in previous work, violating these assumptions degrades performance of the rehearsal learning approaches, especially the unique target assumption. Although optimality is not guaranteed in these settings, our method still significantly outperforms the prior works.

Finally, we would like to respectfully clarify that whenever Ass. 3.3 holds, our method is theoretically guaranteed to perform well, as established by Thms 3.4, 3.5, and 3.6.


W3. Concerns about the scalability.

A3. Thanks for your thoughtful question. Indeed, in sequential decision-making problems, both dimensionality and the time horizon can be large. We would like to highlight that, as shown in Fig. 4(mid) of our paper, our method is significantly faster than existing rehearsal learning approaches [1, 2] for a long time horizon when the dimensionality is fixed. For further reducing the executing time in decision-making, we propose two feasible directions: (i) In high-dimensional settings, heuristic methods can be developed to select a key subset of variables ZsZZ_s \subset Z and approximately solve Eq. (5) over ZsZ_s; (ii) For long horizons, the matrices (MM, NN, HH in Eq. (5)) that depend on the horizon length can be precomputed and cached offline to accelerate real-time inference.

We will incorporate this discussion into the revised paper. Thanks!


Q1. Potential distribution shift over time.

A4. Thanks for your question. The rehearsal learning process involves two sequential steps:

  • Step 1. Estimate/Update the underlying system parameters θ\theta from historical observational data.
  • Step 2. Use θ\theta to construct system matrices (e.g., AA, BB in Eq. (2)), and solve Eq. (5) for the decision alterations.

The model uncertainty and potential distribution shifts primarily affect the parameter estimation in Step 1. To handle model uncertainty (e.g., a time-varying graph structure), one could segment the data based on different structures and learn a separate set of parameters for each. To address distribution shifts of the data, if the shift is gradual, some online learning methods can be employed to update the parameters [5]; and if the shift is abrupt, a restart mechanism can be implemented [6], i.e., discard the old parameters upon detecting a major shift and re-estimate them from new samples.


Q2. How to deal with dynamic environments.

A5. Thanks for your question. As noted in A4, our framework follows a two-step sequential process. The challenge of dynamic environments, where system parameters θ\theta change over time, primarily pertains to the parameter estimation in Step 1. This issue has been explored in prior work on rehearsal learning [2], and similar adaptive strategies could potentially be incorporated into our framework as well.


Q3. How to integrate inter additional constraints.

A6. Thanks for the insightful suggestion. Prior works in rehearsal learning [1, 2] have considered integrating additional constraints (e.g., decision cost). If such constraints define a convex feasible set for the decision variables, they can be naturally incorporated into our QP reformulation (Eq. (5)) without compromising convexity, allowing for efficient optimization.

We will incorporate this discussion into the revised paper. Thanks!


We also take this opportunity to sincerely thank you for the careful review. Your suggestions are very important for further improving the paper. Thanks again!


References:

[1] Rehearsal Learning for Avoiding Undesired Future. NeurIPS 2023.

[2] Avoiding Undesired Future with Minimal Cost in Non-Stationary Environments. NeurIPS 2024.

[3] Kernel-based conditional independence test and application in causal discovery. UAI 2011.

[4] Distinguishing Cause from Effect Using Observational Data: Methods and Benchmarks. JMLR 2016.

[5] Adaptive online learning in dynamic environments. NIPS 2018.

[6] Learning to optimize under non-stationarity. AISTATS 2019.

评论

Dear Reviewer eWH7,

We sincerely thank you for your time and feedback on reviewing our paper. As the discussion period is nearing its end, we would like to politely remind you and would greatly appreciate it if you could take a moment to consider our responses.

If there are any remaining concerns about our work, we would be happy to engage in further discussion.

Yours sincerely,
The Authors

评论

Thank you for your response and clarifications! I think this is a strong paper, and will continue to recommend acceptance.

评论

Dear Reviewer eWH7,

We sincerely thank you for your time and valuable feedback, and we truly appreciate your positive evaluation and recognition of our work.

Sincerely yours, The Authors

审稿意见
4

The paper proposes to solve the sequential AUF problem of linear systems. The algorithm is to extend the current rehearsal algorithm to the sequential setting and is reduced to a QP formulation to as the core optimization problem. The work proves that it has relative tight risk bound and the variance is also well bounded by the time horizon. Empirical result show a proof-of concept numerical experiments on the algorithm on multiple linear systems.

优缺点分析

Strenghts:

  1. the problem is an natural extension of previuos work and the formation reduction to QP is also reasonable.
  2. Though i cannot understand the proofs in detail, it looks correct to me intuitively.
  3. The experiments support the results on different linear systems, include the common assumptions on the noise distribution.

Weekness:

  1. the work only focuses on the linear system, which limits its applicability. If this is not easy, empirical evidence on non-linear system should also be useful.
  2. the graph structure definition of the sequential problem is not straightforward, which potentially cannot cover all problem definitions.
  3. no practical application being mentioned. it is unclear why this setting and problem is useful and worth research.

问题

  1. Can the formulation and equivalent results been extended to non-linear systems?
  2. Why the dynamics have the specific graph structure?
  3. What is the potential application of this problem?

局限性

None

格式问题

None

作者回复

We sincerely thank you for your detailed feedback and for recognizing the strengths of our work! We hope that our responses can adequately address your concerns.


W1&Q1. Extension to nonlinear systems.

A1. Thanks for your insightful question. The core ideas of our proposed algorithm can be extended to general nonlinear settings. In principle, as long as the relationship between (i) the mean of the aggregated outcome iYi/T\sum_i Y_i/T and (ii) the selected alteration sequence zi\\{z_i\\} can be captured, e.g., via a parameterized functional form ff such that iYi/T=f(θ;zi)+ε~,\sum_i Y_i/T = f(\theta; \\{z_i\\}) + \tilde{\varepsilon}, for some parameter θ\theta and residual ε~\tilde{\varepsilon}, then we can turn to optimize argminzif(θ;zi)sk22\arg\min_{\\{z_i\\}}\\|f(\theta; \\{z_i\\}) - s_k\\|_2^2 at decision round kk, where sks_k denotes the adjusted center of the desired region.

Moreover, our main theoretical results can be adapted to such nonlinear settings. Taking the above example, if the surrogate loss f(θ;zi)sk22\\|f(\theta; \\{z_i\\}) - s_k\\|_2^2 is convex with respect to zi\\{z_i\\}, then Thm. 3.4 still applies. This is because the proof of Thm. 3.4 is based on residual analysis, with a key step showing that the aggregated error ε~\tilde{\varepsilon} remains within the log-concave distribution family (Prop. 3.2). Therefore, as long as the nonlinearity preserves this property, our main results remain valid.

We will incorporate this discussion in the revised version. Thanks!


W2&Q2. Concern about the graph structure.

A2. Thanks for the thoughtful question. We are not entirely sure if we have fully understood your concern; if our response does not address your point accurately, we would be happy to further clarify during the upcoming author-reviewer discussion period.

Our framework focuses on a specific class of decision problems in which structural relations among variables can be leveraged to improve decision-making, particularly when opportunities for interaction are limited. This line of work falls under the umbrella of rehearsal learning [1, 2, 3], with connections to causal bandits and causal RL [4, 5].

Furthermore, even when structural information is not explicitly available in the raw data, it can often be inferred using well-established structure learning methods [6], including in sequential [7] or cyclic [8] settings. Our method can then be applied based on the learned structure.

We will incorporate this clarification into the revised paper. Thanks again!


W3&Q3. Potential applications.

A3. Thanks for the question. We would like to address it in two parts:

  • First, our proposed decision-making method leverages structural information. Leveraging structural information to aid decision-making has broad applications in domains such as healthcare [9], finance [10], and environmental science [11]. For instance, Simpson’s paradox [12] illustrates that accurately evaluating the effect of a treatment XX on an outcome YY requires accounting for confounding variables, highlighting the need for causal structure in certain decision-making scenario. Recent advances in causal bandits [4, 5] and rehearsal learning [2, 3] similarly emphasize the importance of structure-aware decision-making in complex environments.

  • Second, our work extends prior rehearsal learning frameworks by explicitly incorporating sequential information, which is crucial in many real-world scenarios. For example, a physician prescribing medication over time for a chronic disease, or a financial institution regularly updating trading strategies. In such cases, methods that do not guarantee single-step optimality may accumulate errors over time [2, 3]. In contrast, our method guarantees optimality at each step (Thm. 3.4) and controls variance across the full sequence (Thm. 3.5). Our experimental results further support the effectiveness of our method in these settings.


Your suggestions are constructive and important for further improving the paper. Thanks again!


References:

[1] Rehearsal: Learning from prediction to decision. FCS 2022.

[2] Rehearsal Learning for Avoiding Undesired Future. NeurIPS 2023.

[3] Avoiding Undesired Future with Minimal Cost in Non-Stationary Environments. NeurIPS 2024.

[4] Bandits with unobserved confounders: A causal approach. NIPS 2015.

[5] Causal bandits: Learning good interventions via causal inference. NIPS 2016.

[6] DAGs with NO TEARS: Continuous Optimization for Structure Learning. NIPS 2018.

[7] DYNOTEARS: Structure learning from time-series data. AISTATS 2020.

[8] Discovering Cyclic Causal Models by Independent Components Analysis. UAI 2008.

[9] Causal inference and counterfactual prediction in machine learning for actionable healthcare. Nature Machine Intelligence 2020.

[10] The causal impact of media in financial markets. Journal of Finance 2011.

[11] Environmental controls on modern scleractinian coral and reef-scale calcification. Science Advances 2017.

[12] Causation, Prediction, and Search. MIT Press 2000.

评论

Dear Reviewer JXfA,

We sincerely thank you for your time and feedback on reviewing our paper. As the discussion period is nearing its end, we would like to politely remind you and would greatly appreciate it if you could take a moment to consider our responses.

If there are any remaining concerns about our work, we would be happy to engage in further discussion.

Yours sincerely,
The Authors

最终决定

This work introduces a new approach for the “Avoiding Undesired Future” problems in linear systems. The authors propose a variance reduced rehearsal learning framework based on quadratic programming that can be used to make tractable the optimization problem at the heart of AUF. The paper goes on to prove a tight risk bound and empirical results show that their methods work well via numerical experiments in linear systems. The reviewers agreed this work meets the requirements for publication at Neurips.