PaperHub
5.5
/10
Poster4 位审稿人
最低2最高4标准差0.7
4
3
2
3
ICML 2025

A Two-Stage Learning-to-Defer Approach for Multi-Task Learning

OpenReviewPDF
提交: 2025-01-17更新: 2025-08-16
TL;DR

Extend Two-Stage Learning-to-Defer to multi-task problems, providing additional theoretical guarantees.

摘要

The Two-Stage Learning-to-Defer (L2D) framework has been extensively studied for classification and, more recently, regression tasks. However, many real-world applications require solving both tasks jointly in a multi-task setting. We introduce a novel Two-Stage L2D framework for multi-task learning that integrates classification and regression through a unified deferral mechanism. Our method leverages a two-stage surrogate loss family, which we prove to be both Bayes-consistent and $(\mathcal{G}, \mathcal{R})$-consistent, ensuring convergence to the Bayes-optimal rejector. We derive explicit consistency bounds tied to the cross-entropy surrogate and the $L_1$-norm of agent-specific costs, and extend minimizability gap analysis to the multi-expert two-stage regime. We also make explicit how shared representation learning—commonly used in multi-task models—affects these consistency guarantees. Experiments on object detection and electronic health record analysis demonstrate the effectiveness of our approach and highlight the limitations of existing L2D methods in multi-task scenarios.
关键词
Learning to deferlearning from abstentionlearning theorymulti task

评审与讨论

审稿意见
4

The paper presents a new application of learning to defer to the context of multitask, where the task's target consists of both a regression and a classification task. The authors provide a theoretical analysis of two-stage L2D, showing that the proposed surrogate loss is both Bayes-consistent and G,R\mathcal{G}, \mathcal{R} consistent. Empirical results showcase the effectiveness of the approach.

给作者的问题

I think the paper is sound enough, with good motivation and adequate evaluation. Hence, I am prone to suggesting acceptance of the paper.

I have a couple of questions for the authors:

  • Have the authors considered how to directly model the coverage for experts, as considered for instance in [Okati et al.,2021; Mozannar et al., 2023; Palomba et al., 2025]?
  • Could the authors comment on my concern for the choice of subsampling the dataset? I guess adding a weighted loss is a safe option, but I do not want to overlook consistency concerns.

Update After Rebuttal

I am satisfied with the answers from the authors. I keep my positive score.

论据与证据

I think the paper is convincing enough in terms of contribution and impact of the approach in real-life scenarios. The theoretical analysis correctly supports the paper's claims.

方法与评估标准

The proposed datasets and evaluation criteria make sense. The authors consider two datasets, one that is new in the setting of learning to defer, and one used also in previous works [Mao et al., 2023a, 2024e]. Hence, the empirical evaluation makes sense overall.

理论论述

The proofs seem correct, but I might have missed some details.

实验设计与分析

The analyses provided are sufficiently sound, with enough details to reproduce results. Overall, the main concern I have is regarding the choice of subsampling the dataset for MIMIC IV, as detailed in Appendix G.2. Other options could be available, e.g., weighting the loss between classes differently, and the authors could have commented a bit more this aspect.

补充材料

I quickly passed through the proofs and looked at the details for the dataset creation.

与现有文献的关系

The paper is correctly positioned in the literature, with most works in learning to defer correctly referenced. Moreover, the work extends existing literature, considering cases where regression and classification occur simultaneously. I think this is a valuable contribution with sound proof and analysis.

遗漏的重要参考文献

The essential literature is correctly discussed, the only reference I think must be necessarily added is the work by Okati et al., 2021, which considers a formulation with explicit constraints in terms of coverage.

[Okati et al., 2021] - Okati, N., De, A., & Rodriguez, M. (2021). Differentiable learning under triage. NeurIPS 2021

其他优缺点

The paper is clearly written, with convincing motivations for the proposed approach.

其他意见或建议

I provide here for completeness a few recent articles, which could enlarge the current related work section (as they expand different aspects of Learning to Defer):

  • In [Wei et al., 2024], the authors provide a refinement of Bayes consistency, called Bayes-dependent consistency;
  • In [Palomba et al., 2025], the authors bridge causal inference and learning to defer for improved evaluation of such systems;
  • In [Strong et al., 2025], the authors present an application of L2D for healthcare using LLMs

Finally, according to ICML guidelines, the caption for tables should be placed above. Please consider adjusting the ones in the paper.

[Wei et al., 2024] - Wei, Z., Cao, Y., & Feng, L. (2024). Exploiting human-AI dependence for learning to defer. ICML '24

[Palomba et al., 2025] - Palomba, F., Pugnana, A., Alvarez, J., Ruggieri, S. (2025). A Causal Framework for Evaluating Deferring Systems. AISTATS '25

[Strong et al., 2025] - Strong, J., Men, Q., & Noble, A. (2025). Towards Human-AI Collaboration in Healthcare: Guided Deferral Systems with Large Language Models. AAAI '25

作者回复

We sincerely thank the reviewer for their constructive feedback. We appreciate their positive assessment of the strength of our theoretical and empirical contributions.

Below, we address the potential connection to coverage constraints and provide further details on specific aspects of our experimental design.

The analyses provided are sufficiently sound, with enough details to reproduce results. Overall, the main concern I have is regarding the choice of subsampling the dataset for MIMIC IV, as detailed in Appendix G.2. Other options could be available, e.g., weighting the loss between classes differently, and the authors could have commented a bit more this aspect.

Could the authors comment on my concern for the choice of subsampling the dataset? I guess adding a weighted loss is a safe option, but I do not want to overlook consistency.

Thank you for raising this point. Indeed, alternative heuristics such as class-weighted losses could have been employed. However, our primary objective was not necessarily to optimize overall performance metrics, but rather to highlight the relative improvement and effectiveness of our approach. Nevertheless, your suggestion of using weighted losses is valid and aligns with our theoretical framework, as our consistency proofs hold for any positive cost cjc_j. A promising direction for future experiments would involve adapting agent-specific cost functions in a cost-sensitive manner to address class imbalance explicitly.

The essential literature is correctly discussed, the only reference I think must be necessarily added is the work by Okati et al., 2021, which considers a formulation with explicit constraints in terms of coverage.

This is correct, we will add this paper in the related work.

Have the authors considered how to directly model the coverage for experts, as considered for instance in [Okati et al.,2021; Mozannar et al., 2023; Palomba et al., 2025]?

This is an excellent question. In the single-expert setting considered in [5,6,7], modeling coverage explicitly is relatively straightforward. Consider a rejector r:XRr:\mathcal{X}\rightarrow\mathbb{R}, where r(x)0r(x)\geq 0 indicates deferring to the expert, and r(x)<0r(x)<0 indicates no deferral. To directly model coverage, one can introduce a coverage level kR+k \in \mathbb{R}^+ corresponding to the kth\overline{k}^{\,\text{th}} percentile of the rejector scores r(x)r(x) computed over a validation set. This yields an adjusted decision rule: 1_r(x)k1\_{r(x)\geq k}, instead of 1_r(x)01\_{r(x)\geq 0}, resulting in an expected coverage of E[1_r(x)k]=1k\mathbb{E}[1\_{r(x)\geq k}] = 1-\overline{k}.

However, in the multi-expert setting, extending this idea is nontrivial because deferral decisions involve complex allocations across multiple experts. This complexity likely explains why [5,6,7] primarily focus on single-expert scenarios. One possible direction to handle coverage explicitly in multi-expert settings is to consider margin-based rejectors. Let A={0,,J}\mathcal{A}= \lbrace 0,\dots,J \rbrace denote the set of JJ experts plus a main model, and define a rejector r:XAr:\mathcal{X}\rightarrow\mathcal{A}. We then define the margin-based score as

ρ_r(x,j)=r(x,j)max_jjr(x,j),\rho\_r(x,j)=r(x,j)-\max\_{j'\neq j} r(x,j'),

where a positive margin ρ_r(x,j)0\rho\_r(x,j)\geq 0 leads to deferring to expert jj. Coverage can be incorporated by setting thresholds k~_j\tilde{k}\_j separately for each expert jj, corresponding to specific percentiles of the margin r(x,j)r(x,0)r(x,j)-r(x,0). The decision rule then becomes 1_ρ_r(x,j)k~_j1\_{\rho\_r(x,j)\geq \tilde{k}\_j}. Importantly, this approach naturally reduces to the single-expert coverage definition used in [5,6,7], since deferral occurs when ρ(x,1)0\rho(x,1)\geq 0, i.e., r(x,1)r(x,0)+k~_1r(x,1)\geq r(x,0)+\tilde{k}\_1.

We hope this aligns with your question and would be glad to discuss it further. We will include this clarification and an expanded discussion in the revised manuscript.

References

[1] Mao, et al. (2023). Two-Stage Learning to Defer with Multiple Experts. NeurIPS23

[2] Mao, et al. (2024). Regression with multi-expert deferral. NeurIPS24

[3] Narasimhan et al. (2022) Post-hoc estimators for learning to defer to an expert. NeurIPS22

[4] Mao et al. (2024) Theoretically grounded loss functions and algorithms for score-based multi-class abstention. AISTATS24

[5] Okati et al. (2021). Differentiable learning under triage. NeurIPS21

[6] Mozannar et al. (2023, April). Who should predict? exact algorithms for learning to defer to humans. AISTATS23

[7] Palomba et al. (2025). A Causal Framework for Evaluating Deferring Systems. AISTATS25

审稿人评论

I thank the authors for their rebuttal and for addressing my questions. I am satisfied with their answers. Overall, I think this is a valuable contribution. Thus, I am keeping my acceptance score.

审稿意见
3

This paper presents a two-stage learning-to-defer (L2D) approach for the multi-task setting involving both classification and regression. The authors provide theoretical justification in the form of consistency bounds, as well as empirical justification in the form of validation on (multi-task) object detection and EHR analysis.

Update after rebuttal

As stated in my rebuttal comment, I thank the authors for the clarifications and will maintain my recommendation of Weak accept.

给作者的问题

I want to be transparent that this is outside of my area of expertise, so I may have elementary questions regarding the background of L2D and interpretation of results.

  1. When describing a medical use case of L2D, the authors write, “If the model is sufficiently confident, its diagnosis is accepted; otherwise, the decision is deferred to a medical expert who provides the final diagnosis.” I take it this is an illustrative example of the concept of deference rather than a concrete description of L2D as manifested in empirical experiments, correct? For example, in the empirical experiments, it seems that “deference” refers to accepting/rejecting (weighting) the predictions of task experts rather than deferring to, say, a human expert. Perhaps this is an obvious point to researchers like yourselves, but based on this description in the Introduction, I was expecting some sort of human-in-the-loop evaluation where low-confidence cases were deferred to an actual human expert. Is my understanding correct, and could the authors provide some way to make this distinction clearer for non-experts?
  2. In the object detection setting, why do the authors not compare performance to other approaches? Also, can the authors explain why this approach is practically useful if it never outperforms the largest model (Expert 2)?

论据与证据

Regarding theoretical claims, I will have to defer to domain experts to verify accuracy. Empirical claims appear sound, though it is somewhat difficult to interpret these results without being intimately familiar with L2D literature and evaluation.

方法与评估标准

The two validation datasets were appropriate for this unique setting.

理论论述

I did not verify proofs due to time constraints.

实验设计与分析

Experimental design of empirical experiments seems appropriate – performance metrics were appropriate, and results (+ variability) were presented over 4 trials. Source code was provided in order to reproduce experimental results.

补充材料

I reviewed Sections A and G, but did not evaluate proofs closely.

与现有文献的关系

The authors clearly lay out how, despite recent progress in L2D, previous two-stage L2D approaches do not address multi-task “classifier-regressor models”, which are relatively common in more complex tasks such as object detection.

遗漏的重要参考文献

If there are key references missing, then I am not aware of them nor qualified to propose them.

其他优缺点

Strengths

  • The paper is very well-written and carefully organized to logically walk the reader through each section.
  • The related work is thorough, with unique contributions of this study clearly laid out relative to prior work.

Weaknesses

  • Empirical results may be difficult to interpret for non-experts in L2D.

其他意见或建议

References to tables/figures are unusual (ex: “Figure 5.1” [L372] and “Table 5.2” [L435]). These should refer to the tables/figures themselves, not the section.

作者回复

We thank the reviewer for their careful and constructive feedback. We are pleased that they found the paper well-written, the empirical setup appropriate, and the contributions clearly positioned in relation to prior work.

Below, we provide clarifications on several points raised in the review.

Empirical results may be difficult to interpret for non-experts in L2D.

To clarify, we conducted two distinct sets of experiments. In the object detection task, we considered three models of varying complexity (a lightweight Faster R-CNN, a medium-sized Faster R-CNN, and a large Faster R-CNN). To reflect their varying complexity and computational cost, we assigned different consultation scores to each agent: β0=0\beta_0=0, β1=β1\beta_1=\beta_1, and β2=RGβ1\beta_2=R_G\beta_1 with RGR_G being a ratio of GLOPs. This setup explicitly illustrates the trade-off between computational efficiency and prediction correctness, as the largest the model is, the better is its performance. Figure 1 shows that, for higher consultation costs β1\beta_1, our approach allocates queries mostly to the main lightweight model (cost β0\beta_0) or the first expert (cost β1\beta_1). Conversely, when β1\beta_1 is lower, the rejector strategically allocates queries across both experts (expert 1 and expert 2).

In the EHR task, baseline approaches fail to achieve strong performance because they base allocation decisions solely on individual tasks (classification or regression separately). In contrast, our method jointly considers both tasks, resulting in more balanced and effective query allocation. We will clearly emphasize these interpretations in the revised manuscript to improve clarity.

References to tables/figures are unusual (ex: “Figure 5.1” [L372] and “Table 5.2” [L435]). These should refer to the tables/figures themselves, not the section.

Thank you for pointing this out. We have taken note and will update this in the revised manuscript.

When describing a medical use case of L2D, the authors write, “If the model is sufficiently confident [...] I was expecting some sort of human-in-the-loop evaluation where low-confidence cases were deferred to an actual human expert. Is my understanding correct, and could the authors provide some way to make this distinction clearer for non-experts?

Thank you for raising this important point. To clarify, the medical use-case mentioned in the Introduction is intended solely as an illustrative example of deference rather than a description of the concrete mechanism used in our empirical experiments. In our experiments, “deference” refers to the process of automatically selecting predictive agents (e.g models, humans) based solely on their predictions and associated costs.

A key strength is its agent-agnostic nature: it does not rely on any specific internal structure, training paradigm, or decision process of the agents. As long as we have access to an agent’s predictions, our framework can learn a rejector rRr\in\mathcal{R} that, for each input xXx\in\mathcal{X}, computes confidence scores and allocates the query to the agent deemed most reliable. Importantly, during inference, we only query the selected expert.

In our empirical implementation, we have used automated black-box models or synthetic distribution as the experts. However, the framework is fully flexible and can incorporate human experts or any other decision-making system, as long as their predictions are available. We will update the manuscript to explicitly clarify this distinction, thereby reducing any potential confusion.

In the object detection setting, why do the authors not compare performance to other approaches? Also, can the authors explain why this approach is practically useful if it never outperforms the largest model (Expert 2)?

The primary reason we did not include comparisons with other object detection approaches is that our paper focuses explicitly on evaluating the effectiveness of the allocation mechanism within the Learning-to-Defer framework, rather than competing directly with state-of-the-art detection methods. Importantly, our method remains practically valuable even when it does not surpass the performance of the largest expert model (Expert 2). This is because Expert 2, while accurate, is typically too computationally expensive or slow to deploy on every query in realistic, resource-constrained environments. Our L2D approach provides a principled solution to this critical challenge by strategically routing queries to less computationally intensive models whenever appropriate, significantly reducing resource usage without severely impacting accuracy. Furthermore, when expert models have specialized strengths or operate on slightly different distributions—as demonstrated clearly in our EHR experiments—our method can exploit these discrepancies to effectively improve overall system performance beyond what any single agent could achieve alone.

审稿人评论

I appreciate the authors' clarifications and will likely maintain my recommendation of Weak accept.

审稿意见
2

The paper developed a Two-Stage Learning-to-Defer framework for multi-task problems, enabling joint classification and regression. This framework features a novel two-stage surrogate loss family that is both (G,R)(\mathcal{G}, \mathcal{R})-consistent and Bayes-consistent for cross-entropy-based surrogates. The authors derived tight consistency bounds and established minimizability gaps, extending prior work on Learning-to-Defer. Their learning bounds improve with richer hypothesis spaces and more confident experts. The authors validated the approach on object detection and electronic health record analysis, demonstrating its superiority over existing methods.

给作者的问题

  1. Is it possible to view classification as a specific instance of regression, where the zero-one loss serves as the evaluation metric? If this perspective is valid, would the current framework be adequately addressed by the regression-based deferral framework presented in previous research?

  2. What specific challenges arise when adapting the learning-to-defer framework for multi-task learning scenarios?

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

Yes.

实验设计与分析

Yes.

补充材料

No.

与现有文献的关系

Section 2 provides a sound discussion.

遗漏的重要参考文献

N/A.

其他优缺点

Strengths: The paper is well-written, clearly presented, and provides a detailed discussion of prior work, making it accessible to non-experts.

Weaknesses: The results, while sound, seem to offer incremental progress rather than significant novelty compared to existing work. Furthermore, I question whether classification can be viewed as a specific instance of regression, with the zero-one loss as the evaluation metric. If so, would the proposed framework be adequately addressed by existing regression-based deferral frameworks?

其他意见或建议

N/A.

作者回复

We thank the reviewer for their careful reading and thoughtful comments. We are pleased that they found our framework well-presented and the theoretical analysis rigorous. In the following, we clarify the technical challenges that arise in the multi-task setting.

First, we emphasize that our paper provides theoretical novelty. Our framework is generalized to accommodate any positive cost c_jc\_j, enabling a unified treatment of classification, regression, and multi-task problems within Learning-to-Defer. We derived the Bayes-optimal rejector and established consistency guarantees for multi-task L2D. In particular, we prove consistency bounds that hold for any surrogate in the comp-sum family—parameterized by ν\nu as defined in Equation 1—which encompasses commonly used surrogates such as the logistic, MAE, and exponential losses. These bounds are both tighter and more interpretable than those in [1, 2, 3], accounting for both the parameter ν\nu and the L_1L\_1 norm of the aggregated cost vector τ\boldsymbol{\tau} (see discussion with reviewer @oatj). Additionally, we present a novel analysis of minimizability gaps in Theorem 4.6, demonstrating that the optimal conditional risk critically depends on both the norm of the aggregated costs τ\boldsymbol{\tau} and the choice of the multiclass surrogate Φ_01ν\Phi\_{01}^\nu (Equation 1). Notably, our results hold without any assumptions on the underlying distribution D\mathcal{D}—in contrast to [4]—and apply to any surrogate Φ_01ν\Phi\_{01}^\nu within the comp-sum family. Furthermore, we provide new generalization bounds specifically tailored to this setting.

While classification can conceptually be viewed as a special case of regression under the zero-one loss, such a simplification overlooks important theoretical distinctions that arise when analyzing consistency and optimality in deferral decisions. To highlight this, we give some additional analysis:

Let g(x)=(hw(x),fw(x))g(x) = (h \circ w(x), f \circ w(x)) be any multi‐head model where wWw \in \mathcal{W} is a shared representation and hHh \in \mathcal{H}, fFf \in \mathcal{F} are heads for classification and regression, respectively. Define L(w,h,f)=E_y,tx[c_0(h(w(x)),f(w(x)),z)]L(w,h,f) = \mathbb{E}\_{y,t \mid x}[c\_0(h(w(x)),f(w(x)),z)] as the conditional risk for the triple (w,h,f)(w,h,f). Let L=inf_(w,h,f)L(w,h,f)L^\ast = \inf\_{(w,h,f)} L(w,h,f) be the optimal multi‐head conditional risk in the hypothesis class G={x(h(w(x)),f(w(x)))wW,hH,fF}\mathcal{G} = \lbrace x \mapsto (h(w(x)), f(w(x))) \mid w \in \mathcal{W}, h \in \mathcal{H}, f \in \mathcal{F} \rbrace. We then write L(w,h,f)LL(w,h,f) - L^\ast to denote the conditional excess risk of (w,h,f)(w,h,f) above the best multi‐head. Observe that for a chosen triple (w,h,f)(w,h,f), we can decompose: L(w,h,f)L=[L(w,h,f)inf_h,fL(w,h,f)]+[inf_h,fL(w,h,f)L],L(w,h,f) - L^\ast = \bigl[L(w,h,f) - \inf\_{h',f'} L(w,h',f')\bigr] + \bigl[\inf\_{h',f'} L(w,h',f') - L^\ast\bigr],

with hh' and ff' acting directly on xx. Define the heads gap Δ_heads(w,h,f)=L(w,h,f)inf_h,fL(w,h,f)\Delta\_{\mathrm{heads}}(w,h,f) = L(w,h,f) - \inf\_{h',f'} L(w,h',f'), which measures how well the specific heads (h,f)(h,f) perform given a fixed representation ww. Then define the representation gap Δ_repr(w)=inf_h,fL(w,h,f)L\Delta\_{\mathrm{repr}}(w) = \inf\_{h',f'} L(w,h',f') - L^\ast, which captures how close ww is to the best possible representation for both tasks. From this notation, we rewrite

L(w,h,f)L=Δ_heads(w,h,f)+Δ_repr(w).L(w,h,f) - L^\ast = \Delta\_{\mathrm{heads}}(w,h,f) + \Delta\_{\mathrm{repr}}(w).

We next introduce a multi‐task comparison by defining a separate‐training conditional risk. Let L_sep(h,f)=E_y,tx[c_0((h(x),f(x)),z)]L\_{\mathrm{sep}}(h',f') = \mathbb{E}\_{y,t \mid x}[c\_0((h'(x),f'(x)),z)] be the conditional risk of using completely separate models hh' and ff' without any shared representation. Let L_sep=inf_h,fL_sep(h,f)L\_{\mathrm{sep}}^\ast = \inf\_{h',f'} L\_{\mathrm{sep}}(h',f'). We measure the quality of forcing a single shared ww via

Δ_MTL=LL_sep.\Delta\_{\mathrm{MTL}} = L^\ast - L\_{\mathrm{sep}}^\ast.

If Δ_MTL<0\Delta\_{\mathrm{MTL}} < 0, then joint training (i.e. a shared ww) is strictly better than separate solutions; if Δ_MTL>0\Delta\_{\mathrm{MTL}} > 0, it is worse. Finally, we can link Δ_MTL\Delta\_{\mathrm{MTL}} to the multi‐head conditional excess risk by noting

L(w,h,f)L=[L(w,h,f)L_sep]Δ_MTL.L(w,h,f) - L^\ast = \bigl[L(w,h,f) - L\_{\mathrm{sep}}^\ast\bigr] - \Delta\_{\mathrm{MTL}}.

Hence, if Δ_MTL<0\Delta\_{\mathrm{MTL}} < 0, it follows that L(w,h,f)L<L(w,h,f)L_sepL(w,h,f) - L^\ast < L(w,h,f) - L\_{\mathrm{sep}}^\ast, meaning the multi‐head optimum is easier to approach than the separate‐training optimum. Conversely, if Δ_MTL>0\Delta\_{\mathrm{MTL}} > 0, we get L(w,h,f)L>L(w,h,f)L_sepL(w,h,f) - L^\ast > L(w,h,f) - L\_{\mathrm{sep}}^\ast, so it is harder to reach the best multi‐head risk than to reach the best separate‐training solution.

Following the discussion with Reviewer @oXmZ, we will also incorporate a discussion on how our approach can be adapted to account for coverage constraints.

We hope this clarifies your question. We will add those novel analysis and revise the manuscript to make this point clearer in the final version.

Please refer to the discussion with reviewer @oXmZ for references.

审稿意见
3

In this paper, the authors analyze the learning-to-differ (L2D) problem in the two-staged multi-task (classification and regression) setting. The paper introduces the pointwise Bayes rejector for the mult-task deferral and introduces a surrogate differal loss that is Bayes consistent. The paper further provides generalization guarantees for the learned rejector, which provide insights into the conditions of the problem setting that can lead to better generalization. Numerical experiments are provided to validate the proposed method, and a comparison of the proposed multi-task rejector with prior single-task rejectors shows that the proposed method can have a balanced trade-off between the tasks compared to prior methods.

给作者的问题

  • Please refer to the questions/concerns raised in previous sections.

论据与证据

Yes.

方法与评估标准

Yes.

理论论述

Did not check all the proofs, minor typos in the proof of Lemma 4.2 (mentioned in Other Comments Or Suggestions).

实验设计与分析

  • It is unclear how the coefficients λcla,λreg\lambda^\text{cla}, \lambda^\text{reg} used in the agents' cost were determined. These coefficients seem crucial in attaining the balanced performance described in the discussion.
  • It is unclear how the rejection is implemented for the prior methods in the experiments in Section 5.2 in the multi-task setting. For example, when the classification rejector is triggered, does the model consult for the expertise on both classification and regression tasks from the expert, or only the classification task?

补充材料

Yes, Sections A, B, and G.

与现有文献的关系

This paper extends the literature in the area of L2D, by providing theoretical insights into L2D in the two-staged multi-task setting.

遗漏的重要参考文献

None.

其他优缺点

Strengths

  • The paper provides theoretical insights into L2D in the two-staged multi-task setting, which seems novel in the related literature.
  • The paper is mostly self-contained and easy to follow
  • The problem is well-motivated, and the theoretical insights are sufficiently discussed.

Weaknesses

  • The definition of the deferral function class LdefL_{def} (column 1 line 319) to be a mapping to [0,1] seems unrealistic, given that the def\ell_{def} is made up of c0c_0, which in turn is assumed to be a summation of 01\ell_{01} and reg\ell_{reg}. Since reg\ell_{reg} is not necessarily bounded between 0 and 1, it is unclear why the definition of LdefL_{def} is correct.
  • Given the possibly different scales of classification and regression losses, more discussion seems to be needed regarding designing a balanced/meaningful rejector in this setting.

其他意见或建议

  • It seems like in proof of Lemma 4.2, CdefBC^B_{\ell_{\text{def}}} should not depend on g,rg, r.
  • τˉ\bar{\tau} is not defined in Theorem 4.4, and it is not clear where T1,ν,Tν=1T^{-1, \nu}, T^{\nu=1} appear in the Theorem.
  • The clusters CclaM1,CclaM2C_{cla}^{M1}, C_{cla}^{M2} defined in Experiments in 5.2 seem not distinct as described.
作者回复

We thank the reviewer for their detailed and thoughtful feedback. We are glad that they found our theoretical contributions novel, the paper well-motivated, and the analysis of the two-stage multi-task Learning-to-Defer setting valuable to the literature. Please find some clarification below:

It is unclear how the coefficients λcla\lambda^{cla} and λreg\lambda^{reg} [...] balanced performance described in the discussion.

Given the possibly different scales of classification and regression losses, more discussion seems to be needed regarding designing a balanced/meaningful rejector in this setting.

We agree that the choice of the coefficients λcla\lambda^{cla} and λreg\lambda^{reg} is crucial for achieving balanced and meaningful performance. In our experiments, we considered both classification and regression tasks equally important; thus, we set λcla=λreg=1\lambda^{cla} = \lambda^{reg} = 1. This choice represents a balanced and task-agnostic baseline, ensuring no implicit bias toward either component. In practical scenarios, these coefficients should indeed be tuned according to task-specific priorities and performance requirements. For instance, if the classification is more important in this setting, we should set λcla>λreg\lambda^{cla}>\lambda^{reg} to prioritize agent with more proficiency on the classification task. We will explicitly discuss this and highlight the importance of tuning λcla\lambda^{cla} and λreg\lambda^{reg} in the revised manuscript.

It is unclear how the rejection is implemented for the prior methods [...] or only the classification task?

In the EHR task, we treat classification and regression as independent but simultaneously allocated tasks. Specifically, when the classification rejector (as described in [1]) is triggered, we allocate both the classification and regression queries jointly to the selected agent. Similarly, when the regression rejector is triggered, it also allocates both tasks together [2]. We will clarify this joint-allocation implementation explicitly in the revised manuscript.

The definition of the deferral function class (column 1 line 319) to be a mapping to [0,1][0,1] seems unrealistic.

It seems like in proof of Lemma 4.2, C__defB\mathcal{C}\_{\ell\_{def}}^B should not depend on g,rg,r.

Thank you for pointing this out. You are correct—these are typos and do not affect the correctness of the proofs. We will fix them in the revised manuscript.

τ\overline{\tau} is not defined in Theorem 4.4, and it is not clear where T1νT^{1-\nu} appear in the Theorem.

Indeed, τ_j\overline{\tau}\_j represents the expected aggregated cost of agent jAj \in \mathcal{A}, formally defined as \overline{\tau}\_j = \mathbb{E}_\{y,t|x}[\tau\_j]. Consequently, we define the vector τ=(τ_0,τ_1,,τ_J)\overline{\boldsymbol{\tau}} = (\overline{\tau}\_0, \overline{\tau}\_1, \dots, \overline{\tau}\_J) and τ_1||\boldsymbol{\overline{\tau}}||\_1 its L1 norm.

The transformation function Tν(u)\mathcal{T}^{\nu}(u) is introduced to determine the function Γν(u)\Gamma^{\nu}(u) corresponding to a given multiclass surrogate loss Φν_01\Phi^\nu\_{01}. We explicitly show that the consistency bounds of the deferral loss (Lemma 4.3) inherently depends on the consistency bounds of the multiclass surrogate loss Φ01ν\Phi^\nu_{01} from the comp-sum family (Equation 1). For example, choosing the log-softmax surrogate corresponds to setting ν=1\nu = 1. In this case, the relevant transformation function is explicitly given by T1(u)=1+u2log[1+u]+1u2log[1u]\mathcal{T}^1(u)=\frac{1+u}{2} \log[1+u] + \frac{1-u}{2} \log[1-u] [8]. Then, by taking the inverse of this transformation, we obtain the function Γν(u)=(Tν)1(u)\Gamma^\nu(u)=(\mathcal{T}^\nu)^{-1}(u) leading to a bound with the function Γν(u)=τ_1Γν(uτ_1)\overline{\Gamma}^\nu(u)=||\overline{\boldsymbol{\tau}}||\_1 \Gamma^\nu(\frac{u}{||\overline{\boldsymbol{\tau}}||\_1}) depending on both the surrogate transformation Tν(u)\mathcal{T}^{\nu}(u) and the L1 norm τ_1||\overline{\boldsymbol{\tau}}||\_1.

We will explicitly include these definitions and clarifications in the revised manuscript.

The clusters CclaM1,CclaM2C_{cla}^{M_1}, C_{cla}^{M_2} defined in Experiments in 5.2 seem not distinct as described.

We confirm that clusters C_claM1C\_{cla}^{M_1} and C_claM2C\_{cla}^{M_2} were explicitly constructed to represent distinct regions of expertise, although they do indeed exhibit some overlap. Upon review, we acknowledge a minor inconsistency in the original manuscript regarding their definitions (but not on experiments). The correct cluster assignments are CM_1_cla={C_1,C_2,C_4}C^{M\_1}\_{cla} = \lbrace C\_1, C\_2, C\_4\rbrace and CM_2_cla={C_1,C_5,C_6}C^{M\_2}\_{cla} = \lbrace C\_1, C\_5, C\_6 \rbrace. We will correct this in the revised manuscript.

References

[1] Mao, et al. (2023). Two-Stage Learning to Defer with Multiple Experts. NeurIPS23

[2] Mao, et al. (2024). Regression with multi-expert deferral. NeurIPS24

[8] Mao et al. (2023). Cross-entropy loss functions: theoretical analysis and applications. ICML23

最终决定

This paper examines two-stage learning to defer under the multi-task setting, meaning that there are supervised losses for both classification and regression, providing a theoretical analysis of the consistency of the proposed loss as well as empirical validation on the Pascal VOC and MIMIC datasets. The reviewers generally had positive opinions of the work (1 x accept, 2 x weak accept, 1 x weak reject). The reviewer who leaned toward rejection's primary criticism is that the extension to multi-task learning had marginal novelty. The authors provided a detailed rebuttal, nicely explaining the paper's theoretical contributions, which I found convincing.

Small comment: I noticed two errors in the years of the citations---one was that Verma et al. should be AIStats 2023, not 2022. I can't remember the other one