PaperHub
7.1
/10
Poster5 位审稿人
最低4最高5标准差0.5
4
4
4
5
5
3.0
置信度
创新性2.8
质量2.8
清晰度2.8
重要性3.0
NeurIPS 2025

Discretization-free Multicalibration through Loss Minimization over Tree Ensembles

OpenReviewPDF
提交: 2025-05-12更新: 2025-10-29

摘要

关键词
Algorithmic FairnessMulticalibration

评审与讨论

审稿意见
4

The paper presents a novel multi-calibration method: given a base predictor f0f_0, the paper proposed to optimize the MSE by adding a series of depth-2 trees. A single depth-2 tree would partition the dataset into four subgroups, based on 1) whether f0(x)>=vf_0(x) >= v and 2) whether the instances belong to a certain pre-defined group ii.

Theoretical results show that such a algorithm "automatically" guarantees bounded multi-calibration error. Thus, the method does not depend on how we discretize the range of YY; although this is still needed to evaluate the multi-calibration error afterwards. Emprical experiments also domenstrate that the multi-calibration errors are lower than competitor methods.

优缺点分析

I like the proposed direction in the paper; however I don't think the paper is mature enough to be published at a venue like NeurIPS.

Strenghs:

  • The paper is well-structured. The story line is in general clear.
  • The report empirical experiment results look quite strong.
  • The paper takes a quite novel approach to tackle this problem (to the best of my knowledge).

However, I do have the following concerns:

  1. In the descriptions of the technical part in Section 4, some parts are contradictory to each other.
  • In eq.3, it does not make sense for S1(f0)S_1(f_0) and S2(G)S_2(\mathcal{G}) to take the union among all possible vv and ii respectively. I assume you want to use them to denote the collection of the subsets, not the union of the subsets?
  • Eq.4 is contradictory to the illustration in Fig.1: in eq.4, s2s_2 is any FIXED element from S2(G)S_2(G), yet in Fig.1, it seems that s2s_2 can be different (you can take g_1(.) and g_2(.) for different leaves. )
  1. The source code is not shared making the reproducibility of the results quite low. Any specific reason why the code is not shared to the reviewers?

  2. It seems in the paper that the method can work with any model f0f_0; however, I am a bit skeptical in whether the proposed method would encounter a pitfall: e.g., when f0f_0 is trained with a tree boosting method (which is quite common for tabular data), and e.q.(5) is optimized with a tree boosting as well, then will the proposed method still learn to multi-calibrate? I would appreciate if this is at least discussed.

问题

Besides my comments above, I have the following questions:

  1. why using linear models for the tabular data in the experiment? This is related to my last point above: what would happen if you use, e.g., lightGBM, or another tree-boosting method like XGboost here?

  2. Why some of the points are missing in Figure 2? Is it because of the too long runtime or something else?

  3. I would appreciate an empirical verification of Thm 4.6, to further justify the assumptions used for proof. That is, to empirically compare the multi-calibration error with the sqrt(ϵround+ϵloss)sqrt(\epsilon_{round} + \epsilon_{loss}).

局限性

Yes

最终评判理由

All my major concerns are addressed. The only minor issues left are about

  1. the "overshooting" in the empirical validation of the theorem.
  2. the presentation of the method in the figure, but this is minor (though it might cause some big confusions so please revise it).

格式问题

NA

作者回复

We thank the reviewer for the careful review. We’re happy to respond to the questions raised by the reviewer.

  1. Notation error in eq 3.

    Indeed, according to eq 4, s1s_1 and s2s_2 should be sets, so S\mathcal{S} should be a collection of sets. The correct version of Eq. 3 should be S1(f0)=xX:f0(x)v:vR(f0)0\mathcal{S}_1(f_0) = \\{ \\{ x \in \mathcal{X} : f_0(x) \geq v \\} : v \in R(f_0) \cup \\{0\\} \\} and S2(G)=xX:gi(x)=1:i[G].\mathcal{S}_2(G) = \\{ \\{ x \in \mathcal{X} : g_i(x) = 1 \\} : i \in [|G|] \\} . Thank you for noting this technical inaccuracy, and we’ll correct it in our manuscript.

  2. Difference between Fig. 1 and Eq. 4

    The two representations are different but theoretically equivalent. On one hand, the formulation in Eq. 4 can be represented by the one in Fig. 1 simply by setting the groups to be the same in a tree. On the other hand, a tree that has two different groups as features can be written as the sum of two trees in Eq. 4, one with c1=c2=0c_1=c_2=0 and the other with c3=c4=0c_3=c_4=0. In the illustration, we used different groups in the first place, as that’s what is implemented by the solver. We intend to change this to avoid confusion.

  3. Source code availability

    We have already made the codebase available online. Thank you for your interest in our implementation.

  4. On the model architecture of the baseline algorithm for tabular data Linear models are commonly used for tabular data in fairness literature, as evidenced by [1,2,3]. However, we recognize tree-boosting methods are also prevalent. In our derivation, we didn’t restrict the architecture of the uncalibrated baseline. Also, our method operates on different input features, and therefore does not contradict the baseline architecture.

    We further conducted an experiment here in which tree-boosting architecture is used for the uncalibrated baseline:

    • Experiment setting: instead of linear models, we use LGBMClassifier(num_leaves=8) and LGBMRegressor(num_leaves=8) for the classification and regression tasks respectively. The other hyperparameters are chosen by default.

    • Results: We present the loss in Table 3 (complementing Table 1 in the paper) and worst-group smECE in Table 4 (complementing Table 2 in the paper).

      Table 3: The empirical marginal improvement of running Algorithm 1 the second time.

      Task(f0)\ell(f_0)(fcal)\ell(f^{cal})(pG(fcal))\ell(p_G(f^{cal}))ϵ^loss=(fcal)(pG(fcal))\hat{\epsilon}_{loss} = \ell(f^{cal}) - \ell(p_G(f^{cal}))
      Income Classification0.12850.12770.12762×1052 \times 10^{-5}
      Income Regression0.03250.03220.03229×1079 \times 10^{-7}
      Travel Time Regression0.02760.02760.02763×106-3 \times 10^{-6}

      Table 4: The worst-group smECE of the evaluated algorithms. The values have been multiplied by 1000 for readability.

      MethodDiscr.Income Class.Income Reg.Travel Time
      Uncalibrated Baseline/93.2063.2724.93
      Multiaccurate Baseline/67.50 ± 12.7641.13 ± 1.9224.88 ± 2.23
      LSBoost1091.06 ± 3.2950.56 ± 7.2742.86 ± 0.07
      2093.48 ± 1.2251.16 ± 7.7626.56 ± 0.22
      3086.89 ± 6.1850.78 ± 5.0528.04 ± 0.75
      5078.52 ± 8.0156.51 ± 4.2127.26 ± 2.65
      7584.49 ± 7.4656.99 ± 3.5326.36 ± 1.50
      10087.58 ± 4.2157.69 ± 3.1928.97 ± 3.17
      MCBoost1094.57 ± 0.0062.99 ± 0.0042.84 ± 0.00
      2090.41 ± 3.2164.95 ± 0.4526.48 ± 0.00
      3089.62 ± 2.4963.78 ± 0.5127.82 ± 0.71
      5091.54 ± 2.0262.26 ± 0.4928.62 ± 1.89
      7592.33 ± 1.8263.06 ± 0.5025.67 ± 1.39
      10093.64 ± 1.6262.70 ± 1.0127.13 ± 0.36
      Ours/64.08 ± 13.6538.34 ± 2.3324.75 ± 2.20

    Table 3 shows that Assumption 1 is still empirically valid in the new setting. Table 4 shows that using a tree-boosting baseline also produces remarkable multicalibration, even surpassing what we originally reported with linear base models. Therefore, our algorithm still works when the uncalibrated model is a tree ensemble.

  5. Clarification of Fig. 2:

    During discretization, we set the size of the codomain to be 10, 20, 30, 50, 75, 100, and there are always six points on each line in Fig. 2. From this perspective, no point is missing. We think what the reviewer meant by missing is that some lines don’t extend to 100. This is because the x axis in Fig. 2 is the size of the range, which can be smaller than the size of the codomain.

  6. Empirical evaluation of the theorem

    First, we would like to point out that it is not possible to directly evaluate ϵloss\epsilon_{loss}, as discussed in 5.3.1. We use ϵ^loss\hat \epsilon_{loss}, the empirical loss improvement instead as a proxy, with ϵloss=ϵ^loss+ϵoptϵ^loss\epsilon_{loss}=\hat \epsilon_{loss} + \epsilon_{opt} \geq \hat \epsilon_{loss}, where ϵopt\epsilon_{opt} is the optimization error.

    Denoting the multicalibration error as α\alpha, we show that, empirically, α<ϵ^loss+ϵroundϵloss+ϵround\alpha < \sqrt{\hat \epsilon_{loss} + \epsilon_{round}} \le \sqrt{\epsilon_{loss} + \epsilon_{round}} in most cases. We calculate αϵ^loss+ϵround\alpha - \sqrt{\hat \epsilon_{loss} + \epsilon_{round}} and present it in the following table.

    Table: αϵ^loss+ϵround\alpha - \sqrt{\hat \epsilon_{loss} + \epsilon_{round}} of our algorithm on different tasks and different discretization, multiplied by 1000 for readability.

    DiscretizationIncome Class.Comment Toxicity Class.Age Reg.Income Reg.Travel Time Reg.
    10-19.69 ± 1.14-35.78 ± 0.76-19.15 ± 0.37-25.30 ± 0.87-17.55 ± 0.72
    20-6.63 ± 0.98-16.86 ± 0.69-7.00 ± 0.70-11.93 ± 0.83-11.41 ± 0.62
    30-2.22 ± 1.71-10.74 ± 1.07-4.84 ± 0.75-6.33 ± 0.90-7.86 ± 0.89
    501.62 ± 1.58-5.75 ± 1.66-2.66 ± 0.54-1.39 ± 1.33-2.49 ± 1.01
    752.68 ± 1.70-3.75 ± 1.98-1.31 ± 0.810.72 ± 1.62-0.63 ± 0.62
    1005.34 ± 2.02-2.03 ± 2.27-0.39 ± 0.752.11 ± 1.290.80 ± 0.61

    Note that the task the task Skin Lesion Classification is not included here. This is because the empirical evaluation of ϵ^loss+ϵround\hat \epsilon_{loss} + \epsilon_{round} is negative for this task.

    As is shown here, we have α<ϵ^loss+ϵround\alpha < \sqrt{\hat \epsilon_{loss} + \epsilon_{round}} in 80% of the different settings. Given ϵlossϵ^loss\epsilon_{loss}\geq \hat \epsilon_{loss}, we can assert that ϵloss+ϵround\sqrt{\epsilon_{loss} + \epsilon_{round}} is empirically larger than the multicalibration error in most cases, thus validating the theorem empirically.

We hope this helps answer your questions.

[1] Jung, Christopher et al. “Batch Multivalid Conformal Prediction.” International Conference on Learning Representations(2023).

[2] Xian, Ruicheng et al. “Fair and Optimal Classification via Post-Processing.” International Conference on Machine Learning (2022).

[3] Chen, Yatong et al. “Fairness Transferability Subject to Bounded Distribution Shift.” Advances in Neural Information Processing Systems (2022)

评论

Thanks a lot for the comments and the additional experiments..

A few follow-up questions:

  • you wrote (Line 112) "Each decision tree of depth 2 in the ensemble assigns a real value ci 112 : i ∈ [4] to each of its 4 leaves", and each "tree-illustrating subfigure" (like the one under the title Tree 1) in figure 4 should correpond to a single tree. Hence, in my opinion, g1 must be equal to g2. Could you explain what do you mean by, here I quote, "In the illustration, we used different groups in the first place, as that’s what is implemented by the solver"?

  • For the point 4 in your rebuttal letter, about the experiment, when you use lightGBM for f0f_0, did you use LightGBM or XGboost to optimzie Eq (5) in the paper? If you use the LightGBM, optimizing (5) is somewhat like adding more iterations in the boosting, with different attributes for the trees of course, but the attributes like f0>3f_0 > 3 in figure 1 can theoretically represented by sums of trees constructed by original features (i.e., the X's).. Thus, I find it hard to intuitively understand why it works. That's also why I want to check your implementation to see if something is missing in the descriptions.

  • Regarding the point 6 in your rebuttal letter, I agree that negative values mean good, but I also saw a disturbing trend that as the number of discretization increases, αϵ^loss+ϵround\alpha - \sqrt{\hat \epsilon_{loss} + \epsilon_{round}} also increases. For 3 out 5 datasets you show, the increase does not stop at 0 but seem to overshoot and keep increasing to larger positive numbers... Have you noticed this and it there any explanation?

评论
  • On the clarification of our rebuttal point 2

    We are aware that, according to line 112 and Eq. 4, we require that each tree should only address a single group; yet our graph contradicts this. Since our graph is an illustration of our framework, we agree that it should be consistent with Eq. 4, and we’ll change it.

    As for the quoted sentence, we were trying to explain why this contradiction was brought to our manuscript in the first place. Many tree ensemble solvers, like XGBoost and lightGBM, allow different features on the same level, and we mistakenly include this solver-specific feature into our illustration of our solver-independent framework. We did explain why these two formulations are equivalent in our previous response though.

    Hope that this answers your confusion!

  • On why thresholding on f0f_0 cannot be replaces by additional trees

    We did use LightGBM for the base predictor, but it’s not simply equivalent to adding more iterations (or, more trees). The feature f0(x)f_0(x) that is exclusively for our algorithm is the key difference, even when f0f_0 itself is an ensemble of trees. We would like to point out that (a) learning an ensemble of depth-m trees and splitting on its outputs, is not equivalent to (b) learning an ensemble of depth-(m+1) trees. The former has a larger capacity than the latter.

    Let’s illustrate this point by setting m=1m=1. Say there are nn binary features. It’s possible for a ensemble of nn depth-1 tree to output 2n2^n different outputs, and with an additional ensemble of depth-1 trees that take the thresholds on such outputs, it is theoretically possible for (a) to assign different values to all 2n2^n configurations of the nn binary features. However, the number of unique trees in ensemble (b) is at most polynomial in nn.

    Of course in practice, to ensure generalization, the output f0(x)f_0(x) is treated as a single continuous feature instead of an indicator for every single datapoint; but we hope that through this example, you can see that even in the specific case where tabular data is already fitted with an ensemble and where the group indicators can be represented by the original features, our post-processing algorithm still cannot be absorbed into f0f_0 and be written as a larger ensemble with extended iterations.

  • On the explanation of the empirical results in the additional experiment

    Building on our previous discussion, the empirical result serves as a loose upper bound, which explains the positive values we observe. As for the trend, while it’s hard to provide a formal analysis, we can try to understand it intuitively. Let’s assume that the discretizing [0,1][0,1] to mm level sets adds a perturbation δ\delta to the output f(x)f(x), where δ\delta is a uniform random variable in [1/2m,1/2m][-1/2m, 1/2m]. If the mean of the predictor output E[f(x)]E[f(x)] is the same as the mean of the ground truths E[y]E[y], then the discretization error is E[(f(x)+δy)2(f(x)y)2]=E[δ2]=112m2E[(f(x)+\delta-y)^2-(f(x)-y)^2]=E[\delta^2]=\frac{1}{12m^2}, which increases fast as mm increase. However, the multicalibration error remains the same as E[δ]=0E[\delta]=0. Then, the subtracted value will become more negative as δ\delta increases, namely as mm decreases.

We appreciate your feedback, and hope this discussion addresses your concerns!

评论

Thanks for the further feedback. Regarding your three points:

  • I don't agree it is equivalent. As said, it is only equivalent if g1 is equal to g2.

  • Please do notice that in my previous reply, I was aware that using f0>3f_0 > 3 as the splitting condition in the tree is not the same as using original features. Your reply hence does not answer my question. My question was about that, since f0f_0 is a tree ensemble as well, the condition f0>3f_0 > 3 should be able to be represented by the sum of the trees (precisely leaves, actually). In that sense, it is hard to intuitively understand whether you could obtain "new information" to further decrease the loss. I'd appreciate if you could elaborate here, but I do understand that it might work in practice indeed, as shown by your experiments.

  • I think your reply contains errors, which make it hard to follow. For instance, E[(f(x)+δy)2(f(x)y)2]=E[δ2]=112m2E[(f(x)+\delta-y)^2-(f(x)-y)^2]=E[\delta^2]=\frac{1}{12m^2}, which increases fast as mm increase. I believe as mm increases 112m2\frac{1}{12m^2} will only decrease. Further, I don't see why "loose upper bound" can explain "positive results" here.

评论

We agree that there were some wording issues in our previous response for point 3, so we’d like to explain it again in a more straightforward way.

First, it’s hard to formally analyze this trend, so we try to intuitively understand why αϵloss+ϵround\alpha-\sqrt{\epsilon_{loss}+\epsilon_{round}} is increasing as mm increases.

  • ϵloss\epsilon_{loss}, as defined in Assumption 4.5, is the additional loss decrease by optimizing Eq. 5 a second time, and this does not change with discretization.
  • For an intuitive understanding, we consider the simplified case where the predictions are uniformly distributed in [0,1]. Then, discretizing to mm outputs can be viewed as adding a perturbation δ\delta to the predictions, with δ\delta being a uniform random variable in [1/2m,1/2m][-1/2m, 1/2m]. For the discretization error, ϵround=E[(f(x)+δy)2(f(x)y)2]=E[δ2]=112m2\epsilon_{round}=E[(f(x)+\delta-y)^2-(f(x)-y)^2]=E[\delta^2]=\frac{1}{12m^2} decreases fast as mm increases (previously a typo here).
  • On the other hand, the multicalibration error is calculated using sums of E[f(x)]E[y]f(x),g(x)]\left| E[f(x)]-E[y] \mid f(x), g(x) \right\|] where the expectation is taken within the absolute values. Intuitively, since E[f(x)+δ]=E[f(x)]E[f(x)+\delta]=E[f(x)], α\alpha can remain unchanged as mm changes.

This constitutes an intuitive understanding of why αϵloss+ϵround\alpha-\sqrt{\epsilon_{loss}+\epsilon_{round}} increases as mm increases.

Moreover, we reported αϵ^loss+ϵround\alpha-\sqrt{\hat \epsilon_{loss}+\epsilon_{round}}, which is an empirical upper bound of αϵloss+ϵround\alpha-\sqrt{ \epsilon_{loss}+\epsilon_{round}}, as we leave out the optimization error ϵopt\epsilon_{opt} in our evaluation. We know for sure that αϵloss+ϵround\alpha-\sqrt{ \epsilon_{loss}+\epsilon_{round}} will be smaller than the reported values, but unfortunately, we do not have enough tools at this moment to quantify such differences.

We hope that this explanation could be of help!

评论

Hi,

Thanks a lot for the very detailed explanations.

I think if including f0f_0 as new features can further optimize the loss is true in general for tree boostings, this observation deserves some more discussion. Of course for this paper, the tabular data is only part of the experiments, hence not super important for now.

I am still a bit concerned about the overshooting, but I am willing to consider this as a issue for future work, as the contribution now seems enough.

I will increase my score accordingly.

评论

Thank you for the letting us know that we agree that splitting condition in the tree is not the same as using original features. So you're asking: since f0f_0 is itself a tree ensemble: f0(x)=T1(x)+T2(x)++TM(x),f_0(x) = T_1(x) + T_2(x) + \ldots + T_M(x), shouldn't the threshold condition, like f0(x)>3f_0(x) > 3, be representable through some combination of the existing leaves in those trees? And if so, why does our post-processing provide any benefit—what "new information" are we gaining?

  • First, we'd like to point out that for decision trees ensembles, like LightGBM and XGBoost, all the trees use the same set of features that are fixed at the beginning of the algorithm; usually it does not allow taking the intermediate outputs, like the constructed leaves, or the output of the trees constructed so far, as the feature. Therefore, although f0f_0 can be represented by the sum of the trees, taking the threshold on f0f_0 is not a valid operation in tree ensembles. We believe this might be the major source of confusion, because this is actually different from the notion of boosting as in the paper [1], where the output of the model in the current step can be used as the input feature of the next timestep. Therefore, our method is very different from adding more iterations in the boosting.

  • Second, we'd like to explain why the loss can be further decreased. You're correct that no new information about the underlying data distribution is gained, and that threshold conditions can theoretically be derived using the original features, though not with standard tree ensembles only. The key insight is, instead of gaining new information, we are gaining access to decision boundaries that lie outside the representational capacity of standard tree ensembles. More precisely, we'll show that: given nn binary features xx, there exists an ensemble f0f_0 of trees with depth d0d_0 such that representing all threshold functions on f0f_0 requires trees of depth nn.

    • A simple example to see why this is true is to set f0=i=1n2ixif_0=\sum_{i=1}^n 2^{-i} x_i. Then, f0f_0 is an one-to-one function that has 2n2^n discrete outputs. With two thresholding functions If(x)cI\\{f(x)\ge c\\} and If(x)c+2nI\\{f(x)\ge c+2^{-n}\\}, we can obtain any indicator Ix=x0I\\{x=x_0\\} for any x0x_0, with Ix=x0=j=1n((xj)(x0)j(1xj)1(x0)j),I\\{x=x_0\\}=\prod_{j=1}^{n}\left((x_j)^{(x_0)_j}(1-x_j)^{1-(x_0)_j}\right), where we assume 00=10^0=1.

      Now ensembles of depth-dd trees can be written as a polynomial of degree dd:

      lvljTlxjbl,j(1xj)1bl,j\sum_{l} v_l \prod_{j \in T_l} x_j^{b_{l,j}}(1-x_j)^{1-b_{l,j}}

      where TlT_l contains the indices jj of features xjx_j that are split on in the path to leaf ll, Tld|T_l| \le d is the depth constraint, and bl,j0,1b_{l,j} \in \\{0,1\\} indicates whether feature xjx_j appears in its original form (bl,j=1b_{l,j}=1) or negated form (bl,j=0b_{l,j}=0) in the leaf condition.

      If two polynomials of degree dd subtracts to an irreducible polynomial of degree nn, the only possibility is that dnd\ge n. Since we only have nn binary features, we got d=nd=n. In fact, this argument holds as long as f0f_0 is one-to-one.

    That is to say, if we want all threshold functions to be represented by an ensemble of trees as well, this ensemble has to consist of depth-nn trees. Indeed, if we can learn an f0f_0 that consists of trees of depth d0=nd_0=n, there will not be any improvement on the representation power by post-processing, and indeed our algorithm will not be beneficial.

    And for the general case where d0<nd_0<n, thresholding functions on f0f_0, which is an ensemble of trees of depth d0d_0, cannot be written as an ensemble of depth-dd trees for any d<nd<n in general. This is where the gain in loss comes from, as we learn the same data with a more powerful model.

  • Third, the model could also benefit from feature engineering even when no new information is present. We'd like to further illustrate the effect of the groups here. Even the group indicators are obtainable from the original features, explicitly including them in the input of our algorithm serves the role of feature engineering. This doesn't introduce new information neither, but it makes the model more focused on the multicalibration task, which also contributes to the decreased multicalibration error.

[1] Ira Globus-Harris, Declan Harrison, Michael Kearns, Aaron Roth, and Jessica Sorrell. 2023. Multicalibration as boosting for regression. In Proceedings of the 40th International Conference on Machine Learning (ICML'23), Vol. 202. JMLR.org, Article 460, 11459–11492.

评论

Thank you for your continued interest in the details of our paper and for your patience with us to discuss them.

The difference between the ensembles is a very important question on the ensemble of decision trees that our proposed algorithm optimizes over, which is {tTt:TT(f0,G)}\{\sum_{t \in T }t : T\subseteq \mathcal{T}(f_0,\mathcal{G}) \} in Equation (5) of our paper.

In the theoretical analysis of the paper, for simplicity, we define T(f0,G)\mathcal{T}(f_0,\mathcal{G}) as in Equation (4), hereafter we call this T1(f0,G)\mathcal{T_1}(f_0,\mathcal{G}). In practice, many tree ensemble solvers, like XGBoost and lightGBM, allow different features on the same level. So the actual decision trees to be ensembled are in the form illustrated in Figure 1, where g1g_1 and g2g_2 are not necessarily the same. We call this class of decision trees T2(f0,G)\mathcal{T_2}(f_0,\mathcal{G}): T2(f0,G)=c1Ixs1xs2+c2Ixs1xs2+c3Ixs1xs3+c4Ixs1xs3:c1,c2,c3,c4R,s1S1(f0),s2,s3S2(G).\mathcal{T_2}(f_0,\mathcal{G})= \\{c_1\cdot I\\{x\in s_1 \wedge x\in s_2\\}+c_2\cdot I\\{x\in s_1 \wedge x\notin s_2\\}+ c_3\cdot I\\{x\notin s_1 \wedge x\in s_3\\} +c_4\cdot I\\{x\notin s_1 \wedge x\notin s_3\\} :c_1,c_2,c_3,c_4\in \mathbb{R}, s_1\in \mathcal{S}_1(f_0),s_2, s_3\in \mathcal{S}_2(\mathcal{G})\\}.

Below we show tTt:TT1(f0,G)=tTt:TT2(f0,G). \\{\sum_{t \in T }t : T\subseteq \mathcal{T_1}(f_0,\mathcal{G}) \\} = \\{\sum_{t \in T }t : T\subseteq \mathcal{T}_2(f_0,\mathcal{G}) \\} .

It is obvious that T1(f0,G)T2(f0,G)\mathcal{T_1}(f_0,\mathcal{G}) \subseteq \mathcal{T_2}(f_0,\mathcal{G}). Therefore tTt:TT1(f0,G)tTt:TT2(f0,G)\\{\sum_{t \in T }t : T\subseteq \mathcal{T_1}(f_0,\mathcal{G}) \\} \subseteq \\{\sum_{t \in T }t : T\subseteq \mathcal{T_2}(f_0,\mathcal{G}) \\}.

On the other hand, for any tT2(f0,G)t\in \mathcal{T_2}(f_0,\mathcal{G}), it can be written as the sum of two trees in T1(f0,G)\mathcal{T_1}(f_0,\mathcal{G}). One tree has g1g_1 on the second level and c3=c4=0c_3 = c_4 = 0. The other tree has g2g_2 on the second level and c1=c2=0c_1 = c_2 = 0.

More concretely, let t=c1Ixs1xs2+c2Ixs1xs2+c3Ixs1xs3+c4Ixs1xs3t = c_1\cdot I\\{x\in s_1 \wedge x\in s_2\\}+c_2\cdot I\\{x\in s_1 \wedge x\notin s_2\\}+ c_3\cdot I\\{x\notin s_1 \wedge x\in s_3\\} +c_4\cdot I\\{x\notin s_1 \wedge x\notin s_3\\}, where c1,c2,c3,c4R,s1S1(f0),s2,s3S2(G)c_1,c_2,c_3,c_4\in \mathbb{R}, s_1\in \mathcal{S}_1(f_0),s_2, s_3\in \mathcal{S}_2(\mathcal{G}). t=t1+t2t = t_1 + t_2, where t1=c1Ixs1xs2+c2Ixs1xs2+0Ixs1xs2+0Ixs1xs2t_1 = c_1\cdot I\\{x\in s_1 \wedge x\in s_2\\}+c_2\cdot I\\{x\in s_1 \wedge x\notin s_2\\}+ 0\cdot I\\{x\notin s_1 \wedge x\in s_2\\} +0\cdot I\\{x\notin s_1 \wedge x\notin s_2\\} and t2=0Ixs1xs3+0Ixs1xs3+c3Ixs1xs3+c4Ixs1xs3t_2 = 0\cdot I\\{x\in s_1 \wedge x\in s_3\\}+0\cdot I\\{x\in s_1 \wedge x\notin s_3\\}+ c_3\cdot I\\{x\notin s_1 \wedge x\in s_3\\} +c_4\cdot I\\{x\notin s_1 \wedge x\notin s_3\\}. t1,t2T1(f0,G)t_1, t_2 \in \mathcal{T_1}(f_0,\mathcal{G}).

Therefore, tTt:TT1(f0,G)=tTt:TT2(f0,G)\\{\sum_{t \in T }t : T\subseteq \mathcal{T_1}(f_0,\mathcal{G}) \\} =\\{\sum_{t \in T }t : T\subseteq \mathcal{T_2}(f_0,\mathcal{G}) \\}. The two ensembles are equivalent.

审稿意见
4

This paper studies the problem of multicalibration, where a predictor is calibrated for a large family of groups. It proposes a one-shot, discretization-free method, which treats the uncalibrated score as another feature, fits an L2-boosted forest so that once it cannot lower the loss further, the model is multicalibrated across all groups without choosing bins by hand. The paper provides theoretical support and experiments across six datasets, achieving the lowest worst-group Smooth-ECE on each dataset.

优缺点分析

Strengths:

  • The paper seems to provide a nice theoretical link between loss and multicalibration.
  • The results show that the method improves worst-group calibration.

Weaknesses:

  • Does the algorithm affect accuracy of the predictors? I may have missed this but it looks like only calibration error is reported.
  • How scalable is this method to many groups? Can this approach extend to implicit groups defined by unlabeled features?
  • Does the method rely on square loss? Would it work to minimize cross-entropy instead?
  • Can you include a baseline that is not multicalibration-aware just to see how much calibration error remains?
  • The method requires explicit group features, which may be difficult to obtain.

问题

See weaknesses.

局限性

yes

格式问题

None

作者回复

We thank the reviewer for the careful review, and we’re happy to answer the questions raised.

  1. On the accuracy of the predictors

    This is a major aspect that practitioners might wonder about, and we thank the reviewer for pointing this out. Apart from calibration, other metrics might be of interest when evaluating an algorithm. Here we present the accuracy (for classification tasks) and MSE error (for regression tasks)

    Table 1: The accuracy (%) of the predictors on classification tasks. The higher, the better.

    MethodDiscr.Skin Lesion Class.Income Class.Comment Toxicity Class.
    Uncalibrated Baseline/70.2176.7192.88
    Multiaccurate Baseline/73.19 ± 0.3178.97 ± 0.0692.89 ± 0.00
    LSBoost1077.22 ± 1.0077.96 ± 0.2592.81 ± 0.08
    2076.55 ± 0.9578.09 ± 0.3392.78 ± 0.06
    3076.86 ± 0.4678.16 ± 0.2192.78 ± 0.06
    5077.06 ± 0.5278.11 ± 0.2092.71 ± 0.07
    7576.52 ± 0.7277.81 ± 0.2792.65 ± 0.04
    10076.12 ± 0.5077.60 ± 0.3092.81 ± 0.05
    MCBoost1071.00 ± 1.0176.71 ± 0.0092.88 ± 0.00
    2072.15 ± 1.6377.10 ± 0.3492.88 ± 0.00
    3073.54 ± 1.2477.23 ± 0.2092.88 ± 0.00
    5072.07 ± 1.0177.20 ± 0.4092.88 ± 0.00
    7573.12 ± 1.0076.89 ± 0.1792.88 ± 0.00
    10072.29 ± 1.0876.70 ± 0.2292.87 ± 0.01
    Ours/78.90 ± 0.1779.27 ± 0.0792.93 ± 0.01

    Table 2: The MSE loss (×103\times 10^{-3}) of the predictors on regression tasks. The lower, the better.

    MethodDiscr.Age Reg.Income Reg.Travel Time Reg.
    Uncalibrated Baseline/5.1141.3930.19
    Multiaccurate Baseline/3.51 ± 0.0237.92 ± 0.0229.76 ± 0.02
    LSBoost102.01 ± 0.2439.66 ± 0.3531.06 ± 0.07
    201.42 ± 0.1638.98 ± 0.3130.29 ± 0.09
    301.36 ± 0.1339.03 ± 0.2330.25 ± 0.04
    501.52 ± 0.1739.51 ± 0.2230.16 ± 0.03
    751.53 ± 0.1938.83 ± 0.1930.13 ± 0.03
    1001.72 ± 0.2038.80 ± 0.2030.17 ± 0.03
    MCBoost105.08 ± 0.2542.18 ± 0.0131.23 ± 0.00
    204.69 ± 0.3041.10 ± 0.2330.42 ± 0.00
    304.92 ± 0.1740.70 ± 0.2630.31 ± 0.00
    504.40 ± 0.3840.93 ± 0.2730.22 ± 0.00
    753.81 ± 0.3640.72 ± 0.3230.22 ± 0.01
    1003.88 ± 0.4140.95 ± 0.2930.18 ± 0.03
    Ours/1.08 ± 0.0236.11 ± 0.1129.80 ± 0.04

    By optimizing for the loss directly, our algorithm performs slightly better on standard metrics (MSE error / accuracy) on all the tasks evaluated.

  2. On the scalability w.r.t. number of groups.

    Our approach scales well, which has been supported by both theoretical and empirical analysis. In theory, we selected a specific boosting algorithm that can be used in our framework and conducted finite-sample analysis in Appendix E. We showed that the sample complexity scales logarithmically with the number of groups. Empirically, we evaluated on approximately 50 groups for tabular data, substantially exceeding prior empirical multicalibration studies (e.g., Hansen et al. utilized at most 20 groups).

  3. On the selection of loss function.

    Unfortunately we didn’t support other loss functions for now. We acknowledge that cross entropy is more prevalent than Brier score loss for classification tasks, and this could be left for future investigations.

  4. On baselines that are not multicalibration aware.

    We did include non-multicalibration-aware baselines in our evaluation. The Multiaccurate Baseline achieves multiaccuracy (a weaker condition than multicalibration) and consistently exhibits higher calibration error than other algorithms.

  5. On the requirement of explicit group features

    Indeed, we require the group features to be explicitly given, and the model's performance on unseen groups is not guaranteed. This represents a limitation of our current approach. However, as established in the algorithmic fairness literature, many works in algorithmic fairness require that “variables are explicitly defined as ‘sensitive’ by specific legal frameworks”. [1] While extending to implicit group discovery constitutes important future work, explicit group definition remains the predominant paradigm in fairness-aware machine learning applications.

    We would also like to add that our response point 3 to Reviewer 3k17 might be related, in which we discussed how certain group structures can be implicitly defined.

[1] Simon Caton and Christian Haas. 2024. Fairness in Machine Learning: A Survey. ACM Comput. Surv. 56, 7, Article 166 (July 2024), 38 pages. https://doi.org/10.1145/3616865

审稿意见
4

The paper proposes a novel multicalibration algorithm that relaxes the discretization preprocessing typically required in existing multicalibration methods. The theoretical results hold for any Discretization Operation as defined by the authors. The core idea is to use depth-two decision trees trained via empirical risk minimization (ERM). Under the loss saturation assumption, the algorithm achieves multicalibration guarantees.

To demonstrate its practicality, the authors conduct empirical evaluations on multiple datasets, showing improved performance over existing multicalibration algorithms.

优缺点分析

Strengths

  1. The method generalizes previous approaches by relaxing the discretization requirement, enabling application to a broader class of discretization strategies.
  2. The use of simple depth-two decision trees enhances both practical usability and interpretability.
  3. The paper is clearly written and easy to follow.

Weaknesses The loss saturation assumption is essential to the theoretical guarantee. While it often holds in practice, the authors provide a specific counterexample where it fails.

问题

In Figure 2, what exactly is meant by “different discretization”? What is the significance of using different discretizations in this plot?

局限性

Yes

最终评判理由

The authors have fully answered to my questions.

格式问题

N/A

作者回复

We thank the reviewer for the careful review. We are happy to respond to the question raised by the reviewer.

By discretization we mean restricting the codomain of the predictor to a discrete set, and a different discretization means a different codomain. For uncalibrated baseline, multiaccurate baseline and our algorithm, it's done by rounding the real-valued output to the nearest value in the codomain; and for LSBoost and MCBoost, it's done by setting the hyperparameter of the algorithm before running it.

We vary the discretization for these two reasons:

  1. Evaluation requirement: Multicalibration error, as defined in Definition 3.2, can only be computed for predictors with finite output ranges, so even our continuous predictor must be discretized for evaluation purposes. This metric can be sensitive to discretization, so we vary it for a more robust evaluation.

  2. Practical significance: Different downstream applications may require different precision levels or thresholds for decision-making. Varying the discretization strengthens the point that a single trained predictor maintains low multicalibration error across various discretization schemes, making it more flexible for diverse applications without retraining.

评论

I appreciate the response from the authors.

审稿意见
5

The paper proposes a new multicalibration algorithm which does not require specifying the granularity of the level sets at training time. This is because the algorithm works by finding a subset of decision trees (on particular features such as group membership / prediction thresholds) which minimize the squared loss. This stands in contrast to most previous iterative-patching style algorithms.

The main theoretical result is a statement on the multicalibration error achieved by the post-processing algorithm. In particular, assuming that the post-processing algorithm does not improve the loss by much, and also that the discretization error is low, then the multicalibration error is also low.

Finally, experiments are conducted in comparing the proposed post-processing algorithm to LSBoost and MCBoost, two previous multicalibration algorithms.

优缺点分析

I previously served as an emergency reviewer for this paper (and hence, re-used the summary above).

In my opinion, the most important strength of the paper is that the multicalibration post-processing algorithm it proposes seems to be extremely effective (conditioned on the evaluations run by the paper).

At the time of my previous review, my main concerns / weaknesses were the following:

  1. Motivation for focusing on the “discretization-free” approach of the paper.
  2. Not evaluating worst-group calibration error within the experiments, which is a natural metric that practitioners may be interested in.

The first point (1) seems to be better addressed in this version of the paper. In particular, the second paragraph of the introduction motivates why one would want to get rid of the multicalibration discretization parameter with five concrete reasons:

  • Introduces rounding error which leads to prediction distortion
  • Adds a (sensitive) additional hyperparameter to model tuning
  • The parameter introduces a bias / variance tradeoff
  • Instability under different parameter choices
  • Downstream decision makers with different utility functions may not get predictions with enough granularity in order to make a correct decision.

I think all of these are reasonable and effective motivations to get rid of the discretization parameter, and hence, I feel point (1) is no longer a major weakness of the paper. I would appreciate additional citations for the first and fifth bullets if possible!

The second weakness I had, point (2), also seems to have been addressed. In particular, the paper evaluates on the worst group smECE in Table 2, where they show that their proposed post-processing method outperforms all other baselines. In the prior version, the paper evaluated on the multicalibration error of a rounded predictor, which was a bit of an odd metric to evaluate on.

Since both of the major weaknesses were addressed, I am happy to advocate for accepting the paper.

问题

I had some questions on the evaluations run in the paper.

  1. When reporting the results of the uncalibrated baseline, is the baseline predictor trained on the entire train + calib set, or just the train set? The reason I ask is because Hansen et al. argue that we should evaluate mcb post-processing algorithms w.r.t. the baseline which uses all available training data (since that is what a practitioner would probably try first).
  2. Similarly, it seems that the baseline DistillBert model for CivilComments used in the submitted paper seems quite weak. In particular, Hansen et al. seem to use the same groups and model architecture, and achieve a much better baseline worst-group smECE of around 0.06 (compared to the 0.119 reported in this paper). Therefore, all multicalibration post-processing algorithms “appear” less useful, since the baseline is already so good. Improving the baseline model may decrease the gap / improvement between standard mcb post-processing methods and the proposed decision tree ensemble approach. (If you are struggling to improve the worst-group smECE of the Distilbert model, one tip is to leave it training for longer!)

I think that these are minor quibbles given that the proposed post-processing method seems to clearly improve on the tested baselines. In particular, the fact that one could get better baselines in certain regimes --- say, for the experiments with NNs --- does not detract from the overall results or message, which holds for all model families, including simple models on tabular baselines.

局限性

Yes

最终评判理由

The authors answered the few questions I had. The work remains a strong contribution to the empirical multicalibration literature, and may eventually become the de-facto multicalibration algorithm for practitioners.

格式问题

N/A

作者回复

We thank the reviewer for the careful review. We are happy to respond to the questions raised by the reviewer.

  • On the dataset partitioning for the training of the baseline:

    Indeed, our uncalibrated baseline models were trained only on the training set, not on the combined training + calibration data. While training a separate model with the combined dataset may decrease the performance gap between the uncalibrated models and the multicalibration ones, this won’t affect the gap among the multicalibration models, as they all post-processed the uncalibrated model (trained on only the training set) using a held-out calibration dataset.

    The difference in this setting stems from the purpose of the experiments: Hansen et al. focused on whether multicalibration algorithms are necessary or not, i.e., if setting aside some data for different multicalibration algorithms will be beneficial compared to the baseline algorithms; while we focused on, given the same baseline algorithm and a calibration set, how well our algorithm perform compared to others. If considered this way, our setting should make sense as well.

  • On the discrepancy of the performance of the uncalibrated model on CivilComments dataset:

    Thank you for pointing this out! Due to the aforementioned difference in data partitioning, we used 225,000 samples for baseline training, while Hansen et al. used 270,000 for the baseline algorithm. We also followed the suggested setting of finetuning for 5 epochs suggested by Koh et al. Indeed, it seems that training for additional 5 epochs might increase the performance of the uncalibrated predictor, as indicated by the results of Hansen et al., and we thank the reviewer for the kind tip.

  • Additional citations

    As for the bibliographyic reference of the motivation, we find this work [1] relevant. Our motivation is quite similar to the motivating example of this paper, namely, down-stream decision-makers hope that the same model's predictions "lead to beneficial decisions according to their own loss functions".

    Another work [2] might be relevant as well, as it discussed how information loss "caused by rounding can influence the predictive power of the machine learning models".

[1] Zhao et al. (2021). Calibrating Predictions to Decisions: A Novel Approach to Multi-Class Calibration. NeurIPS 2021.

[2] Senavirathne, N., Torra, V. Rounding based continuous data discretization for statistical disclosure control. J Ambient Intell Human Comput 14, 15139–15157 (2023)

评论

Thank you for the response! I will keep my score.

审稿意见
5

This paper tackles multicalibration without the discretization step required by prior methods. The authors proposed a simple one short learning algorithm that fits a LightGBM with depth 2, which effectively models the interaction of base predictor and the grouping indicators. They also proved that, if the learned ensemble is ”loss-saturated”, then the multicalibration error is bounded. Empirically, across 6 datasets, the method has the lowest worst group smECE in general compared to baselines.

优缺点分析

Strengths:

  1. This paper developed a novel technique that got rid of the discretization step widely used by previous methods.
  2. This paper used a single Ensemble tree fit to capture the interactions between the base prediction and group indicators, which is concise and easy to implement.
  3. The empirical evaluation showed improved performance across 6 datasets over existing approaches on multicalibration error and worst group smECE.
  4. The paper is well-written with theoretical justifications.

Weaknesses:

  1. Unlike LSBoost that calibrates with respect to a function class, the method this paper introduced only guarantees the application to pre-specified groups. The method could suffer from unspecified important subgroups or high-dimension grouping structure.
  2. Section 5.3.2 reports the discretized version of the proposed predictor, however, the author doesn’t show the undiscretized version, but only showed the worst group smECE in Section 5.3.3 instead. I understand that comparison in 5.3.2 is already fair, but adding the undiscretized version would give readers a more complete picture of practical performance.

问题

Questions:

  1. Could you comment on how the algorithm behaves when relevant sub-groups are not explicitly encoded? How sensitive is it to “hidden” groups that fall outside of the provided indicators?
  2. Have you experimented with tree depth greater than 2? Could deeper trees help capture interactions among explicit groups and extend to a broader group family?
  3. Are there any empirical details on the number of boosting rounds you used, as well as that used by LSBoost, at the same tree depth? I think comparing the number of round convergence of “one-shot” and “recursive” algorithms is quite interesting.

局限性

Yes.

最终评判理由

All the main concerns are solved. W1 is partially addressed and could be an interesting further direction.

格式问题

N/A

作者回复

We thank the reviewer for the careful review, and we’re happy to answer the questions raised by the reviewer.

  1. On the binary group features

    As pointed out by the reviewers, LSBoost extends the binary groups to a function class that is closed under affine transformations, while currently, we do not fully support such a function class. This extends the scope of our algorithm beyond group fairness, and we intend to leave this for future investigation. We’ll also discuss more about this in point 3.

  2. On the evaluation in section 5.3.2

    We’d like to point out that the metric we use in 5.3.2 is the Multicalibration Error in Definition 3.2, which only works with discretized predictors. Therefore, it’s not possible to directly calculate this metric using the undiscretized version of the predictor. Indeed in practice, there may be other metrics that practitioners might be interested in, and we provided a smooth indicator of multicalibration in Section 5.3.3. For other practical metrics beyond multicalibration that works on continuous outputs, we reported the accuracy and loss in our response to Reviewer zBZu, and we kindly refer you to our answer there.

  3. On the behaviors of hidden groups, and extensions to deeper trees

    This is a great question! While we cannot guarantee performance on general unspecified "hidden" groups, it might be possible to extend our approach if these groups are assumed to have certain structures, which include:

    • Intervals on continuous features. Trees can learn to take thresholds on continuous features, and two trees of depth one on some feature can represent an interval on this feature. If we assume that there are some unknown sensitive groups that can be represented by intervals on some continuous feature, ensembles of trees of depth 2 could still learn to be multicalibrated wrt to that group.
    • Intersections on identifiable groups. If hidden groups are intersections of existing binary indicators or continuous feature intervals, then using trees of depth three instead (one level for f0f_0, two for group intersections) could enable multicalibration for these composite groups.

    Note that the tree ensemble will try to be multicalibrated w.r.t. all possible groups in this manner. For instance, the continuous age feature can be passed if we want the model to be multicalibrated to all age intervals. It's important to note that explicit group definition is generally preferable when possible, because extending to these complex structures increases model complexity, potentially leading to decreased performance due to bias-variance trade-offs. For instance, using depth-3 trees requires multicalibration with respect to all sets described by intersections of groups, which is a much more challenging task.

  4. On the empirical details of the number of trees

    We evaluate the number of trees of our best performing model on each task, averaged over 10 folds, and present the result in the column “Ours”.

    As for LSBoost, in most of the cases, trees of depth 1 performs better than trees of depth 2 as a weak learner, and these cases are removed from the evaluation. We record the number of refinement rounds of LSBoost as well as the total number of trees used in the remaining cases, and present the results in the column "LSBoost".

    Table: Number of trees used for each task.

    OursLSBoost
    Skin Lesion Classification423.9 ± 97.5
    Income Classification830.8 ± 133.4#Grid = 10: 64.2 ± 13.5 trees in 17.6 ± 4.5 rounds
    #Grid = 30: 441.9 ± 83.1 trees in 58.5 ± 16.0 rounds
    Comments Toxicity Classification558.0 ± 103.3
    Age Regression2496.1 ± 677.5#Grid = 20: 94.9 ± 14.0 trees in 8.9 ± 1.4 rounds
    #Grid = 30: 90.3 ± 30.6 trees in 7.0 ± 1.9 rounds
    Income Regression1201.3 ± 171.4
    Travel Time Regression108.2 ± 48.7#Grid = 10: 36.0 ± 32.2 trees in 15.0 ± 12.6 rounds
    #Grid = 30: 11.6 ± 6.0 trees in 6.2 ± 3.0 rounds
    #Grid = 100: 8.0 ± 3.8 trees in 4.6 ± 2.8 rounds

    The number of trees actually varies a lot across datasets, but in general, our models consist of more trees.

评论

Thank you for your comments and the additional experimental details. I have raised my score to 5.

最终决定

This paper proposes a discretization-free approach to multicalibration based on optimizing a loss objective over ensembles of depth-two decision trees. The work is motivated by the limitations of prior discretization-based methods, which introduce rounding errors, additional hyperparameters, and potential distortions in downstream use. The authors provide theoretical guarantees under a "loss saturation" condition, and present empirical evaluations across multiple datasets showing improvements over existing baselines.

The reviewers found the paper to be technically solid, novel, and relevant to the community. They noted that the method elegantly eliminates the discretization bottleneck, while retaining multicalibration guarantees. The approach is simple to implement (using standard tree ensemble solvers) and empirically performs competitively or better than previous multicalibration methods, especially on worst-group calibration metrics. The theoretical contribution was found to be well-motivated and rigorous. Reviewers also appreciated the clear presentation of the results.